Skip to main content

Workspace Settings Page

The following settings can be customized for the user experience in your workspace. When you modify a setting, the change is immediately applied to the workspace. To access the page, select User menu > Admin console > Workspace settings.

Note

Users may not experience the changed environment until each user refreshes the application page or logs out and in again.

Enablement Options:

Note

Any values specified in the Workspace Settings page applies exclusively to the specific workspace and override any system-level defaults.

Option

Description

Default

The default value is applied. This value may be inherited from higher-level configuration.

Tip

You can review the default value as part of the help text.

Enabled

The setting is enabled.

Note

If the setting applies to a feature, the feature is enabled. Additional configuration may be required. See below.

Disabled

The setting is disabled.

Edit

Click Edit to enter a specific value for the setting.

General

Filter Job History

Set the default number of days of jobs that are displayed in the Job History page. Default value is 180 days.

Tip

You can filter the dates of the jobs displayed in the Job History page.

For more information, see Job History Page.

Hide underlying file system to users

When enabled, workspace users cannot see locations in the default storage layer.

Locale

Set the locale to use for inferring or validating data in the application, such as numeric values or dates. The default is United States.

Note

After saving changes to your locale, refresh your page. Subsequent executions of the data inference service use the new locale settings.

For more information, see Locale Settings.

Self service password reset

When enabled, workspace users can reset their own passwords via link on the login page.

Session Tags: Enable the use of session tags when assuming an IAM role

If you are using IAM roles to request temporary credentials for access to AWS resources, you can enable the use of session tags to make those requests. When a session tag is submitted, the Alteryx user is provided access to AWS resources based on the user's corresponding permissions within AWS, instead of having to specify those permissions in the Designer Cloud Powered by Trifacta platform. This method leverages the existing permission infrastructure in your enterprise and simplifies the use of IAM roles in the Trifacta Application.

Note

After enabling the use of session tags, you must spin up a new EMR cluster, which forces EMR to use the newly deployed credential provider JAR file.

Note

Additional configuration is required. For more information, seeConfigure AWS Per-User Auth for Temporary Credentials.

Session Tags: The name of the session tag that holds the username as its value

When Session Tags: Enable the use of session tags when assuming an IAM role is enabled, you must specify the name of the session tag to be submitted to AWS containing the username of the Alteryx user requesting resources. Default value is trifacta-user.

For more information, see Configure AWS Per-User Auth for Temporary Credentials.

Session duration

Specify the length of time in minutes before a session expires. Default is 10080 (one week).

Show file location

When enabled, workspace users can see the locations of source and output files within the application.

Storage directories

Allow members of the workspace to change paths to their upload and output results locations through their user profile.

For more information, see Storage Config Page.

User messaging

When enabled, workspace users can explore content through the Trifacta Application .

API

API Access Token

When accessing the REST APIs, you can optionally use a token for simpler use and enhanced security.

Note

This feature may not be available in all environments.

Note

API access tokens must be enabled to use the API reference documentation available through the User menu.

For more information, see Access Tokens Page.

Allow users to generate access tokens

When enabled, individual workspace users can generate their own personal access tokens, which enable access to REST APIs. For more information, see Manage API Access Tokens.

Maximum lifetime for user generated access tokens (days)

Defines the maximum number of days that a user-generated access token is permitted for use in the product.

Tip

To permit generation of access tokens that never expire, set this value to -1.

For more information, see Manage API Access Tokens.

Connectivity

Connectivity feature

When enabled, workspace users can create connections to relational datasources.

Note

Disabling this feature hides existing relational connections.

See Relational Access.

Custom SQL query

When enabled, users can create custom SQL queries to import datasets from relational tables.For more information, see Enable Custom SQL Query.

Enable S3 connectivity

When enabled, base connectivity to S3 is enabled for workspace users.

Note

Additional platform configuration is required. See S3 Access.

Enable conversion of standard JSON files via conversion service

When enabled, the Trifacta Application utilizes the conversion service to ingest JSON files and convert them to a tabular format that is easier to import into the application.

Note

This feature is enabled by default but can be disabled as needed. The conversion process performs cleanup and re-organization of the ingested data for display in tabular format.

When disabled, the Trifacta Application uses the old version of JSON import, which does not restructure the data and may require additional recipe steps to manually structure it into tabular format.

Note

The legacy version of JSON import is required if you are working with compressed JSON files or only Newline JSON files.

Note

Although imported datasets and recipes created under v1 of the JSON importer continue to work without interruption, the v1 version is likely to be deprecated in a future release. You should switch your old imported datasets and recipes to using the new version. Instructions to migrate are provided at the link below.

Max endpoints per JDBC REST connection

For a REST API connection to a JDBC source, this parameter defines the maximum number of endpoints that can be defined to use the connection.

Avoid modifying this value unless you are experiencing timeouts or failures to connect.

For more information, see REST API Connections.

Upgrade to connectivity

When enabled, workspace users are presented with the option to upgrade to a plan that supports connection to external data sources, if the feature is current disabled.

Upgrade to upload large files

When enabled, workspace users are presented with the option to upgrade to a plan that supports uploading large files, if the feature is current disabled.

Flows, recipes and plans

Collaborative suggestions

If desired, you can enable the inclusion of suggestion cards that are generated from recent use of the Trifacta Application. As the application gathers more information about how you or members of your workspace apply transformations to your data, the suggestions become more meaningful for the data that you are processing.

Note

No data is shared with Alteryx or any system outside of the Designer Cloud Powered by Trifacta platform.

These collaborative suggestion cards can be generated from individual usage or from workspace level usage. These suggestions appear under the Recently Used heading in the side panel.

Note

This feature requires the machine learning service, which is enabled by default. For more information, see Miscellaneous Configuration.

When this feature is enabled, individual users can still choose to opt-out of sharing their usage data with this feature. See User Profile Page.

Option

Description

disabled

Collaborative suggestions are not surfaced in the application.

personal

Collaborative suggestions are based on the individual user's previous transformations.

workspace

Collaborative suggestions are based on the transformations from all users in the workspace.

Default

The default setting for the workspace is applied.

Column from examples

When enabled, users can access a tool through the column menus that enables creation of new columns based on example mappings from the selected column.

Editor Scheduling

When enabled, flow editors are also permitted to create and edit schedules. For more information, see Flow View Page.

Note

The Scheduling feature may need to be enabled in your environment. When enabled, flow owners can always create and edit schedules.

When this feature is enabled, plan collaborators are also permitted to create and edit schedules. For more information, see Plan View Page.

Export

When enabled, workspace users are permitted to export their flows and plans. Exported flows can be imported into other workspaces or product editions.

Note

If plans are been enabled in your workspace, enabling this flag applies to flows and plans.

Import

When enabled, workspace users are permitted to import exported flows and plans.

Note

If plans have been enabled in your workspace, enabling this flag applies to flows and plans.

Maximum number of files to read in a directory for the initial sample

When the Trifacta Application is generating an initial sample of data for your dataset from a set of source files, you can define the maximum number of files in a directory from which the sample is generated. This limit is applied to reduce the overhead of reading in a new file, which improves performance in the Transformer page.

Tip

The initial sample type for files is generated by reading one file after another from the source. If the source is multiple files or a directory, this limit caps the maximum number of files that can be scanned for sampling purposes.

Note

If the files in the directory are small, the initial sample may contain the maximum number of files and less than the maximum size permitted for a sample. You may see fewer rows that expected.

If the generated sample is unsatisfactory, you can generate a new sample using a different method. In that case, this limit no longer applies. For more information, see Overview of Sampling.

Plan feature

When enabled, users can create plans to execute sequences of recipes across one or more flows. For more information, see Plans Page.

For more information on plans and orchestration, see Overview of Operationalization.

Sample downloads

When enabled, members can download the contents of the Transformer page at any time. For an individual step, a member can download the current sample, as modified by the current recipe up to the point of the current step. For more information, see Recipe Panel.

Schematized output

When enabled, all output columns for all types of outputs are typecast to their annotated types. This feature is enabled by default.

For non-schematized outputs, the Designer Cloud Powered by Trifacta platform enforces casting of all values to the annotated data type of the column by default. For example, if the output value is -3.4 and the data type for the output column is Integer, the platform enforces Integer type casting and writes a null value instead.

  • true: All output values must match the data type of the output columns, or a null value is written.

  • false: All output values are written in their output form, regardless of the column's data type.

Sharing

When enabled, workspace users are permitted to share flows and plans with other users in the workspace.

Note

If plans have has been enabled in your workspace, enabling this flag applies to flows and plans.

UI for range join

When enabled, workspace users can specify join key matching across a range of values. For more information, see Configure Range Join.

Webhooks

When enabled, webhook notification tasks can be configured on a per-flow basis in Flow View page. Webhook notifications allow you to deliver messages to third-party applications based on the success or failure of your job executions. For more information, see Create Flow Webhook Task.

Note

Additional configuration may be required. See Configure Webhooks.

Job execution

Combine Spark Transform and Profile jobs into one.

When enabled, the transform and profiling tasks of a job executed on the Spark running environment are combined. The profiling task is executed as a part of the transform task, which eliminates any time spent orchestrating the profiling task and accessing the profiler input file on storage.

Note

When these two tasks are combined, publishing actions are not undertaken if the profiling task fails.

For more information on these actions, see Configure for Spark.

In the Job Details page, combined jobs appear in a Transform with profile card. See Job Details Page.

Custom Spark Options Feature

When enabled, users can override Spark configuration options for output objects before running Spark jobs.

Tip

When enabled, a default set of Spark configuration options is available for users. Additional properties can be specified through the Spark Whitelist Properties setting.

See Enable Spark Job Overrides.

Databricks Cluster Policies

When enabled, this feature allows the Designer Cloud Powered by Trifacta platform to leverage cluster policies that you have created for use when creating new Databricks clusters for job execution.

Note

You must create cluster policies before enabling this feature. Each user may select a cluster policy to use. Additional configuration and considerations may be required. If no policy is selected, jobs may still be executed.

For more information:

Databricks Job Management

Enables job execution on Databricks through a secondary method. When enabled, Databricks jobs are executed via the run/submit API endpoint, which avoids the job quota limitation imposed by Databricks clusters. This flag also enables deletion of Databricks jobs from the Databricks workspace.

For more information on these options:

Databricks Job Runs Submit Fallback

When this flag is enabled, users can execute Databricks jobs using the runs/submit API method as a fallback when the job quota limit is reached for a Databricks workspace.

For more information:

Ignore publishing warnings for running jobs

When enabled, a user may execute a job if the previously saved location is not available for the current IAM permissions used to run the job. Default is Enabled .

Tip

Setting this value to Enabled is helpful for resolving changes in IAM permissions.

When disabled, the Run Job button is disabled if the previously saved location is not available through IAM permissions.

Tip

Setting this value to Disabled prevents execution of jobs that are going to fail at publication time, which can be expensive in terms of time and compute costs.

Logical and physical optimization of jobs

When enabled, the Trifacta Application attempts to optimize job execution through logical optimizations of your recipe and physical optimizations of your recipes interactions with data.

Note

This feature requires the optimizer service, which is enabled by default, and the optimizer service database, which is installed by default. For more information on installing the database, see Install Databases.

This workspace setting can be overridden for individual flows. For more information, see Flow Optimization Settings Dialog.

SQL Scripts

When enabled, users may define SQL scripts to execute as part of a job's run. Scripts can be executed before data ingestion, after output publication, or both through any write-supported relational connection to which the user has access.

For more information, see Create Output SQL Scripts.

Schema validation feature

When enabled, by default the structure and ordering of columns in your import datasets are checked for changes before data is ingested for job execution.

Tip

Schema validation can be overridden for individual jobs when the schema validation option is enabled in the job settings. See below.

Errors are immediately reported in the Job Details page. See Job Details Page.

For more information on schema validation, see Overview of Schema Management.

Schema validation option in job settings

When the schema validation feature and this setting are enabled, users can make choices on how individual jobs are managed when schema changes are detected. This setting is enabled by default.

For more information on schema validation, see Overview of Schema Management.

Schema validation option to fail job

When schema validation is enabled, this setting specifies the default behavior when schema changes are found.

  • When enabled, jobs are failed when schema changes are found, and error messages are surfaced in the Trifacta Application.

  • When disabled, jobs are permitted to continue.

    • Jobs may ultimately fail due to schema changes.

    • Jobs may result in bad data being written in outputs.

    • Job failures may be more challenging to debug.

      Tip

      Setting this value to Disabled matches the behavior of the Trifacta Application from before schema validation was possible.

Tip

This setting can be overridden for individual jobs, even if it is disabled.

Errors are immediately reported in the Job Details page. See Job Details Page.

For more information on schema validation, see Overview of Schema Management.

Spark Whitelist Properties

Comma-separated list of additional Spark properties to be whitelisted for configuration of output objects while running Spark jobs.

Note

The Custom Spark Options feature must be enabled.

See Enable Spark Job Overrides.

Trifacta Photon execution

When enabled, users can choose to execute their jobs on Trifacta Photon, a proprietary running environment built for execution of small- to medium-sized jobs in memory on the Trifacta node.

Tip

When enabled, you can select to run jobs on Photon through the Run Job page. The default running environment is the one that is best for the size of your job.

When Trifacta Photon is disabled:

  • You cannot run jobs on the local running environment. All jobs must be executed on a clustered running environment.

  • Trifacta Photon is used for Quick Scan sampling jobs. If Trifacta Photon is disabled, the Trifacta Application attempts to run the Quick Scan job on another available running environment. If that job fails or no suitable running environment is available, the Quick Scan sampling job fails.

Scheduling and parameterization

Include Hidden Files in Parameterization

When enabled, hidden files and hidden directories can be searched for matches for wildcard- or pattern-based parameters when importing datasets.

Tip

This can be useful for importing data from generated profiles, which are stored in the .profiler folder in a job output directory.

Note

Scanning hidden folders may impact performance. For existing imported datasets with parameters, you should enable the inclusion of hidden folders on individual datasets and run a test job to evaluate impact.

For more information, see Parameterize Files for Import.

Parameterization

By default, the Designer Cloud Powered by Trifacta platformsupports the application of parameters to imported datasets. Datetime, wildcard, or variable parameters can be used to operationalize execution of jobs on different versions of the same dataset.

When enabled, users can create parameters, which can be applied to import, creating sample, and outputs. For more information, see Overview of Parameterization.

Schedule list

When enabled, administrators and workspace administrators can see a list of all schedules in the workspace.

Scheduling feature

When enabled, workspace users can schedule the execution of flows. See Add Schedule Dialog.

Publishing

Avro output format

When enabled, members can generate outputs in Avro format.

CSV output format

When enabled, members can generate outputs in CSV format.

Hyper output format

When enabled, members can generate outputs in Hyper format for publication and use on Tableau Server.

JSON output format

When enabled, members can generate outputs in JSON format.

Parquet output format

When enabled, members can generate outputs in Parquet format.

Publish job results

When enabled, workspace users are permitted to publish results through the Output Destinations tab in the Job Details page to external datastores.

Note

These external datastores must be enabled and configured. See Connection Types.

For more information, see Job Details Page.

Publishing actions options

When enabled, users are permitted to create custom publishing actions for their jobs.

When disabled, users must accept the default publishing actions.

Notifications

Email notification feature

When enabled, the Designer Cloud Powered by Trifacta platform can send email notifications to users based on the success or failure of jobs. By default, this feature is Enabled.

Note

This feature requires access to an SMTP server to send emails. For more information, see Enable SMTP Email Server Integration.

Email notification trigger when flow jobs fail

When email notifications are enabled, you can configure the default setting for the types of failed jobs that generate an email to interested stakeholders. The value set here is the default value for each flow in the workspace.

Settings:

Setting

Description

Default (any jobs)

By default, email notifications are sent on failure of any job.

Never send

Email notifications are never sent for job failures.

Scheduled jobs

Notifications are sent only when scheduled jobs fail.

Manual jobs

Notifications are sent only when ad-hoc (manually executed) jobs fail.

Tip

Jobs executed via API are Ad hoc jobs.

Any

Notifications are sent for all job failures.

Individual users can opt out of receiving notifications or configure a different email address. See Email Notifications Page.

Emailed stakeholders are configured by individual flow. For more information, see Manage Flow Notifications Dialog.

Email notification trigger when flow jobs succeed

When email notifications are enabled, you can configure the default setting for the types of successful jobs that generate an email to interested stakeholders. The value set here is the default value for each flow in the workspace.

For more information on the settings, see the previous section. Default setting is Default (any jobs).

Individual users can opt out of receiving notifications or configure a different email address. See Email Notifications Page.

Emailed stakeholders are configured by individual flow. For more information, see Manage Flow Notifications Dialog.

Email notification trigger when plans run

You can configure the default trigger for email notifications when a plan runs. Default setting is Default (all runs).

Setting

Description

Default (all runs)

By default, email notifications are sent to users for all plan runs.

All runs

Emails are sent for all runs.

Failed runs

Emails are sent for failed runs only.

Success runs

Emails are sent for successful runs only.

Sharing email notifications

When email notifications are enabled, users automatically receive notifications whenever an owner shares the plan or flow with the user.

Individual users can opt out of receiving notifications. For more information, see Preferences Page.

Experimental features

These experimental features are not supported.

Warning

Experimental features are in active development. Their functionality may change from release to release, and they may be removed from the product at any time. Do not use experimental features in a production environment.

These settings may or may not change application behavior.

Cache data in the Transformer intelligently

Note

NOTE: This feature is in Beta release.

When enabled, this feature allows the Trifacta Application to cache data from the Transformer page periodically based on Trifacta Photon execution time. This feature enables users to move faster between recipe steps.

Default language

Select the default language to use in the Trifacta Application.

Edit recipes without loading sample

When enabled, you can perform edits in the Transformer page without loading a sample in the data grid.

Tip

This feature can be helpful when you know the edits that need to be performed and do not need sample data to perform the corrections. You can also use it to switch the active sample without loading.

In Flow View, select Edit recipe without datagrid from the context menu on the right side when the recipe is selected.

Enable/Disable data grid from view options

When enabled, you can enable or disable live previewing in the data grid of the Transformer page. Disabling can improve performance. These options are available in the Show/hide data grid options drop-down in the status bar at the bottom of the Transformer page:

  • Edit with data grid

    • When the data grid is disabled, you may not be able to edit some recipe steps. For steps that you can edit, select Preview to see the effects of the step on the data. When you select Preview, the data grid is re-enabled.

  • Show column histogram

    • When the data grid is enabled, you can choose to disable the column histograms in the data grid, which can improve performance.

For more information, see Data Grid Panel.

Execution time threshold (in milliseconds) to control caching in the Transformer

Note

NOTE: This feature is in Beta release.

When intelligent caching in the Transformer is enabled, you can set the threshold time in milliseconds for when Trifacta Photon updates the cache. At each threshold of execution time in Trifacta Photon, the output of the intermediate recipe (CDF) steps are cached in-memory, which speeds up movements between recipe steps in the Trifacta Application.

Language localization

When enabled, the Trifacta Application is permitted to display text in the selected language.

Show user language preference

When enabled, individual users can select a preferred language in which to display text in the Trifacta Application.

Note

This experimental feature requires installation of a language resource file on the Trifacta node. For this release, only U.S. English (default) and Korean are supported. For more information, please contact Alteryx Support.

Users can make personal language selections through their preferences. See Account Settings Page.

Wrangle to Python Conversion

Note

As of Release 9.7, Wrangle to Python conversion has been deprecated. For more information, please see End of Life and Deprecated Features.

Alpha feature: When enabled, you can use an API endpoint to generate Python Pandas code that completes the steps required to generate an output in Python.

Warning

This feature may be modified or removed in a future release without warning. It is intended for demonstration purposes only and should not be enabled in a production environment.

For more information, see API Task - Wrangle Output to Python.

Tip

You can download and install the Python SDK to integrate use of the Trifacta Application in your Python environment. Use the visual tools of the Trifacta Application to build your transformations, and then generate Python Pandas code for use in your Python data pipelines. For more information, see Python SDK.