Skip to main content

Product Limitations

This section covers key known limitations of Designer Cloud Powered by Trifacta Enterprise Edition.

Note

This list of limitations should not be considered complete.

General Limitations

Data Volume

The Trifacta Application applies no fixed limits to the number of columns or rows that can be handled during transformation.

Note

During transformation, Designer Cloud Powered by Trifacta Enterprise Edition is designed to process data volumes of any size.

However, some important considerations:

Soft row limits

  • The number of rows that you see within the Trifacta Application in the currently selected sample is determined by:

    • Maximum permitted sample size stored on the base storage layer

    • Currently configured sample size for the current recipe

See Sampling below.

Soft column limits

  • Soft row limits do not affect the number of columns that are displayed. All available and visible columns are displayed. The number of rows may be affected by the number of columns.

    Tip

    Avoid creating and working with datasets that are wider than 1000 columns. Datasets that are wider than this recommendation may result in performance impacts in the Trifacta Application.

  • The number of columns may be limited by:

    • Number of columns permitted in the source datastore.

    • For SQL-based datastores, limits may be placed on the length of individual queries.

Sampling

  • Sample sizes are defined by parameter for each available running environment. See Sample Size Limits below.

  • All values displayed or generated in the application are based on the currently displayed sample.

    • Transforms that generate new data may not factor values that are not present in the current sample.

    • When the job is executed, transforms are applied across all rows and values in the source data.

    • Transforms that make changes based on data values, such as header and valuestocols, will still be configured according to sample data at the time of that the step was added, instead at execution time. For example, all of the values detected in the sample are used to determine the columns of a valuestocols transform step based on the selected sample when the step was added.

  • Random samples are derived from up to the first 1 GB of the source file.

    • Data from later parts of a multi-part file may not be included in the sample.

Internationalization

  • The product supports a variety of global file encoding types for import.

    For more information, see Configure Global File Encoding Type in the Configuration Guide.

  • Within the application, UTF-8 encodings are displayed.

    • Limited set of characters allowed in column names.

    • Header does not support all UTF-8 characters.

    • Emoji are not supported in data wrangling operations.

    • Umlauts and other international characters are not supported when filtering datasets in browsers of external datastores.

  • States and Zip Code Column Types and the corresponding maps in visual profiling apply only to the United States.

  • UTF-8 is generated in output.

  • UTF-32 encoding is not supported

Note

Some functions do not correctly account for multi-byte characters. Multi-byte metadata values may not be consistently managed.

Size Limits

Job Size Limits

Execution on a Spark running environment is recommended for any files over 5GB in net data size, including join keys.

Limitations by Integration

General

The product requires definition of a base storage layer, which can be HDFS or S3 for this version. This base storage layer must be defined during install and cannot be changed after installation. See Set Base Storage Layerin the Configuration Guide.

LDAP

  • If LDAP integration is enabled, the LDAP user [ldap.user (default=trifacta)] should be created in the same realm.

  • SeeConfigure SSO for AD-LDAPin the Configuration Guide.

Hadoop

  • Designer Cloud Powered by Trifacta Enterprise Editionrequires an integration with a working Hadoop cluster.

  • See Running Environment Optionsin the Configuration Guide.

Amazon AMI

  • For more information, see product documentation on the Amazon Marketplace.

Amazon EMR

  • For more information, see product documentation on the Amazon Marketplace.

Microsoft Azure

  • For more information, see product documentation on the Azure Marketplace.

Redshift

None.

S3

  • S3 integration is supported only over AWS-hosted instances of S3.

  • Oracle Java Runtime 1.8 must be installed on the node hosting the product.

  • Writing to S3 requires use of S3 as the base storage layer. For more information, see Set Base Storage Layer in the Configuration Guide.

  • When publishing single files to S3, you cannot apply an append publishing action.

Hive

  • Only HiveServer2 is supported.

  • You can create only one connection of this type.

  • When reading from a partitioned table, the product reads from all partitions, which impacts performance.

  • For more information, see Configure for Hivein the Configuration Guide.

Spark

  • None.

JDBC

  • The product supports explicit versions of each JDBC source. See Connection Types in the Configuration Guide.

  • Additional installation may be required. Additional limitations may apply.

  • See Relational Access in the Configuration Guide.

Limitations by Product Edition

Limitations for subscription editions

For more specific limitations on each product edition, see Pricing and Editions.

Other Limitations

  • File Formats:Limitations may apply to individual file formats. SeeSupported File Formats.

  • Data Type Conversions: There are some limitations on how data types are converted during import or export/publication. See Type Conversions.