Page tree

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 4 Current »

Release 5.0.1



NOTE: To work with formats that are proprietary to a desktop application, such as Microsoft Excel, you do not need the supporting application installed on your desktop.


NOTE: Filenames that include special characters can cause problems during import or when publishing to a file-based datastore. Do not use the slash (/) character in your filenames.

Native Input File Formats:

Designer Cloud powered by Trifacta® Enterprise Edition can read and import directly these file formats:

  • Excel (XLS/XLSX), upload only

    Tip: You may import multiple worksheets from a single workbook at one time. See Import Excel Data.

  • CSV
  • JSON, including nested

    NOTE: Designer Cloud powered by Trifacta Enterprise Edition requires that JSON files be submitted with one valid JSON object per line. Consistently malformed JSON objects or objects that overlap linebreaks might cause import to fail. See Initial Parsing Steps.

  • Plain Text
  • LOG
  • TSV


    NOTE: XML files can be ingested as unstructured text. XML support is not enabled by default. For more information, please contact Alteryx Customer Success and Services.


  • Avro

NOTE: Designer Cloud powered by Trifacta Enterprise Edition supports Hive connectivity, which can be used to read data for Hadoop file formats that are not listed here, such as Parquet. For more information, please view the documentation for your Hive version.

For more information on data is handled initially, see Initial Parsing Steps.


Native Output File Formats:

Designer Cloud powered by Trifacta Enterprise Edition can write to these file formats:

  • CSV
  • JSON


  • Tableau (TDE)
  • Avro

    NOTE: The Photon and Spark running environments apply Snappy compression to this format.

  • Parquet

    NOTE: The Photon and Spark running environments apply Snappy compression to this format.



Compression Algorithms:



NOTE: Importing a compressed file with a high compression ratio can overload the available memory for the application. In such cases, you can uncompress the file before uploading.

Or, if that fails, you should contact your administrator about increasing the Java Heap Size memory.

NOTE: GZIP files on Hadoop are not split across multiple nodes. As a result, jobs can crash when processing it through a single Hadoop task. This is a known issue with GZIP on Hadoop.

Where possible, limit the size of your GZIP files to 100 MB of less, or use BZIP2 as an alternative compression method. As a workaround, you can try to run the job on the unzipped file. You may also disable profiling for the job. See Run Job Page.

Read Native File Formats:

CSV SupportedSupportedSupported
Avro  Supported
Hive  Supported

Write Native File Formats:

Avro  Supported; always on
Hive  Supported; always on




  • No labels

This page has no comments.