Page tree

Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Info

NOTE: To work with formats that are proprietary to a desktop application, such as Microsoft Excel, you do not need the supporting application installed on your desktop.

Filenames

Info

NOTE: Filenames that include special characters can cause problems during import or when publishing to a file-based datastore. Do not use the slash (/) character in your filenames.

Native Input File Formats

...

  • XML

    Info

    NOTE: XML files can be ingested as unstructured text. XML support is not enabled by default. For more information, please contact

    D s proserv
    .



  • Avro


For more information on data is handled initially, see Initial Parsing Steps.


Native Output File Formats

...

  • Parquet

    Info

    NOTE: The

    D s photon
    and Spark running environments apply Snappy compression to this format.



Compression Algorithms


Info

NOTE: Importing a compressed file with a high compression ratio can overload the available memory for the application. In such cases, you can decompress the file before uploading.

If decompression fails, you should contact your administrator about increasing the Java Heap Size memory.


Info

NOTE: Publication of results in Snappy format may require additional configuration. See below.


Info

NOTE: GZIP files on Hadoop are not split across multiple nodes. As a result, jobs can crash when processing it through a single Hadoop task. This is a known issue with GZIP on Hadoop.

Where possible, limit the size of your GZIP files to 100 MB of less, or use BZIP2 as an alternative compression method. As a workaround, you can try to run the job on the unzipped file. You may also disable profiling for the job. See Run Job Page.



Read Native File Formats


 GZIPBZIPSnappy
CSV SupportedSupportedSupported
JSONSupportedSupportedSupported
Avro  Supported
Hive  Supported

...