...
- This feature must be enabled by the workspace admin. See below.
must be integrated with Snowflake. See Snowflake Connections.D s webapp - The permission to execute jobs in Snowflake must be enabled.
- All sources and outputs must reside in Snowflake.
- Permissions are required to run the Snowflake to file jobs.
Spark + Snowflake
must be selected as running environment. See Run Job Page.Jobs are executed in the virtual warehouse that is specified as part of the Snowflake connection.
Info NOTE: Job execution requires significantly more resources than ingest or publish jobs on Snowflake. Before you begin using Snowflake, you should verify that your Snowflake virtual warehouse has sufficient resources to handle the expected load. For more information, see Snowflake Connections.
- In your flow, you must enable all general and Snowflake-specific flow optimizations. When all of these optimizations are enabled, the job can be pushed down to Snowflake for execution. See "Flow Optimizations" below.
...
The following setting must be enabled in the workspace. Select User menu > Admin console > Workspace settingsSettings.
Optimization | Description | |
---|---|---|
Logical and physical optimization of jobs | When enabled, the
|
...
Optimization | Description | ||||||
---|---|---|---|---|---|---|---|
Snowflake > Column pruning from source | When enabled, job execution performance is improved by removing any unused or redundant columns from the source database. | ||||||
Snowflake > Filter pushdown | When this setting is enabled, the
| ||||||
Snowflake > Full pushdown | When this setting is enabled, all supported pushdown operations, including full transformation and profiling job execution, is pushed down to
| ||||||
Full execution for S3 | If requirements are met for data sourced from S3, you can enable execution of your S3 datasources in
| ||||||
Source to Files | When this setting is enabled, Snowflake table that meet all pushdown requirements can be executed through Snowflake and published to S3. |
For more information, see Flow Optimization Settings Dialog.
...
Tip |
---|
Tip: After launching the job, you can monitor job execution through the Job Details page, which includes a link to the corresponding job in the Snowflake console. |
S3 File Support
D s ed | ||
---|---|---|
|
In addition to
sources, you can execute jobs in D s conntype type snowflake
on source files from S3. D s conntype type snowflake
The Full execution for S3 file option must enabled in your flow. For more information, see Flow Optimization Settings Dialog.
Tip | ||||||||
---|---|---|---|---|---|---|---|---|
Tip: The |
Requirements
...
S3 or
D s tfs |
---|
...
D s conntype | ||
---|---|---|
|
...
D s webapp |
---|
...
optimizationD s conntype type snowflake - Full execution for S3 file
- See Flow Optimization Settings Dialog.
In the Run Job page, the Spark +
running environment must be selected. D s conntype type snowflake
Tip | ||||
---|---|---|---|---|
Tip: If this option is not available, one or more requirements for S3 file execution on |
...
Execution requirements
Info | ||||||||
---|---|---|---|---|---|---|---|---|
NOTE: For execution of S3 jobs in |
...
. |
...
Supported file formats from S3
...
CSV: Files that fail to meet the following requirements may cause job failures when executed in
, even though they can be imported into D s conntype type snowflake
. Requirements: D s product
For job execution of CSV files in
, source CSV files must be well-formatted.D s conntype type snowflake Newlines must be inserted.
Fields must be demarcated with quotes and commas.
Info NOTE: Escaped quotes in field values must be represented as double quotes (
""
). Escaped quotes with a backslash is not supported.- Each row must have the same number of columns.
...
TSV
JSON (newline-delimited)
Info | ||||
---|---|---|---|---|
NOTE:
|
...
gzip and bz2 compressed formats are supported.
Info | ||||
---|---|---|---|---|
NOTE: Snappy compression is not supported for S3 execution on
|
Supported file encodings:
- UTF-8
- ISO-8859-1
Supported delimiters:
- Comma
- Tab
- Pipe
Supported quote characters:
...
Uploaded File Support
When a file is uploaded from your desktop, ingested, and stored in a storage layer that is supported for file pushdown, jobs that reference datasets created from that file are eligible for execution in
D s conntype | ||
---|---|---|
|
D s conntype | ||
---|---|---|
|
...