See Job Results Page.
You can add, remove, or edit the outputs generated from this job. By default, a CSV output for your home directory on the selected datastore is included in the list of destinations, which can be removed if needed. You must include at least one output destination.
From the available datastores in the left column, select the target for your publication.
Add Publishing Action
NOTE: Do not create separate publishing actions that apply to the same file or database table.
New/Edit: You can create new or modify existing connections. See Create Connection Window.
- Select the publishing target. Click an icon in the left column.
If Hive publishing is enabled, you must select or specify a database table to which to publish.
Depending on the running environment, results are generated in Avro or Parquet format. See below for details on specifying the action and the target table.
If you are publishing a wide dataset to Hive, you should generate results using Parquet.
For more information on how data is written to Hive, see Hive Data Type Conversions.
Locate a publishing destination: Do one of the following.
NOTE: The publishing location must already exist before you can publish to it. The publishing user must have write permissions to the location.
NOTE: If your HDFS environment is encrypted, the default output home directory for your user and the output directory where you choose to generate results must be in the same encryption zone. Otherwise, writing the job results fails with a
Publish Job Failederror. For more information on your default output home directory, see User Profile Page.
- To sort the listings in the current directory, click the carets next to any column name.
- For larger directories, browse using the paging controls.
- Use the breadcrumb trail to explore the target datastore. Navigate folders as needed.
- Search: Use the search bar to search for specific locations in the current folder only.
- Manual entry: Click the Edit icon to manually edit or paste in a destination.
Choose an existing file or folder: When the location is found, select the file to overwrite or the folder into which to write the results.
NOTE: You must have write permissions to the folder or file that you select.
- To write to a new file, click Create a new file.
Create a new file: See below.
- Create Folder: Depending on the storage destination, you can click it to create a new folder for the job inside the currently selected one. Do not include spaces in your folder name.
To save the publishing destination, click Save Settings.
When you generate file-based results, you can configure the filename, storage format, compression, number of files, and the updating actions in the right-hand panel.
Output File Settings
- Select location: Navigate the Redshift browser to select the schema and table to which to publish.
- To create a new table, click Create a new table.
- Select table options:
- Table name:
- New table: enter a name for it. You may use a pre-existing table name, and schema checks are performed against it.
- Existing table: you cannot modify the name.
Output database: To change the database to which you are publishing, click the Redshift icon in the sidebar. Select a different database.
- Publish actions: Select one of the following.
- Create new table every run: Each run generates a new table with a timestamp appended to the name.
- Append to this table every run: Each run adds any new results to the end of the table.
- Truncate the table every run: With each run, all data in the table is truncated and replaced with any new results.
- Drop the table every run: With each run, the table is dropped (deleted), and all data is deleted. A new table with the same name is created, and any new results are added to it.
- Table name:
- To save the publishing action, click Save Settings.
To execute the job as configured, click Run Job. The job is queued for execution.