Page tree

Release 6.0.2


Contents:

   

For the latest updates on available API endpoints and documentation, see api.trifacta.com.

This is the latest version of the APIs.

Contents:


For a specified jobGroup, this endpoint performs an ad-hoc publish of the results to the designated target.

  • Target information is based on the specified connection.
  • Job results to published are based on the specified jobGroup. 

You can specify:

  • Database and table to which to publish
  • Type of action to be applied to the target table. Details are below.

Supported targets:

  • Hive
  • Redshift

For more information on jobGroups, see API JobGroups Get v4.

For additional examples, see API Workflow - Publish Results.

Version:  v4

Required Permissions

NOTE: Each request to the Designer Cloud Powered by Trifacta® platform must include authentication credentials. See https://api.trifacta.com/ee/6.0/index.html#section/Authentication.

Request

Request Type: PUT

Endpoint:

/v4/jobGroups/<id>/publish

where:

ParameterDescription
<id>Internal identifier for the job group

Request URI - Example:

/v4/jobGroups/31/publish

Request Body - Hive:

{
  "connection": { 
    "id": 1 
  },
  "path": ["default"],
  "table": "test_table",
  "action": "create",
  "inputFormat": "pqt",
  "flowNodeId": 27
}

Response

Response Status Code - Success:  200 - OK

Response Body Example:

{
    "jobgroupId": 31,
    "reason": "JobStarted",
    "sessionId": "f6c5f350-2102-11e9-bb80-9faf7b15f235"
}

Reference

Request Reference:

PropertyDescription
connectionInternal identifier of the connection to use to write the results.
pathName of database to which to write the results. This value must be enclosed in square brackets.
tableName of table in the database to which to write the results.
action

Type of writing action to perform with the results. Supported actions:

  • create - Create a new table with each publication. This table is empty except for the schema, which is taken from the results. A new table receives a timestamp extension to its name.
  • load - Append a pre-existing table with the results of the data. The schema of the results and the table must match.
  • createAndLoad - Create a new table with each publication and load it with the results data. A new table receives a timestamp extension to its name.
  • truncateAndLoad - Truncate a pre-existing table and load it with fresh data from the results.
  • dropAndLoad - Drop the target table and load a new table with the schema and data from the results.
inputFormat

Source format of the results. Supported values:

Hive:

  • avro
  • pqt

Redshift:

NOTE: For results to be written to Redshift, the source must be stored in S3 and accessed through an S3 connection.

NOTE: By default, data is published to Redshift using the public schema. To publish using a different schema, preface the table value with the name of the schema to use: MySchema.MyTable.

 

  • csv
  • json
  • avro
flowNodeIdThe internal identifier for the recipe (wrangledDataset) from which the job was executed.

 

For more information on the available status messages, see API JobGroups Put Publish v4.

This page has no comments.