Page tree

Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

  1. Acquire the target URL for the datastore through the 
    D s webapp
    rtrue
     or through the datastore itself. Examples URLs:
    1. HDFS (file):

      Code Block
      hdfs:///user/warehouse/campaign_data/d000001_01.csv


    2. S3 (directory):

      Code Block
      s3:///3fad-demo/data/biosci/source/


  2. Navigate the browser to the appropriate URL in the 

    D s platform
    . The following example applies to the HDFS file example from above. It must be preceded by the base URL for the platform. For more information, see API - UI Integrations.

    Code Block
    <base_url>/import/data?uri=hdfs:///user/warehouse/campaign_data/d000001_01.csv


  3. For file-based URLs, the file is selected automatically.
  4. For directory-based URLs, the user can select which ones to include through the browser. Click the Add Datasets to a Flow. Add the dataset to an existing flow or create a new one for it.  
  5. After the datasets have been imported, open the flow in which your import is located. For the datasets that you wish to execute, you should do the following in the Flow View page:
    1. Click the icon for the dataset.
    2. From the URL, retrieve the identifiers for the flow and the dataset. These values are needed for later execution through the command line interface.
    3. Example:

      Dataset URLflowIddatasetId

      http://latest-dev.trifacta.netexample.com:3005/flows/31#dataset=186

      31186

      The flowId is consistent across all datasets that you imported through the above steps.

  6. You can open the datasets and wrangle them as needed.

  7. Complete any required actions from within your source application.   

...