An exported flow can be imported into .

Limitations

You cannot import flows that were exported before Release 6.8.


NOTE: You cannot import flows into a version of the product that is earlier than the one from which you exported it. For example, if you develop a flow on free , which is updated frequently, you may not be able to import it into other editions of the product, which are updated less frequently.

Imported flows do not contain the following objects:

NOTE: Depending on the import environment, some objects in the flow definition may be incompatible. For example, the connection type may not be valid, or a datasource may not be reachable. In these cases, the objects may be removed from the flow, or you may have to fix a reference in the object definition. After import, you should review the objects in the flow to verify.


Imported datasets that are ingested into backend storage for  may be broken after the flow has been imported into another instance. These datasets must be reconnected to their source. You cannot use import mapping rules to reconnect these data sources. This issue applies to the following data sources:



Import into a new Google Cloud Platform project

When a flow is imported from one Google Cloud Platform (GCP) project into another GCP project, the underlying data must be accessible through the imported flow in the new GCP project. This means:

NOTE: If the above requirements are not met for flows imported into a different project, users may experience 403 Access Forbidden errors when attempting to connect to the flow or its underlying assets.

NOTE: If you import a flow from into , you may encounter soft validation errors during job execution. If the imported flow uses custom VPC mode, then the job execution may fail, since network may be inaccessible. The workaround is 1) to set the VPC mode to Auto or 2) set accessible VPC options before you run your job. See Dataflow Execution Settings.

NOTE: If your imported flow contains an output object where dataflow execution overrides have been specified, then the overrides are applied to any jobs executed in the project where the flow was imported. Property values that do not appear in the imported output object are taken from your execution settings. See Dataflow Execution Settings.


Import

NOTE: If the exported ZIP file contains a single JSON file, you can just import the JSON file. If the export ZIP also contains other artifact files, you must import the whole flow definition as a ZIP file. For best results, import the entire ZIP file.

Steps:

  1. Export the flow from the source system. See Export Flow.
  2. Login to the import system, if needed.
  3. Click Flows.
  4. From the context menu in the Flow page, select Import Flow.

    Tip: You can import multiple flows (ZIP files) through the file browser or through drag-and-drop. Press CTRL/COMMAND + click or SHIFT + click to select multiple files for import.

  5. Select the ZIP file containing the exported flow. Click Open.


Tip: When you import the flow, click the Warnings link to review the list of objects that must be remapped.

The flow is imported and available for use in the Flows page. After import: