Orchestrating the Data Product Pipeline

Optionally, you can orchestrate the data product pipeline either using the Job API create operations or by manually running jobs in the Tamr Cloud UI.

See the Jobs API documentation for more information on running each of the jobs below. Additionally, see Coordinating Data Product Refresh via the Jobs API for a sample script to perform these tasks end-to-end.

To orchestrate the pipeline via the Jobs API:

  1. Refresh the sources in the data product. This job requires the sourceId for the source, which is available on the Configurations > Sources page.
  2. Update (refresh) the data product. This job requires the dataProductId for the data product, which is available on the Configure Data Product page for that data product.
  3. Stage, then Apply updates to the Tamr RealTime SOR. Run the Stage and Apply workflows using their respective workflowIds, which are available on the System of Record Management page and the Configurations > Workflows page.
  4. Publish the data product outputs. This job requires both the dataProductId and the destinationId. The destinationId is available on the Publish page for that data product.
  5. (Optional) Publish the SOR mapping table. This job requires the destinationId for the destination, which is available on the Configurations > Destinations page.