runPipeline

Runs a Data Integration pipeline (orchestration of a series of jobs as a single process) based on the variables defined for the pipeline. For detailed information on pipelines, see Using the Pipeline in Administering Data Integration for Oracle Enterprise Performance Management Cloud.

Applies to

Planning, Planning Modules, FreeForm, Financial Consolidation and Close, Enterprise Profitability and Cost Management, Tax Reporting, Sales Planning, and Strategic Workforce Planning

Required Roles

Service Administrator or any user set as a proxy user to run a pipeline definition (by using location security or by specifying the Enable Pipeline Execution for Non-Admin Users setting of System Setting to Yes).

Usage

epmautomate runPipeline PIPELINE_CODE [PARAMETER=VALUE], where:
  • PIPELINE_CODE is a code defined while creating the pipeline in Data Integration. For detailed information, see Pipeline Process Description in Administering Data Integration.
  • PARAMETER=VALUE, optionally, identifies the parameters and their values to run the pipeline. Specify as many parameter and value pairs as the process requires.

    The list of parameters depends on the number of variables defined in the Pipeline variable screen of Data Integration. Default pipeline parameters and acceptable values:

    • STARTPERIOD The first period for which data is to be loaded. This period name must be defined in Data Integration period mapping.
    • ENDPERIOD The last period for which data is to be loaded. This period name must be defined in Data Integration period mapping.
    • IMPORTMODE determines how the data is imported into Data Integration. Acceptable values are:
      • Append Adds to the existing POV data in Data Integration.
      • Replace Deletes the POV data and replace it with the data from the file specified in the pipeline definition or as a parameter.
      • Map and Validate Reprocesses the data with updated mappings and logic accounts without importing the data from the file specified in the pipeline definition or as a parameter.
      • No Import Skips data import into Data Integration staging table.
    • EXPORTMODE determines how the data is exported into Data Integration. Acceptable export modes are
      • Merge Merges the data in the Data Integration staging table with the existing data in the business process.
      • Accumulate Adds the data in the Data Integration staging table to the business process.

        This export mode applies only to Planning, Planning Modules, FreeForm, Sales Planning, and Strategic Workforce Planning.

      • Replace Deletes the POV data in the business process and replaces it with data from the Data Integration staging table. The data is cleared for Scenario, Version, Year, Period, and Entity dimensions.
      • No Export Skips data export from Data Integration into the business process.
    • ATTACH_LOGS specifies whether to zip and attach log files to notifications related to the pipeline execution. Acceptable values are Y for Yes and N for No.
    • SEND_MAIL specifies whether to a send an email on completion of the pipeline run. Acceptable values are: Always, No (default), On Failure, and On Success.
    • SEND_TO is a comma separated list of email IDs to which the email is to be sent.

Example

epmautomate runPipeline DAILYLOAD "STARTPERIOD=Jan-24" "ENDPERIOD=Jan-24" "IMPORTMODE=Replace" "EXPORTMODE=Merge" "SEND_MAIL=Always" "SEND_TO=John.Doe@example.com, Jane.Doe@example.com" "ATTACH_LOGS=Y"