About Data Pipelines on Autonomous AI Database

Autonomous AI Database data pipelines are either load pipelines or export pipelines.

Load pipelines provide continuous incremental data loading from external sources (as data arrives on object store it is loaded to a database table). Export pipelines provide continuous incremental data exporting to object store (as new data appears in a database table it is exported to object store). Pipelines use database scheduler to continuously load or export incremental data.

Autonomous AI Database data pipelines provide the following:

Data Pipeline Lifecycle

The DBMS_CLOUD_PIPELINE package provides procedures for creating, configuring, testing, and starting a pipeline. The pipeline lifecycle and procedures are the same for both load and export pipelines.

Description of the illustration pipeline_lifecycle.png

For either pipeline type you perform the following steps to create and use a pipeline:

  1. Create and configure the pipeline. See Create and Configure Pipelines for more information.

  2. Test a new pipeline. See Test Pipelines for more information.

  3. Start a pipeline. See Start a Pipeline for more information.

In addition, you can monitor, stop, or drop pipelines:

Load Pipelines

Use a load pipeline for continuous incremental data loading from external files in object store into a database table. A load pipeline periodically identifies new files in object store and loads the new data into the database table.

A load pipeline operates as follows (some of these features are configurable using pipeline attributes):

Description of image follows

Description of the illustration load-pipeline.svg

Migration from non-Oracle databases is one possible use case for a load pipeline. When you need to migrate your data from a non-Oracle database to Oracle Autonomous AI Database on Dedicated Exadata Infrastructure, you can extract the data and load it into Autonomous AI Database (Oracle Data Pump format cannot be used for migrations from non-Oracle databases). By using a generic file format such as CSV to export data from a non-Oracle database, you can save your data to files and upload the files to object store. Next, create a pipeline to load the data to Autonomous AI Database. Using a load pipeline to load a large set of CSV files provides important benefits such as fault tolerance, and resume and retry operations. For a migration with a large data set you can create multiple pipelines, one per table for the non-Oracle database files, to load data into Autonomous AI Database.

Export Pipelines

Use an export pipeline for continuous incremental export of data from the database to object store. An export pipeline periodically identifies candidate data and uploads the data to object store.

There are three export pipeline options (the export options are configurable using pipeline attributes):

Export pipelines have the following features (some of these are configurable using pipeline attributes):

Oracle Maintained Pipelines

Autonomous AI Database on Dedicated Exadata Infrastructure provides built-in pipelines to export specific logs to an Object Store in JSON format. These pipelines are preconfigured and are started and owned by the ADMIN user.

The Oracle Maintained pipelines are:

To configure and start an Oracle Managed pipeline:

  1. Determine the Oracle Managed Pipeline you want to use: ORA$AUDIT_EXPORT or ORA$APEX_ACTIVITY_EXPORT.

  2. Set the credential_name and location attributes.

    Note: The credential_name is a mandatory value on Autonomous AI Database on Dedicated Exadata Infrastructure.

    For example:

     BEGIN
       DBMS_CLOUD_PIPELINE.SET_ATTRIBUTE(
         pipeline_name => 'ORA$AUDIT_EXPORT',
         attribute_name => 'credential_name',
         attribute_value => 'DEF_CRED_OBJ_STORE'
       );
       DBMS_CLOUD_PIPELINE.SET_ATTRIBUTE(
         pipeline_name => 'ORA$AUDIT_EXPORT',
         attribute_name => 'location',
         attribute_value => 'https://objectstorage.us-phoenix-1.oraclecloud.com/n/namespace-string/b/bucketname/o/'
       );
     END;
     /
    

    The log data from database is exported to the object store location you specify.

    See SET_ATTRIBUTE for more information.

  3. Optionally, set the interval, format, or priority attributes.

    See SET_ATTRIBUTE for more information.

  4. Start the pipeline.

    See START_PIPELINE for more information.

Related Content

DBMS_CLOUD_PIPELINE Package