Adding a Subledger Application

This section provides information for adding a Subledger Application (SLA).

To add a Subledger, do as follows:

1.     Click Add to create a new Subledger. A New Page is displayed.

2.     In Details,

a.     Enter the following details fields:

o       Specify a unique Subledger Application Name.

o       Describe the Subledger Application.

o       Specify a unique code for Subledger Application.

b.     Select the Event Class. The options displayed are Passthrough and Transaction.

c.     Select the Products / Entities  based on the Event selected click Next.

Based on the selected Event Class, the Product / Entities list is updated.

The  "Save Successful" Confirmation Message is displayed. The SLA details are successfully saved.

The status of the Subledger is changed to Draft and Created By details are added.

3.     In Event Types,

a.     From the list of event types available in the application, select the required event type.

b.     Click Add Event Type to add a new event type if not available.

A New window is displayed.

c.     Enter the following details:

o       Name

o       Code

o       Description

d.     Select the Event Class. The options displayed are Passthrough and Transaction.

e.     Click Save.

A new event type is added to the list of Event Types.

f.       Search for an Event Type in the Search Field Box.

g.     Click Next.

4.     In Attributes,

a.     From the list of attributes available in the application, select the required attribute.

b.     Click Add Attribute to add a new attribute if not available.

A New window is displayed.

c.     Enter the following details:

o       Name

o       Physical Name

o       Domain

d.     Select the Attribute Type. The options displayed are Header and Line.

e.     Select the Event Class. The options displayed are Passthrough and Transaction.

f.       Click Save.

A new attribute is added to the list of Attributes.

g.     Search for an Attribute in the Search Field Box.

h.     Click Next.

 

NOTE:   

Before proceeding with Header Mapping, you must register the SLA.

For more information, see the Registering Accounting Content with your ERP Service Section.

 

 

5.     In Header Mapping,

a.     Select the required Data Mapping from the Left Menu.

b.     Enter the Header Data Mapping Name in Specify the Header Data Mapping Name Field. This is auto-populated field. You can change as required

c.     Define the mapping from list of Subledger Attributes and Account Attributes using the corresponding check-box. Only selected Subledger attribute will be used.

d.     Click Next.

 

NOTE:   

Here, you can see the list of Product tables and respective columns. You must ensure that the data selected in the Product and Transaction Tables is available in the Staging Tables too.

 

6.     In Line Mapping,

a.     Select the required Data Mapping from the Left Menu. Here, you can see the list of Transaction Tables and respective columns. You must ensure that the data selected in the Product and Transaction Tables is available in the Staging Tables too.

b.     Specify the header data mapping name in Specify the line data mapping name Field. This is auto-populated field. You can change as required.

c.     Define the mapping from list of Subledger Attributes and Account Attributes using the corresponding check-box. Only selected Subledger attribute will be used.

7.     Click Next.

In Ledger Name Assignment, select the ledgers relevant to the Sub Ledger Application and the Assignment attributes to perform Ledger Name assignment. Also, select the Default Ledger for the SLA. The Ledger names are fetched from ERP. Each SLA can now have multiple ledger names configured based on the attributes.

For more information, see the Configuring a SLA section.

8.     Click Next.

In Review, you can view the details of the SLAs. In the Review Tab, the details of the connectors defined for the SLA (current SLA) can be viewed.

9.     Click Publish.

The SLAs which have connectors are auto-generated. You cannot modify the Seeded Connector.

10.  Click Yes to continue when the "Do you want to Proceed? Subledger application will not be available for any modification and will be locked for execution." confirmation message is displayed.

A pipeline is automatically created for the SLA in Process Management. Execute the Pipeline using the As of Date parameter.

For more details, see the Executing the SLA Pipelines section.

11.  Click Yes to continue.

The "Publish Successful" Confirmation Message is displayed after the SLA is published. The status changes to Available.

To Unpublish SLAs, see Unpublishing SLAs.

12.  Click the SLA, navigate to Product Information > Review to view the SLA connectors.

For more details, see the Creating Connectors section.

Accounting Balance Ingestion

The AFCS has a predefined process to extract the General Ledger Balance from your ERP Financials Cloud Service and ingest it into AFCS.  This is achieved by executing the GL Balances Process.

AFCS will support the run-time setting of As-Of date across the Journal and GL balance ingestion from Financials Cloud. In the case of journal data, the AFCS Posting Date will also be set based on the As-Of date set at run-time and ERP Accounting Date.

The following logic will apply:

·        As-Of Date is provided as a run-time parameter.

·        When it concerns journals data:

§       If ERP Accounting Date is on or after run-time As-Of Date.

§       As-Of Date and Posting Date are both set to ERP Accounting Date.

§       If ERP Accounting Date is prior to run-time As-Of Date.

§       As-Of Date is set to the provided As-Of Date.

§       Posting Date is set to ERP Accounting Date (indicating a back-dated journal line).

·        When it concerns GL Balance data:

§       AFCS As-Of Date is recorded as the run-rime As-Of Date provided.

§       GL Balance extract jobs will not permit extraction of balance ‘as of a prior date.

Prerequisite

ERP Journal Ingestion and GL Balance Ingestion connectors now refer to Ledger Dimension sourced from Ledger Master which should contain information on consolidation and accounting standard.

Before executing the GL Balances Process, perform the following to ensure that the files are available on the ERP Financial Cloud Service.

Schedule the GL_SR_EXTRACT_JOB Extract Job that is created in the BI Cloud Console while saving the AHCS Instance Details and ensure that the Job is completed.

GL Balances Process

The GL Balances Process consists of components that perform the following tasks.

1.     General Ledger Balance Extraction - Downloads extracted data files from the ERP Financial Cloud Service.

2.     Balance Ingestion - Loads data from the downloaded files to AFCS.

Executing GL Balances Process

To execute the GL Balances Process, follow these steps:

1.     Navigate to the Subledger Application Window.

2.     Select an SLA.

3.     Click Process Management to display the List of Processes linked to the selected SLA.

4.     To execute, click the GL Balances Process and open it in the Process Flow Canvas.

5.     Click Execute using the As of Date parameter.

For more details on how to view the execution status in the Process Monitor Window, see the Managing Processes Section.

NOTE:   

Irrespective of the As of Date parameter selected in the PMF run, the application will always extract the latest files for the respective job scheduled in the BI Cloud Console.

Journal Ingestion

AFCS has a configured mechanism to extract Journal Entries from your ERP Financials Cloud Service and ingest into AFCS. This is achieved via a pre-configured PMF Process (ERP Journal Ingestion). Oracle recommends that you do not alter the ERP Journal Ingestion Process.

 

NOTE:  Decomposition of Journals

AFCS supports decomposition of journals as of this release.

The decomposition of journals refers to the service's built-in ability to receive journal lines from ERP prepared off events or journal lines processed or grouped or published by AFCS and generate movement information for instrument-grain balance computation based on the following:

1.      Only journal lines that meet the following criteria can be decomposed by AFCS:

·            Extracted through the standard ERP Financials PVOs (see the ERP Financials Cloud Documentation for details).

·            Ingested through the AFCS-provided ingestion mechanism.

·            Prepared through Accounting Hub off events or journal lines processed or grouped or published by AFCS.

2.      Decomposition logic and execution are not user-configurable.

3.      While managing process sequence / scheduling, you must first run the BICC extract jobs followed by corresponding AFCS ingest processes. Once JOURNAL_INGESTION_EXTRACT_JOB is executed in BICC, 'ERP Journal Ingestion' process should be executed in AFCS before the former is executed again.

 

Prerequisite

Before executing the ERP Journal Ingestion Process, perform the following to ensure that the files are available on the ERP Financial Cloud Service:

Schedule the JOURNAL_INGESTION_EXTRACT_JOB Extract Job created in the BI Cloud Console while saving the AHCS Instance Details and ensure that the Job is completed.

ERP Journal Ingestion Process

The ERP Journal Ingestion Process consists of components that perform the following tasks.

1.     Journal Entries Extraction - Downloads the extracted data files from the ERP Financial Cloud Service.

2.     Journal Entries Ingestion - Loads data from the downloaded files to AFCS.

3.     Run DQ Checks - Performs DQ Checks on data loaded to the Accounting Entries in AFCS.

4.     Journal Entries Decomposition - Decomposes Journal Entries into the respective transactions.

Executing ERP Journal Ingestion Process

To view and execute the ERP Journal Ingestion Process, follow these steps:

1.     Navigate to the Subledger Application Window.

2.     Select an SLA.

3.     Click Process Management to display the List of Processes linked to the selected SLA.

4.     To execute, click the ERP Journal Ingestion Process and open it in the Process Flow Canvas.

5.     Click Execute using the As of Date parameter.

For more details on how to view the execution status in the Process Monitor Window, see the Managing Processes Section.

NOTE:   

Irrespective of the As of Date parameter selected in the PMF run, the application will always extract the latest files for the respective job scheduled in the BI Cloud Console.