Configure Application Settings
You can configure the settings for all your applications.
In the Oracle Integration navigation pane, click My Tasks, click Workspace, and then click Administration.
Option | Description |
---|---|
Under Services, create a connection to Oracle Content Management so that end users can collaborate using documents. |
|
Under Services, create a connection to Oracle Storage Service to store your archived data as objects that are stored in a container. |
|
Under Services, configure e-mail (human task) notifications. |
|
Runtime Settings |
Under Runtime Settings, configure the process audit level and schedule a daily recovery for process instances that encountered a remote fault while invoking a web service or expired timer messages and edit the logger’s settings. |
Archive and Purge |
Schedule instances archive or analytics archive to back up your data from one or more applications. |
Update time zone settings. |
|
Assign roles to users and groups. If needed, create, modify, and remove roles. |
|
Manage runtime keystore credentials for web and REST services. Upload, update, or delete credentials as needed. |
|
View details and resend email notifications sent for human tasks. |
Configure Oracle Content Management
Before users can access the documents feature, an administrator must configure settings in both Oracle Content Management and Process.
-
To read about the benefits of integrating with Oracle Content Management, see Why should I integrate documents?
-
To learn more about how to configure and set up your services, see How do I integrate with Oracle Content Management?
Configure Audit and Log Levels
You can select the type of messages you want to store in the audit trail and schedule a daily auto recovery for process instances that encounter a remote fault while invoking a web service or expired timer messages. You can also recover the invoke or callback messages that were not delivered and resubmit them. You can use the logger setting to change the log levels of different loggers.
Configure the Process Audit Level
Auto recovery is performed once each 24 hours to recover any faulted (non-service task related faults) or stranded process instances. Note that auto recovery doesn’t pick up process instances that fault on a service task and then wait for manual recovery on the same service task. By default, auto recovery starts at 00:00 hours and stops at 04:00 hours (server time zone), or earlier if there are no instances pending recovery. By default, it recovers in batches of 50 instances.
To configure the process audit level:
-
In the Oracle Integration navigation pane, click My Tasks, click Workspace, and then click Administration.
-
Click Runtime Settings.
-
In the Process Runtime section, select an audit level from the Audit Level drop-down list.
Available audit levels are:-
Production: logs all events but doesn't log the values used when assigning input or output values to data objects.
-
Development: log all events and also logs the values used when assigning input or output values to data objects.
-
Off: doesn't log any events.
Note:
If the audit level is Production or Development, the following activities will record the payload details:-
USER_TASK
-
SERVICE_TASK
-
RECEIVE_TASK
-
SEND_TASK
-
THROW_INTERMEDIATE_EVENT
-
CATCH_INTERMEDIATE_EVENT
-
BUSINESS_RULE_TASK
-
START_EVENT END_EVENT
-
-
Select a time to start and stop the scheduled recovery using the time editor.
The runtime environment uses the selected time and configuration to perform a daily scheduled recovery. It is recommended that you set the recovery window to off-peak hours.
-
Enter the maximum number of instances to recover.
It is recommended that you specify a small batch size (maximum number of instances to recover) such as 50.
-
Click Save.
Configure Logger Settings
As an administrator, you can use the Logger Settings section to change the log levels of different loggers and send an error report to Oracle if an error occurs. Logger levels include Incident_Error, Error (Severe), Warning, Notification (Info), Notification (Config), and Trace.
Configure Oracle Storage Service
Oracle Storage Service is an object store that is used to save objects that are identifiable by names within containers. You must request access to an Oracle Storage Service account before provisioning and then create a container to be able to export your data to Oracle Storage Service. After the account and container are created, use the Oracle Storage Service URL, container name, and login credentials to create a connection between the two services. Only when the connection has been made and tested successfully can you enable archiving.
Enable Email Notifications
You can configure Process to use emails for human workflow notifications that are sent to the task assignee when events such as assignment and reassignment occur.
Note:
Before enabling email notifications in runtime, make sure you customize the email notifications, for example, their content, template, attachments, and subject lines in design time. See Customize Notification Emails for Human Tasks. After enabling email notifications, you can view the notification logs and resend emails to all or some of the original recipients. See View and Resend Email Notifications.Email Notifications from Comments in Tasks
You can send email notifications to other task assignees and also to a specific user while entering a comment in the Comments field of a task.
Email notifications can be sent to task assignees or to a specific user by tagging them in the comment entered for a task. End users can tag comments with @assignees to notify all task assignees and @userid to notify a specific user.
An email with the specified comment is sent accordingly. The subject of the email notifies that a comment has been entered by the sender for a particular task. The recipient can view the comment entered for the task in the email without actually opening the task. A link to the task for which the comment was entered is provided in the email. The recipient can click the link to access the task and view details.
Note that before you can send email notifications from a comment, you have to enable email notifications in runtime. You can also view the notification logs and resend emails to recipients. See View and Resend Email Notifications.
Archive and Purge Data
You can submit requests for scheduled instances archive and purge or scheduled analytics archive to back up your process instances from one or more applications. The data you archive gets saved in Oracle Storage Service.
You can also schedule the Auto Purge from the Schedule Instances Archive page and be able to remove the BPM runtime information. The purge runs as a separate job on the database and ensures optimal performance.
Note:
Make sure you have configured the Oracle Storage Service settings before you archive. See Configure Oracle Storage Service.Schedule Instances Archive and Purge
Specify whether and when to archive your data automatically by creating a schedule based on either a time you select or a CRON expression. Also specify the number of days to retain data before it is purged. (You do not need to enable archiving to set purging retention.)
Note:
Data retention settings configured in the Oracle Integration Settings area can affect Process instance retention. For example, the Purge When Low Space Reached field determines whether Integration and Process runtime instances are automatically purged without reclaiming database space. In addition, clicking the Perform Manual Purge button performs a manual purge of all integration and process instances. See Set Data Retention in Provisioning and Administering Oracle Integration Generation 2.To schedule instance archive and purge:
-
In the Oracle Integration navigation pane, click My Tasks, click Workspace, and then click Administration.
-
Click Archive and Purge.
-
Click the Schedule Instances Archive tab.
-
To schedule archiving, select the Enable Archive check box and select archive settings.
Note:
Disabling Enable Archive and clicking Save terminates all archive schedules.-
Configure a schedule for the archive. You can use either the time fields or the advanced scheduling interval.
-
Use the time fields to specify when (how often, what day, and what time) the archive occurs. For example, to schedule an archive for every week on Friday at 03:20, for Every, select Week, for on select Friday, and for at, select 03 for hour and 20 for minutes.
-
Use the advanced scheduling option to specify a CRON expression. Select the Use Advanced Scheduling Interval check box and enter a valid CRON expression. Click the adjacent help icon for CRON details and examples.
Note:
Enabling the Use Advanced Scheduling Interval check box automatically disables the settings in the time fields.
-
-
In the Configure Archive Content fields, select items to include in the archive.
-
In the Archive Job Timeout field, enter the maximum number of minutes that the archive is allowed to run.
-
In the Failure Notification Address field, enter an email address to send archive error notifications to.
-
-
Schedule purging in the Purge Retention field.
Enter the number of days that data should be retained. Once the number of days is complete, the data gets purged. (The default number of days is 7.) You can set the purge retention without enabling archiving.
-
Click Save.
Review the confirmation message for details about your request.
You can expand the Archive Requests section to view the status of your scheduled archives. See View Archive Requests.
Schedule Analytics Archive and Purge
Specify whether and when to automatically archive your analytics data by creating a schedule based on either a time you select or a CRON expression. Also specify the number of days to retain analytics data before it is purged. (You do not need to enable archiving to set purging retention.)
To schedule analytics archive and purge:
- In the Oracle Integration navigation pane, click My Tasks, click Workspace, and then click Administration.
- Click Archive and Purge.
- To schedule analytics archiving, select the Schedule Analytics
Archive tab, select the Enable Archive
check box and select archive settings.
Note:
Disabling Enable Archive and clicking Save terminates all analytics archive schedules.- Configure a schedule for this archive. You can use either the time
fields or the advanced scheduling interval.
- Use the time fields to specify when (how often, what day, and what time) the archive occurs. For example, to schedule an archive for every week on Friday at 03:20, for Every, select Week, for on select Friday, and for at, select 03 for hour and 20 for minutes.
- Use the advanced scheduling option to specify a CRON expression. Select the Use Advanced Scheduling Interval check box and enter a valid CRON expression. Click the adjacent help icon for CRON details and examples. Enabling the Use Advanced Scheduling Interval check box automatically disables the settings in the time fields.
- In the Failure Notification Address field, enter the email address that all error notifications will be sent to.
- Configure a schedule for this archive. You can use either the time
fields or the advanced scheduling interval.
- Schedule analytics purging.
Click the Schedule Analytics Purge tab and enter the number of days that data should be retained in the Purge Retention field. Once the number of days is complete, the data gets purged. (The default number of days is 7.) You can set the purge retention without enabling archiving.
- Click Save.
Review the confirmation message for details about your request.
You can expand the Archive Requests section to view the status of your scheduled archives. See View Archive Requests.
After you schedule an analytics archive, some records may not be archived to the Oracle Storage Service. In such cases, the missing records are picked up during the next archive cycle.
View Archive Requests
When you submit a request for a scheduled instance archive or a scheduled analytics archive, the Archive Requests table displays the current status of the job.
To view the Archive Requests table:
-
In the Oracle Integration navigation pane, click My Tasks, click Workspace, and then click Administration.
-
In the Administration pane, click Archive and Purge.
-
Expand the Archive Requests section on the appropriate archive tab.
State | Description |
---|---|
Pending |
The archive request has been submitted for processing, but it hasn't been picked up by the scheduler. |
Running |
The archive is in progress. |
Completed |
The archive is complete. The archived data is stored in Oracle Storage Service. |
Terminated |
The request was canceled or stopped, most likely due to a server restart. May also occur when the schedule gets changed. |
Failed |
An exception occurred while executing the job. To get details, view the server log file or use the REST APIs. |
Note:
A scheduled request maintains its job ID and one row, and the state alternates between Pending and Running. The scheduled request gets Terminated only when the schedule gets changed. In that case, the scheduled request gets a new job ID. The request will again go into the Pending and Running states.Work with Archive Data
The data that you archive from one or more applications can be retrieved and used for audit and other purposes.
The key steps to archive and then extract the archived data for use are:
Want to learn more? Start by exploring the archive structure and then learn about the various components of the archive.
Note the following about archive jobs:
- Users can set the limit for archive job timeout.
- The maximum allowed export size is 500 MB. If the file system usage exceeds the maximum threshold of 500 MB, whatever has been exported till that time will be packaged and uploaded to the object storage.
- In an archive job, when the file system usage exceeds the maximum threshold (500 MB), but the timeout limit set by the user for the archive job has not yet been exceeded, the archive job will attempt another batch. In such a case the last flow id to be exported may become the first flow id exported in the next batch, resulting in duplicate exports.
Explore Archive Structure
The data you archive gets saved as a ZIP file. The root directory is bpm-archive. There are a number of sub-directories under this root directory. The two main sub-directories are – metadata and runtime.
The metadata directory provides details about the deployed applications.
The runtime directory contains the exported runtime data. This runtime data is distributed across sub-directories that identify the application, its version, and the Flow_ID for the application.
The Flow_ID is an internal identifier that unites related process instances into a common flow. Each time a process gets executed a process instance ID is created. For example, two processes will have unique process instance Ids. And if one of the process calls the other process, another unique process instance Id is created. But all the three will share a common Flow_ID. The runtime data is exported based on this common Flow_ID so that related processes get exported together and contained in the same export sub-directory.
Components of Archive
You can use the components of the archive to display and examine the extracted information.
The archive contains the following components:
- MetaData
- Runtime
- Audit Diagram
- Audit Trace
- Task
Metadata
The metadata directory contains information about the process applications that corresponds to the process instances contained in the archive. The application files are stored in their own directory structure that identifies the partition they are deployed to – oracleinternalpcs, application name, and its version. The files are stored in this way to ensure separation if there are multiple deployments of the same application.
Runtime
The runtime directory contains the exported runtime information. The runtime information is distributed in several sub-directories to keep the overview of the applications, their versions, and the related process instances.
It contains an important directory called the FlowInstance directory. See FlowInstance.
Audit Diagram
The Audit Diagram is a PNG file that contains a snapshot of the process instance and the process flow, that is, the path taken by the process during its execution.
The file is named as audit-diagram-${processInstanceId}.png, where, ${processInstanceId} is the process instance identifier that the diagram represents.
Audit Trace
The audit trace file is an export of the information collected by the audit system. This contains a root level audit log element and then any number of audit instance elements. Each audit instance element represents some operation that occurred in the execution of the process. The name of the file follows a similar structure to the audit diagram in that it is appended with the Process Instance Id, for example, audit-trace-${processInstanceId}.xml.
Note that if more than one process instance have been invoked, then there will be more than one audit trace file.
See Audit Trace Elements.
Audit Log
The audit log is a structured object that contains the detailed information about the audit instance.
The components of the audit log are:
- Data State
- Gateway Execution
- Service Input
- Service Output
See Audit Trace Elements.
Task
Task provides information about the human tasks executed during the process execution. It also includes the comments and attachment details. When external Social Network or Oracle Docs is not configured, then the comments are included in-line of the task export and the attachment are included in the archive content.
See Tasks Elements.
Audit Trace Elements
Audit trace contains the following:
ActivityName
This is the internal generic name for an activity. It is of type string, and the allowed values are:
- CATCH_INTERMEDIATE_EVENT
- PARALLEL_GATEWAY
- SERVICE_TASK
- START_EVENT
- USER_TASK
AuditInstance
Element | Type | Description |
---|---|---|
DNApplicationName |
string |
The name of the partition where the application is deployed into. |
DNCompositeName |
string |
The name of the Composite, that is, the application name. |
DNLabel |
string |
The unique label for an application deployment. |
DNRevision |
string |
The revision number as given by the user during the deployment of the application. |
ECID |
string |
An internal identifier for a group of related process instances. This identifier has given way to the Flow ID. |
activityId |
string |
The identifier for an application with a process definition. |
activityName |
string - enumerated |
The BPMN name for the activity. |
auditInstanceType |
string - enumerated |
The audit instance type determines at what stage of an activities life the audit message was recorded. Traditionally, this was START and END to represent pre-audit and post-audit messages. |
auditLevel |
integer |
The audit level when the event was recorded. |
auditLog |
object |
The audit log records more detailed information about the audit instance such as the state of the output data associations. |
componentName |
string |
The name of the process. |
componentType |
string |
The type of the process engine that executed the component. This will always be BPMN. |
compositeDn |
string |
The Domain Name identifier for the deployed application. This is based on four parts:
|
compositeInstanceId |
integer |
The composite instance ID is an identifier for the deployed application. This is largely replaced by the Flow ID and is provided for backward compatibility. |
compositeName |
string |
The name of the application process to which this audit record belongs. |
createTime |
dateTime |
The timestamp at which the audit instance record was created. |
dn |
string |
Domain name identifier for the deployed application. This is based on four parts:
|
faultIsRecoverable |
boolean |
If a fault was recorded on this audit event, then this flag will indicate if this was a recoverable fault. |
flowId |
integer |
The unique identifier that binds related process instances. |
instanceCount |
integer |
The instance number for a process activity that has been replicated multiple times. |
label |
string |
The user provided name for an activity in the process model. |
loopCount |
integer |
During a looping construct, an activity can be visited more than once. This counter tracks the number of times an activity is executed. |
operation |
string - enumerated |
This contains the audit instance operation. |
parentThread |
integer |
Each branch of a BPMN process is called a thread. Each thread is assigned a unique number. By default, this value will be -1 meaning that this is the primary or initial thread. As the process branches, the thread ID will increase and this will contain the value of the threads parent thread ID. |
partitionDate |
dateTime |
Internal use for the purge function where partition tables are used. |
processName |
string |
The name of the actual process being executed within the application. |
processTitle |
string |
The display name for the process being executed within the application. |
queryId |
integer |
The identifier for the audit record. |
scaPartitionId |
integer |
For internal use only. Used by the purge system. |
scopeId |
string |
The scope contains some run time information such as variables and data. These are tracked using a scope ID. |
sourceActivity |
string |
The previously executed activity in relation to the current activity. |
targetActivity |
string |
The next activity to be executed in relation to this activity. |
tenantId |
integer |
The identifier used for a multi-tenant install. This is for internal use only. |
threadId |
integer |
Each branch of a BPMN process is called a thread. Each thread is assigned a unique number. By default, this value will be -1 meaning that this is the primary or initial thread. As the process branches, the thread ID will increase. |
variables |
array |
This tracks the Analytics Variables. |
AuditInstanceType
The audit instance record can be classified into several types based on where they appear in the project model. It is of type string. The allowed values are:
- AFTER_INPUT_DATA_ASSOCIATION
- AFTER_INSTANCE_EXECUTION
- AFTER_OUTPUT_DATA_ASSOCIATION
- BEFORE_INPUT_DATA_ASSOCIATION
- BEFORE_INSTANCE_EXECUTION
- BEFORE_ITERATION
- BEFORE_OUTPUT_DATA_ASSOCIATION
- COMPONENT_STATUS
- END
- EXECUTION_LOGGING
- INTERMEDIATE
- START
AuditOperation
It is of type string. The allowed values are:
- AFTER_INPUT_DATA_ASSOCIATION
- AFTER_INSTANCE_EXECUTION
- AFTER_OUTPUT_DATA_ASSOCIATION
- BEFORE_INPUT_DATA_ASSOCIATION
- BEFORE_INSTANCE_EXECUTION
- BEFORE_ITERATION
- BEFORE_OUTPUT_DATA_ASSOCIATION
- COMPONENT_DEPLOYED
- COMPONENT_RETIRED
- COMPONENT_SUSPENDED
- COMPONENT_UNDEPLOYED
- EXECUTION_LOGGING
- FLOW_NODE_CANCELLED
- FLOW_NODE_DATA_CHANGED
- FLOW_NODE_IN
- FLOW_NODE_MOVED
- FLOW_NODE_OUT
- INSTANCE_ABORTED
- INSTANCE_CREATED
- INSTANCE_FAULT
- INSTANCE_RECOVERY_REQUESTED
- INSTANCE_RESUMED
- INSTANCE_SUSPENDED
- INSTANCE_SYSTEM_FAULT
- INSTANCE_TERMINATED
- INSTANCE_UPDATED
- MEASUREMENT_COUNTER
- MEASUREMENT_START
- MEASUREMENT_START_STOP
- MEASUREMENT_STOP
- STALE_ABORTED
- STALE_COMPLETED
AuditQueryPayload
Element | Type | Description |
---|---|---|
auditId |
integer |
The instance ID for the current audit event. |
ciKey |
integer |
The identifier for the process instance, that is, the Process Instance ID. |
BPMAudit
BPMAudit is the root element for the Audit Trace document. It contains an element auditInstance of type array. The auditinstance is the instance of the audit message. This will either be a pre-audit, post-audit or an ad-hoc message generated but the system to record such things as updates to the process flow (alter flow).
BPMNActivityNType
The type of activity that is being described. This can be a MESSAGE. It is of type string.
DataObject
The details of a data object. This can be a primitive, simple type or complex type. The structure of the value depends on the definition of the object within the process model design.
Element | Type | Description |
---|---|---|
detailId |
integer |
The identifier for the data object value where it has had to be moved the overflow storage. For internal use only. |
isBusinessIndicator |
boolean |
The flag to indicate if the data object is associated with an Analytics Business Indicator. |
name |
string |
The name of the data object. |
value |
anyType |
The raw value of the data object. |
DataState
It has the element dataObject. dataObject contains details of a data object. This can be a primitive, simple type or complex type. The structure of the value depends on the definition of the object within the process model design.
Element
It contains the details of a data object. This can be a primitive, simple type or complex type. The structure of the value depends on the definition of the object within the process model design.
Element | Type | Description |
---|---|---|
isBusinessIndicator |
boolean |
The flag to indicate if the data object is associated with an Analytics Business Indicator. |
name |
string |
The name of the data object. |
value |
anyType |
The raw value of the data object. |
FlowElementType
The flow element type refers to the basic BPMN type of the element. It of type string. The allowed values are:
- ACTIVITY
- EVENT
- GATEWAY
- PROCESS
ServiceInput
The details of a data object. This can be a primitive, simple type or complex type. The structure of the value depends on the definition of the object within the process model design.
ServiceOutput
The details of a data object. This can be a primitive, simple type or complex type. The structure of the value depends on the definition of the object within the process model design.
Variable
It contains the elements - name and variableInstance.
name is the name of the variable.
variableInstance contains the details of an analytics variable that are being tracked by audit.
VariableDirection
It is of type string. Allowed value: DIRECTION_STATE.
VariableInstance
Element | Type | Description |
---|---|---|
attribute |
integer |
The slot number that the variable occupies. |
name |
string |
The name of the variable. |
table |
string |
The name of the database table where the data value is stored. |
type |
string |
The type of variable. |
value |
anyType |
The value of the variable for this instance of the audit. The value is of type Object and can represent xsd:anyType. |
VariableModelType
The data type of the variable as it is declared in the application process model. It is of type string. The allowed values are Integer and String.
VariablePrimitives
The data type of the variable. It is of type string. The allowed values are:
- DATE
- NUMBER
- STRING
FlowInstance
The flowInstance directory is present immediately under the runtime directory.
Under the flowInstance directory is the partition directory – oracleinternalpcs. This is present to distinguish applications that may be deployed to another partition. The sub-directories for individual applications can be found under the partition directory. The name of the sub-directories are based on the application name and its version.
Under the sub-directories (named based on the application name and version) are the individual process instance directories. The name of these directories are derived from the Flow_ID of the process instance.
- The Flow_ID maintains reference to all related process instances.
- The Flow_ID is the internal reference that ties all related processes together. For example, when a process calls another process, a Process Instance ID is created for the new process. The two processes will each have their own process instance IDs. In this case all of them will share the same Flow_ ID. So the Flow_ID can contain runtime information of more than one related process.
- Flow_ID provides the sub-directory name for the exported runtime information. The sub-directories contain the individual export files that are all related to the same Flow_ID.
ProcessInstance
The process instance file contains the summary of the process instance. Note that the information contained in the file is at the time of export.
Element | Type | Description |
---|---|---|
applicationContext |
string |
The application name. |
creator |
string |
The creator of the application instance. |
cubeInstanceId |
string |
The cube instance identifier. |
identityContext |
string |
The scheme used for identity |
priority |
integer |
The priority designator of the deployment. |
processDN |
string |
The Domain Name for the process. |
title |
string |
The display name for the process instance |
Element | Type | Description |
---|---|---|
componentInstanceId |
integer |
The instance ID of the component. This should also represent the process instance ID. |
componentName |
string |
The name of the process within the component/application. |
compositeDN |
string |
The Domain Name for the composite/application. |
compositeInstanceId |
integer |
The identifier of the composite or application. This Id is not exposed to users. |
compositeVersion |
string |
The revision number of the deployed application. |
ecId |
string |
The ECID is an identifier to bind several related process instances together. This has been replaced with the FLOW_ID. |
flowId |
integer |
The unique identifier that binds related process instances. |
scaPartitionId |
string |
The deployment partition for the application. |
Tasks Elements
Tasks contains the following:
Action
It is of type string. The allowed values are:
- VIEW_PROCESS_HISTORY
- VIEW_SUB_TASKS
- VIEW_TASK
- VIEW_TASK_HISTORY
Attachment
Element | Type |
---|---|
attachmentScope |
string |
mimeType |
string |
name |
string |
size |
integer |
taskId |
string |
updatedBy |
string |
updatedByDisplayName |
string |
updatedDate |
dateTime |
version |
integer |
Callback
Element | Type |
---|---|
conversationId |
string |
id |
string |
Conversations
It contains the element conversationsEnabled which is of type boolean.
CustomAttributes
Element | Type |
---|---|
customAttributeNumber1 |
number |
customAttributeNumber2 |
number |
DocumentationDetails
It contains the element DocsEnabled which is of type boolean.
Payload
It contains the element IntakeWebForm which is of anyType.
ProcessInfo
Element | Type |
---|---|
instanceId |
integer |
processId |
string |
processName |
string |
Sca
Element | Type |
---|---|
applicationName |
string |
componentInstanceId |
integer |
componentName |
string |
compositeCreatedTime |
dateTime |
compositeDN |
string |
compositeInstanceId |
integer |
compositeName |
string |
compositeVersion |
string |
ecId |
string |
flowId |
integer |
parentComponentInstanceId |
string |
parentComponentInstanceRefId |
string |
scaPartitionId |
integer |
ShortHistoryTask
Element | Type |
---|---|
state |
string |
updatedDate |
dateTime |
version |
integer |
versionReason |
string |
SystemActions
Element | Type |
---|---|
displayName |
string |
SystemAttributes
Elements | Type |
---|---|
actionDisplayName |
string |
activityId |
string |
activityName |
string |
approvalDuration |
integer |
approvers |
string |
approversDisplayNames |
string |
assignedDate |
dateTime |
componentType |
string |
createdDate |
dateTime |
digitalSignatureRequired |
boolean |
endDate |
dateTime |
formName |
string |
hasSubTask |
boolean |
imageUrl |
string |
inShortHistory |
boolean |
isDecomposedTask |
boolean |
isGroup |
boolean |
isTemplateTask |
boolean |
isTestTask |
boolean |
numberOfTimesModified |
integer |
outcome |
string |
parentThread |
integer |
participantName |
string |
passwordRequiredOnUpdate |
boolean |
pushbackSequence |
string |
rootTaskId |
string |
secureNotifications |
boolean |
state |
string |
swimlaneRole |
string |
systemActions |
array |
systemStringActions |
string |
task |
array |
taskDefinitionId |
string |
taskDefinitionName |
string |
taskId |
string |
taskNamespace |
string |
taskNumber |
integer |
thread |
integer |
timersSuspended |
boolean |
updatedDate |
dateTime |
version |
integer |
versionReason |
string |
workflowPattern |
string |
SystemMessasgeAttributes
Element | Type |
---|---|
numberAttribute1 |
number |
numberAttribute10 |
number |
numberAttribute2 |
number |
numberAttribute3 |
number |
numberAttribute4 |
number |
numberAttribute5 |
number |
numberAttribute6 |
number |
numberAttribute7 |
number |
numberAttribute8 |
number |
numberAttribute9 |
number |
protectedNumberAttribute1 |
number |
protectedNumberAttribute10 |
number |
protectedNumberAttribute2 |
number |
protectedNumberAttribute3 |
number |
protectedNumberAttribute4 |
number |
protectedNumberAttribute5 |
number |
protectedNumberAttribute6 |
number |
protectedNumberAttribute7 |
number |
protectedNumberAttribute8 |
number |
protectedNumberAttribute9 |
number |
Task
Element | Type |
---|---|
applicationContext |
string |
correlationId |
string |
creator |
string |
creatorDisplayName |
string |
identityContext |
integer |
isPublic |
boolean |
mdsLabel |
string |
ownerRole |
string |
ownerRoleDisplayName |
string |
percentageComplete |
number |
priority |
integer |
taskDefinitionId |
string |
taskDefinitionURI |
string |
title |
string |
userComment |
array |
UpdatedBy
Element | Type |
---|---|
displayName |
string |
id |
string |
type |
string |
Configure Credentials for Web Services
You can use a credential to securely call a web service. You can add new credentials, make changes to existing credentials, or delete them.
You can manage credentials for a web service in both the environments. However, it’s more common to create a credential when you’re creating the connector to the web service. See Create a Web Service Connector.
Add a Runtime Credential
Manage Security Certificates during Runtime
Certificates are used to validate an application’s external web service connections when message security is applied. You can replace expired certificates and maintain a separate set of certificates for production use only.
Show Dates in User Time Zone
You can configure Process to show dates based on the user’s time zone.
Hide Comments and Attachments
You can configure Process to hide comments and attachments in tasks and start forms.
To restore the default settings, click Reset, and then click Reset to Default. To revert your changes to the last saved version, click Reset, and then click Reset to Last Saved.
Set the Default View for Process History
You can set the default view for process history from the UI Customization page.
The option you select will be set as the default view for Process history in all available processes in Workspace.
For example, if you selected Graphical View in the Process History View Type Setting, the process history of all available processes in Workspace will appear in the graphical view. If required, you can change the process history view of a particular process by choosing another option such as Tree View or List View from the drop-down list available under the History section of that process.