Submit a Spark Compute Job

post

/bdcsce/api/v1.1/clustermgmt/{identityDomainId}/instances/{clusterId}/jobs/spark

Request

Path Parameters
Body ()
Definition of the Spark Job that you want to submit.
Root Schema : SparkJobHttpRequest
Type: object
Show Source
Nested Schema : SparkJob
Type: object
Show Source
Nested Schema : applicationArchives
Type: array
Archives to be uncompressed in the executor working directory (YARN mode only)
Show Source
Nested Schema : applicationArguments
Type: array
Arguments that need to be sent to the _Application Main_ or the _Application Script_.
Show Source
Nested Schema : applicationJarFiles
Type: array
List of jars that will be downloaded and collocated for the given Job.
Show Source
Nested Schema : applicationPyFiles
Type: array
List of python files that will be downloaded and collocated for the given Job.
Show Source
Nested Schema : applicationSupportFiles
Type: array
List of files that will be downloaded and collocated for the given Job.
Show Source
Nested Schema : driverClasspath
Type: array
Driver classpath.
Show Source
Nested Schema : excludePackages
Type: array
Comma-separated list of groupId:artifactId, to exclude while resolving the dependencies provided in `packages`.
Show Source
Nested Schema : executorEnv
Type: object
Additional Properties Allowed
Show Source
Map of environment variable that should be available to the Executor process. For example, the maximum waiting period that the requester is willing to wait for the job to actually start running inside a cluster before the Job is considered failed, REQUEST_TIMEOUT.
Nested Schema : executorExtraClasspath
Type: array
Extra classpath entries to prepend to the classpath of executors.This exists primarily for backwards-compatibility with older versions of Spark.Users typically should not need to set this option.
Show Source
Nested Schema : extraJavaOptions
Type: array
A map of extra JVM options to pass to executors and driver. For instance, GC settings or other logging. Note that it is illegal to set Spark properties or heap size settings with this option. Spark properties should be set using a SparkConf object or the spark-defaults.conf file used with the spark-submit script. Heap size settings can be set with spark.executor.memory.
Show Source
Nested Schema : extraListeners
Type: array
Extra listeners
Show Source
Nested Schema : maxJobDurationInSecs
Type: object
Maximum duration of a Job in seconds. When a Job exceeds the amount of time the Job will be terminated.
Nested Schema : maxSubmissionLatencyInSecs
Type: object
The maximum waiting period that the requester is willing to wait for the job to actually start running inside a cluster before the Job is considered failed, REQUEST_TIMEOUT.
Nested Schema : packages
Type: array
Comma-separated list of maven coordinates of jars to include on the driver and executor classpaths. Will search the local maven repo, then maven central and any additional remote repository The format for the coordinates should be `groupId:artifactId:version`.
Show Source
Nested Schema : repositories
Type: array
Comma-separated list of additional remote repositories to search for the maven coordinates given with `packages`.
Show Source
Nested Schema : sparkConf
Type: object
Additional Properties Allowed
Show Source
Map of Spark configuration options that were used to submit the Job.
Security
Back to Top

Response

200 Response

Body ()
Root Schema : AsyncJobStatus
Type: object
Show Source
Nested Schema : JobAggregatedLogs
Type: object
Show Source
Nested Schema : JobAggregatedStoredLogs
Type: object
Show Source
Nested Schema : allocatedOCores
Type: object
Read Only: true
Sum of virtual cores allocated to the job???s running containers.
Nested Schema : applicationTags
Type: array
A collection of tags or labels associated with this job.
Show Source
Nested Schema : completed
Type: object
Read Only: true
Specifies whether the job has been completed. It can be used to determine whether the client needs to keep polling the -progress- status.
Nested Schema : connectors
Type: array
Connecting interfaces that this Job offers. An example of this connector will be the Spark UI.
Show Source
Nested Schema : containerLogs
Type: array
Location of log file(s) for a running job.
Show Source
Nested Schema : Error
Type: object
Show Source
Nested Schema : files
Type: array
Show Source
Nested Schema : JobLogFileReference
Type: object
Show Source
Nested Schema : files
Type: array
Show Source
Nested Schema : JobConnectorReference
Type: object
Show Source
Nested Schema : uris
Type: array
The URI of the linked resource
Show Source
Nested Schema : JobConnectorResourceIdentifier
Type: object
Show Source
Nested Schema : JobContainerLogs
Type: object
Show Source
Nested Schema : files
Type: array
Log files with path relative to the Container.
Show Source
Nested Schema : o:errorDetails
Type: array
Read Only: true
error details
Show Source

400 Response

List of errors related to the request.
Body ()
Root Schema : schema
Type: array
Show Source

404 Response

The Compute Job was not found.

500 Response

An internal error occurred.
Body ()
Root Schema : Error
Type: object
Show Source
Nested Schema : o:errorDetails
Type: array
Read Only: true
error details
Show Source
Back to Top