CRD reference
Below are listed the CRD fields that can be defined by the user:
CRD field | Remarks |
---|---|
|
|
|
|
|
Application name |
|
Application version |
|
|
|
User-supplied image containing spark-job dependencies that will be copied to the specified volume mount |
|
Spark image which will be deployed to driver and executor pods, which must contain spark environment needed by the job e.g. |
|
Optional Enum (one of |
|
An optional list of references to secrets in the same namespace to use for pulling any of the images used by a |
|
The actual application file that will be called by |
|
The main class i.e. entry point for JVM artifacts |
|
Arguments passed directly to the job artifact |
|
S3 connection specification. See the S3 resources for more details. |
|
A map of key/value strings that will be passed directly to |
|
A list of python packages that will be installed via |
|
A list of packages that is passed directly to |
|
A list of excluded packages that is passed directly to |
|
A list of repositories that is passed directly to |
|
A list of volumes |
|
The volume name |
|
The persistent volume claim backing the volume |
|
Resources specification for the initiating Job |
|
Resources specification for the driver Pod |
|
A list of mounted volumes for the driver |
|
Name of mount |
|
Volume mount path |
|
Driver Pod placement affinity. See Pod Placement for details |
|
Logging aggregation for the driver Pod. See Logging for details |
|
Resources specification for the executor Pods |
|
Number of executor instances launched for this job |
|
A list of mounted volumes for each executor |
|
Name of mount |
|
Volume mount path |
|
Driver Pod placement affinity. See Pod Placement for details. |
|
Logging aggregation for the executor Pods. See Logging for details |
|
S3 bucket definition where applications should publish events for the Spark History server. |
|
Prefix to use when storing events for the Spark History server. |