CreateJobCreateJob
CreateJobCertified

Create a Databricks job and run it.

Create a Databricks job and run it.

Set waitForCompletion to the desired maximum duration if you want the task to wait for the job completion (e.g., PT1H to wait up to one hour).

yaml
type: "io.kestra.plugin.databricks.job.CreateJob"

Create a Databricks job, run it, and wait for completion for five minutes.

yaml
id: databricks_job_create
namespace: company.team

tasks:
  - id: create_job
    type: io.kestra.plugin.databricks.job.CreateJob
    authentication:
      token: "{{ secret('DATABRICKS_TOKEN') }}"
    host: "{{ secret('DATABRICKS_HOST') }}"
    jobTasks:
      - existingClusterId: <your-cluster>
        taskKey: taskKey
        sparkPythonTask:
          pythonFile: /Shared/hello.py
          sparkPythonTaskSource: WORKSPACE
    waitForCompletion: PT5M
Properties
Min items1

The job tasks, if multiple tasks are defined you must set dependsOn on each task.

Definitions
dbtTask

DBT task settings.

catalogstring
commandsarray
SubTypestring
schemastring
warehouseIdstring
dependsOnarray
SubTypestring

Task dependencies, set this if multiple tasks are defined on the job.

descriptionstring

Task description.

existingClusterIdstring

The identifier of the cluster.

librariesarray

Task libraries.

cran
_packagestring
repostring
eggstring
jarstring
maven
coordinatesstring
exclusionsarray
SubTypestring
repostring
pypi
_packagestring
repostring
whlstring
notebookTask

Notebook task settings.

baseParametersstringobject
SubTypestring

Map of task base parameters.

notebookPathstring
sourcestring
Possible Values
GITWORKSPACE
pipelineTask

Pipeline task settings.

fullRefreshbooleanstring
pipelineIdstring
pythonWheelTask

Python Wheel task settings.

entryPointstring
namedParametersstringobject
SubTypestring

Map of task named parameters.

packageNamestring
parametersstringarray
runJobTask

Run job task settings.

jobIdstring
jobParametersobject
sparkJarTask

Spark JAR task settings.

jarUristring
mainClassNamestring
parametersstringarray
sparkPythonTask

Spark Python task settings.

pythonFile*Requiredstring
sparkPythonTaskSource*Requiredstring
Possible Values
GITWORKSPACE
parametersstringarray
sparkSubmitTask

Spark Submit task settings.

parametersstringarray

List of task parameters.

Can be a list of strings or a variable that binds to a JSON array of strings.

sqlTask

SQL task settings.

parametersstringobject
SubTypestring

Map of task parameters.

Can be a map of string/string or a variable that binds to a JSON object.

queryIdstring
warehouseIdstring
taskKeystring

Task key.

timeoutSecondsintegerstring

Task timeout in seconds.

Databricks account identifier.

Databricks authentication configuration.

This property allows to configure the authentication to Databricks, different properties should be set depending on the type of authentication and the cloud provider. All configuration options can also be set using the standard Databricks environment variables. Check the Databricks authentication guide for more information.

Definitions
authTypestring
azureClientIdstring
azureClientSecretstring
azureTenantIdstring
clientIdstring
clientSecretstring
googleCredentialsstring
googleServiceAccountstring
passwordstring
tokenstring
usernamestring

Databricks configuration file, use this if you don't want to configure each Databricks account properties one by one.

Databricks host.

The name of the job.

Formatduration

If set, the task will wait for the job run completion for up to the waitForCompletion duration before timing out.

The job identifier.

Formaturi

The job URI on the Databricks console.

The run identifier.

Formaturi

The run URI on the Databricks console.

Create automations with Databricks JOB CreateJob