
CreateJob
Create a Databricks job and run it.
Create a Databricks job and run it.
Create a Databricks job and run it.
Set waitForCompletion to the desired maximum duration if you want the task to wait for the job completion (e.g., PT1H to wait up to one hour).
type: "io.kestra.plugin.databricks.job.CreateJob"Examples
Create a Databricks job, run it, and wait for completion for five minutes.
id: databricks_job_create
namespace: company.team
tasks:
- id: create_job
type: io.kestra.plugin.databricks.job.CreateJob
authentication:
token: "{{ secret('DATABRICKS_TOKEN') }}"
host: "{{ secret('DATABRICKS_HOST') }}"
jobTasks:
- existingClusterId: <your-cluster>
taskKey: taskKey
sparkPythonTask:
pythonFile: /Shared/hello.py
sparkPythonTaskSource: WORKSPACE
waitForCompletion: PT5M
Properties
jobTasks*Requiredarray
1The job tasks, if multiple tasks are defined you must set dependsOn on each task.
io.kestra.plugin.databricks.job.CreateJob-JobTaskSetting
DBT task settings.
io.kestra.plugin.databricks.job.task.DbtTaskSetting
Task dependencies, set this if multiple tasks are defined on the job.
Task description.
The identifier of the cluster.
Task libraries.
io.kestra.plugin.databricks.job.task.LibrarySetting
io.kestra.plugin.databricks.job.task.LibrarySetting-CranSetting
io.kestra.plugin.databricks.job.task.LibrarySetting-MavenSetting
io.kestra.plugin.databricks.job.task.LibrarySetting-PypiSetting
Notebook task settings.
io.kestra.plugin.databricks.job.task.NotebookTaskSetting
Map of task base parameters.
GITWORKSPACEPipeline task settings.
io.kestra.plugin.databricks.job.task.PipelineTaskSetting
Python Wheel task settings.
io.kestra.plugin.databricks.job.task.PythonWheelTaskSetting
Map of task named parameters.
Run job task settings.
io.kestra.plugin.databricks.job.task.RunJobTaskSetting
Spark JAR task settings.
io.kestra.plugin.databricks.job.task.SparkJarTaskSetting
Spark Python task settings.
io.kestra.plugin.databricks.job.task.SparkPythonTaskSetting
GITWORKSPACESpark Submit task settings.
io.kestra.plugin.databricks.job.task.SparkSubmitTaskSetting
List of task parameters.
Can be a list of strings or a variable that binds to a JSON array of strings.
SQL task settings.
io.kestra.plugin.databricks.job.task.SqlTaskSetting
Map of task parameters.
Can be a map of string/string or a variable that binds to a JSON object.
Task key.
Task timeout in seconds.
accountIdstring
Databricks account identifier.
authentication
Databricks authentication configuration.
This property allows to configure the authentication to Databricks, different properties should be set depending on the type of authentication and the cloud provider. All configuration options can also be set using the standard Databricks environment variables. Check the Databricks authentication guide for more information.
io.kestra.plugin.databricks.AbstractTask-AuthenticationConfig
configFilestring
Databricks configuration file, use this if you don't want to configure each Databricks account properties one by one.
hoststring
Databricks host.
jobNamestring
The name of the job.
waitForCompletionstring
durationIf set, the task will wait for the job run completion for up to the waitForCompletion duration before timing out.
Outputs
jobIdinteger
The job identifier.
jobURIstring
uriThe job URI on the Databricks console.
runIdinteger
The run identifier.
runURIstring
uriThe run URI on the Databricks console.