Azure Databricks · Schema

TaskSettings

AnalyticsApache SparkBig DataData EngineeringMachine Learning

Properties

Name Type Description
task_key string Unique key for the task within the job. Must be unique among all tasks in the same job.
description string Optional description for the task
depends_on array Tasks this task depends on. The task will not start until all dependent tasks complete successfully.
existing_cluster_id string ID of an existing cluster to use for this task
job_cluster_key string Key of a job cluster specification to use. Refers to a cluster defined in the job's job_clusters field.
notebook_task object
spark_jar_task object
spark_python_task object
spark_submit_task object
pipeline_task object
python_wheel_task object
sql_task object
dbt_task object
run_if string Condition for running this task based on dependencies
timeout_seconds integer Timeout in seconds for this individual task
max_retries integer Maximum number of retries for this task
min_retry_interval_millis integer Minimum interval between retries in milliseconds
retry_on_timeout boolean Whether to retry the task if it times out
libraries array Libraries to install on the task cluster
View JSON Schema on GitHub