databricks-sdk 0.32.3__py3-none-any.whl → 0.34.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of databricks-sdk might be problematic. Click here for more details.
- databricks/sdk/__init__.py +10 -0
- databricks/sdk/_base_client.py +323 -0
- databricks/sdk/core.py +25 -292
- databricks/sdk/mixins/files.py +9 -9
- databricks/sdk/service/apps.py +366 -113
- databricks/sdk/service/catalog.py +269 -7
- databricks/sdk/service/compute.py +68 -18
- databricks/sdk/service/dashboards.py +34 -11
- databricks/sdk/service/jobs.py +104 -86
- databricks/sdk/service/pipelines.py +58 -1
- databricks/sdk/service/serving.py +326 -10
- databricks/sdk/service/settings.py +573 -1
- databricks/sdk/service/sharing.py +1 -1
- databricks/sdk/service/sql.py +10 -246
- databricks/sdk/service/workspace.py +268 -107
- databricks/sdk/version.py +1 -1
- {databricks_sdk-0.32.3.dist-info → databricks_sdk-0.34.0.dist-info}/METADATA +1 -1
- {databricks_sdk-0.32.3.dist-info → databricks_sdk-0.34.0.dist-info}/RECORD +22 -21
- {databricks_sdk-0.32.3.dist-info → databricks_sdk-0.34.0.dist-info}/LICENSE +0 -0
- {databricks_sdk-0.32.3.dist-info → databricks_sdk-0.34.0.dist-info}/NOTICE +0 -0
- {databricks_sdk-0.32.3.dist-info → databricks_sdk-0.34.0.dist-info}/WHEEL +0 -0
- {databricks_sdk-0.32.3.dist-info → databricks_sdk-0.34.0.dist-info}/top_level.txt +0 -0
|
@@ -31,7 +31,11 @@ class CreateDashboardRequest:
|
|
|
31
31
|
|
|
32
32
|
serialized_dashboard: Optional[str] = None
|
|
33
33
|
"""The contents of the dashboard in serialized string form. This field is excluded in List
|
|
34
|
-
Dashboards responses.
|
|
34
|
+
Dashboards responses. Use the [get dashboard API] to retrieve an example response, which
|
|
35
|
+
includes the `serialized_dashboard` field. This field provides the structure of the JSON string
|
|
36
|
+
that represents the dashboard's layout and components.
|
|
37
|
+
|
|
38
|
+
[get dashboard API]: https://docs.databricks.com/api/workspace/lakeview/get"""
|
|
35
39
|
|
|
36
40
|
warehouse_id: Optional[str] = None
|
|
37
41
|
"""The warehouse ID used to run the dashboard."""
|
|
@@ -165,12 +169,16 @@ class Dashboard:
|
|
|
165
169
|
trailing slash. This field is excluded in List Dashboards responses."""
|
|
166
170
|
|
|
167
171
|
path: Optional[str] = None
|
|
168
|
-
"""The workspace path of the dashboard asset, including the file name.
|
|
169
|
-
List Dashboards responses."""
|
|
172
|
+
"""The workspace path of the dashboard asset, including the file name. Exported dashboards always
|
|
173
|
+
have the file extension `.lvdash.json`. This field is excluded in List Dashboards responses."""
|
|
170
174
|
|
|
171
175
|
serialized_dashboard: Optional[str] = None
|
|
172
176
|
"""The contents of the dashboard in serialized string form. This field is excluded in List
|
|
173
|
-
Dashboards responses.
|
|
177
|
+
Dashboards responses. Use the [get dashboard API] to retrieve an example response, which
|
|
178
|
+
includes the `serialized_dashboard` field. This field provides the structure of the JSON string
|
|
179
|
+
that represents the dashboard's layout and components.
|
|
180
|
+
|
|
181
|
+
[get dashboard API]: https://docs.databricks.com/api/workspace/lakeview/get"""
|
|
174
182
|
|
|
175
183
|
update_time: Optional[str] = None
|
|
176
184
|
"""The timestamp of when the dashboard was last updated by the user. This field is excluded in List
|
|
@@ -382,8 +390,9 @@ class GenieMessage:
|
|
|
382
390
|
|
|
383
391
|
status: Optional[MessageStatus] = None
|
|
384
392
|
"""MesssageStatus. The possible values are: * `FETCHING_METADATA`: Fetching metadata from the data
|
|
385
|
-
sources. * `
|
|
386
|
-
`
|
|
393
|
+
sources. * `FILTERING_CONTEXT`: Running smart context step to determine relevant context. *
|
|
394
|
+
`ASKING_AI`: Waiting for the LLM to respond to the users question. * `EXECUTING_QUERY`:
|
|
395
|
+
Executing AI provided SQL query. Get the SQL query result by calling
|
|
387
396
|
[getMessageQueryResult](:method:genie/getMessageQueryResult) API. **Important: The message
|
|
388
397
|
status will stay in the `EXECUTING_QUERY` until a client calls
|
|
389
398
|
[getMessageQueryResult](:method:genie/getMessageQueryResult)**. * `FAILED`: Generating a
|
|
@@ -615,8 +624,9 @@ class MessageErrorType(Enum):
|
|
|
615
624
|
|
|
616
625
|
class MessageStatus(Enum):
|
|
617
626
|
"""MesssageStatus. The possible values are: * `FETCHING_METADATA`: Fetching metadata from the data
|
|
618
|
-
sources. * `
|
|
619
|
-
`
|
|
627
|
+
sources. * `FILTERING_CONTEXT`: Running smart context step to determine relevant context. *
|
|
628
|
+
`ASKING_AI`: Waiting for the LLM to respond to the users question. * `EXECUTING_QUERY`:
|
|
629
|
+
Executing AI provided SQL query. Get the SQL query result by calling
|
|
620
630
|
[getMessageQueryResult](:method:genie/getMessageQueryResult) API. **Important: The message
|
|
621
631
|
status will stay in the `EXECUTING_QUERY` until a client calls
|
|
622
632
|
[getMessageQueryResult](:method:genie/getMessageQueryResult)**. * `FAILED`: Generating a
|
|
@@ -632,6 +642,7 @@ class MessageStatus(Enum):
|
|
|
632
642
|
EXECUTING_QUERY = 'EXECUTING_QUERY'
|
|
633
643
|
FAILED = 'FAILED'
|
|
634
644
|
FETCHING_METADATA = 'FETCHING_METADATA'
|
|
645
|
+
FILTERING_CONTEXT = 'FILTERING_CONTEXT'
|
|
635
646
|
QUERY_RESULT_EXPIRED = 'QUERY_RESULT_EXPIRED'
|
|
636
647
|
SUBMITTED = 'SUBMITTED'
|
|
637
648
|
|
|
@@ -1028,7 +1039,11 @@ class UpdateDashboardRequest:
|
|
|
1028
1039
|
|
|
1029
1040
|
serialized_dashboard: Optional[str] = None
|
|
1030
1041
|
"""The contents of the dashboard in serialized string form. This field is excluded in List
|
|
1031
|
-
Dashboards responses.
|
|
1042
|
+
Dashboards responses. Use the [get dashboard API] to retrieve an example response, which
|
|
1043
|
+
includes the `serialized_dashboard` field. This field provides the structure of the JSON string
|
|
1044
|
+
that represents the dashboard's layout and components.
|
|
1045
|
+
|
|
1046
|
+
[get dashboard API]: https://docs.databricks.com/api/workspace/lakeview/get"""
|
|
1032
1047
|
|
|
1033
1048
|
warehouse_id: Optional[str] = None
|
|
1034
1049
|
"""The warehouse ID used to run the dashboard."""
|
|
@@ -1308,7 +1323,11 @@ class LakeviewAPI:
|
|
|
1308
1323
|
slash. This field is excluded in List Dashboards responses.
|
|
1309
1324
|
:param serialized_dashboard: str (optional)
|
|
1310
1325
|
The contents of the dashboard in serialized string form. This field is excluded in List Dashboards
|
|
1311
|
-
responses.
|
|
1326
|
+
responses. Use the [get dashboard API] to retrieve an example response, which includes the
|
|
1327
|
+
`serialized_dashboard` field. This field provides the structure of the JSON string that represents
|
|
1328
|
+
the dashboard's layout and components.
|
|
1329
|
+
|
|
1330
|
+
[get dashboard API]: https://docs.databricks.com/api/workspace/lakeview/get
|
|
1312
1331
|
:param warehouse_id: str (optional)
|
|
1313
1332
|
The warehouse ID used to run the dashboard.
|
|
1314
1333
|
|
|
@@ -1723,7 +1742,11 @@ class LakeviewAPI:
|
|
|
1723
1742
|
not been modified since the last read. This field is excluded in List Dashboards responses.
|
|
1724
1743
|
:param serialized_dashboard: str (optional)
|
|
1725
1744
|
The contents of the dashboard in serialized string form. This field is excluded in List Dashboards
|
|
1726
|
-
responses.
|
|
1745
|
+
responses. Use the [get dashboard API] to retrieve an example response, which includes the
|
|
1746
|
+
`serialized_dashboard` field. This field provides the structure of the JSON string that represents
|
|
1747
|
+
the dashboard's layout and components.
|
|
1748
|
+
|
|
1749
|
+
[get dashboard API]: https://docs.databricks.com/api/workspace/lakeview/get
|
|
1727
1750
|
:param warehouse_id: str (optional)
|
|
1728
1751
|
The warehouse ID used to run the dashboard.
|
|
1729
1752
|
|
databricks/sdk/service/jobs.py
CHANGED
|
@@ -505,7 +505,11 @@ class CreateJob:
|
|
|
505
505
|
well as when this job is deleted."""
|
|
506
506
|
|
|
507
507
|
environments: Optional[List[JobEnvironment]] = None
|
|
508
|
-
"""A list of task execution environment specifications that can be referenced by tasks
|
|
508
|
+
"""A list of task execution environment specifications that can be referenced by serverless tasks
|
|
509
|
+
of this job. An environment is required to be present for serverless tasks. For serverless
|
|
510
|
+
notebook tasks, the environment is accessible in the notebook environment panel. For other
|
|
511
|
+
serverless tasks, the task environment is required to be specified using environment_key in the
|
|
512
|
+
task settings."""
|
|
509
513
|
|
|
510
514
|
format: Optional[Format] = None
|
|
511
515
|
"""Used to tell what is the format of the job. This field is ignored in Create/Update/Reset calls.
|
|
@@ -553,12 +557,11 @@ class CreateJob:
|
|
|
553
557
|
"""The queue settings of the job."""
|
|
554
558
|
|
|
555
559
|
run_as: Optional[JobRunAs] = None
|
|
556
|
-
"""Write-only setting
|
|
557
|
-
|
|
558
|
-
created the job.
|
|
560
|
+
"""Write-only setting. Specifies the user, service principal or group that the job/pipeline runs
|
|
561
|
+
as. If not specified, the job/pipeline runs as the user who created the job/pipeline.
|
|
559
562
|
|
|
560
|
-
|
|
561
|
-
is thrown."""
|
|
563
|
+
Exactly one of `user_name`, `service_principal_name`, `group_name` should be specified. If not,
|
|
564
|
+
an error is thrown."""
|
|
562
565
|
|
|
563
566
|
schedule: Optional[CronSchedule] = None
|
|
564
567
|
"""An optional periodic schedule for this job. The default behavior is that the job only runs when
|
|
@@ -1462,7 +1465,8 @@ class JobEditMode(Enum):
|
|
|
1462
1465
|
@dataclass
|
|
1463
1466
|
class JobEmailNotifications:
|
|
1464
1467
|
no_alert_for_skipped_runs: Optional[bool] = None
|
|
1465
|
-
"""If true, do not send email to recipients specified in `on_failure` if the run is skipped.
|
|
1468
|
+
"""If true, do not send email to recipients specified in `on_failure` if the run is skipped. This
|
|
1469
|
+
field is `deprecated`. Please use the `notification_settings.no_alert_for_skipped_runs` field."""
|
|
1466
1470
|
|
|
1467
1471
|
on_duration_warning_threshold_exceeded: Optional[List[str]] = None
|
|
1468
1472
|
"""A list of email addresses to be notified when the duration of a run exceeds the threshold
|
|
@@ -1720,12 +1724,11 @@ class JobPermissionsRequest:
|
|
|
1720
1724
|
|
|
1721
1725
|
@dataclass
|
|
1722
1726
|
class JobRunAs:
|
|
1723
|
-
"""Write-only setting
|
|
1724
|
-
|
|
1725
|
-
created the job.
|
|
1727
|
+
"""Write-only setting. Specifies the user, service principal or group that the job/pipeline runs
|
|
1728
|
+
as. If not specified, the job/pipeline runs as the user who created the job/pipeline.
|
|
1726
1729
|
|
|
1727
|
-
|
|
1728
|
-
is thrown."""
|
|
1730
|
+
Exactly one of `user_name`, `service_principal_name`, `group_name` should be specified. If not,
|
|
1731
|
+
an error is thrown."""
|
|
1729
1732
|
|
|
1730
1733
|
service_principal_name: Optional[str] = None
|
|
1731
1734
|
"""Application ID of an active service principal. Setting this field requires the
|
|
@@ -1773,7 +1776,11 @@ class JobSettings:
|
|
|
1773
1776
|
well as when this job is deleted."""
|
|
1774
1777
|
|
|
1775
1778
|
environments: Optional[List[JobEnvironment]] = None
|
|
1776
|
-
"""A list of task execution environment specifications that can be referenced by tasks
|
|
1779
|
+
"""A list of task execution environment specifications that can be referenced by serverless tasks
|
|
1780
|
+
of this job. An environment is required to be present for serverless tasks. For serverless
|
|
1781
|
+
notebook tasks, the environment is accessible in the notebook environment panel. For other
|
|
1782
|
+
serverless tasks, the task environment is required to be specified using environment_key in the
|
|
1783
|
+
task settings."""
|
|
1777
1784
|
|
|
1778
1785
|
format: Optional[Format] = None
|
|
1779
1786
|
"""Used to tell what is the format of the job. This field is ignored in Create/Update/Reset calls.
|
|
@@ -1821,12 +1828,11 @@ class JobSettings:
|
|
|
1821
1828
|
"""The queue settings of the job."""
|
|
1822
1829
|
|
|
1823
1830
|
run_as: Optional[JobRunAs] = None
|
|
1824
|
-
"""Write-only setting
|
|
1825
|
-
|
|
1826
|
-
created the job.
|
|
1831
|
+
"""Write-only setting. Specifies the user, service principal or group that the job/pipeline runs
|
|
1832
|
+
as. If not specified, the job/pipeline runs as the user who created the job/pipeline.
|
|
1827
1833
|
|
|
1828
|
-
|
|
1829
|
-
is thrown."""
|
|
1834
|
+
Exactly one of `user_name`, `service_principal_name`, `group_name` should be specified. If not,
|
|
1835
|
+
an error is thrown."""
|
|
1830
1836
|
|
|
1831
1837
|
schedule: Optional[CronSchedule] = None
|
|
1832
1838
|
"""An optional periodic schedule for this job. The default behavior is that the job only runs when
|
|
@@ -2472,6 +2478,7 @@ class RepairRun:
|
|
|
2472
2478
|
[dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html"""
|
|
2473
2479
|
|
|
2474
2480
|
pipeline_params: Optional[PipelineParams] = None
|
|
2481
|
+
"""Controls whether the pipeline should perform a full refresh"""
|
|
2475
2482
|
|
|
2476
2483
|
python_named_params: Optional[Dict[str, str]] = None
|
|
2477
2484
|
|
|
@@ -3175,6 +3182,7 @@ class RunJobTask:
|
|
|
3175
3182
|
[dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html"""
|
|
3176
3183
|
|
|
3177
3184
|
pipeline_params: Optional[PipelineParams] = None
|
|
3185
|
+
"""Controls whether the pipeline should perform a full refresh"""
|
|
3178
3186
|
|
|
3179
3187
|
python_named_params: Optional[Dict[str, str]] = None
|
|
3180
3188
|
|
|
@@ -3334,6 +3342,7 @@ class RunNow:
|
|
|
3334
3342
|
[dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html"""
|
|
3335
3343
|
|
|
3336
3344
|
pipeline_params: Optional[PipelineParams] = None
|
|
3345
|
+
"""Controls whether the pipeline should perform a full refresh"""
|
|
3337
3346
|
|
|
3338
3347
|
python_named_params: Optional[Dict[str, str]] = None
|
|
3339
3348
|
|
|
@@ -3543,6 +3552,7 @@ class RunParameters:
|
|
|
3543
3552
|
[dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html"""
|
|
3544
3553
|
|
|
3545
3554
|
pipeline_params: Optional[PipelineParams] = None
|
|
3555
|
+
"""Controls whether the pipeline should perform a full refresh"""
|
|
3546
3556
|
|
|
3547
3557
|
python_named_params: Optional[Dict[str, str]] = None
|
|
3548
3558
|
|
|
@@ -3617,9 +3627,11 @@ class RunResultState(Enum):
|
|
|
3617
3627
|
reached. * `EXCLUDED`: The run was skipped because the necessary conditions were not met. *
|
|
3618
3628
|
`SUCCESS_WITH_FAILURES`: The job run completed successfully with some failures; leaf tasks were
|
|
3619
3629
|
successful. * `UPSTREAM_FAILED`: The run was skipped because of an upstream failure. *
|
|
3620
|
-
`UPSTREAM_CANCELED`: The run was skipped because an upstream task was canceled.
|
|
3630
|
+
`UPSTREAM_CANCELED`: The run was skipped because an upstream task was canceled. * `DISABLED`:
|
|
3631
|
+
The run was skipped because it was disabled explicitly by the user."""
|
|
3621
3632
|
|
|
3622
3633
|
CANCELED = 'CANCELED'
|
|
3634
|
+
DISABLED = 'DISABLED'
|
|
3623
3635
|
EXCLUDED = 'EXCLUDED'
|
|
3624
3636
|
FAILED = 'FAILED'
|
|
3625
3637
|
MAXIMUM_CONCURRENT_RUNS_REACHED = 'MAXIMUM_CONCURRENT_RUNS_REACHED'
|
|
@@ -5034,7 +5046,8 @@ class TaskDependency:
|
|
|
5034
5046
|
@dataclass
|
|
5035
5047
|
class TaskEmailNotifications:
|
|
5036
5048
|
no_alert_for_skipped_runs: Optional[bool] = None
|
|
5037
|
-
"""If true, do not send email to recipients specified in `on_failure` if the run is skipped.
|
|
5049
|
+
"""If true, do not send email to recipients specified in `on_failure` if the run is skipped. This
|
|
5050
|
+
field is `deprecated`. Please use the `notification_settings.no_alert_for_skipped_runs` field."""
|
|
5038
5051
|
|
|
5039
5052
|
on_duration_warning_threshold_exceeded: Optional[List[str]] = None
|
|
5040
5053
|
"""A list of email addresses to be notified when the duration of a run exceeds the threshold
|
|
@@ -5128,36 +5141,36 @@ class TaskNotificationSettings:
|
|
|
5128
5141
|
|
|
5129
5142
|
class TerminationCodeCode(Enum):
|
|
5130
5143
|
"""The code indicates why the run was terminated. Additional codes might be introduced in future
|
|
5131
|
-
releases. * `SUCCESS`: The run was completed successfully. * `
|
|
5132
|
-
during execution by
|
|
5133
|
-
|
|
5134
|
-
|
|
5135
|
-
|
|
5136
|
-
|
|
5137
|
-
|
|
5138
|
-
|
|
5139
|
-
|
|
5140
|
-
|
|
5141
|
-
|
|
5142
|
-
|
|
5143
|
-
`FEATURE_DISABLED`: The run failed because it
|
|
5144
|
-
workspace. * `CLUSTER_REQUEST_LIMIT_EXCEEDED`: The
|
|
5145
|
-
requests have exceeded the allotted rate limit.
|
|
5146
|
-
larger time frame. * `STORAGE_ACCESS_ERROR`: The run
|
|
5147
|
-
customer blob storage. Refer to the state message for
|
|
5148
|
-
The run was completed with task failures. For more
|
|
5149
|
-
output. * `UNAUTHORIZED_ERROR`: The run failed due to
|
|
5150
|
-
resource. Refer to the state message for further details. *
|
|
5151
|
-
run failed while installing the user-requested library. Refer
|
|
5152
|
-
details. The causes might include, but are not limited to: The
|
|
5153
|
-
there are insufficient permissions to install the library, and so
|
|
5154
|
-
`MAX_CONCURRENT_RUNS_EXCEEDED`: The scheduled run exceeds the limit of maximum
|
|
5155
|
-
set for the job. * `MAX_SPARK_CONTEXTS_EXCEEDED`: The run is scheduled on a
|
|
5156
|
-
already reached the maximum number of contexts it is configured to create. See:
|
|
5157
|
-
`RESOURCE_NOT_FOUND`: A resource necessary for run execution does not exist. Refer to
|
|
5158
|
-
message for further details. * `INVALID_RUN_CONFIGURATION`: The run failed due to an
|
|
5159
|
-
configuration. Refer to the state message for further details. * `CLOUD_FAILURE`: The
|
|
5160
|
-
due to a cloud provider issue. Refer to the state message for further details. *
|
|
5144
|
+
releases. * `SUCCESS`: The run was completed successfully. * `USER_CANCELED`: The run was
|
|
5145
|
+
successfully canceled during execution by a user. * `CANCELED`: The run was canceled during
|
|
5146
|
+
execution by the Databricks platform; for example, if the maximum run duration was exceeded. *
|
|
5147
|
+
`SKIPPED`: Run was never executed, for example, if the upstream task run failed, the dependency
|
|
5148
|
+
type condition was not met, or there were no material tasks to execute. * `INTERNAL_ERROR`: The
|
|
5149
|
+
run encountered an unexpected error. Refer to the state message for further details. *
|
|
5150
|
+
`DRIVER_ERROR`: The run encountered an error while communicating with the Spark Driver. *
|
|
5151
|
+
`CLUSTER_ERROR`: The run failed due to a cluster error. Refer to the state message for further
|
|
5152
|
+
details. * `REPOSITORY_CHECKOUT_FAILED`: Failed to complete the checkout due to an error when
|
|
5153
|
+
communicating with the third party service. * `INVALID_CLUSTER_REQUEST`: The run failed because
|
|
5154
|
+
it issued an invalid request to start the cluster. * `WORKSPACE_RUN_LIMIT_EXCEEDED`: The
|
|
5155
|
+
workspace has reached the quota for the maximum number of concurrent active runs. Consider
|
|
5156
|
+
scheduling the runs over a larger time frame. * `FEATURE_DISABLED`: The run failed because it
|
|
5157
|
+
tried to access a feature unavailable for the workspace. * `CLUSTER_REQUEST_LIMIT_EXCEEDED`: The
|
|
5158
|
+
number of cluster creation, start, and upsize requests have exceeded the allotted rate limit.
|
|
5159
|
+
Consider spreading the run execution over a larger time frame. * `STORAGE_ACCESS_ERROR`: The run
|
|
5160
|
+
failed due to an error when accessing the customer blob storage. Refer to the state message for
|
|
5161
|
+
further details. * `RUN_EXECUTION_ERROR`: The run was completed with task failures. For more
|
|
5162
|
+
details, refer to the state message or run output. * `UNAUTHORIZED_ERROR`: The run failed due to
|
|
5163
|
+
a permission issue while accessing a resource. Refer to the state message for further details. *
|
|
5164
|
+
`LIBRARY_INSTALLATION_ERROR`: The run failed while installing the user-requested library. Refer
|
|
5165
|
+
to the state message for further details. The causes might include, but are not limited to: The
|
|
5166
|
+
provided library is invalid, there are insufficient permissions to install the library, and so
|
|
5167
|
+
forth. * `MAX_CONCURRENT_RUNS_EXCEEDED`: The scheduled run exceeds the limit of maximum
|
|
5168
|
+
concurrent runs set for the job. * `MAX_SPARK_CONTEXTS_EXCEEDED`: The run is scheduled on a
|
|
5169
|
+
cluster that has already reached the maximum number of contexts it is configured to create. See:
|
|
5170
|
+
[Link]. * `RESOURCE_NOT_FOUND`: A resource necessary for run execution does not exist. Refer to
|
|
5171
|
+
the state message for further details. * `INVALID_RUN_CONFIGURATION`: The run failed due to an
|
|
5172
|
+
invalid configuration. Refer to the state message for further details. * `CLOUD_FAILURE`: The
|
|
5173
|
+
run failed due to a cloud provider issue. Refer to the state message for further details. *
|
|
5161
5174
|
`MAX_JOB_QUEUE_SIZE_EXCEEDED`: The run was skipped due to reaching the job level queue size
|
|
5162
5175
|
limit.
|
|
5163
5176
|
|
|
@@ -5183,6 +5196,7 @@ class TerminationCodeCode(Enum):
|
|
|
5183
5196
|
STORAGE_ACCESS_ERROR = 'STORAGE_ACCESS_ERROR'
|
|
5184
5197
|
SUCCESS = 'SUCCESS'
|
|
5185
5198
|
UNAUTHORIZED_ERROR = 'UNAUTHORIZED_ERROR'
|
|
5199
|
+
USER_CANCELED = 'USER_CANCELED'
|
|
5186
5200
|
WORKSPACE_RUN_LIMIT_EXCEEDED = 'WORKSPACE_RUN_LIMIT_EXCEEDED'
|
|
5187
5201
|
|
|
5188
5202
|
|
|
@@ -5190,36 +5204,36 @@ class TerminationCodeCode(Enum):
|
|
|
5190
5204
|
class TerminationDetails:
|
|
5191
5205
|
code: Optional[TerminationCodeCode] = None
|
|
5192
5206
|
"""The code indicates why the run was terminated. Additional codes might be introduced in future
|
|
5193
|
-
releases. * `SUCCESS`: The run was completed successfully. * `
|
|
5194
|
-
during execution by
|
|
5195
|
-
|
|
5196
|
-
|
|
5197
|
-
|
|
5198
|
-
|
|
5199
|
-
|
|
5200
|
-
|
|
5201
|
-
|
|
5202
|
-
|
|
5203
|
-
|
|
5204
|
-
|
|
5205
|
-
`FEATURE_DISABLED`: The run failed because it
|
|
5206
|
-
workspace. * `CLUSTER_REQUEST_LIMIT_EXCEEDED`: The
|
|
5207
|
-
requests have exceeded the allotted rate limit.
|
|
5208
|
-
larger time frame. * `STORAGE_ACCESS_ERROR`: The run
|
|
5209
|
-
customer blob storage. Refer to the state message for
|
|
5210
|
-
The run was completed with task failures. For more
|
|
5211
|
-
output. * `UNAUTHORIZED_ERROR`: The run failed due to
|
|
5212
|
-
resource. Refer to the state message for further details. *
|
|
5213
|
-
run failed while installing the user-requested library. Refer
|
|
5214
|
-
details. The causes might include, but are not limited to: The
|
|
5215
|
-
there are insufficient permissions to install the library, and so
|
|
5216
|
-
`MAX_CONCURRENT_RUNS_EXCEEDED`: The scheduled run exceeds the limit of maximum
|
|
5217
|
-
set for the job. * `MAX_SPARK_CONTEXTS_EXCEEDED`: The run is scheduled on a
|
|
5218
|
-
already reached the maximum number of contexts it is configured to create. See:
|
|
5219
|
-
`RESOURCE_NOT_FOUND`: A resource necessary for run execution does not exist. Refer to
|
|
5220
|
-
message for further details. * `INVALID_RUN_CONFIGURATION`: The run failed due to an
|
|
5221
|
-
configuration. Refer to the state message for further details. * `CLOUD_FAILURE`: The
|
|
5222
|
-
due to a cloud provider issue. Refer to the state message for further details. *
|
|
5207
|
+
releases. * `SUCCESS`: The run was completed successfully. * `USER_CANCELED`: The run was
|
|
5208
|
+
successfully canceled during execution by a user. * `CANCELED`: The run was canceled during
|
|
5209
|
+
execution by the Databricks platform; for example, if the maximum run duration was exceeded. *
|
|
5210
|
+
`SKIPPED`: Run was never executed, for example, if the upstream task run failed, the dependency
|
|
5211
|
+
type condition was not met, or there were no material tasks to execute. * `INTERNAL_ERROR`: The
|
|
5212
|
+
run encountered an unexpected error. Refer to the state message for further details. *
|
|
5213
|
+
`DRIVER_ERROR`: The run encountered an error while communicating with the Spark Driver. *
|
|
5214
|
+
`CLUSTER_ERROR`: The run failed due to a cluster error. Refer to the state message for further
|
|
5215
|
+
details. * `REPOSITORY_CHECKOUT_FAILED`: Failed to complete the checkout due to an error when
|
|
5216
|
+
communicating with the third party service. * `INVALID_CLUSTER_REQUEST`: The run failed because
|
|
5217
|
+
it issued an invalid request to start the cluster. * `WORKSPACE_RUN_LIMIT_EXCEEDED`: The
|
|
5218
|
+
workspace has reached the quota for the maximum number of concurrent active runs. Consider
|
|
5219
|
+
scheduling the runs over a larger time frame. * `FEATURE_DISABLED`: The run failed because it
|
|
5220
|
+
tried to access a feature unavailable for the workspace. * `CLUSTER_REQUEST_LIMIT_EXCEEDED`: The
|
|
5221
|
+
number of cluster creation, start, and upsize requests have exceeded the allotted rate limit.
|
|
5222
|
+
Consider spreading the run execution over a larger time frame. * `STORAGE_ACCESS_ERROR`: The run
|
|
5223
|
+
failed due to an error when accessing the customer blob storage. Refer to the state message for
|
|
5224
|
+
further details. * `RUN_EXECUTION_ERROR`: The run was completed with task failures. For more
|
|
5225
|
+
details, refer to the state message or run output. * `UNAUTHORIZED_ERROR`: The run failed due to
|
|
5226
|
+
a permission issue while accessing a resource. Refer to the state message for further details. *
|
|
5227
|
+
`LIBRARY_INSTALLATION_ERROR`: The run failed while installing the user-requested library. Refer
|
|
5228
|
+
to the state message for further details. The causes might include, but are not limited to: The
|
|
5229
|
+
provided library is invalid, there are insufficient permissions to install the library, and so
|
|
5230
|
+
forth. * `MAX_CONCURRENT_RUNS_EXCEEDED`: The scheduled run exceeds the limit of maximum
|
|
5231
|
+
concurrent runs set for the job. * `MAX_SPARK_CONTEXTS_EXCEEDED`: The run is scheduled on a
|
|
5232
|
+
cluster that has already reached the maximum number of contexts it is configured to create. See:
|
|
5233
|
+
[Link]. * `RESOURCE_NOT_FOUND`: A resource necessary for run execution does not exist. Refer to
|
|
5234
|
+
the state message for further details. * `INVALID_RUN_CONFIGURATION`: The run failed due to an
|
|
5235
|
+
invalid configuration. Refer to the state message for further details. * `CLOUD_FAILURE`: The
|
|
5236
|
+
run failed due to a cloud provider issue. Refer to the state message for further details. *
|
|
5223
5237
|
`MAX_JOB_QUEUE_SIZE_EXCEEDED`: The run was skipped due to reaching the job level queue size
|
|
5224
5238
|
limit.
|
|
5225
5239
|
|
|
@@ -5649,7 +5663,10 @@ class JobsAPI:
|
|
|
5649
5663
|
An optional set of email addresses that is notified when runs of this job begin or complete as well
|
|
5650
5664
|
as when this job is deleted.
|
|
5651
5665
|
:param environments: List[:class:`JobEnvironment`] (optional)
|
|
5652
|
-
A list of task execution environment specifications that can be referenced by tasks of
|
|
5666
|
+
A list of task execution environment specifications that can be referenced by serverless tasks of
|
|
5667
|
+
this job. An environment is required to be present for serverless tasks. For serverless notebook
|
|
5668
|
+
tasks, the environment is accessible in the notebook environment panel. For other serverless tasks,
|
|
5669
|
+
the task environment is required to be specified using environment_key in the task settings.
|
|
5653
5670
|
:param format: :class:`Format` (optional)
|
|
5654
5671
|
Used to tell what is the format of the job. This field is ignored in Create/Update/Reset calls. When
|
|
5655
5672
|
using the Jobs API 2.1 this value is always set to `"MULTI_TASK"`.
|
|
@@ -5686,12 +5703,11 @@ class JobsAPI:
|
|
|
5686
5703
|
:param queue: :class:`QueueSettings` (optional)
|
|
5687
5704
|
The queue settings of the job.
|
|
5688
5705
|
:param run_as: :class:`JobRunAs` (optional)
|
|
5689
|
-
Write-only setting
|
|
5690
|
-
|
|
5691
|
-
job.
|
|
5706
|
+
Write-only setting. Specifies the user, service principal or group that the job/pipeline runs as. If
|
|
5707
|
+
not specified, the job/pipeline runs as the user who created the job/pipeline.
|
|
5692
5708
|
|
|
5693
|
-
|
|
5694
|
-
thrown.
|
|
5709
|
+
Exactly one of `user_name`, `service_principal_name`, `group_name` should be specified. If not, an
|
|
5710
|
+
error is thrown.
|
|
5695
5711
|
:param schedule: :class:`CronSchedule` (optional)
|
|
5696
5712
|
An optional periodic schedule for this job. The default behavior is that the job only runs when
|
|
5697
5713
|
triggered by clicking “Run Now” in the Jobs UI or sending an API request to `runNow`.
|
|
@@ -6075,6 +6091,7 @@ class JobsAPI:
|
|
|
6075
6091
|
[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables
|
|
6076
6092
|
[dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html
|
|
6077
6093
|
:param pipeline_params: :class:`PipelineParams` (optional)
|
|
6094
|
+
Controls whether the pipeline should perform a full refresh
|
|
6078
6095
|
:param python_named_params: Dict[str,str] (optional)
|
|
6079
6096
|
:param python_params: List[str] (optional)
|
|
6080
6097
|
A list of parameters for jobs with Python tasks, for example `"python_params": ["john doe", "35"]`.
|
|
@@ -6264,6 +6281,7 @@ class JobsAPI:
|
|
|
6264
6281
|
[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables
|
|
6265
6282
|
[dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html
|
|
6266
6283
|
:param pipeline_params: :class:`PipelineParams` (optional)
|
|
6284
|
+
Controls whether the pipeline should perform a full refresh
|
|
6267
6285
|
:param python_named_params: Dict[str,str] (optional)
|
|
6268
6286
|
:param python_params: List[str] (optional)
|
|
6269
6287
|
A list of parameters for jobs with Python tasks, for example `"python_params": ["john doe", "35"]`.
|