databricks-sdk 0.70.0__py3-none-any.whl → 0.72.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of databricks-sdk might be problematic. Click here for more details.
- databricks/sdk/__init__.py +25 -25
- databricks/sdk/mixins/files.py +51 -15
- databricks/sdk/service/agentbricks.py +2 -0
- databricks/sdk/service/apps.py +10 -0
- databricks/sdk/service/billing.py +13 -3
- databricks/sdk/service/catalog.py +149 -46
- databricks/sdk/service/cleanrooms.py +11 -3
- databricks/sdk/service/compute.py +55 -0
- databricks/sdk/service/dashboards.py +11 -0
- databricks/sdk/service/database.py +12 -0
- databricks/sdk/service/dataquality.py +4 -0
- databricks/sdk/service/files.py +7 -72
- databricks/sdk/service/iam.py +26 -36
- databricks/sdk/service/iamv2.py +6 -0
- databricks/sdk/service/jobs.py +22 -122
- databricks/sdk/service/marketplace.py +18 -0
- databricks/sdk/service/ml.py +321 -17
- databricks/sdk/service/oauth2.py +10 -18
- databricks/sdk/service/pipelines.py +44 -13
- databricks/sdk/service/provisioning.py +9 -0
- databricks/sdk/service/qualitymonitorv2.py +2 -0
- databricks/sdk/service/serving.py +16 -21
- databricks/sdk/service/settings.py +43 -72
- databricks/sdk/service/settingsv2.py +2 -0
- databricks/sdk/service/sharing.py +23 -31
- databricks/sdk/service/sql.py +48 -24
- databricks/sdk/service/tags.py +2 -0
- databricks/sdk/service/vectorsearch.py +11 -1
- databricks/sdk/service/workspace.py +18 -91
- databricks/sdk/version.py +1 -1
- {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/METADATA +1 -1
- {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/RECORD +36 -36
- {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/WHEEL +0 -0
- {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/licenses/LICENSE +0 -0
- {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/licenses/NOTICE +0 -0
- {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/top_level.txt +0 -0
databricks/sdk/service/oauth2.py
CHANGED
|
@@ -194,24 +194,6 @@ class DeletePublishedAppIntegrationOutput:
|
|
|
194
194
|
return cls()
|
|
195
195
|
|
|
196
196
|
|
|
197
|
-
@dataclass
|
|
198
|
-
class DeleteResponse:
|
|
199
|
-
def as_dict(self) -> dict:
|
|
200
|
-
"""Serializes the DeleteResponse into a dictionary suitable for use as a JSON request body."""
|
|
201
|
-
body = {}
|
|
202
|
-
return body
|
|
203
|
-
|
|
204
|
-
def as_shallow_dict(self) -> dict:
|
|
205
|
-
"""Serializes the DeleteResponse into a shallow dictionary of its immediate attributes."""
|
|
206
|
-
body = {}
|
|
207
|
-
return body
|
|
208
|
-
|
|
209
|
-
@classmethod
|
|
210
|
-
def from_dict(cls, d: Dict[str, Any]) -> DeleteResponse:
|
|
211
|
-
"""Deserializes the DeleteResponse from a dictionary."""
|
|
212
|
-
return cls()
|
|
213
|
-
|
|
214
|
-
|
|
215
197
|
@dataclass
|
|
216
198
|
class FederationPolicy:
|
|
217
199
|
create_time: Optional[str] = None
|
|
@@ -1013,6 +995,7 @@ class AccountFederationPolicyAPI:
|
|
|
1013
995
|
|
|
1014
996
|
:returns: :class:`FederationPolicy`
|
|
1015
997
|
"""
|
|
998
|
+
|
|
1016
999
|
body = policy.as_dict()
|
|
1017
1000
|
query = {}
|
|
1018
1001
|
if policy_id is not None:
|
|
@@ -1112,6 +1095,7 @@ class AccountFederationPolicyAPI:
|
|
|
1112
1095
|
|
|
1113
1096
|
:returns: :class:`FederationPolicy`
|
|
1114
1097
|
"""
|
|
1098
|
+
|
|
1115
1099
|
body = policy.as_dict()
|
|
1116
1100
|
query = {}
|
|
1117
1101
|
if update_mask is not None:
|
|
@@ -1169,6 +1153,7 @@ class CustomAppIntegrationAPI:
|
|
|
1169
1153
|
|
|
1170
1154
|
:returns: :class:`CreateCustomAppIntegrationOutput`
|
|
1171
1155
|
"""
|
|
1156
|
+
|
|
1172
1157
|
body = {}
|
|
1173
1158
|
if confidential is not None:
|
|
1174
1159
|
body["confidential"] = confidential
|
|
@@ -1301,6 +1286,7 @@ class CustomAppIntegrationAPI:
|
|
|
1301
1286
|
|
|
1302
1287
|
|
|
1303
1288
|
"""
|
|
1289
|
+
|
|
1304
1290
|
body = {}
|
|
1305
1291
|
if redirect_urls is not None:
|
|
1306
1292
|
body["redirect_urls"] = [v for v in redirect_urls]
|
|
@@ -1386,6 +1372,7 @@ class PublishedAppIntegrationAPI:
|
|
|
1386
1372
|
|
|
1387
1373
|
:returns: :class:`CreatePublishedAppIntegrationOutput`
|
|
1388
1374
|
"""
|
|
1375
|
+
|
|
1389
1376
|
body = {}
|
|
1390
1377
|
if app_id is not None:
|
|
1391
1378
|
body["app_id"] = app_id
|
|
@@ -1486,6 +1473,7 @@ class PublishedAppIntegrationAPI:
|
|
|
1486
1473
|
|
|
1487
1474
|
|
|
1488
1475
|
"""
|
|
1476
|
+
|
|
1489
1477
|
body = {}
|
|
1490
1478
|
if token_access_policy is not None:
|
|
1491
1479
|
body["token_access_policy"] = token_access_policy.as_dict()
|
|
@@ -1561,6 +1549,7 @@ class ServicePrincipalFederationPolicyAPI:
|
|
|
1561
1549
|
|
|
1562
1550
|
:returns: :class:`FederationPolicy`
|
|
1563
1551
|
"""
|
|
1552
|
+
|
|
1564
1553
|
body = policy.as_dict()
|
|
1565
1554
|
query = {}
|
|
1566
1555
|
if policy_id is not None:
|
|
@@ -1677,6 +1666,7 @@ class ServicePrincipalFederationPolicyAPI:
|
|
|
1677
1666
|
|
|
1678
1667
|
:returns: :class:`FederationPolicy`
|
|
1679
1668
|
"""
|
|
1669
|
+
|
|
1680
1670
|
body = policy.as_dict()
|
|
1681
1671
|
query = {}
|
|
1682
1672
|
if update_mask is not None:
|
|
@@ -1726,6 +1716,7 @@ class ServicePrincipalSecretsAPI:
|
|
|
1726
1716
|
|
|
1727
1717
|
:returns: :class:`CreateServicePrincipalSecretResponse`
|
|
1728
1718
|
"""
|
|
1719
|
+
|
|
1729
1720
|
body = {}
|
|
1730
1721
|
if lifetime is not None:
|
|
1731
1722
|
body["lifetime"] = lifetime
|
|
@@ -1836,6 +1827,7 @@ class ServicePrincipalSecretsProxyAPI:
|
|
|
1836
1827
|
|
|
1837
1828
|
:returns: :class:`CreateServicePrincipalSecretResponse`
|
|
1838
1829
|
"""
|
|
1830
|
+
|
|
1839
1831
|
body = {}
|
|
1840
1832
|
if lifetime is not None:
|
|
1841
1833
|
body["lifetime"] = lifetime
|
|
@@ -556,8 +556,8 @@ class IngestionGatewayPipelineDefinition:
|
|
|
556
556
|
|
|
557
557
|
gateway_storage_name: Optional[str] = None
|
|
558
558
|
"""Optional. The Unity Catalog-compatible name for the gateway storage location. This is the
|
|
559
|
-
destination to use for the data that is extracted by the gateway.
|
|
560
|
-
automatically create the storage location under the catalog and schema."""
|
|
559
|
+
destination to use for the data that is extracted by the gateway. Spark Declarative Pipelines
|
|
560
|
+
system will automatically create the storage location under the catalog and schema."""
|
|
561
561
|
|
|
562
562
|
def as_dict(self) -> dict:
|
|
563
563
|
"""Serializes the IngestionGatewayPipelineDefinition into a dictionary suitable for use as a JSON request body."""
|
|
@@ -832,6 +832,9 @@ class IngestionSourceType(Enum):
|
|
|
832
832
|
DYNAMICS365 = "DYNAMICS365"
|
|
833
833
|
FOREIGN_CATALOG = "FOREIGN_CATALOG"
|
|
834
834
|
GA4_RAW_DATA = "GA4_RAW_DATA"
|
|
835
|
+
GOOGLE_ADS = "GOOGLE_ADS"
|
|
836
|
+
GUIDEWIRE = "GUIDEWIRE"
|
|
837
|
+
HUBSPOT = "HUBSPOT"
|
|
835
838
|
MANAGED_POSTGRESQL = "MANAGED_POSTGRESQL"
|
|
836
839
|
META_MARKETING = "META_MARKETING"
|
|
837
840
|
MYSQL = "MYSQL"
|
|
@@ -840,12 +843,16 @@ class IngestionSourceType(Enum):
|
|
|
840
843
|
POSTGRESQL = "POSTGRESQL"
|
|
841
844
|
REDSHIFT = "REDSHIFT"
|
|
842
845
|
SALESFORCE = "SALESFORCE"
|
|
846
|
+
SALESFORCE_MARKETING_CLOUD = "SALESFORCE_MARKETING_CLOUD"
|
|
843
847
|
SERVICENOW = "SERVICENOW"
|
|
844
848
|
SHAREPOINT = "SHAREPOINT"
|
|
845
849
|
SQLDW = "SQLDW"
|
|
846
850
|
SQLSERVER = "SQLSERVER"
|
|
847
851
|
TERADATA = "TERADATA"
|
|
852
|
+
TIKTOK_ADS = "TIKTOK_ADS"
|
|
853
|
+
WORKDAY_HCM = "WORKDAY_HCM"
|
|
848
854
|
WORKDAY_RAAS = "WORKDAY_RAAS"
|
|
855
|
+
ZENDESK = "ZENDESK"
|
|
849
856
|
|
|
850
857
|
|
|
851
858
|
@dataclass
|
|
@@ -2024,6 +2031,9 @@ class PipelineSpec:
|
|
|
2024
2031
|
trigger: Optional[PipelineTrigger] = None
|
|
2025
2032
|
"""Which pipeline trigger to use. Deprecated: Use `continuous` instead."""
|
|
2026
2033
|
|
|
2034
|
+
usage_policy_id: Optional[str] = None
|
|
2035
|
+
"""Usage policy of this pipeline."""
|
|
2036
|
+
|
|
2027
2037
|
def as_dict(self) -> dict:
|
|
2028
2038
|
"""Serializes the PipelineSpec into a dictionary suitable for use as a JSON request body."""
|
|
2029
2039
|
body = {}
|
|
@@ -2081,6 +2091,8 @@ class PipelineSpec:
|
|
|
2081
2091
|
body["target"] = self.target
|
|
2082
2092
|
if self.trigger:
|
|
2083
2093
|
body["trigger"] = self.trigger.as_dict()
|
|
2094
|
+
if self.usage_policy_id is not None:
|
|
2095
|
+
body["usage_policy_id"] = self.usage_policy_id
|
|
2084
2096
|
return body
|
|
2085
2097
|
|
|
2086
2098
|
def as_shallow_dict(self) -> dict:
|
|
@@ -2140,6 +2152,8 @@ class PipelineSpec:
|
|
|
2140
2152
|
body["target"] = self.target
|
|
2141
2153
|
if self.trigger:
|
|
2142
2154
|
body["trigger"] = self.trigger
|
|
2155
|
+
if self.usage_policy_id is not None:
|
|
2156
|
+
body["usage_policy_id"] = self.usage_policy_id
|
|
2143
2157
|
return body
|
|
2144
2158
|
|
|
2145
2159
|
@classmethod
|
|
@@ -2173,6 +2187,7 @@ class PipelineSpec:
|
|
|
2173
2187
|
tags=d.get("tags", None),
|
|
2174
2188
|
target=d.get("target", None),
|
|
2175
2189
|
trigger=_from_dict(d, "trigger", PipelineTrigger),
|
|
2190
|
+
usage_policy_id=d.get("usage_policy_id", None),
|
|
2176
2191
|
)
|
|
2177
2192
|
|
|
2178
2193
|
|
|
@@ -2964,8 +2979,8 @@ class TableSpecificConfig:
|
|
|
2964
2979
|
"""The SCD type to use to ingest the table."""
|
|
2965
2980
|
|
|
2966
2981
|
sequence_by: Optional[List[str]] = None
|
|
2967
|
-
"""The column names specifying the logical order of events in the source data.
|
|
2968
|
-
uses this sequencing to handle change events that arrive out of order."""
|
|
2982
|
+
"""The column names specifying the logical order of events in the source data. Spark Declarative
|
|
2983
|
+
Pipelines uses this sequencing to handle change events that arrive out of order."""
|
|
2969
2984
|
|
|
2970
2985
|
workday_report_parameters: Optional[IngestionPipelineDefinitionWorkdayReportParameters] = None
|
|
2971
2986
|
"""(Optional) Additional custom parameters for Workday Report"""
|
|
@@ -3239,16 +3254,17 @@ class UpdateStateInfoState(Enum):
|
|
|
3239
3254
|
|
|
3240
3255
|
|
|
3241
3256
|
class PipelinesAPI:
|
|
3242
|
-
"""The
|
|
3257
|
+
"""The Lakeflow Spark Declarative Pipelines API allows you to create, edit, delete, start, and view details
|
|
3258
|
+
about pipelines.
|
|
3243
3259
|
|
|
3244
|
-
|
|
3245
|
-
pipelines. You define the transformations to perform on your data, and
|
|
3246
|
-
orchestration, cluster management, monitoring, data quality, and error handling.
|
|
3260
|
+
Spark Declarative Pipelines is a framework for building reliable, maintainable, and testable data
|
|
3261
|
+
processing pipelines. You define the transformations to perform on your data, and Spark Declarative
|
|
3262
|
+
Pipelines manages task orchestration, cluster management, monitoring, data quality, and error handling.
|
|
3247
3263
|
|
|
3248
|
-
Instead of defining your data pipelines using a series of separate Apache Spark tasks,
|
|
3249
|
-
manages how your data is transformed based on a target schema you define for each processing
|
|
3250
|
-
also enforce data quality with
|
|
3251
|
-
data quality and specify how to handle records that fail those expectations."""
|
|
3264
|
+
Instead of defining your data pipelines using a series of separate Apache Spark tasks, Spark Declarative
|
|
3265
|
+
Pipelines manages how your data is transformed based on a target schema you define for each processing
|
|
3266
|
+
step. You can also enforce data quality with Spark Declarative Pipelines expectations. Expectations allow
|
|
3267
|
+
you to define expected data quality and specify how to handle records that fail those expectations."""
|
|
3252
3268
|
|
|
3253
3269
|
def __init__(self, api_client):
|
|
3254
3270
|
self._api = api_client
|
|
@@ -3318,6 +3334,7 @@ class PipelinesAPI:
|
|
|
3318
3334
|
tags: Optional[Dict[str, str]] = None,
|
|
3319
3335
|
target: Optional[str] = None,
|
|
3320
3336
|
trigger: Optional[PipelineTrigger] = None,
|
|
3337
|
+
usage_policy_id: Optional[str] = None,
|
|
3321
3338
|
) -> CreatePipelineResponse:
|
|
3322
3339
|
"""Creates a new data processing pipeline based on the requested configuration. If successful, this
|
|
3323
3340
|
method returns the ID of the new pipeline.
|
|
@@ -3388,9 +3405,12 @@ class PipelinesAPI:
|
|
|
3388
3405
|
for pipeline creation in favor of the `schema` field.
|
|
3389
3406
|
:param trigger: :class:`PipelineTrigger` (optional)
|
|
3390
3407
|
Which pipeline trigger to use. Deprecated: Use `continuous` instead.
|
|
3408
|
+
:param usage_policy_id: str (optional)
|
|
3409
|
+
Usage policy of this pipeline.
|
|
3391
3410
|
|
|
3392
3411
|
:returns: :class:`CreatePipelineResponse`
|
|
3393
3412
|
"""
|
|
3413
|
+
|
|
3394
3414
|
body = {}
|
|
3395
3415
|
if allow_duplicate_names is not None:
|
|
3396
3416
|
body["allow_duplicate_names"] = allow_duplicate_names
|
|
@@ -3452,6 +3472,8 @@ class PipelinesAPI:
|
|
|
3452
3472
|
body["target"] = target
|
|
3453
3473
|
if trigger is not None:
|
|
3454
3474
|
body["trigger"] = trigger.as_dict()
|
|
3475
|
+
if usage_policy_id is not None:
|
|
3476
|
+
body["usage_policy_id"] = usage_policy_id
|
|
3455
3477
|
headers = {
|
|
3456
3478
|
"Accept": "application/json",
|
|
3457
3479
|
"Content-Type": "application/json",
|
|
@@ -3605,7 +3627,7 @@ class PipelinesAPI:
|
|
|
3605
3627
|
order_by: Optional[List[str]] = None,
|
|
3606
3628
|
page_token: Optional[str] = None,
|
|
3607
3629
|
) -> Iterator[PipelineStateInfo]:
|
|
3608
|
-
"""Lists pipelines defined in the
|
|
3630
|
+
"""Lists pipelines defined in the Spark Declarative Pipelines system.
|
|
3609
3631
|
|
|
3610
3632
|
:param filter: str (optional)
|
|
3611
3633
|
Select a subset of results based on the specified criteria. The supported filters are:
|
|
@@ -3699,6 +3721,7 @@ class PipelinesAPI:
|
|
|
3699
3721
|
|
|
3700
3722
|
:returns: :class:`PipelinePermissions`
|
|
3701
3723
|
"""
|
|
3724
|
+
|
|
3702
3725
|
body = {}
|
|
3703
3726
|
if access_control_list is not None:
|
|
3704
3727
|
body["access_control_list"] = [v.as_dict() for v in access_control_list]
|
|
@@ -3741,6 +3764,7 @@ class PipelinesAPI:
|
|
|
3741
3764
|
|
|
3742
3765
|
:returns: :class:`StartUpdateResponse`
|
|
3743
3766
|
"""
|
|
3767
|
+
|
|
3744
3768
|
body = {}
|
|
3745
3769
|
if cause is not None:
|
|
3746
3770
|
body["cause"] = cause.value
|
|
@@ -3815,6 +3839,7 @@ class PipelinesAPI:
|
|
|
3815
3839
|
tags: Optional[Dict[str, str]] = None,
|
|
3816
3840
|
target: Optional[str] = None,
|
|
3817
3841
|
trigger: Optional[PipelineTrigger] = None,
|
|
3842
|
+
usage_policy_id: Optional[str] = None,
|
|
3818
3843
|
):
|
|
3819
3844
|
"""Updates a pipeline with the supplied configuration.
|
|
3820
3845
|
|
|
@@ -3888,9 +3913,12 @@ class PipelinesAPI:
|
|
|
3888
3913
|
for pipeline creation in favor of the `schema` field.
|
|
3889
3914
|
:param trigger: :class:`PipelineTrigger` (optional)
|
|
3890
3915
|
Which pipeline trigger to use. Deprecated: Use `continuous` instead.
|
|
3916
|
+
:param usage_policy_id: str (optional)
|
|
3917
|
+
Usage policy of this pipeline.
|
|
3891
3918
|
|
|
3892
3919
|
|
|
3893
3920
|
"""
|
|
3921
|
+
|
|
3894
3922
|
body = {}
|
|
3895
3923
|
if allow_duplicate_names is not None:
|
|
3896
3924
|
body["allow_duplicate_names"] = allow_duplicate_names
|
|
@@ -3952,6 +3980,8 @@ class PipelinesAPI:
|
|
|
3952
3980
|
body["target"] = target
|
|
3953
3981
|
if trigger is not None:
|
|
3954
3982
|
body["trigger"] = trigger.as_dict()
|
|
3983
|
+
if usage_policy_id is not None:
|
|
3984
|
+
body["usage_policy_id"] = usage_policy_id
|
|
3955
3985
|
headers = {
|
|
3956
3986
|
"Accept": "application/json",
|
|
3957
3987
|
"Content-Type": "application/json",
|
|
@@ -3970,6 +4000,7 @@ class PipelinesAPI:
|
|
|
3970
4000
|
|
|
3971
4001
|
:returns: :class:`PipelinePermissions`
|
|
3972
4002
|
"""
|
|
4003
|
+
|
|
3973
4004
|
body = {}
|
|
3974
4005
|
if access_control_list is not None:
|
|
3975
4006
|
body["access_control_list"] = [v.as_dict() for v in access_control_list]
|
|
@@ -1765,6 +1765,7 @@ class CredentialsAPI:
|
|
|
1765
1765
|
|
|
1766
1766
|
:returns: :class:`Credential`
|
|
1767
1767
|
"""
|
|
1768
|
+
|
|
1768
1769
|
body = {}
|
|
1769
1770
|
if aws_credentials is not None:
|
|
1770
1771
|
body["aws_credentials"] = aws_credentials.as_dict()
|
|
@@ -1876,6 +1877,7 @@ class EncryptionKeysAPI:
|
|
|
1876
1877
|
|
|
1877
1878
|
:returns: :class:`CustomerManagedKey`
|
|
1878
1879
|
"""
|
|
1880
|
+
|
|
1879
1881
|
body = {}
|
|
1880
1882
|
if aws_key_info is not None:
|
|
1881
1883
|
body["aws_key_info"] = aws_key_info.as_dict()
|
|
@@ -1996,6 +1998,7 @@ class NetworksAPI:
|
|
|
1996
1998
|
|
|
1997
1999
|
:returns: :class:`Network`
|
|
1998
2000
|
"""
|
|
2001
|
+
|
|
1999
2002
|
body = {}
|
|
2000
2003
|
if gcp_network_info is not None:
|
|
2001
2004
|
body["gcp_network_info"] = gcp_network_info.as_dict()
|
|
@@ -2111,6 +2114,7 @@ class PrivateAccessAPI:
|
|
|
2111
2114
|
|
|
2112
2115
|
:returns: :class:`PrivateAccessSettings`
|
|
2113
2116
|
"""
|
|
2117
|
+
|
|
2114
2118
|
body = {}
|
|
2115
2119
|
if allowed_vpc_endpoint_ids is not None:
|
|
2116
2120
|
body["allowed_vpc_endpoint_ids"] = [v for v in allowed_vpc_endpoint_ids]
|
|
@@ -2205,6 +2209,7 @@ class PrivateAccessAPI:
|
|
|
2205
2209
|
|
|
2206
2210
|
:returns: :class:`PrivateAccessSettings`
|
|
2207
2211
|
"""
|
|
2212
|
+
|
|
2208
2213
|
body = customer_facing_private_access_settings.as_dict()
|
|
2209
2214
|
headers = {
|
|
2210
2215
|
"Accept": "application/json",
|
|
@@ -2247,6 +2252,7 @@ class StorageAPI:
|
|
|
2247
2252
|
|
|
2248
2253
|
:returns: :class:`StorageConfiguration`
|
|
2249
2254
|
"""
|
|
2255
|
+
|
|
2250
2256
|
body = {}
|
|
2251
2257
|
if role_arn is not None:
|
|
2252
2258
|
body["role_arn"] = role_arn
|
|
@@ -2356,6 +2362,7 @@ class VpcEndpointsAPI:
|
|
|
2356
2362
|
|
|
2357
2363
|
:returns: :class:`VpcEndpoint`
|
|
2358
2364
|
"""
|
|
2365
|
+
|
|
2359
2366
|
body = {}
|
|
2360
2367
|
if aws_vpc_endpoint_id is not None:
|
|
2361
2368
|
body["aws_vpc_endpoint_id"] = aws_vpc_endpoint_id
|
|
@@ -2595,6 +2602,7 @@ class WorkspacesAPI:
|
|
|
2595
2602
|
Long-running operation waiter for :class:`Workspace`.
|
|
2596
2603
|
See :method:wait_get_workspace_running for more details.
|
|
2597
2604
|
"""
|
|
2605
|
+
|
|
2598
2606
|
body = {}
|
|
2599
2607
|
if aws_region is not None:
|
|
2600
2608
|
body["aws_region"] = aws_region
|
|
@@ -2768,6 +2776,7 @@ class WorkspacesAPI:
|
|
|
2768
2776
|
Long-running operation waiter for :class:`Workspace`.
|
|
2769
2777
|
See :method:wait_get_workspace_running for more details.
|
|
2770
2778
|
"""
|
|
2779
|
+
|
|
2771
2780
|
body = customer_facing_workspace.as_dict()
|
|
2772
2781
|
query = {}
|
|
2773
2782
|
if update_mask is not None:
|
|
@@ -151,6 +151,7 @@ class QualityMonitorV2API:
|
|
|
151
151
|
|
|
152
152
|
:returns: :class:`QualityMonitor`
|
|
153
153
|
"""
|
|
154
|
+
|
|
154
155
|
body = quality_monitor.as_dict()
|
|
155
156
|
headers = {
|
|
156
157
|
"Accept": "application/json",
|
|
@@ -237,6 +238,7 @@ class QualityMonitorV2API:
|
|
|
237
238
|
|
|
238
239
|
:returns: :class:`QualityMonitor`
|
|
239
240
|
"""
|
|
241
|
+
|
|
240
242
|
body = quality_monitor.as_dict()
|
|
241
243
|
headers = {
|
|
242
244
|
"Accept": "application/json",
|
|
@@ -1036,24 +1036,6 @@ class DataframeSplitInput:
|
|
|
1036
1036
|
return cls(columns=d.get("columns", None), data=d.get("data", None), index=d.get("index", None))
|
|
1037
1037
|
|
|
1038
1038
|
|
|
1039
|
-
@dataclass
|
|
1040
|
-
class DeleteResponse:
|
|
1041
|
-
def as_dict(self) -> dict:
|
|
1042
|
-
"""Serializes the DeleteResponse into a dictionary suitable for use as a JSON request body."""
|
|
1043
|
-
body = {}
|
|
1044
|
-
return body
|
|
1045
|
-
|
|
1046
|
-
def as_shallow_dict(self) -> dict:
|
|
1047
|
-
"""Serializes the DeleteResponse into a shallow dictionary of its immediate attributes."""
|
|
1048
|
-
body = {}
|
|
1049
|
-
return body
|
|
1050
|
-
|
|
1051
|
-
@classmethod
|
|
1052
|
-
def from_dict(cls, d: Dict[str, Any]) -> DeleteResponse:
|
|
1053
|
-
"""Deserializes the DeleteResponse from a dictionary."""
|
|
1054
|
-
return cls()
|
|
1055
|
-
|
|
1056
|
-
|
|
1057
1039
|
@dataclass
|
|
1058
1040
|
class EmailNotifications:
|
|
1059
1041
|
on_update_failure: Optional[List[str]] = None
|
|
@@ -1141,15 +1123,15 @@ class EmbeddingsV1ResponseEmbeddingElementObject(Enum):
|
|
|
1141
1123
|
|
|
1142
1124
|
@dataclass
|
|
1143
1125
|
class EndpointCoreConfigInput:
|
|
1126
|
+
name: str
|
|
1127
|
+
"""The name of the serving endpoint to update. This field is required."""
|
|
1128
|
+
|
|
1144
1129
|
auto_capture_config: Optional[AutoCaptureConfigInput] = None
|
|
1145
1130
|
"""Configuration for Inference Tables which automatically logs requests and responses to Unity
|
|
1146
1131
|
Catalog. Note: this field is deprecated for creating new provisioned throughput endpoints, or
|
|
1147
1132
|
updating existing provisioned throughput endpoints that never have inference table configured;
|
|
1148
1133
|
in these cases please use AI Gateway to manage inference tables."""
|
|
1149
1134
|
|
|
1150
|
-
name: Optional[str] = None
|
|
1151
|
-
"""The name of the serving endpoint to update. This field is required."""
|
|
1152
|
-
|
|
1153
1135
|
served_entities: Optional[List[ServedEntityInput]] = None
|
|
1154
1136
|
"""The list of served entities under the serving endpoint config."""
|
|
1155
1137
|
|
|
@@ -4109,6 +4091,7 @@ class ServingEndpointsAPI:
|
|
|
4109
4091
|
Long-running operation waiter for :class:`ServingEndpointDetailed`.
|
|
4110
4092
|
See :method:wait_get_serving_endpoint_not_updating for more details.
|
|
4111
4093
|
"""
|
|
4094
|
+
|
|
4112
4095
|
body = {}
|
|
4113
4096
|
if ai_gateway is not None:
|
|
4114
4097
|
body["ai_gateway"] = ai_gateway.as_dict()
|
|
@@ -4196,6 +4179,7 @@ class ServingEndpointsAPI:
|
|
|
4196
4179
|
Long-running operation waiter for :class:`ServingEndpointDetailed`.
|
|
4197
4180
|
See :method:wait_get_serving_endpoint_not_updating for more details.
|
|
4198
4181
|
"""
|
|
4182
|
+
|
|
4199
4183
|
body = {}
|
|
4200
4184
|
if ai_gateway is not None:
|
|
4201
4185
|
body["ai_gateway"] = ai_gateway.as_dict()
|
|
@@ -4366,6 +4350,7 @@ class ServingEndpointsAPI:
|
|
|
4366
4350
|
|
|
4367
4351
|
:returns: :class:`HttpRequestResponse`
|
|
4368
4352
|
"""
|
|
4353
|
+
|
|
4369
4354
|
body = {}
|
|
4370
4355
|
if connection_name is not None:
|
|
4371
4356
|
body["connection_name"] = connection_name
|
|
@@ -4436,6 +4421,7 @@ class ServingEndpointsAPI:
|
|
|
4436
4421
|
|
|
4437
4422
|
:returns: :class:`EndpointTags`
|
|
4438
4423
|
"""
|
|
4424
|
+
|
|
4439
4425
|
body = {}
|
|
4440
4426
|
if add_tags is not None:
|
|
4441
4427
|
body["add_tags"] = [v.as_dict() for v in add_tags]
|
|
@@ -4459,6 +4445,7 @@ class ServingEndpointsAPI:
|
|
|
4459
4445
|
|
|
4460
4446
|
:returns: :class:`PutResponse`
|
|
4461
4447
|
"""
|
|
4448
|
+
|
|
4462
4449
|
body = {}
|
|
4463
4450
|
if rate_limits is not None:
|
|
4464
4451
|
body["rate_limits"] = [v.as_dict() for v in rate_limits]
|
|
@@ -4501,6 +4488,7 @@ class ServingEndpointsAPI:
|
|
|
4501
4488
|
|
|
4502
4489
|
:returns: :class:`PutAiGatewayResponse`
|
|
4503
4490
|
"""
|
|
4491
|
+
|
|
4504
4492
|
body = {}
|
|
4505
4493
|
if fallback_config is not None:
|
|
4506
4494
|
body["fallback_config"] = fallback_config.as_dict()
|
|
@@ -4593,6 +4581,7 @@ class ServingEndpointsAPI:
|
|
|
4593
4581
|
|
|
4594
4582
|
:returns: :class:`QueryEndpointResponse`
|
|
4595
4583
|
"""
|
|
4584
|
+
|
|
4596
4585
|
body = {}
|
|
4597
4586
|
if client_request_id is not None:
|
|
4598
4587
|
body["client_request_id"] = client_request_id
|
|
@@ -4656,6 +4645,7 @@ class ServingEndpointsAPI:
|
|
|
4656
4645
|
|
|
4657
4646
|
:returns: :class:`ServingEndpointPermissions`
|
|
4658
4647
|
"""
|
|
4648
|
+
|
|
4659
4649
|
body = {}
|
|
4660
4650
|
if access_control_list is not None:
|
|
4661
4651
|
body["access_control_list"] = [v.as_dict() for v in access_control_list]
|
|
@@ -4701,6 +4691,7 @@ class ServingEndpointsAPI:
|
|
|
4701
4691
|
Long-running operation waiter for :class:`ServingEndpointDetailed`.
|
|
4702
4692
|
See :method:wait_get_serving_endpoint_not_updating for more details.
|
|
4703
4693
|
"""
|
|
4694
|
+
|
|
4704
4695
|
body = {}
|
|
4705
4696
|
if auto_capture_config is not None:
|
|
4706
4697
|
body["auto_capture_config"] = auto_capture_config.as_dict()
|
|
@@ -4753,6 +4744,7 @@ class ServingEndpointsAPI:
|
|
|
4753
4744
|
|
|
4754
4745
|
:returns: :class:`UpdateInferenceEndpointNotificationsResponse`
|
|
4755
4746
|
"""
|
|
4747
|
+
|
|
4756
4748
|
body = {}
|
|
4757
4749
|
if email_notifications is not None:
|
|
4758
4750
|
body["email_notifications"] = email_notifications.as_dict()
|
|
@@ -4779,6 +4771,7 @@ class ServingEndpointsAPI:
|
|
|
4779
4771
|
|
|
4780
4772
|
:returns: :class:`ServingEndpointPermissions`
|
|
4781
4773
|
"""
|
|
4774
|
+
|
|
4782
4775
|
body = {}
|
|
4783
4776
|
if access_control_list is not None:
|
|
4784
4777
|
body["access_control_list"] = [v.as_dict() for v in access_control_list]
|
|
@@ -4807,6 +4800,7 @@ class ServingEndpointsAPI:
|
|
|
4807
4800
|
Long-running operation waiter for :class:`ServingEndpointDetailed`.
|
|
4808
4801
|
See :method:wait_get_serving_endpoint_not_updating for more details.
|
|
4809
4802
|
"""
|
|
4803
|
+
|
|
4810
4804
|
body = {}
|
|
4811
4805
|
if config is not None:
|
|
4812
4806
|
body["config"] = config.as_dict()
|
|
@@ -4931,6 +4925,7 @@ class ServingEndpointsDataPlaneAPI:
|
|
|
4931
4925
|
|
|
4932
4926
|
:returns: :class:`QueryEndpointResponse`
|
|
4933
4927
|
"""
|
|
4928
|
+
|
|
4934
4929
|
body = {}
|
|
4935
4930
|
if client_request_id is not None:
|
|
4936
4931
|
body["client_request_id"] = client_request_id
|