databricks-sdk 0.70.0__py3-none-any.whl → 0.72.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of databricks-sdk might be problematic. Click here for more details.

Files changed (36) hide show
  1. databricks/sdk/__init__.py +25 -25
  2. databricks/sdk/mixins/files.py +51 -15
  3. databricks/sdk/service/agentbricks.py +2 -0
  4. databricks/sdk/service/apps.py +10 -0
  5. databricks/sdk/service/billing.py +13 -3
  6. databricks/sdk/service/catalog.py +149 -46
  7. databricks/sdk/service/cleanrooms.py +11 -3
  8. databricks/sdk/service/compute.py +55 -0
  9. databricks/sdk/service/dashboards.py +11 -0
  10. databricks/sdk/service/database.py +12 -0
  11. databricks/sdk/service/dataquality.py +4 -0
  12. databricks/sdk/service/files.py +7 -72
  13. databricks/sdk/service/iam.py +26 -36
  14. databricks/sdk/service/iamv2.py +6 -0
  15. databricks/sdk/service/jobs.py +22 -122
  16. databricks/sdk/service/marketplace.py +18 -0
  17. databricks/sdk/service/ml.py +321 -17
  18. databricks/sdk/service/oauth2.py +10 -18
  19. databricks/sdk/service/pipelines.py +44 -13
  20. databricks/sdk/service/provisioning.py +9 -0
  21. databricks/sdk/service/qualitymonitorv2.py +2 -0
  22. databricks/sdk/service/serving.py +16 -21
  23. databricks/sdk/service/settings.py +43 -72
  24. databricks/sdk/service/settingsv2.py +2 -0
  25. databricks/sdk/service/sharing.py +23 -31
  26. databricks/sdk/service/sql.py +48 -24
  27. databricks/sdk/service/tags.py +2 -0
  28. databricks/sdk/service/vectorsearch.py +11 -1
  29. databricks/sdk/service/workspace.py +18 -91
  30. databricks/sdk/version.py +1 -1
  31. {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/METADATA +1 -1
  32. {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/RECORD +36 -36
  33. {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/WHEEL +0 -0
  34. {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/licenses/LICENSE +0 -0
  35. {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/licenses/NOTICE +0 -0
  36. {databricks_sdk-0.70.0.dist-info → databricks_sdk-0.72.0.dist-info}/top_level.txt +0 -0
@@ -194,24 +194,6 @@ class DeletePublishedAppIntegrationOutput:
194
194
  return cls()
195
195
 
196
196
 
197
- @dataclass
198
- class DeleteResponse:
199
- def as_dict(self) -> dict:
200
- """Serializes the DeleteResponse into a dictionary suitable for use as a JSON request body."""
201
- body = {}
202
- return body
203
-
204
- def as_shallow_dict(self) -> dict:
205
- """Serializes the DeleteResponse into a shallow dictionary of its immediate attributes."""
206
- body = {}
207
- return body
208
-
209
- @classmethod
210
- def from_dict(cls, d: Dict[str, Any]) -> DeleteResponse:
211
- """Deserializes the DeleteResponse from a dictionary."""
212
- return cls()
213
-
214
-
215
197
  @dataclass
216
198
  class FederationPolicy:
217
199
  create_time: Optional[str] = None
@@ -1013,6 +995,7 @@ class AccountFederationPolicyAPI:
1013
995
 
1014
996
  :returns: :class:`FederationPolicy`
1015
997
  """
998
+
1016
999
  body = policy.as_dict()
1017
1000
  query = {}
1018
1001
  if policy_id is not None:
@@ -1112,6 +1095,7 @@ class AccountFederationPolicyAPI:
1112
1095
 
1113
1096
  :returns: :class:`FederationPolicy`
1114
1097
  """
1098
+
1115
1099
  body = policy.as_dict()
1116
1100
  query = {}
1117
1101
  if update_mask is not None:
@@ -1169,6 +1153,7 @@ class CustomAppIntegrationAPI:
1169
1153
 
1170
1154
  :returns: :class:`CreateCustomAppIntegrationOutput`
1171
1155
  """
1156
+
1172
1157
  body = {}
1173
1158
  if confidential is not None:
1174
1159
  body["confidential"] = confidential
@@ -1301,6 +1286,7 @@ class CustomAppIntegrationAPI:
1301
1286
 
1302
1287
 
1303
1288
  """
1289
+
1304
1290
  body = {}
1305
1291
  if redirect_urls is not None:
1306
1292
  body["redirect_urls"] = [v for v in redirect_urls]
@@ -1386,6 +1372,7 @@ class PublishedAppIntegrationAPI:
1386
1372
 
1387
1373
  :returns: :class:`CreatePublishedAppIntegrationOutput`
1388
1374
  """
1375
+
1389
1376
  body = {}
1390
1377
  if app_id is not None:
1391
1378
  body["app_id"] = app_id
@@ -1486,6 +1473,7 @@ class PublishedAppIntegrationAPI:
1486
1473
 
1487
1474
 
1488
1475
  """
1476
+
1489
1477
  body = {}
1490
1478
  if token_access_policy is not None:
1491
1479
  body["token_access_policy"] = token_access_policy.as_dict()
@@ -1561,6 +1549,7 @@ class ServicePrincipalFederationPolicyAPI:
1561
1549
 
1562
1550
  :returns: :class:`FederationPolicy`
1563
1551
  """
1552
+
1564
1553
  body = policy.as_dict()
1565
1554
  query = {}
1566
1555
  if policy_id is not None:
@@ -1677,6 +1666,7 @@ class ServicePrincipalFederationPolicyAPI:
1677
1666
 
1678
1667
  :returns: :class:`FederationPolicy`
1679
1668
  """
1669
+
1680
1670
  body = policy.as_dict()
1681
1671
  query = {}
1682
1672
  if update_mask is not None:
@@ -1726,6 +1716,7 @@ class ServicePrincipalSecretsAPI:
1726
1716
 
1727
1717
  :returns: :class:`CreateServicePrincipalSecretResponse`
1728
1718
  """
1719
+
1729
1720
  body = {}
1730
1721
  if lifetime is not None:
1731
1722
  body["lifetime"] = lifetime
@@ -1836,6 +1827,7 @@ class ServicePrincipalSecretsProxyAPI:
1836
1827
 
1837
1828
  :returns: :class:`CreateServicePrincipalSecretResponse`
1838
1829
  """
1830
+
1839
1831
  body = {}
1840
1832
  if lifetime is not None:
1841
1833
  body["lifetime"] = lifetime
@@ -556,8 +556,8 @@ class IngestionGatewayPipelineDefinition:
556
556
 
557
557
  gateway_storage_name: Optional[str] = None
558
558
  """Optional. The Unity Catalog-compatible name for the gateway storage location. This is the
559
- destination to use for the data that is extracted by the gateway. Delta Live Tables system will
560
- automatically create the storage location under the catalog and schema."""
559
+ destination to use for the data that is extracted by the gateway. Spark Declarative Pipelines
560
+ system will automatically create the storage location under the catalog and schema."""
561
561
 
562
562
  def as_dict(self) -> dict:
563
563
  """Serializes the IngestionGatewayPipelineDefinition into a dictionary suitable for use as a JSON request body."""
@@ -832,6 +832,9 @@ class IngestionSourceType(Enum):
832
832
  DYNAMICS365 = "DYNAMICS365"
833
833
  FOREIGN_CATALOG = "FOREIGN_CATALOG"
834
834
  GA4_RAW_DATA = "GA4_RAW_DATA"
835
+ GOOGLE_ADS = "GOOGLE_ADS"
836
+ GUIDEWIRE = "GUIDEWIRE"
837
+ HUBSPOT = "HUBSPOT"
835
838
  MANAGED_POSTGRESQL = "MANAGED_POSTGRESQL"
836
839
  META_MARKETING = "META_MARKETING"
837
840
  MYSQL = "MYSQL"
@@ -840,12 +843,16 @@ class IngestionSourceType(Enum):
840
843
  POSTGRESQL = "POSTGRESQL"
841
844
  REDSHIFT = "REDSHIFT"
842
845
  SALESFORCE = "SALESFORCE"
846
+ SALESFORCE_MARKETING_CLOUD = "SALESFORCE_MARKETING_CLOUD"
843
847
  SERVICENOW = "SERVICENOW"
844
848
  SHAREPOINT = "SHAREPOINT"
845
849
  SQLDW = "SQLDW"
846
850
  SQLSERVER = "SQLSERVER"
847
851
  TERADATA = "TERADATA"
852
+ TIKTOK_ADS = "TIKTOK_ADS"
853
+ WORKDAY_HCM = "WORKDAY_HCM"
848
854
  WORKDAY_RAAS = "WORKDAY_RAAS"
855
+ ZENDESK = "ZENDESK"
849
856
 
850
857
 
851
858
  @dataclass
@@ -2024,6 +2031,9 @@ class PipelineSpec:
2024
2031
  trigger: Optional[PipelineTrigger] = None
2025
2032
  """Which pipeline trigger to use. Deprecated: Use `continuous` instead."""
2026
2033
 
2034
+ usage_policy_id: Optional[str] = None
2035
+ """Usage policy of this pipeline."""
2036
+
2027
2037
  def as_dict(self) -> dict:
2028
2038
  """Serializes the PipelineSpec into a dictionary suitable for use as a JSON request body."""
2029
2039
  body = {}
@@ -2081,6 +2091,8 @@ class PipelineSpec:
2081
2091
  body["target"] = self.target
2082
2092
  if self.trigger:
2083
2093
  body["trigger"] = self.trigger.as_dict()
2094
+ if self.usage_policy_id is not None:
2095
+ body["usage_policy_id"] = self.usage_policy_id
2084
2096
  return body
2085
2097
 
2086
2098
  def as_shallow_dict(self) -> dict:
@@ -2140,6 +2152,8 @@ class PipelineSpec:
2140
2152
  body["target"] = self.target
2141
2153
  if self.trigger:
2142
2154
  body["trigger"] = self.trigger
2155
+ if self.usage_policy_id is not None:
2156
+ body["usage_policy_id"] = self.usage_policy_id
2143
2157
  return body
2144
2158
 
2145
2159
  @classmethod
@@ -2173,6 +2187,7 @@ class PipelineSpec:
2173
2187
  tags=d.get("tags", None),
2174
2188
  target=d.get("target", None),
2175
2189
  trigger=_from_dict(d, "trigger", PipelineTrigger),
2190
+ usage_policy_id=d.get("usage_policy_id", None),
2176
2191
  )
2177
2192
 
2178
2193
 
@@ -2964,8 +2979,8 @@ class TableSpecificConfig:
2964
2979
  """The SCD type to use to ingest the table."""
2965
2980
 
2966
2981
  sequence_by: Optional[List[str]] = None
2967
- """The column names specifying the logical order of events in the source data. Delta Live Tables
2968
- uses this sequencing to handle change events that arrive out of order."""
2982
+ """The column names specifying the logical order of events in the source data. Spark Declarative
2983
+ Pipelines uses this sequencing to handle change events that arrive out of order."""
2969
2984
 
2970
2985
  workday_report_parameters: Optional[IngestionPipelineDefinitionWorkdayReportParameters] = None
2971
2986
  """(Optional) Additional custom parameters for Workday Report"""
@@ -3239,16 +3254,17 @@ class UpdateStateInfoState(Enum):
3239
3254
 
3240
3255
 
3241
3256
  class PipelinesAPI:
3242
- """The Delta Live Tables API allows you to create, edit, delete, start, and view details about pipelines.
3257
+ """The Lakeflow Spark Declarative Pipelines API allows you to create, edit, delete, start, and view details
3258
+ about pipelines.
3243
3259
 
3244
- Delta Live Tables is a framework for building reliable, maintainable, and testable data processing
3245
- pipelines. You define the transformations to perform on your data, and Delta Live Tables manages task
3246
- orchestration, cluster management, monitoring, data quality, and error handling.
3260
+ Spark Declarative Pipelines is a framework for building reliable, maintainable, and testable data
3261
+ processing pipelines. You define the transformations to perform on your data, and Spark Declarative
3262
+ Pipelines manages task orchestration, cluster management, monitoring, data quality, and error handling.
3247
3263
 
3248
- Instead of defining your data pipelines using a series of separate Apache Spark tasks, Delta Live Tables
3249
- manages how your data is transformed based on a target schema you define for each processing step. You can
3250
- also enforce data quality with Delta Live Tables expectations. Expectations allow you to define expected
3251
- data quality and specify how to handle records that fail those expectations."""
3264
+ Instead of defining your data pipelines using a series of separate Apache Spark tasks, Spark Declarative
3265
+ Pipelines manages how your data is transformed based on a target schema you define for each processing
3266
+ step. You can also enforce data quality with Spark Declarative Pipelines expectations. Expectations allow
3267
+ you to define expected data quality and specify how to handle records that fail those expectations."""
3252
3268
 
3253
3269
  def __init__(self, api_client):
3254
3270
  self._api = api_client
@@ -3318,6 +3334,7 @@ class PipelinesAPI:
3318
3334
  tags: Optional[Dict[str, str]] = None,
3319
3335
  target: Optional[str] = None,
3320
3336
  trigger: Optional[PipelineTrigger] = None,
3337
+ usage_policy_id: Optional[str] = None,
3321
3338
  ) -> CreatePipelineResponse:
3322
3339
  """Creates a new data processing pipeline based on the requested configuration. If successful, this
3323
3340
  method returns the ID of the new pipeline.
@@ -3388,9 +3405,12 @@ class PipelinesAPI:
3388
3405
  for pipeline creation in favor of the `schema` field.
3389
3406
  :param trigger: :class:`PipelineTrigger` (optional)
3390
3407
  Which pipeline trigger to use. Deprecated: Use `continuous` instead.
3408
+ :param usage_policy_id: str (optional)
3409
+ Usage policy of this pipeline.
3391
3410
 
3392
3411
  :returns: :class:`CreatePipelineResponse`
3393
3412
  """
3413
+
3394
3414
  body = {}
3395
3415
  if allow_duplicate_names is not None:
3396
3416
  body["allow_duplicate_names"] = allow_duplicate_names
@@ -3452,6 +3472,8 @@ class PipelinesAPI:
3452
3472
  body["target"] = target
3453
3473
  if trigger is not None:
3454
3474
  body["trigger"] = trigger.as_dict()
3475
+ if usage_policy_id is not None:
3476
+ body["usage_policy_id"] = usage_policy_id
3455
3477
  headers = {
3456
3478
  "Accept": "application/json",
3457
3479
  "Content-Type": "application/json",
@@ -3605,7 +3627,7 @@ class PipelinesAPI:
3605
3627
  order_by: Optional[List[str]] = None,
3606
3628
  page_token: Optional[str] = None,
3607
3629
  ) -> Iterator[PipelineStateInfo]:
3608
- """Lists pipelines defined in the Delta Live Tables system.
3630
+ """Lists pipelines defined in the Spark Declarative Pipelines system.
3609
3631
 
3610
3632
  :param filter: str (optional)
3611
3633
  Select a subset of results based on the specified criteria. The supported filters are:
@@ -3699,6 +3721,7 @@ class PipelinesAPI:
3699
3721
 
3700
3722
  :returns: :class:`PipelinePermissions`
3701
3723
  """
3724
+
3702
3725
  body = {}
3703
3726
  if access_control_list is not None:
3704
3727
  body["access_control_list"] = [v.as_dict() for v in access_control_list]
@@ -3741,6 +3764,7 @@ class PipelinesAPI:
3741
3764
 
3742
3765
  :returns: :class:`StartUpdateResponse`
3743
3766
  """
3767
+
3744
3768
  body = {}
3745
3769
  if cause is not None:
3746
3770
  body["cause"] = cause.value
@@ -3815,6 +3839,7 @@ class PipelinesAPI:
3815
3839
  tags: Optional[Dict[str, str]] = None,
3816
3840
  target: Optional[str] = None,
3817
3841
  trigger: Optional[PipelineTrigger] = None,
3842
+ usage_policy_id: Optional[str] = None,
3818
3843
  ):
3819
3844
  """Updates a pipeline with the supplied configuration.
3820
3845
 
@@ -3888,9 +3913,12 @@ class PipelinesAPI:
3888
3913
  for pipeline creation in favor of the `schema` field.
3889
3914
  :param trigger: :class:`PipelineTrigger` (optional)
3890
3915
  Which pipeline trigger to use. Deprecated: Use `continuous` instead.
3916
+ :param usage_policy_id: str (optional)
3917
+ Usage policy of this pipeline.
3891
3918
 
3892
3919
 
3893
3920
  """
3921
+
3894
3922
  body = {}
3895
3923
  if allow_duplicate_names is not None:
3896
3924
  body["allow_duplicate_names"] = allow_duplicate_names
@@ -3952,6 +3980,8 @@ class PipelinesAPI:
3952
3980
  body["target"] = target
3953
3981
  if trigger is not None:
3954
3982
  body["trigger"] = trigger.as_dict()
3983
+ if usage_policy_id is not None:
3984
+ body["usage_policy_id"] = usage_policy_id
3955
3985
  headers = {
3956
3986
  "Accept": "application/json",
3957
3987
  "Content-Type": "application/json",
@@ -3970,6 +4000,7 @@ class PipelinesAPI:
3970
4000
 
3971
4001
  :returns: :class:`PipelinePermissions`
3972
4002
  """
4003
+
3973
4004
  body = {}
3974
4005
  if access_control_list is not None:
3975
4006
  body["access_control_list"] = [v.as_dict() for v in access_control_list]
@@ -1765,6 +1765,7 @@ class CredentialsAPI:
1765
1765
 
1766
1766
  :returns: :class:`Credential`
1767
1767
  """
1768
+
1768
1769
  body = {}
1769
1770
  if aws_credentials is not None:
1770
1771
  body["aws_credentials"] = aws_credentials.as_dict()
@@ -1876,6 +1877,7 @@ class EncryptionKeysAPI:
1876
1877
 
1877
1878
  :returns: :class:`CustomerManagedKey`
1878
1879
  """
1880
+
1879
1881
  body = {}
1880
1882
  if aws_key_info is not None:
1881
1883
  body["aws_key_info"] = aws_key_info.as_dict()
@@ -1996,6 +1998,7 @@ class NetworksAPI:
1996
1998
 
1997
1999
  :returns: :class:`Network`
1998
2000
  """
2001
+
1999
2002
  body = {}
2000
2003
  if gcp_network_info is not None:
2001
2004
  body["gcp_network_info"] = gcp_network_info.as_dict()
@@ -2111,6 +2114,7 @@ class PrivateAccessAPI:
2111
2114
 
2112
2115
  :returns: :class:`PrivateAccessSettings`
2113
2116
  """
2117
+
2114
2118
  body = {}
2115
2119
  if allowed_vpc_endpoint_ids is not None:
2116
2120
  body["allowed_vpc_endpoint_ids"] = [v for v in allowed_vpc_endpoint_ids]
@@ -2205,6 +2209,7 @@ class PrivateAccessAPI:
2205
2209
 
2206
2210
  :returns: :class:`PrivateAccessSettings`
2207
2211
  """
2212
+
2208
2213
  body = customer_facing_private_access_settings.as_dict()
2209
2214
  headers = {
2210
2215
  "Accept": "application/json",
@@ -2247,6 +2252,7 @@ class StorageAPI:
2247
2252
 
2248
2253
  :returns: :class:`StorageConfiguration`
2249
2254
  """
2255
+
2250
2256
  body = {}
2251
2257
  if role_arn is not None:
2252
2258
  body["role_arn"] = role_arn
@@ -2356,6 +2362,7 @@ class VpcEndpointsAPI:
2356
2362
 
2357
2363
  :returns: :class:`VpcEndpoint`
2358
2364
  """
2365
+
2359
2366
  body = {}
2360
2367
  if aws_vpc_endpoint_id is not None:
2361
2368
  body["aws_vpc_endpoint_id"] = aws_vpc_endpoint_id
@@ -2595,6 +2602,7 @@ class WorkspacesAPI:
2595
2602
  Long-running operation waiter for :class:`Workspace`.
2596
2603
  See :method:wait_get_workspace_running for more details.
2597
2604
  """
2605
+
2598
2606
  body = {}
2599
2607
  if aws_region is not None:
2600
2608
  body["aws_region"] = aws_region
@@ -2768,6 +2776,7 @@ class WorkspacesAPI:
2768
2776
  Long-running operation waiter for :class:`Workspace`.
2769
2777
  See :method:wait_get_workspace_running for more details.
2770
2778
  """
2779
+
2771
2780
  body = customer_facing_workspace.as_dict()
2772
2781
  query = {}
2773
2782
  if update_mask is not None:
@@ -151,6 +151,7 @@ class QualityMonitorV2API:
151
151
 
152
152
  :returns: :class:`QualityMonitor`
153
153
  """
154
+
154
155
  body = quality_monitor.as_dict()
155
156
  headers = {
156
157
  "Accept": "application/json",
@@ -237,6 +238,7 @@ class QualityMonitorV2API:
237
238
 
238
239
  :returns: :class:`QualityMonitor`
239
240
  """
241
+
240
242
  body = quality_monitor.as_dict()
241
243
  headers = {
242
244
  "Accept": "application/json",
@@ -1036,24 +1036,6 @@ class DataframeSplitInput:
1036
1036
  return cls(columns=d.get("columns", None), data=d.get("data", None), index=d.get("index", None))
1037
1037
 
1038
1038
 
1039
- @dataclass
1040
- class DeleteResponse:
1041
- def as_dict(self) -> dict:
1042
- """Serializes the DeleteResponse into a dictionary suitable for use as a JSON request body."""
1043
- body = {}
1044
- return body
1045
-
1046
- def as_shallow_dict(self) -> dict:
1047
- """Serializes the DeleteResponse into a shallow dictionary of its immediate attributes."""
1048
- body = {}
1049
- return body
1050
-
1051
- @classmethod
1052
- def from_dict(cls, d: Dict[str, Any]) -> DeleteResponse:
1053
- """Deserializes the DeleteResponse from a dictionary."""
1054
- return cls()
1055
-
1056
-
1057
1039
  @dataclass
1058
1040
  class EmailNotifications:
1059
1041
  on_update_failure: Optional[List[str]] = None
@@ -1141,15 +1123,15 @@ class EmbeddingsV1ResponseEmbeddingElementObject(Enum):
1141
1123
 
1142
1124
  @dataclass
1143
1125
  class EndpointCoreConfigInput:
1126
+ name: str
1127
+ """The name of the serving endpoint to update. This field is required."""
1128
+
1144
1129
  auto_capture_config: Optional[AutoCaptureConfigInput] = None
1145
1130
  """Configuration for Inference Tables which automatically logs requests and responses to Unity
1146
1131
  Catalog. Note: this field is deprecated for creating new provisioned throughput endpoints, or
1147
1132
  updating existing provisioned throughput endpoints that never have inference table configured;
1148
1133
  in these cases please use AI Gateway to manage inference tables."""
1149
1134
 
1150
- name: Optional[str] = None
1151
- """The name of the serving endpoint to update. This field is required."""
1152
-
1153
1135
  served_entities: Optional[List[ServedEntityInput]] = None
1154
1136
  """The list of served entities under the serving endpoint config."""
1155
1137
 
@@ -4109,6 +4091,7 @@ class ServingEndpointsAPI:
4109
4091
  Long-running operation waiter for :class:`ServingEndpointDetailed`.
4110
4092
  See :method:wait_get_serving_endpoint_not_updating for more details.
4111
4093
  """
4094
+
4112
4095
  body = {}
4113
4096
  if ai_gateway is not None:
4114
4097
  body["ai_gateway"] = ai_gateway.as_dict()
@@ -4196,6 +4179,7 @@ class ServingEndpointsAPI:
4196
4179
  Long-running operation waiter for :class:`ServingEndpointDetailed`.
4197
4180
  See :method:wait_get_serving_endpoint_not_updating for more details.
4198
4181
  """
4182
+
4199
4183
  body = {}
4200
4184
  if ai_gateway is not None:
4201
4185
  body["ai_gateway"] = ai_gateway.as_dict()
@@ -4366,6 +4350,7 @@ class ServingEndpointsAPI:
4366
4350
 
4367
4351
  :returns: :class:`HttpRequestResponse`
4368
4352
  """
4353
+
4369
4354
  body = {}
4370
4355
  if connection_name is not None:
4371
4356
  body["connection_name"] = connection_name
@@ -4436,6 +4421,7 @@ class ServingEndpointsAPI:
4436
4421
 
4437
4422
  :returns: :class:`EndpointTags`
4438
4423
  """
4424
+
4439
4425
  body = {}
4440
4426
  if add_tags is not None:
4441
4427
  body["add_tags"] = [v.as_dict() for v in add_tags]
@@ -4459,6 +4445,7 @@ class ServingEndpointsAPI:
4459
4445
 
4460
4446
  :returns: :class:`PutResponse`
4461
4447
  """
4448
+
4462
4449
  body = {}
4463
4450
  if rate_limits is not None:
4464
4451
  body["rate_limits"] = [v.as_dict() for v in rate_limits]
@@ -4501,6 +4488,7 @@ class ServingEndpointsAPI:
4501
4488
 
4502
4489
  :returns: :class:`PutAiGatewayResponse`
4503
4490
  """
4491
+
4504
4492
  body = {}
4505
4493
  if fallback_config is not None:
4506
4494
  body["fallback_config"] = fallback_config.as_dict()
@@ -4593,6 +4581,7 @@ class ServingEndpointsAPI:
4593
4581
 
4594
4582
  :returns: :class:`QueryEndpointResponse`
4595
4583
  """
4584
+
4596
4585
  body = {}
4597
4586
  if client_request_id is not None:
4598
4587
  body["client_request_id"] = client_request_id
@@ -4656,6 +4645,7 @@ class ServingEndpointsAPI:
4656
4645
 
4657
4646
  :returns: :class:`ServingEndpointPermissions`
4658
4647
  """
4648
+
4659
4649
  body = {}
4660
4650
  if access_control_list is not None:
4661
4651
  body["access_control_list"] = [v.as_dict() for v in access_control_list]
@@ -4701,6 +4691,7 @@ class ServingEndpointsAPI:
4701
4691
  Long-running operation waiter for :class:`ServingEndpointDetailed`.
4702
4692
  See :method:wait_get_serving_endpoint_not_updating for more details.
4703
4693
  """
4694
+
4704
4695
  body = {}
4705
4696
  if auto_capture_config is not None:
4706
4697
  body["auto_capture_config"] = auto_capture_config.as_dict()
@@ -4753,6 +4744,7 @@ class ServingEndpointsAPI:
4753
4744
 
4754
4745
  :returns: :class:`UpdateInferenceEndpointNotificationsResponse`
4755
4746
  """
4747
+
4756
4748
  body = {}
4757
4749
  if email_notifications is not None:
4758
4750
  body["email_notifications"] = email_notifications.as_dict()
@@ -4779,6 +4771,7 @@ class ServingEndpointsAPI:
4779
4771
 
4780
4772
  :returns: :class:`ServingEndpointPermissions`
4781
4773
  """
4774
+
4782
4775
  body = {}
4783
4776
  if access_control_list is not None:
4784
4777
  body["access_control_list"] = [v.as_dict() for v in access_control_list]
@@ -4807,6 +4800,7 @@ class ServingEndpointsAPI:
4807
4800
  Long-running operation waiter for :class:`ServingEndpointDetailed`.
4808
4801
  See :method:wait_get_serving_endpoint_not_updating for more details.
4809
4802
  """
4803
+
4810
4804
  body = {}
4811
4805
  if config is not None:
4812
4806
  body["config"] = config.as_dict()
@@ -4931,6 +4925,7 @@ class ServingEndpointsDataPlaneAPI:
4931
4925
 
4932
4926
  :returns: :class:`QueryEndpointResponse`
4933
4927
  """
4928
+
4934
4929
  body = {}
4935
4930
  if client_request_id is not None:
4936
4931
  body["client_request_id"] = client_request_id