cribl-control-plane 0.4.0a12__py3-none-any.whl → 0.4.0b12__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +4 -4
- cribl_control_plane/acl.py +14 -6
- cribl_control_plane/branches.py +24 -8
- cribl_control_plane/commits.py +98 -42
- cribl_control_plane/commits_files.py +28 -12
- cribl_control_plane/configs_versions.py +14 -10
- cribl_control_plane/destinations.py +70 -30
- cribl_control_plane/destinations_pq.py +28 -12
- cribl_control_plane/groups_sdk.py +84 -60
- cribl_control_plane/health.py +10 -2
- cribl_control_plane/hectokens.py +28 -32
- cribl_control_plane/lakedatasets.py +70 -86
- cribl_control_plane/models/__init__.py +266 -438
- cribl_control_plane/models/addhectokenrequest.py +1 -7
- cribl_control_plane/models/configgroup.py +1 -0
- cribl_control_plane/models/{getversionbranchop.py → countedbranchinfo.py} +2 -6
- cribl_control_plane/models/countedconfiggroup.py +20 -0
- cribl_control_plane/models/countedcribllakedataset.py +20 -0
- cribl_control_plane/models/counteddistributedsummary.py +20 -0
- cribl_control_plane/models/countedgitcommitsummary.py +20 -0
- cribl_control_plane/models/countedgitcountresult.py +20 -0
- cribl_control_plane/models/{createinputop.py → countedgitdiffresult.py} +5 -9
- cribl_control_plane/models/countedgitfilesresponse.py +20 -0
- cribl_control_plane/models/{getversioninfoop.py → countedgitinfo.py} +2 -6
- cribl_control_plane/models/countedgitlogresult.py +20 -0
- cribl_control_plane/models/countedgitrevertresult.py +20 -0
- cribl_control_plane/models/countedgitshowresult.py +20 -0
- cribl_control_plane/models/countedgitstatusresult.py +20 -0
- cribl_control_plane/models/{listinputop.py → countedinput.py} +2 -6
- cribl_control_plane/models/countedinputsplunkhec.py +20 -0
- cribl_control_plane/models/countedjobinfo.py +20 -0
- cribl_control_plane/models/countedmasterworkerentry.py +20 -0
- cribl_control_plane/models/countednumber.py +19 -0
- cribl_control_plane/models/countedobject.py +19 -0
- cribl_control_plane/models/{listoutputop.py → countedoutput.py} +2 -6
- cribl_control_plane/models/{createoutputop.py → countedoutputsamplesresponse.py} +5 -9
- cribl_control_plane/models/countedoutputtestresponse.py +20 -0
- cribl_control_plane/models/countedpackinfo.py +20 -0
- cribl_control_plane/models/{createpacksop.py → countedpackinstallinfo.py} +2 -6
- cribl_control_plane/models/{listpipelineop.py → countedpipeline.py} +2 -6
- cribl_control_plane/models/{listroutesop.py → countedroutes.py} +2 -6
- cribl_control_plane/models/{createversionpushop.py → countedstring.py} +2 -6
- cribl_control_plane/models/countedteamaccesscontrollist.py +20 -0
- cribl_control_plane/models/counteduseraccesscontrollist.py +20 -0
- cribl_control_plane/models/createauthloginop.py +18 -0
- cribl_control_plane/models/createconfiggroupbyproductop.py +1 -20
- cribl_control_plane/models/createcribllakedatasetbylakeidop.py +1 -19
- cribl_control_plane/models/createinputhectokenbyidop.py +1 -20
- cribl_control_plane/models/createoutputtestbyidop.py +1 -20
- cribl_control_plane/models/createroutesappendbyidop.py +2 -20
- cribl_control_plane/models/createversioncommitop.py +1 -19
- cribl_control_plane/models/createversionrevertop.py +1 -19
- cribl_control_plane/models/createversionundoop.py +1 -18
- cribl_control_plane/models/criblevent.py +15 -0
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +1 -20
- cribl_control_plane/models/deletecribllakedatasetbylakeidandidop.py +1 -20
- cribl_control_plane/models/deleteinputbyidop.py +1 -20
- cribl_control_plane/models/deleteoutputbyidop.py +1 -20
- cribl_control_plane/models/deleteoutputpqbyidop.py +1 -19
- cribl_control_plane/models/deletepacksbyidop.py +1 -20
- cribl_control_plane/models/deletepipelinebyidop.py +1 -20
- cribl_control_plane/models/difffiles.py +5 -46
- cribl_control_plane/models/diffline.py +26 -0
- cribl_control_plane/models/difflinecontext.py +28 -0
- cribl_control_plane/models/difflinedelete.py +25 -0
- cribl_control_plane/models/difflineinsert.py +25 -0
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +1 -19
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +1 -19
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -19
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +1 -19
- cribl_control_plane/models/getcribllakedatasetbylakeidandidop.py +1 -20
- cribl_control_plane/models/getcribllakedatasetbylakeidop.py +2 -79
- cribl_control_plane/models/getinputbyidop.py +1 -20
- cribl_control_plane/models/getmasterworkerentryop.py +1 -18
- cribl_control_plane/models/getoutputbyidop.py +1 -20
- cribl_control_plane/models/getoutputpqbyidop.py +1 -20
- cribl_control_plane/models/getoutputsamplesbyidop.py +1 -20
- cribl_control_plane/models/getpacksbyidop.py +1 -20
- cribl_control_plane/models/getpacksop.py +1 -19
- cribl_control_plane/models/getpipelinebyidop.py +1 -20
- cribl_control_plane/models/getroutesbyidop.py +1 -20
- cribl_control_plane/models/getsummaryop.py +1 -19
- cribl_control_plane/models/getversioncountop.py +1 -19
- cribl_control_plane/models/getversiondiffop.py +1 -19
- cribl_control_plane/models/getversionfilesop.py +1 -19
- cribl_control_plane/models/getversionop.py +1 -19
- cribl_control_plane/models/getversionshowop.py +1 -19
- cribl_control_plane/models/getversionstatusop.py +1 -19
- cribl_control_plane/models/groupcreaterequest.py +1 -0
- cribl_control_plane/models/hbcriblinfo.py +3 -4
- cribl_control_plane/models/heartbeatmetadata.py +3 -11
- cribl_control_plane/models/inputcloudflarehec.py +9 -4
- cribl_control_plane/models/inputconfluentcloud.py +4 -2
- cribl_control_plane/models/inputcribllakehttp.py +15 -0
- cribl_control_plane/models/inputkafka.py +4 -2
- cribl_control_plane/models/inputmsk.py +4 -2
- cribl_control_plane/models/inputsnmp.py +33 -7
- cribl_control_plane/models/inputsplunkhec.py +8 -4
- cribl_control_plane/models/inputwef.py +23 -0
- cribl_control_plane/models/inputwiz.py +95 -0
- cribl_control_plane/models/inputzscalerhec.py +8 -4
- cribl_control_plane/models/listconfiggroupbyproductop.py +1 -19
- cribl_control_plane/models/listmasterworkerentryop.py +1 -19
- cribl_control_plane/models/masterworkerentry.py +3 -11
- cribl_control_plane/models/nodeprovidedinfo.py +2 -10
- cribl_control_plane/models/outputnetflow.py +0 -7
- cribl_control_plane/models/outputsplunklb.py +10 -0
- cribl_control_plane/models/runnablejobcollection.py +0 -56
- cribl_control_plane/models/runnablejobexecutor.py +0 -28
- cribl_control_plane/models/runnablejobscheduledsearch.py +0 -28
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +1 -19
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +1 -20
- cribl_control_plane/models/updatecribllakedatasetbylakeidandidop.py +1 -20
- cribl_control_plane/models/updatehectokenrequest.py +1 -7
- cribl_control_plane/models/updateinputbyidop.py +1 -19
- cribl_control_plane/models/updateinputhectokenbyidandtokenop.py +1 -20
- cribl_control_plane/models/updateoutputbyidop.py +1 -19
- cribl_control_plane/models/updatepacksbyidop.py +1 -20
- cribl_control_plane/models/updatepipelinebyidop.py +1 -19
- cribl_control_plane/models/updateroutesbyidop.py +1 -19
- cribl_control_plane/nodes.py +28 -20
- cribl_control_plane/packs.py +80 -32
- cribl_control_plane/pipelines.py +70 -30
- cribl_control_plane/routes_sdk.py +56 -28
- cribl_control_plane/samples.py +28 -16
- cribl_control_plane/sources.py +70 -30
- cribl_control_plane/statuses.py +14 -6
- cribl_control_plane/summaries.py +14 -6
- cribl_control_plane/teams.py +14 -6
- cribl_control_plane/tokens.py +26 -8
- cribl_control_plane/versions_configs.py +14 -6
- {cribl_control_plane-0.4.0a12.dist-info → cribl_control_plane-0.4.0b12.dist-info}/METADATA +23 -23
- {cribl_control_plane-0.4.0a12.dist-info → cribl_control_plane-0.4.0b12.dist-info}/RECORD +135 -113
- cribl_control_plane/models/createpipelineop.py +0 -24
- cribl_control_plane/models/lookupversions.py +0 -13
- {cribl_control_plane-0.4.0a12.dist-info → cribl_control_plane-0.4.0b12.dist-info}/WHEEL +0 -0
- {cribl_control_plane-0.4.0a12.dist-info → cribl_control_plane-0.4.0b12.dist-info}/licenses/LICENSE +0 -0
|
@@ -39,7 +39,7 @@ class LakeDatasets(BaseSDK):
|
|
|
39
39
|
server_url: Optional[str] = None,
|
|
40
40
|
timeout_ms: Optional[int] = None,
|
|
41
41
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
42
|
-
) -> models.
|
|
42
|
+
) -> models.CountedCriblLakeDataset:
|
|
43
43
|
r"""Create a Lake Dataset
|
|
44
44
|
|
|
45
45
|
Create a new Lake Dataset in the specified Lake.
|
|
@@ -125,10 +125,14 @@ class LakeDatasets(BaseSDK):
|
|
|
125
125
|
if retries == UNSET:
|
|
126
126
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
127
127
|
retries = self.sdk_configuration.retry_config
|
|
128
|
+
else:
|
|
129
|
+
retries = utils.RetryConfig(
|
|
130
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
131
|
+
)
|
|
128
132
|
|
|
129
133
|
retry_config = None
|
|
130
134
|
if isinstance(retries, utils.RetryConfig):
|
|
131
|
-
retry_config = (retries, ["429"
|
|
135
|
+
retry_config = (retries, ["429"])
|
|
132
136
|
|
|
133
137
|
http_res = self.do_request(
|
|
134
138
|
hook_ctx=HookContext(
|
|
@@ -147,9 +151,7 @@ class LakeDatasets(BaseSDK):
|
|
|
147
151
|
|
|
148
152
|
response_data: Any = None
|
|
149
153
|
if utils.match_response(http_res, "200", "application/json"):
|
|
150
|
-
return unmarshal_json_response(
|
|
151
|
-
models.CreateCriblLakeDatasetByLakeIDResponse, http_res
|
|
152
|
-
)
|
|
154
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
153
155
|
if utils.match_response(http_res, "500", "application/json"):
|
|
154
156
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
155
157
|
raise errors.Error(response_data, http_res)
|
|
@@ -191,7 +193,7 @@ class LakeDatasets(BaseSDK):
|
|
|
191
193
|
server_url: Optional[str] = None,
|
|
192
194
|
timeout_ms: Optional[int] = None,
|
|
193
195
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
194
|
-
) -> models.
|
|
196
|
+
) -> models.CountedCriblLakeDataset:
|
|
195
197
|
r"""Create a Lake Dataset
|
|
196
198
|
|
|
197
199
|
Create a new Lake Dataset in the specified Lake.
|
|
@@ -277,10 +279,14 @@ class LakeDatasets(BaseSDK):
|
|
|
277
279
|
if retries == UNSET:
|
|
278
280
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
279
281
|
retries = self.sdk_configuration.retry_config
|
|
282
|
+
else:
|
|
283
|
+
retries = utils.RetryConfig(
|
|
284
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
285
|
+
)
|
|
280
286
|
|
|
281
287
|
retry_config = None
|
|
282
288
|
if isinstance(retries, utils.RetryConfig):
|
|
283
|
-
retry_config = (retries, ["429"
|
|
289
|
+
retry_config = (retries, ["429"])
|
|
284
290
|
|
|
285
291
|
http_res = await self.do_request_async(
|
|
286
292
|
hook_ctx=HookContext(
|
|
@@ -299,9 +305,7 @@ class LakeDatasets(BaseSDK):
|
|
|
299
305
|
|
|
300
306
|
response_data: Any = None
|
|
301
307
|
if utils.match_response(http_res, "200", "application/json"):
|
|
302
|
-
return unmarshal_json_response(
|
|
303
|
-
models.CreateCriblLakeDatasetByLakeIDResponse, http_res
|
|
304
|
-
)
|
|
308
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
305
309
|
if utils.match_response(http_res, "500", "application/json"):
|
|
306
310
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
307
311
|
raise errors.Error(response_data, http_res)
|
|
@@ -318,28 +322,16 @@ class LakeDatasets(BaseSDK):
|
|
|
318
322
|
self,
|
|
319
323
|
*,
|
|
320
324
|
lake_id: str,
|
|
321
|
-
storage_location_id: Optional[str] = None,
|
|
322
|
-
format_: Optional[str] = None,
|
|
323
|
-
exclude_ddss: Optional[bool] = None,
|
|
324
|
-
exclude_deleted: Optional[bool] = None,
|
|
325
|
-
exclude_internal: Optional[bool] = None,
|
|
326
|
-
exclude_byos: Optional[bool] = None,
|
|
327
325
|
retries: OptionalNullable[utils.RetryConfig] = UNSET,
|
|
328
326
|
server_url: Optional[str] = None,
|
|
329
327
|
timeout_ms: Optional[int] = None,
|
|
330
328
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
331
|
-
) -> models.
|
|
329
|
+
) -> models.CountedCriblLakeDataset:
|
|
332
330
|
r"""List all Lake Datasets
|
|
333
331
|
|
|
334
332
|
Get a list of all Lake Datasets in the specified Lake.
|
|
335
333
|
|
|
336
334
|
:param lake_id: The <code>id</code> of the Lake that contains the Lake Datasets to list.
|
|
337
|
-
:param storage_location_id: Filter datasets by storage location ID. Use <code>default</code> for default storage location.
|
|
338
|
-
:param format_: Filter datasets by format. Set to <code>ddss</code> to return only DDSS datasets.
|
|
339
|
-
:param exclude_ddss: Exclude DDSS format datasets from the response.
|
|
340
|
-
:param exclude_deleted: Exclude deleted datasets from the response.
|
|
341
|
-
:param exclude_internal: Exclude internal datasets (those with IDs starting with <code>cribl_</code>) from the response.
|
|
342
|
-
:param exclude_byos: Exclude BYOS (Bring Your Own Storage) datasets from the response.
|
|
343
335
|
:param retries: Override the default retry configuration for this method
|
|
344
336
|
:param server_url: Override the default server URL for this method
|
|
345
337
|
:param timeout_ms: Override the default request timeout configuration for this method in milliseconds
|
|
@@ -357,12 +349,6 @@ class LakeDatasets(BaseSDK):
|
|
|
357
349
|
|
|
358
350
|
request = models.GetCriblLakeDatasetByLakeIDRequest(
|
|
359
351
|
lake_id=lake_id,
|
|
360
|
-
storage_location_id=storage_location_id,
|
|
361
|
-
format_=format_,
|
|
362
|
-
exclude_ddss=exclude_ddss,
|
|
363
|
-
exclude_deleted=exclude_deleted,
|
|
364
|
-
exclude_internal=exclude_internal,
|
|
365
|
-
exclude_byos=exclude_byos,
|
|
366
352
|
)
|
|
367
353
|
|
|
368
354
|
req = self._build_request(
|
|
@@ -385,10 +371,14 @@ class LakeDatasets(BaseSDK):
|
|
|
385
371
|
if retries == UNSET:
|
|
386
372
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
387
373
|
retries = self.sdk_configuration.retry_config
|
|
374
|
+
else:
|
|
375
|
+
retries = utils.RetryConfig(
|
|
376
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
377
|
+
)
|
|
388
378
|
|
|
389
379
|
retry_config = None
|
|
390
380
|
if isinstance(retries, utils.RetryConfig):
|
|
391
|
-
retry_config = (retries, ["429"
|
|
381
|
+
retry_config = (retries, ["429"])
|
|
392
382
|
|
|
393
383
|
http_res = self.do_request(
|
|
394
384
|
hook_ctx=HookContext(
|
|
@@ -407,9 +397,7 @@ class LakeDatasets(BaseSDK):
|
|
|
407
397
|
|
|
408
398
|
response_data: Any = None
|
|
409
399
|
if utils.match_response(http_res, "200", "application/json"):
|
|
410
|
-
return unmarshal_json_response(
|
|
411
|
-
models.GetCriblLakeDatasetByLakeIDResponse, http_res
|
|
412
|
-
)
|
|
400
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
413
401
|
if utils.match_response(http_res, "500", "application/json"):
|
|
414
402
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
415
403
|
raise errors.Error(response_data, http_res)
|
|
@@ -426,28 +414,16 @@ class LakeDatasets(BaseSDK):
|
|
|
426
414
|
self,
|
|
427
415
|
*,
|
|
428
416
|
lake_id: str,
|
|
429
|
-
storage_location_id: Optional[str] = None,
|
|
430
|
-
format_: Optional[str] = None,
|
|
431
|
-
exclude_ddss: Optional[bool] = None,
|
|
432
|
-
exclude_deleted: Optional[bool] = None,
|
|
433
|
-
exclude_internal: Optional[bool] = None,
|
|
434
|
-
exclude_byos: Optional[bool] = None,
|
|
435
417
|
retries: OptionalNullable[utils.RetryConfig] = UNSET,
|
|
436
418
|
server_url: Optional[str] = None,
|
|
437
419
|
timeout_ms: Optional[int] = None,
|
|
438
420
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
439
|
-
) -> models.
|
|
421
|
+
) -> models.CountedCriblLakeDataset:
|
|
440
422
|
r"""List all Lake Datasets
|
|
441
423
|
|
|
442
424
|
Get a list of all Lake Datasets in the specified Lake.
|
|
443
425
|
|
|
444
426
|
:param lake_id: The <code>id</code> of the Lake that contains the Lake Datasets to list.
|
|
445
|
-
:param storage_location_id: Filter datasets by storage location ID. Use <code>default</code> for default storage location.
|
|
446
|
-
:param format_: Filter datasets by format. Set to <code>ddss</code> to return only DDSS datasets.
|
|
447
|
-
:param exclude_ddss: Exclude DDSS format datasets from the response.
|
|
448
|
-
:param exclude_deleted: Exclude deleted datasets from the response.
|
|
449
|
-
:param exclude_internal: Exclude internal datasets (those with IDs starting with <code>cribl_</code>) from the response.
|
|
450
|
-
:param exclude_byos: Exclude BYOS (Bring Your Own Storage) datasets from the response.
|
|
451
427
|
:param retries: Override the default retry configuration for this method
|
|
452
428
|
:param server_url: Override the default server URL for this method
|
|
453
429
|
:param timeout_ms: Override the default request timeout configuration for this method in milliseconds
|
|
@@ -465,12 +441,6 @@ class LakeDatasets(BaseSDK):
|
|
|
465
441
|
|
|
466
442
|
request = models.GetCriblLakeDatasetByLakeIDRequest(
|
|
467
443
|
lake_id=lake_id,
|
|
468
|
-
storage_location_id=storage_location_id,
|
|
469
|
-
format_=format_,
|
|
470
|
-
exclude_ddss=exclude_ddss,
|
|
471
|
-
exclude_deleted=exclude_deleted,
|
|
472
|
-
exclude_internal=exclude_internal,
|
|
473
|
-
exclude_byos=exclude_byos,
|
|
474
444
|
)
|
|
475
445
|
|
|
476
446
|
req = self._build_request_async(
|
|
@@ -493,10 +463,14 @@ class LakeDatasets(BaseSDK):
|
|
|
493
463
|
if retries == UNSET:
|
|
494
464
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
495
465
|
retries = self.sdk_configuration.retry_config
|
|
466
|
+
else:
|
|
467
|
+
retries = utils.RetryConfig(
|
|
468
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
469
|
+
)
|
|
496
470
|
|
|
497
471
|
retry_config = None
|
|
498
472
|
if isinstance(retries, utils.RetryConfig):
|
|
499
|
-
retry_config = (retries, ["429"
|
|
473
|
+
retry_config = (retries, ["429"])
|
|
500
474
|
|
|
501
475
|
http_res = await self.do_request_async(
|
|
502
476
|
hook_ctx=HookContext(
|
|
@@ -515,9 +489,7 @@ class LakeDatasets(BaseSDK):
|
|
|
515
489
|
|
|
516
490
|
response_data: Any = None
|
|
517
491
|
if utils.match_response(http_res, "200", "application/json"):
|
|
518
|
-
return unmarshal_json_response(
|
|
519
|
-
models.GetCriblLakeDatasetByLakeIDResponse, http_res
|
|
520
|
-
)
|
|
492
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
521
493
|
if utils.match_response(http_res, "500", "application/json"):
|
|
522
494
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
523
495
|
raise errors.Error(response_data, http_res)
|
|
@@ -539,7 +511,7 @@ class LakeDatasets(BaseSDK):
|
|
|
539
511
|
server_url: Optional[str] = None,
|
|
540
512
|
timeout_ms: Optional[int] = None,
|
|
541
513
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
542
|
-
) -> models.
|
|
514
|
+
) -> models.CountedCriblLakeDataset:
|
|
543
515
|
r"""Delete a Lake Dataset
|
|
544
516
|
|
|
545
517
|
Delete the specified Lake Dataset in the specified Lake
|
|
@@ -586,10 +558,14 @@ class LakeDatasets(BaseSDK):
|
|
|
586
558
|
if retries == UNSET:
|
|
587
559
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
588
560
|
retries = self.sdk_configuration.retry_config
|
|
561
|
+
else:
|
|
562
|
+
retries = utils.RetryConfig(
|
|
563
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
564
|
+
)
|
|
589
565
|
|
|
590
566
|
retry_config = None
|
|
591
567
|
if isinstance(retries, utils.RetryConfig):
|
|
592
|
-
retry_config = (retries, ["429"
|
|
568
|
+
retry_config = (retries, ["429"])
|
|
593
569
|
|
|
594
570
|
http_res = self.do_request(
|
|
595
571
|
hook_ctx=HookContext(
|
|
@@ -608,9 +584,7 @@ class LakeDatasets(BaseSDK):
|
|
|
608
584
|
|
|
609
585
|
response_data: Any = None
|
|
610
586
|
if utils.match_response(http_res, "200", "application/json"):
|
|
611
|
-
return unmarshal_json_response(
|
|
612
|
-
models.DeleteCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
613
|
-
)
|
|
587
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
614
588
|
if utils.match_response(http_res, "500", "application/json"):
|
|
615
589
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
616
590
|
raise errors.Error(response_data, http_res)
|
|
@@ -632,7 +606,7 @@ class LakeDatasets(BaseSDK):
|
|
|
632
606
|
server_url: Optional[str] = None,
|
|
633
607
|
timeout_ms: Optional[int] = None,
|
|
634
608
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
635
|
-
) -> models.
|
|
609
|
+
) -> models.CountedCriblLakeDataset:
|
|
636
610
|
r"""Delete a Lake Dataset
|
|
637
611
|
|
|
638
612
|
Delete the specified Lake Dataset in the specified Lake
|
|
@@ -679,10 +653,14 @@ class LakeDatasets(BaseSDK):
|
|
|
679
653
|
if retries == UNSET:
|
|
680
654
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
681
655
|
retries = self.sdk_configuration.retry_config
|
|
656
|
+
else:
|
|
657
|
+
retries = utils.RetryConfig(
|
|
658
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
659
|
+
)
|
|
682
660
|
|
|
683
661
|
retry_config = None
|
|
684
662
|
if isinstance(retries, utils.RetryConfig):
|
|
685
|
-
retry_config = (retries, ["429"
|
|
663
|
+
retry_config = (retries, ["429"])
|
|
686
664
|
|
|
687
665
|
http_res = await self.do_request_async(
|
|
688
666
|
hook_ctx=HookContext(
|
|
@@ -701,9 +679,7 @@ class LakeDatasets(BaseSDK):
|
|
|
701
679
|
|
|
702
680
|
response_data: Any = None
|
|
703
681
|
if utils.match_response(http_res, "200", "application/json"):
|
|
704
|
-
return unmarshal_json_response(
|
|
705
|
-
models.DeleteCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
706
|
-
)
|
|
682
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
707
683
|
if utils.match_response(http_res, "500", "application/json"):
|
|
708
684
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
709
685
|
raise errors.Error(response_data, http_res)
|
|
@@ -725,7 +701,7 @@ class LakeDatasets(BaseSDK):
|
|
|
725
701
|
server_url: Optional[str] = None,
|
|
726
702
|
timeout_ms: Optional[int] = None,
|
|
727
703
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
728
|
-
) -> models.
|
|
704
|
+
) -> models.CountedCriblLakeDataset:
|
|
729
705
|
r"""Get a Lake Dataset
|
|
730
706
|
|
|
731
707
|
Get the specified Lake Dataset in the specified Lake.
|
|
@@ -772,10 +748,14 @@ class LakeDatasets(BaseSDK):
|
|
|
772
748
|
if retries == UNSET:
|
|
773
749
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
774
750
|
retries = self.sdk_configuration.retry_config
|
|
751
|
+
else:
|
|
752
|
+
retries = utils.RetryConfig(
|
|
753
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
754
|
+
)
|
|
775
755
|
|
|
776
756
|
retry_config = None
|
|
777
757
|
if isinstance(retries, utils.RetryConfig):
|
|
778
|
-
retry_config = (retries, ["429"
|
|
758
|
+
retry_config = (retries, ["429"])
|
|
779
759
|
|
|
780
760
|
http_res = self.do_request(
|
|
781
761
|
hook_ctx=HookContext(
|
|
@@ -794,9 +774,7 @@ class LakeDatasets(BaseSDK):
|
|
|
794
774
|
|
|
795
775
|
response_data: Any = None
|
|
796
776
|
if utils.match_response(http_res, "200", "application/json"):
|
|
797
|
-
return unmarshal_json_response(
|
|
798
|
-
models.GetCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
799
|
-
)
|
|
777
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
800
778
|
if utils.match_response(http_res, "500", "application/json"):
|
|
801
779
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
802
780
|
raise errors.Error(response_data, http_res)
|
|
@@ -818,7 +796,7 @@ class LakeDatasets(BaseSDK):
|
|
|
818
796
|
server_url: Optional[str] = None,
|
|
819
797
|
timeout_ms: Optional[int] = None,
|
|
820
798
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
821
|
-
) -> models.
|
|
799
|
+
) -> models.CountedCriblLakeDataset:
|
|
822
800
|
r"""Get a Lake Dataset
|
|
823
801
|
|
|
824
802
|
Get the specified Lake Dataset in the specified Lake.
|
|
@@ -865,10 +843,14 @@ class LakeDatasets(BaseSDK):
|
|
|
865
843
|
if retries == UNSET:
|
|
866
844
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
867
845
|
retries = self.sdk_configuration.retry_config
|
|
846
|
+
else:
|
|
847
|
+
retries = utils.RetryConfig(
|
|
848
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
849
|
+
)
|
|
868
850
|
|
|
869
851
|
retry_config = None
|
|
870
852
|
if isinstance(retries, utils.RetryConfig):
|
|
871
|
-
retry_config = (retries, ["429"
|
|
853
|
+
retry_config = (retries, ["429"])
|
|
872
854
|
|
|
873
855
|
http_res = await self.do_request_async(
|
|
874
856
|
hook_ctx=HookContext(
|
|
@@ -887,9 +869,7 @@ class LakeDatasets(BaseSDK):
|
|
|
887
869
|
|
|
888
870
|
response_data: Any = None
|
|
889
871
|
if utils.match_response(http_res, "200", "application/json"):
|
|
890
|
-
return unmarshal_json_response(
|
|
891
|
-
models.GetCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
892
|
-
)
|
|
872
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
893
873
|
if utils.match_response(http_res, "500", "application/json"):
|
|
894
874
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
895
875
|
raise errors.Error(response_data, http_res)
|
|
@@ -932,7 +912,7 @@ class LakeDatasets(BaseSDK):
|
|
|
932
912
|
server_url: Optional[str] = None,
|
|
933
913
|
timeout_ms: Optional[int] = None,
|
|
934
914
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
935
|
-
) -> models.
|
|
915
|
+
) -> models.CountedCriblLakeDataset:
|
|
936
916
|
r"""Update a Lake Dataset
|
|
937
917
|
|
|
938
918
|
Update the specified Lake Dataset in the specified Lake.
|
|
@@ -1020,10 +1000,14 @@ class LakeDatasets(BaseSDK):
|
|
|
1020
1000
|
if retries == UNSET:
|
|
1021
1001
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
1022
1002
|
retries = self.sdk_configuration.retry_config
|
|
1003
|
+
else:
|
|
1004
|
+
retries = utils.RetryConfig(
|
|
1005
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
1006
|
+
)
|
|
1023
1007
|
|
|
1024
1008
|
retry_config = None
|
|
1025
1009
|
if isinstance(retries, utils.RetryConfig):
|
|
1026
|
-
retry_config = (retries, ["429"
|
|
1010
|
+
retry_config = (retries, ["429"])
|
|
1027
1011
|
|
|
1028
1012
|
http_res = self.do_request(
|
|
1029
1013
|
hook_ctx=HookContext(
|
|
@@ -1042,9 +1026,7 @@ class LakeDatasets(BaseSDK):
|
|
|
1042
1026
|
|
|
1043
1027
|
response_data: Any = None
|
|
1044
1028
|
if utils.match_response(http_res, "200", "application/json"):
|
|
1045
|
-
return unmarshal_json_response(
|
|
1046
|
-
models.UpdateCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
1047
|
-
)
|
|
1029
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
1048
1030
|
if utils.match_response(http_res, "500", "application/json"):
|
|
1049
1031
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
1050
1032
|
raise errors.Error(response_data, http_res)
|
|
@@ -1087,7 +1069,7 @@ class LakeDatasets(BaseSDK):
|
|
|
1087
1069
|
server_url: Optional[str] = None,
|
|
1088
1070
|
timeout_ms: Optional[int] = None,
|
|
1089
1071
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
1090
|
-
) -> models.
|
|
1072
|
+
) -> models.CountedCriblLakeDataset:
|
|
1091
1073
|
r"""Update a Lake Dataset
|
|
1092
1074
|
|
|
1093
1075
|
Update the specified Lake Dataset in the specified Lake.
|
|
@@ -1175,10 +1157,14 @@ class LakeDatasets(BaseSDK):
|
|
|
1175
1157
|
if retries == UNSET:
|
|
1176
1158
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
1177
1159
|
retries = self.sdk_configuration.retry_config
|
|
1160
|
+
else:
|
|
1161
|
+
retries = utils.RetryConfig(
|
|
1162
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
1163
|
+
)
|
|
1178
1164
|
|
|
1179
1165
|
retry_config = None
|
|
1180
1166
|
if isinstance(retries, utils.RetryConfig):
|
|
1181
|
-
retry_config = (retries, ["429"
|
|
1167
|
+
retry_config = (retries, ["429"])
|
|
1182
1168
|
|
|
1183
1169
|
http_res = await self.do_request_async(
|
|
1184
1170
|
hook_ctx=HookContext(
|
|
@@ -1197,9 +1183,7 @@ class LakeDatasets(BaseSDK):
|
|
|
1197
1183
|
|
|
1198
1184
|
response_data: Any = None
|
|
1199
1185
|
if utils.match_response(http_res, "200", "application/json"):
|
|
1200
|
-
return unmarshal_json_response(
|
|
1201
|
-
models.UpdateCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
1202
|
-
)
|
|
1186
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
1203
1187
|
if utils.match_response(http_res, "500", "application/json"):
|
|
1204
1188
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
1205
1189
|
raise errors.Error(response_data, http_res)
|