cribl-control-plane 0.4.0a6__py3-none-any.whl → 0.4.0b12__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +4 -4
- cribl_control_plane/acl.py +16 -6
- cribl_control_plane/basesdk.py +6 -0
- cribl_control_plane/branches.py +28 -8
- cribl_control_plane/commits.py +112 -42
- cribl_control_plane/commits_files.py +32 -12
- cribl_control_plane/configs_versions.py +16 -10
- cribl_control_plane/destinations.py +80 -30
- cribl_control_plane/destinations_pq.py +32 -12
- cribl_control_plane/groups_sdk.py +96 -60
- cribl_control_plane/health.py +12 -2
- cribl_control_plane/hectokens.py +32 -20
- cribl_control_plane/lakedatasets.py +80 -50
- cribl_control_plane/models/__init__.py +266 -438
- cribl_control_plane/models/configgroup.py +1 -0
- cribl_control_plane/models/{getversionbranchop.py → countedbranchinfo.py} +2 -6
- cribl_control_plane/models/countedconfiggroup.py +20 -0
- cribl_control_plane/models/countedcribllakedataset.py +20 -0
- cribl_control_plane/models/counteddistributedsummary.py +20 -0
- cribl_control_plane/models/countedgitcommitsummary.py +20 -0
- cribl_control_plane/models/countedgitcountresult.py +20 -0
- cribl_control_plane/models/{createinputop.py → countedgitdiffresult.py} +5 -9
- cribl_control_plane/models/countedgitfilesresponse.py +20 -0
- cribl_control_plane/models/{getversioninfoop.py → countedgitinfo.py} +2 -6
- cribl_control_plane/models/countedgitlogresult.py +20 -0
- cribl_control_plane/models/countedgitrevertresult.py +20 -0
- cribl_control_plane/models/countedgitshowresult.py +20 -0
- cribl_control_plane/models/countedgitstatusresult.py +20 -0
- cribl_control_plane/models/{listinputop.py → countedinput.py} +2 -6
- cribl_control_plane/models/countedinputsplunkhec.py +20 -0
- cribl_control_plane/models/countedjobinfo.py +20 -0
- cribl_control_plane/models/countedmasterworkerentry.py +20 -0
- cribl_control_plane/models/countednumber.py +19 -0
- cribl_control_plane/models/countedobject.py +19 -0
- cribl_control_plane/models/{listoutputop.py → countedoutput.py} +2 -6
- cribl_control_plane/models/{createoutputop.py → countedoutputsamplesresponse.py} +5 -9
- cribl_control_plane/models/countedoutputtestresponse.py +20 -0
- cribl_control_plane/models/countedpackinfo.py +20 -0
- cribl_control_plane/models/{createpacksop.py → countedpackinstallinfo.py} +2 -6
- cribl_control_plane/models/{listpipelineop.py → countedpipeline.py} +2 -6
- cribl_control_plane/models/{listroutesop.py → countedroutes.py} +2 -6
- cribl_control_plane/models/{createversionpushop.py → countedstring.py} +2 -6
- cribl_control_plane/models/countedteamaccesscontrollist.py +20 -0
- cribl_control_plane/models/counteduseraccesscontrollist.py +20 -0
- cribl_control_plane/models/createauthloginop.py +18 -0
- cribl_control_plane/models/createconfiggroupbyproductop.py +1 -20
- cribl_control_plane/models/createcribllakedatasetbylakeidop.py +1 -19
- cribl_control_plane/models/createinputhectokenbyidop.py +1 -20
- cribl_control_plane/models/createoutputtestbyidop.py +1 -20
- cribl_control_plane/models/createroutesappendbyidop.py +2 -20
- cribl_control_plane/models/createversioncommitop.py +1 -19
- cribl_control_plane/models/createversionrevertop.py +1 -19
- cribl_control_plane/models/createversionundoop.py +1 -18
- cribl_control_plane/models/criblevent.py +15 -0
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +1 -20
- cribl_control_plane/models/deletecribllakedatasetbylakeidandidop.py +1 -20
- cribl_control_plane/models/deleteinputbyidop.py +1 -20
- cribl_control_plane/models/deleteoutputbyidop.py +1 -20
- cribl_control_plane/models/deleteoutputpqbyidop.py +1 -19
- cribl_control_plane/models/deletepacksbyidop.py +1 -20
- cribl_control_plane/models/deletepipelinebyidop.py +1 -20
- cribl_control_plane/models/difffiles.py +5 -46
- cribl_control_plane/models/diffline.py +26 -0
- cribl_control_plane/models/difflinecontext.py +28 -0
- cribl_control_plane/models/difflinedelete.py +25 -0
- cribl_control_plane/models/difflineinsert.py +25 -0
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +1 -19
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +1 -19
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -19
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +1 -19
- cribl_control_plane/models/getcribllakedatasetbylakeidandidop.py +1 -20
- cribl_control_plane/models/getcribllakedatasetbylakeidop.py +1 -20
- cribl_control_plane/models/getinputbyidop.py +1 -20
- cribl_control_plane/models/getmasterworkerentryop.py +1 -18
- cribl_control_plane/models/getoutputbyidop.py +1 -20
- cribl_control_plane/models/getoutputpqbyidop.py +1 -20
- cribl_control_plane/models/getoutputsamplesbyidop.py +1 -20
- cribl_control_plane/models/getpacksbyidop.py +1 -20
- cribl_control_plane/models/getpacksop.py +1 -19
- cribl_control_plane/models/getpipelinebyidop.py +1 -20
- cribl_control_plane/models/getroutesbyidop.py +1 -20
- cribl_control_plane/models/getsummaryop.py +1 -19
- cribl_control_plane/models/getversioncountop.py +1 -19
- cribl_control_plane/models/getversiondiffop.py +1 -19
- cribl_control_plane/models/getversionfilesop.py +1 -19
- cribl_control_plane/models/getversionop.py +1 -19
- cribl_control_plane/models/getversionshowop.py +1 -19
- cribl_control_plane/models/getversionstatusop.py +1 -19
- cribl_control_plane/models/groupcreaterequest.py +1 -0
- cribl_control_plane/models/hbcriblinfo.py +3 -4
- cribl_control_plane/models/heartbeatmetadata.py +3 -11
- cribl_control_plane/models/inputcloudflarehec.py +9 -4
- cribl_control_plane/models/inputconfluentcloud.py +4 -2
- cribl_control_plane/models/inputcribllakehttp.py +15 -0
- cribl_control_plane/models/inputkafka.py +4 -2
- cribl_control_plane/models/inputmsk.py +4 -2
- cribl_control_plane/models/inputsnmp.py +33 -7
- cribl_control_plane/models/inputsplunkhec.py +8 -4
- cribl_control_plane/models/inputwef.py +23 -0
- cribl_control_plane/models/inputwiz.py +95 -0
- cribl_control_plane/models/inputzscalerhec.py +8 -4
- cribl_control_plane/models/listconfiggroupbyproductop.py +1 -19
- cribl_control_plane/models/listmasterworkerentryop.py +1 -19
- cribl_control_plane/models/masterworkerentry.py +3 -11
- cribl_control_plane/models/nodeprovidedinfo.py +2 -10
- cribl_control_plane/models/outputsplunklb.py +10 -0
- cribl_control_plane/models/runnablejobcollection.py +0 -56
- cribl_control_plane/models/runnablejobexecutor.py +0 -28
- cribl_control_plane/models/runnablejobscheduledsearch.py +0 -28
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +1 -19
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +1 -20
- cribl_control_plane/models/updatecribllakedatasetbylakeidandidop.py +1 -20
- cribl_control_plane/models/updateinputbyidop.py +1 -19
- cribl_control_plane/models/updateinputhectokenbyidandtokenop.py +1 -20
- cribl_control_plane/models/updateoutputbyidop.py +1 -19
- cribl_control_plane/models/updatepacksbyidop.py +1 -20
- cribl_control_plane/models/updatepipelinebyidop.py +1 -19
- cribl_control_plane/models/updateroutesbyidop.py +1 -19
- cribl_control_plane/nodes.py +32 -20
- cribl_control_plane/packs.py +92 -32
- cribl_control_plane/pipelines.py +80 -30
- cribl_control_plane/routes_sdk.py +64 -28
- cribl_control_plane/samples.py +32 -16
- cribl_control_plane/sources.py +80 -30
- cribl_control_plane/statuses.py +16 -6
- cribl_control_plane/summaries.py +16 -6
- cribl_control_plane/teams.py +16 -6
- cribl_control_plane/tokens.py +28 -8
- cribl_control_plane/utils/forms.py +21 -10
- cribl_control_plane/utils/queryparams.py +14 -2
- cribl_control_plane/versions_configs.py +16 -6
- {cribl_control_plane-0.4.0a6.dist-info → cribl_control_plane-0.4.0b12.dist-info}/METADATA +1 -1
- {cribl_control_plane-0.4.0a6.dist-info → cribl_control_plane-0.4.0b12.dist-info}/RECORD +135 -113
- cribl_control_plane/models/createpipelineop.py +0 -24
- cribl_control_plane/models/lookupversions.py +0 -13
- {cribl_control_plane-0.4.0a6.dist-info → cribl_control_plane-0.4.0b12.dist-info}/WHEEL +0 -0
- {cribl_control_plane-0.4.0a6.dist-info → cribl_control_plane-0.4.0b12.dist-info}/licenses/LICENSE +0 -0
cribl_control_plane/health.py
CHANGED
|
@@ -49,16 +49,21 @@ class Health(BaseSDK):
|
|
|
49
49
|
user_agent_header="user-agent",
|
|
50
50
|
accept_header_value="application/json",
|
|
51
51
|
http_headers=http_headers,
|
|
52
|
+
allow_empty_value=None,
|
|
52
53
|
timeout_ms=timeout_ms,
|
|
53
54
|
)
|
|
54
55
|
|
|
55
56
|
if retries == UNSET:
|
|
56
57
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
57
58
|
retries = self.sdk_configuration.retry_config
|
|
59
|
+
else:
|
|
60
|
+
retries = utils.RetryConfig(
|
|
61
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
62
|
+
)
|
|
58
63
|
|
|
59
64
|
retry_config = None
|
|
60
65
|
if isinstance(retries, utils.RetryConfig):
|
|
61
|
-
retry_config = (retries, ["429"
|
|
66
|
+
retry_config = (retries, ["429"])
|
|
62
67
|
|
|
63
68
|
http_res = self.do_request(
|
|
64
69
|
hook_ctx=HookContext(
|
|
@@ -131,16 +136,21 @@ class Health(BaseSDK):
|
|
|
131
136
|
user_agent_header="user-agent",
|
|
132
137
|
accept_header_value="application/json",
|
|
133
138
|
http_headers=http_headers,
|
|
139
|
+
allow_empty_value=None,
|
|
134
140
|
timeout_ms=timeout_ms,
|
|
135
141
|
)
|
|
136
142
|
|
|
137
143
|
if retries == UNSET:
|
|
138
144
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
139
145
|
retries = self.sdk_configuration.retry_config
|
|
146
|
+
else:
|
|
147
|
+
retries = utils.RetryConfig(
|
|
148
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
149
|
+
)
|
|
140
150
|
|
|
141
151
|
retry_config = None
|
|
142
152
|
if isinstance(retries, utils.RetryConfig):
|
|
143
|
-
retry_config = (retries, ["429"
|
|
153
|
+
retry_config = (retries, ["429"])
|
|
144
154
|
|
|
145
155
|
http_res = await self.do_request_async(
|
|
146
156
|
hook_ctx=HookContext(
|
cribl_control_plane/hectokens.py
CHANGED
|
@@ -27,7 +27,7 @@ class HecTokens(BaseSDK):
|
|
|
27
27
|
server_url: Optional[str] = None,
|
|
28
28
|
timeout_ms: Optional[int] = None,
|
|
29
29
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
30
|
-
) -> models.
|
|
30
|
+
) -> models.CountedInputSplunkHec:
|
|
31
31
|
r"""Add an HEC token and optional metadata to a Splunk HEC Source
|
|
32
32
|
|
|
33
33
|
Add an HEC token and optional metadata to the specified Splunk HEC Source.
|
|
@@ -84,16 +84,21 @@ class HecTokens(BaseSDK):
|
|
|
84
84
|
"json",
|
|
85
85
|
models.AddHecTokenRequest,
|
|
86
86
|
),
|
|
87
|
+
allow_empty_value=None,
|
|
87
88
|
timeout_ms=timeout_ms,
|
|
88
89
|
)
|
|
89
90
|
|
|
90
91
|
if retries == UNSET:
|
|
91
92
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
92
93
|
retries = self.sdk_configuration.retry_config
|
|
94
|
+
else:
|
|
95
|
+
retries = utils.RetryConfig(
|
|
96
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
97
|
+
)
|
|
93
98
|
|
|
94
99
|
retry_config = None
|
|
95
100
|
if isinstance(retries, utils.RetryConfig):
|
|
96
|
-
retry_config = (retries, ["429"
|
|
101
|
+
retry_config = (retries, ["429"])
|
|
97
102
|
|
|
98
103
|
http_res = self.do_request(
|
|
99
104
|
hook_ctx=HookContext(
|
|
@@ -112,9 +117,7 @@ class HecTokens(BaseSDK):
|
|
|
112
117
|
|
|
113
118
|
response_data: Any = None
|
|
114
119
|
if utils.match_response(http_res, "200", "application/json"):
|
|
115
|
-
return unmarshal_json_response(
|
|
116
|
-
models.CreateInputHecTokenByIDResponse, http_res
|
|
117
|
-
)
|
|
120
|
+
return unmarshal_json_response(models.CountedInputSplunkHec, http_res)
|
|
118
121
|
if utils.match_response(http_res, "500", "application/json"):
|
|
119
122
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
120
123
|
raise errors.Error(response_data, http_res)
|
|
@@ -144,7 +147,7 @@ class HecTokens(BaseSDK):
|
|
|
144
147
|
server_url: Optional[str] = None,
|
|
145
148
|
timeout_ms: Optional[int] = None,
|
|
146
149
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
147
|
-
) -> models.
|
|
150
|
+
) -> models.CountedInputSplunkHec:
|
|
148
151
|
r"""Add an HEC token and optional metadata to a Splunk HEC Source
|
|
149
152
|
|
|
150
153
|
Add an HEC token and optional metadata to the specified Splunk HEC Source.
|
|
@@ -201,16 +204,21 @@ class HecTokens(BaseSDK):
|
|
|
201
204
|
"json",
|
|
202
205
|
models.AddHecTokenRequest,
|
|
203
206
|
),
|
|
207
|
+
allow_empty_value=None,
|
|
204
208
|
timeout_ms=timeout_ms,
|
|
205
209
|
)
|
|
206
210
|
|
|
207
211
|
if retries == UNSET:
|
|
208
212
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
209
213
|
retries = self.sdk_configuration.retry_config
|
|
214
|
+
else:
|
|
215
|
+
retries = utils.RetryConfig(
|
|
216
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
217
|
+
)
|
|
210
218
|
|
|
211
219
|
retry_config = None
|
|
212
220
|
if isinstance(retries, utils.RetryConfig):
|
|
213
|
-
retry_config = (retries, ["429"
|
|
221
|
+
retry_config = (retries, ["429"])
|
|
214
222
|
|
|
215
223
|
http_res = await self.do_request_async(
|
|
216
224
|
hook_ctx=HookContext(
|
|
@@ -229,9 +237,7 @@ class HecTokens(BaseSDK):
|
|
|
229
237
|
|
|
230
238
|
response_data: Any = None
|
|
231
239
|
if utils.match_response(http_res, "200", "application/json"):
|
|
232
|
-
return unmarshal_json_response(
|
|
233
|
-
models.CreateInputHecTokenByIDResponse, http_res
|
|
234
|
-
)
|
|
240
|
+
return unmarshal_json_response(models.CountedInputSplunkHec, http_res)
|
|
235
241
|
if utils.match_response(http_res, "500", "application/json"):
|
|
236
242
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
237
243
|
raise errors.Error(response_data, http_res)
|
|
@@ -261,7 +267,7 @@ class HecTokens(BaseSDK):
|
|
|
261
267
|
server_url: Optional[str] = None,
|
|
262
268
|
timeout_ms: Optional[int] = None,
|
|
263
269
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
264
|
-
) -> models.
|
|
270
|
+
) -> models.CountedInputSplunkHec:
|
|
265
271
|
r"""Update metadata for an HEC token for a Splunk HEC Source
|
|
266
272
|
|
|
267
273
|
Update the metadata for the specified HEC token for the specified Splunk HEC Source.
|
|
@@ -318,16 +324,21 @@ class HecTokens(BaseSDK):
|
|
|
318
324
|
"json",
|
|
319
325
|
models.UpdateHecTokenRequest,
|
|
320
326
|
),
|
|
327
|
+
allow_empty_value=None,
|
|
321
328
|
timeout_ms=timeout_ms,
|
|
322
329
|
)
|
|
323
330
|
|
|
324
331
|
if retries == UNSET:
|
|
325
332
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
326
333
|
retries = self.sdk_configuration.retry_config
|
|
334
|
+
else:
|
|
335
|
+
retries = utils.RetryConfig(
|
|
336
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
337
|
+
)
|
|
327
338
|
|
|
328
339
|
retry_config = None
|
|
329
340
|
if isinstance(retries, utils.RetryConfig):
|
|
330
|
-
retry_config = (retries, ["429"
|
|
341
|
+
retry_config = (retries, ["429"])
|
|
331
342
|
|
|
332
343
|
http_res = self.do_request(
|
|
333
344
|
hook_ctx=HookContext(
|
|
@@ -346,9 +357,7 @@ class HecTokens(BaseSDK):
|
|
|
346
357
|
|
|
347
358
|
response_data: Any = None
|
|
348
359
|
if utils.match_response(http_res, "200", "application/json"):
|
|
349
|
-
return unmarshal_json_response(
|
|
350
|
-
models.UpdateInputHecTokenByIDAndTokenResponse, http_res
|
|
351
|
-
)
|
|
360
|
+
return unmarshal_json_response(models.CountedInputSplunkHec, http_res)
|
|
352
361
|
if utils.match_response(http_res, "500", "application/json"):
|
|
353
362
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
354
363
|
raise errors.Error(response_data, http_res)
|
|
@@ -378,7 +387,7 @@ class HecTokens(BaseSDK):
|
|
|
378
387
|
server_url: Optional[str] = None,
|
|
379
388
|
timeout_ms: Optional[int] = None,
|
|
380
389
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
381
|
-
) -> models.
|
|
390
|
+
) -> models.CountedInputSplunkHec:
|
|
382
391
|
r"""Update metadata for an HEC token for a Splunk HEC Source
|
|
383
392
|
|
|
384
393
|
Update the metadata for the specified HEC token for the specified Splunk HEC Source.
|
|
@@ -435,16 +444,21 @@ class HecTokens(BaseSDK):
|
|
|
435
444
|
"json",
|
|
436
445
|
models.UpdateHecTokenRequest,
|
|
437
446
|
),
|
|
447
|
+
allow_empty_value=None,
|
|
438
448
|
timeout_ms=timeout_ms,
|
|
439
449
|
)
|
|
440
450
|
|
|
441
451
|
if retries == UNSET:
|
|
442
452
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
443
453
|
retries = self.sdk_configuration.retry_config
|
|
454
|
+
else:
|
|
455
|
+
retries = utils.RetryConfig(
|
|
456
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
457
|
+
)
|
|
444
458
|
|
|
445
459
|
retry_config = None
|
|
446
460
|
if isinstance(retries, utils.RetryConfig):
|
|
447
|
-
retry_config = (retries, ["429"
|
|
461
|
+
retry_config = (retries, ["429"])
|
|
448
462
|
|
|
449
463
|
http_res = await self.do_request_async(
|
|
450
464
|
hook_ctx=HookContext(
|
|
@@ -463,9 +477,7 @@ class HecTokens(BaseSDK):
|
|
|
463
477
|
|
|
464
478
|
response_data: Any = None
|
|
465
479
|
if utils.match_response(http_res, "200", "application/json"):
|
|
466
|
-
return unmarshal_json_response(
|
|
467
|
-
models.UpdateInputHecTokenByIDAndTokenResponse, http_res
|
|
468
|
-
)
|
|
480
|
+
return unmarshal_json_response(models.CountedInputSplunkHec, http_res)
|
|
469
481
|
if utils.match_response(http_res, "500", "application/json"):
|
|
470
482
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
471
483
|
raise errors.Error(response_data, http_res)
|
|
@@ -39,7 +39,7 @@ class LakeDatasets(BaseSDK):
|
|
|
39
39
|
server_url: Optional[str] = None,
|
|
40
40
|
timeout_ms: Optional[int] = None,
|
|
41
41
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
42
|
-
) -> models.
|
|
42
|
+
) -> models.CountedCriblLakeDataset:
|
|
43
43
|
r"""Create a Lake Dataset
|
|
44
44
|
|
|
45
45
|
Create a new Lake Dataset in the specified Lake.
|
|
@@ -118,16 +118,21 @@ class LakeDatasets(BaseSDK):
|
|
|
118
118
|
"json",
|
|
119
119
|
models.CriblLakeDataset,
|
|
120
120
|
),
|
|
121
|
+
allow_empty_value=None,
|
|
121
122
|
timeout_ms=timeout_ms,
|
|
122
123
|
)
|
|
123
124
|
|
|
124
125
|
if retries == UNSET:
|
|
125
126
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
126
127
|
retries = self.sdk_configuration.retry_config
|
|
128
|
+
else:
|
|
129
|
+
retries = utils.RetryConfig(
|
|
130
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
131
|
+
)
|
|
127
132
|
|
|
128
133
|
retry_config = None
|
|
129
134
|
if isinstance(retries, utils.RetryConfig):
|
|
130
|
-
retry_config = (retries, ["429"
|
|
135
|
+
retry_config = (retries, ["429"])
|
|
131
136
|
|
|
132
137
|
http_res = self.do_request(
|
|
133
138
|
hook_ctx=HookContext(
|
|
@@ -146,9 +151,7 @@ class LakeDatasets(BaseSDK):
|
|
|
146
151
|
|
|
147
152
|
response_data: Any = None
|
|
148
153
|
if utils.match_response(http_res, "200", "application/json"):
|
|
149
|
-
return unmarshal_json_response(
|
|
150
|
-
models.CreateCriblLakeDatasetByLakeIDResponse, http_res
|
|
151
|
-
)
|
|
154
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
152
155
|
if utils.match_response(http_res, "500", "application/json"):
|
|
153
156
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
154
157
|
raise errors.Error(response_data, http_res)
|
|
@@ -190,7 +193,7 @@ class LakeDatasets(BaseSDK):
|
|
|
190
193
|
server_url: Optional[str] = None,
|
|
191
194
|
timeout_ms: Optional[int] = None,
|
|
192
195
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
193
|
-
) -> models.
|
|
196
|
+
) -> models.CountedCriblLakeDataset:
|
|
194
197
|
r"""Create a Lake Dataset
|
|
195
198
|
|
|
196
199
|
Create a new Lake Dataset in the specified Lake.
|
|
@@ -269,16 +272,21 @@ class LakeDatasets(BaseSDK):
|
|
|
269
272
|
"json",
|
|
270
273
|
models.CriblLakeDataset,
|
|
271
274
|
),
|
|
275
|
+
allow_empty_value=None,
|
|
272
276
|
timeout_ms=timeout_ms,
|
|
273
277
|
)
|
|
274
278
|
|
|
275
279
|
if retries == UNSET:
|
|
276
280
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
277
281
|
retries = self.sdk_configuration.retry_config
|
|
282
|
+
else:
|
|
283
|
+
retries = utils.RetryConfig(
|
|
284
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
285
|
+
)
|
|
278
286
|
|
|
279
287
|
retry_config = None
|
|
280
288
|
if isinstance(retries, utils.RetryConfig):
|
|
281
|
-
retry_config = (retries, ["429"
|
|
289
|
+
retry_config = (retries, ["429"])
|
|
282
290
|
|
|
283
291
|
http_res = await self.do_request_async(
|
|
284
292
|
hook_ctx=HookContext(
|
|
@@ -297,9 +305,7 @@ class LakeDatasets(BaseSDK):
|
|
|
297
305
|
|
|
298
306
|
response_data: Any = None
|
|
299
307
|
if utils.match_response(http_res, "200", "application/json"):
|
|
300
|
-
return unmarshal_json_response(
|
|
301
|
-
models.CreateCriblLakeDatasetByLakeIDResponse, http_res
|
|
302
|
-
)
|
|
308
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
303
309
|
if utils.match_response(http_res, "500", "application/json"):
|
|
304
310
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
305
311
|
raise errors.Error(response_data, http_res)
|
|
@@ -320,7 +326,7 @@ class LakeDatasets(BaseSDK):
|
|
|
320
326
|
server_url: Optional[str] = None,
|
|
321
327
|
timeout_ms: Optional[int] = None,
|
|
322
328
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
323
|
-
) -> models.
|
|
329
|
+
) -> models.CountedCriblLakeDataset:
|
|
324
330
|
r"""List all Lake Datasets
|
|
325
331
|
|
|
326
332
|
Get a list of all Lake Datasets in the specified Lake.
|
|
@@ -358,16 +364,21 @@ class LakeDatasets(BaseSDK):
|
|
|
358
364
|
accept_header_value="application/json",
|
|
359
365
|
http_headers=http_headers,
|
|
360
366
|
security=self.sdk_configuration.security,
|
|
367
|
+
allow_empty_value=None,
|
|
361
368
|
timeout_ms=timeout_ms,
|
|
362
369
|
)
|
|
363
370
|
|
|
364
371
|
if retries == UNSET:
|
|
365
372
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
366
373
|
retries = self.sdk_configuration.retry_config
|
|
374
|
+
else:
|
|
375
|
+
retries = utils.RetryConfig(
|
|
376
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
377
|
+
)
|
|
367
378
|
|
|
368
379
|
retry_config = None
|
|
369
380
|
if isinstance(retries, utils.RetryConfig):
|
|
370
|
-
retry_config = (retries, ["429"
|
|
381
|
+
retry_config = (retries, ["429"])
|
|
371
382
|
|
|
372
383
|
http_res = self.do_request(
|
|
373
384
|
hook_ctx=HookContext(
|
|
@@ -386,9 +397,7 @@ class LakeDatasets(BaseSDK):
|
|
|
386
397
|
|
|
387
398
|
response_data: Any = None
|
|
388
399
|
if utils.match_response(http_res, "200", "application/json"):
|
|
389
|
-
return unmarshal_json_response(
|
|
390
|
-
models.GetCriblLakeDatasetByLakeIDResponse, http_res
|
|
391
|
-
)
|
|
400
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
392
401
|
if utils.match_response(http_res, "500", "application/json"):
|
|
393
402
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
394
403
|
raise errors.Error(response_data, http_res)
|
|
@@ -409,7 +418,7 @@ class LakeDatasets(BaseSDK):
|
|
|
409
418
|
server_url: Optional[str] = None,
|
|
410
419
|
timeout_ms: Optional[int] = None,
|
|
411
420
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
412
|
-
) -> models.
|
|
421
|
+
) -> models.CountedCriblLakeDataset:
|
|
413
422
|
r"""List all Lake Datasets
|
|
414
423
|
|
|
415
424
|
Get a list of all Lake Datasets in the specified Lake.
|
|
@@ -447,16 +456,21 @@ class LakeDatasets(BaseSDK):
|
|
|
447
456
|
accept_header_value="application/json",
|
|
448
457
|
http_headers=http_headers,
|
|
449
458
|
security=self.sdk_configuration.security,
|
|
459
|
+
allow_empty_value=None,
|
|
450
460
|
timeout_ms=timeout_ms,
|
|
451
461
|
)
|
|
452
462
|
|
|
453
463
|
if retries == UNSET:
|
|
454
464
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
455
465
|
retries = self.sdk_configuration.retry_config
|
|
466
|
+
else:
|
|
467
|
+
retries = utils.RetryConfig(
|
|
468
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
469
|
+
)
|
|
456
470
|
|
|
457
471
|
retry_config = None
|
|
458
472
|
if isinstance(retries, utils.RetryConfig):
|
|
459
|
-
retry_config = (retries, ["429"
|
|
473
|
+
retry_config = (retries, ["429"])
|
|
460
474
|
|
|
461
475
|
http_res = await self.do_request_async(
|
|
462
476
|
hook_ctx=HookContext(
|
|
@@ -475,9 +489,7 @@ class LakeDatasets(BaseSDK):
|
|
|
475
489
|
|
|
476
490
|
response_data: Any = None
|
|
477
491
|
if utils.match_response(http_res, "200", "application/json"):
|
|
478
|
-
return unmarshal_json_response(
|
|
479
|
-
models.GetCriblLakeDatasetByLakeIDResponse, http_res
|
|
480
|
-
)
|
|
492
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
481
493
|
if utils.match_response(http_res, "500", "application/json"):
|
|
482
494
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
483
495
|
raise errors.Error(response_data, http_res)
|
|
@@ -499,7 +511,7 @@ class LakeDatasets(BaseSDK):
|
|
|
499
511
|
server_url: Optional[str] = None,
|
|
500
512
|
timeout_ms: Optional[int] = None,
|
|
501
513
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
502
|
-
) -> models.
|
|
514
|
+
) -> models.CountedCriblLakeDataset:
|
|
503
515
|
r"""Delete a Lake Dataset
|
|
504
516
|
|
|
505
517
|
Delete the specified Lake Dataset in the specified Lake
|
|
@@ -539,16 +551,21 @@ class LakeDatasets(BaseSDK):
|
|
|
539
551
|
accept_header_value="application/json",
|
|
540
552
|
http_headers=http_headers,
|
|
541
553
|
security=self.sdk_configuration.security,
|
|
554
|
+
allow_empty_value=None,
|
|
542
555
|
timeout_ms=timeout_ms,
|
|
543
556
|
)
|
|
544
557
|
|
|
545
558
|
if retries == UNSET:
|
|
546
559
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
547
560
|
retries = self.sdk_configuration.retry_config
|
|
561
|
+
else:
|
|
562
|
+
retries = utils.RetryConfig(
|
|
563
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
564
|
+
)
|
|
548
565
|
|
|
549
566
|
retry_config = None
|
|
550
567
|
if isinstance(retries, utils.RetryConfig):
|
|
551
|
-
retry_config = (retries, ["429"
|
|
568
|
+
retry_config = (retries, ["429"])
|
|
552
569
|
|
|
553
570
|
http_res = self.do_request(
|
|
554
571
|
hook_ctx=HookContext(
|
|
@@ -567,9 +584,7 @@ class LakeDatasets(BaseSDK):
|
|
|
567
584
|
|
|
568
585
|
response_data: Any = None
|
|
569
586
|
if utils.match_response(http_res, "200", "application/json"):
|
|
570
|
-
return unmarshal_json_response(
|
|
571
|
-
models.DeleteCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
572
|
-
)
|
|
587
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
573
588
|
if utils.match_response(http_res, "500", "application/json"):
|
|
574
589
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
575
590
|
raise errors.Error(response_data, http_res)
|
|
@@ -591,7 +606,7 @@ class LakeDatasets(BaseSDK):
|
|
|
591
606
|
server_url: Optional[str] = None,
|
|
592
607
|
timeout_ms: Optional[int] = None,
|
|
593
608
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
594
|
-
) -> models.
|
|
609
|
+
) -> models.CountedCriblLakeDataset:
|
|
595
610
|
r"""Delete a Lake Dataset
|
|
596
611
|
|
|
597
612
|
Delete the specified Lake Dataset in the specified Lake
|
|
@@ -631,16 +646,21 @@ class LakeDatasets(BaseSDK):
|
|
|
631
646
|
accept_header_value="application/json",
|
|
632
647
|
http_headers=http_headers,
|
|
633
648
|
security=self.sdk_configuration.security,
|
|
649
|
+
allow_empty_value=None,
|
|
634
650
|
timeout_ms=timeout_ms,
|
|
635
651
|
)
|
|
636
652
|
|
|
637
653
|
if retries == UNSET:
|
|
638
654
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
639
655
|
retries = self.sdk_configuration.retry_config
|
|
656
|
+
else:
|
|
657
|
+
retries = utils.RetryConfig(
|
|
658
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
659
|
+
)
|
|
640
660
|
|
|
641
661
|
retry_config = None
|
|
642
662
|
if isinstance(retries, utils.RetryConfig):
|
|
643
|
-
retry_config = (retries, ["429"
|
|
663
|
+
retry_config = (retries, ["429"])
|
|
644
664
|
|
|
645
665
|
http_res = await self.do_request_async(
|
|
646
666
|
hook_ctx=HookContext(
|
|
@@ -659,9 +679,7 @@ class LakeDatasets(BaseSDK):
|
|
|
659
679
|
|
|
660
680
|
response_data: Any = None
|
|
661
681
|
if utils.match_response(http_res, "200", "application/json"):
|
|
662
|
-
return unmarshal_json_response(
|
|
663
|
-
models.DeleteCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
664
|
-
)
|
|
682
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
665
683
|
if utils.match_response(http_res, "500", "application/json"):
|
|
666
684
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
667
685
|
raise errors.Error(response_data, http_res)
|
|
@@ -683,7 +701,7 @@ class LakeDatasets(BaseSDK):
|
|
|
683
701
|
server_url: Optional[str] = None,
|
|
684
702
|
timeout_ms: Optional[int] = None,
|
|
685
703
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
686
|
-
) -> models.
|
|
704
|
+
) -> models.CountedCriblLakeDataset:
|
|
687
705
|
r"""Get a Lake Dataset
|
|
688
706
|
|
|
689
707
|
Get the specified Lake Dataset in the specified Lake.
|
|
@@ -723,16 +741,21 @@ class LakeDatasets(BaseSDK):
|
|
|
723
741
|
accept_header_value="application/json",
|
|
724
742
|
http_headers=http_headers,
|
|
725
743
|
security=self.sdk_configuration.security,
|
|
744
|
+
allow_empty_value=None,
|
|
726
745
|
timeout_ms=timeout_ms,
|
|
727
746
|
)
|
|
728
747
|
|
|
729
748
|
if retries == UNSET:
|
|
730
749
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
731
750
|
retries = self.sdk_configuration.retry_config
|
|
751
|
+
else:
|
|
752
|
+
retries = utils.RetryConfig(
|
|
753
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
754
|
+
)
|
|
732
755
|
|
|
733
756
|
retry_config = None
|
|
734
757
|
if isinstance(retries, utils.RetryConfig):
|
|
735
|
-
retry_config = (retries, ["429"
|
|
758
|
+
retry_config = (retries, ["429"])
|
|
736
759
|
|
|
737
760
|
http_res = self.do_request(
|
|
738
761
|
hook_ctx=HookContext(
|
|
@@ -751,9 +774,7 @@ class LakeDatasets(BaseSDK):
|
|
|
751
774
|
|
|
752
775
|
response_data: Any = None
|
|
753
776
|
if utils.match_response(http_res, "200", "application/json"):
|
|
754
|
-
return unmarshal_json_response(
|
|
755
|
-
models.GetCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
756
|
-
)
|
|
777
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
757
778
|
if utils.match_response(http_res, "500", "application/json"):
|
|
758
779
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
759
780
|
raise errors.Error(response_data, http_res)
|
|
@@ -775,7 +796,7 @@ class LakeDatasets(BaseSDK):
|
|
|
775
796
|
server_url: Optional[str] = None,
|
|
776
797
|
timeout_ms: Optional[int] = None,
|
|
777
798
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
778
|
-
) -> models.
|
|
799
|
+
) -> models.CountedCriblLakeDataset:
|
|
779
800
|
r"""Get a Lake Dataset
|
|
780
801
|
|
|
781
802
|
Get the specified Lake Dataset in the specified Lake.
|
|
@@ -815,16 +836,21 @@ class LakeDatasets(BaseSDK):
|
|
|
815
836
|
accept_header_value="application/json",
|
|
816
837
|
http_headers=http_headers,
|
|
817
838
|
security=self.sdk_configuration.security,
|
|
839
|
+
allow_empty_value=None,
|
|
818
840
|
timeout_ms=timeout_ms,
|
|
819
841
|
)
|
|
820
842
|
|
|
821
843
|
if retries == UNSET:
|
|
822
844
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
823
845
|
retries = self.sdk_configuration.retry_config
|
|
846
|
+
else:
|
|
847
|
+
retries = utils.RetryConfig(
|
|
848
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
849
|
+
)
|
|
824
850
|
|
|
825
851
|
retry_config = None
|
|
826
852
|
if isinstance(retries, utils.RetryConfig):
|
|
827
|
-
retry_config = (retries, ["429"
|
|
853
|
+
retry_config = (retries, ["429"])
|
|
828
854
|
|
|
829
855
|
http_res = await self.do_request_async(
|
|
830
856
|
hook_ctx=HookContext(
|
|
@@ -843,9 +869,7 @@ class LakeDatasets(BaseSDK):
|
|
|
843
869
|
|
|
844
870
|
response_data: Any = None
|
|
845
871
|
if utils.match_response(http_res, "200", "application/json"):
|
|
846
|
-
return unmarshal_json_response(
|
|
847
|
-
models.GetCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
848
|
-
)
|
|
872
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
849
873
|
if utils.match_response(http_res, "500", "application/json"):
|
|
850
874
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
851
875
|
raise errors.Error(response_data, http_res)
|
|
@@ -888,7 +912,7 @@ class LakeDatasets(BaseSDK):
|
|
|
888
912
|
server_url: Optional[str] = None,
|
|
889
913
|
timeout_ms: Optional[int] = None,
|
|
890
914
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
891
|
-
) -> models.
|
|
915
|
+
) -> models.CountedCriblLakeDataset:
|
|
892
916
|
r"""Update a Lake Dataset
|
|
893
917
|
|
|
894
918
|
Update the specified Lake Dataset in the specified Lake.
|
|
@@ -969,16 +993,21 @@ class LakeDatasets(BaseSDK):
|
|
|
969
993
|
"json",
|
|
970
994
|
models.CriblLakeDatasetUpdate,
|
|
971
995
|
),
|
|
996
|
+
allow_empty_value=None,
|
|
972
997
|
timeout_ms=timeout_ms,
|
|
973
998
|
)
|
|
974
999
|
|
|
975
1000
|
if retries == UNSET:
|
|
976
1001
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
977
1002
|
retries = self.sdk_configuration.retry_config
|
|
1003
|
+
else:
|
|
1004
|
+
retries = utils.RetryConfig(
|
|
1005
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
1006
|
+
)
|
|
978
1007
|
|
|
979
1008
|
retry_config = None
|
|
980
1009
|
if isinstance(retries, utils.RetryConfig):
|
|
981
|
-
retry_config = (retries, ["429"
|
|
1010
|
+
retry_config = (retries, ["429"])
|
|
982
1011
|
|
|
983
1012
|
http_res = self.do_request(
|
|
984
1013
|
hook_ctx=HookContext(
|
|
@@ -997,9 +1026,7 @@ class LakeDatasets(BaseSDK):
|
|
|
997
1026
|
|
|
998
1027
|
response_data: Any = None
|
|
999
1028
|
if utils.match_response(http_res, "200", "application/json"):
|
|
1000
|
-
return unmarshal_json_response(
|
|
1001
|
-
models.UpdateCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
1002
|
-
)
|
|
1029
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
1003
1030
|
if utils.match_response(http_res, "500", "application/json"):
|
|
1004
1031
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
1005
1032
|
raise errors.Error(response_data, http_res)
|
|
@@ -1042,7 +1069,7 @@ class LakeDatasets(BaseSDK):
|
|
|
1042
1069
|
server_url: Optional[str] = None,
|
|
1043
1070
|
timeout_ms: Optional[int] = None,
|
|
1044
1071
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
1045
|
-
) -> models.
|
|
1072
|
+
) -> models.CountedCriblLakeDataset:
|
|
1046
1073
|
r"""Update a Lake Dataset
|
|
1047
1074
|
|
|
1048
1075
|
Update the specified Lake Dataset in the specified Lake.
|
|
@@ -1123,16 +1150,21 @@ class LakeDatasets(BaseSDK):
|
|
|
1123
1150
|
"json",
|
|
1124
1151
|
models.CriblLakeDatasetUpdate,
|
|
1125
1152
|
),
|
|
1153
|
+
allow_empty_value=None,
|
|
1126
1154
|
timeout_ms=timeout_ms,
|
|
1127
1155
|
)
|
|
1128
1156
|
|
|
1129
1157
|
if retries == UNSET:
|
|
1130
1158
|
if self.sdk_configuration.retry_config is not UNSET:
|
|
1131
1159
|
retries = self.sdk_configuration.retry_config
|
|
1160
|
+
else:
|
|
1161
|
+
retries = utils.RetryConfig(
|
|
1162
|
+
"backoff", utils.BackoffStrategy(500, 60000, 1.5, 3600000), True
|
|
1163
|
+
)
|
|
1132
1164
|
|
|
1133
1165
|
retry_config = None
|
|
1134
1166
|
if isinstance(retries, utils.RetryConfig):
|
|
1135
|
-
retry_config = (retries, ["429"
|
|
1167
|
+
retry_config = (retries, ["429"])
|
|
1136
1168
|
|
|
1137
1169
|
http_res = await self.do_request_async(
|
|
1138
1170
|
hook_ctx=HookContext(
|
|
@@ -1151,9 +1183,7 @@ class LakeDatasets(BaseSDK):
|
|
|
1151
1183
|
|
|
1152
1184
|
response_data: Any = None
|
|
1153
1185
|
if utils.match_response(http_res, "200", "application/json"):
|
|
1154
|
-
return unmarshal_json_response(
|
|
1155
|
-
models.UpdateCriblLakeDatasetByLakeIDAndIDResponse, http_res
|
|
1156
|
-
)
|
|
1186
|
+
return unmarshal_json_response(models.CountedCriblLakeDataset, http_res)
|
|
1157
1187
|
if utils.match_response(http_res, "500", "application/json"):
|
|
1158
1188
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
1159
1189
|
raise errors.Error(response_data, http_res)
|