cribl-control-plane 0.2.1rc11__py3-none-any.whl → 0.4.0a12__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +4 -4
- cribl_control_plane/acl.py +6 -4
- cribl_control_plane/basesdk.py +6 -0
- cribl_control_plane/branches.py +8 -4
- cribl_control_plane/commits.py +42 -28
- cribl_control_plane/commits_files.py +12 -8
- cribl_control_plane/configs_versions.py +10 -4
- cribl_control_plane/destinations.py +30 -20
- cribl_control_plane/destinations_pq.py +12 -8
- cribl_control_plane/groups_sdk.py +60 -24
- cribl_control_plane/health.py +2 -0
- cribl_control_plane/hectokens.py +32 -8
- cribl_control_plane/lakedatasets.py +86 -20
- cribl_control_plane/models/__init__.py +568 -214
- cribl_control_plane/models/addhectokenrequest.py +7 -1
- cribl_control_plane/models/createconfiggroupbyproductop.py +20 -1
- cribl_control_plane/models/createcribllakedatasetbylakeidop.py +19 -1
- cribl_control_plane/models/createinputhectokenbyidop.py +20 -1
- cribl_control_plane/models/{countedlistgitdiffresult.py → createinputop.py} +9 -5
- cribl_control_plane/models/{countedlistgitshowresult.py → createoutputop.py} +9 -5
- cribl_control_plane/models/createoutputtestbyidop.py +20 -1
- cribl_control_plane/models/{countedlistpackinstallinfo.py → createpacksop.py} +6 -2
- cribl_control_plane/models/createpipelineop.py +24 -0
- cribl_control_plane/models/createroutesappendbyidop.py +20 -2
- cribl_control_plane/models/createversioncommitop.py +19 -1
- cribl_control_plane/models/{countedliststring.py → createversionpushop.py} +6 -2
- cribl_control_plane/models/createversionrevertop.py +19 -1
- cribl_control_plane/models/createversionundoop.py +18 -1
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +20 -1
- cribl_control_plane/models/deletecribllakedatasetbylakeidandidop.py +20 -1
- cribl_control_plane/models/deleteinputbyidop.py +20 -1
- cribl_control_plane/models/deleteoutputbyidop.py +20 -1
- cribl_control_plane/models/deleteoutputpqbyidop.py +19 -1
- cribl_control_plane/models/deletepacksbyidop.py +20 -1
- cribl_control_plane/models/deletepipelinebyidop.py +20 -1
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +19 -1
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +19 -1
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +19 -1
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +19 -1
- cribl_control_plane/models/getcribllakedatasetbylakeidandidop.py +20 -1
- cribl_control_plane/models/getcribllakedatasetbylakeidop.py +79 -2
- cribl_control_plane/models/getinputbyidop.py +20 -1
- cribl_control_plane/models/getmasterworkerentryop.py +18 -1
- cribl_control_plane/models/getoutputbyidop.py +20 -1
- cribl_control_plane/models/getoutputpqbyidop.py +20 -1
- cribl_control_plane/models/getoutputsamplesbyidop.py +20 -1
- cribl_control_plane/models/getpacksbyidop.py +20 -1
- cribl_control_plane/models/getpacksop.py +19 -1
- cribl_control_plane/models/getpipelinebyidop.py +20 -1
- cribl_control_plane/models/getroutesbyidop.py +20 -1
- cribl_control_plane/models/getsummaryop.py +19 -1
- cribl_control_plane/models/{countedlistbranchinfo.py → getversionbranchop.py} +6 -2
- cribl_control_plane/models/getversioncountop.py +19 -1
- cribl_control_plane/models/getversiondiffop.py +19 -1
- cribl_control_plane/models/getversionfilesop.py +19 -1
- cribl_control_plane/models/{countedlistgitinfo.py → getversioninfoop.py} +6 -2
- cribl_control_plane/models/getversionop.py +19 -1
- cribl_control_plane/models/getversionshowop.py +19 -1
- cribl_control_plane/models/getversionstatusop.py +19 -1
- cribl_control_plane/models/input.py +18 -15
- cribl_control_plane/models/inputcloudflarehec.py +513 -0
- cribl_control_plane/models/inputfile.py +7 -0
- cribl_control_plane/models/listconfiggroupbyproductop.py +19 -1
- cribl_control_plane/models/{countedlistinput.py → listinputop.py} +6 -2
- cribl_control_plane/models/listmasterworkerentryop.py +19 -1
- cribl_control_plane/models/{countedlistoutput.py → listoutputop.py} +6 -2
- cribl_control_plane/models/{countedlistpipeline.py → listpipelineop.py} +6 -2
- cribl_control_plane/models/{countedlistroutes.py → listroutesop.py} +6 -2
- cribl_control_plane/models/output.py +23 -17
- cribl_control_plane/models/outputazureblob.py +14 -0
- cribl_control_plane/models/outputazuredataexplorer.py +7 -0
- cribl_control_plane/models/outputchronicle.py +5 -0
- cribl_control_plane/models/outputcloudflarer2.py +632 -0
- cribl_control_plane/models/outputcribllake.py +14 -0
- cribl_control_plane/models/outputdatabricks.py +19 -0
- cribl_control_plane/models/outputdls3.py +14 -0
- cribl_control_plane/models/outputexabeam.py +7 -0
- cribl_control_plane/models/outputfilesystem.py +14 -0
- cribl_control_plane/models/outputgooglecloudstorage.py +14 -0
- cribl_control_plane/models/outputmicrosoftfabric.py +540 -0
- cribl_control_plane/models/outputminio.py +19 -4
- cribl_control_plane/models/outputnetflow.py +7 -0
- cribl_control_plane/models/outputs3.py +14 -0
- cribl_control_plane/models/outputsecuritylake.py +14 -0
- cribl_control_plane/models/outputsyslog.py +7 -0
- cribl_control_plane/models/runnablejobcollection.py +0 -8
- cribl_control_plane/models/runnablejobexecutor.py +0 -4
- cribl_control_plane/models/runnablejobscheduledsearch.py +0 -4
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +19 -1
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +20 -1
- cribl_control_plane/models/updatecribllakedatasetbylakeidandidop.py +20 -1
- cribl_control_plane/models/updatehectokenrequest.py +7 -1
- cribl_control_plane/models/updateinputbyidop.py +19 -1
- cribl_control_plane/models/updateinputhectokenbyidandtokenop.py +20 -1
- cribl_control_plane/models/updateoutputbyidop.py +19 -1
- cribl_control_plane/models/updatepacksbyidop.py +20 -1
- cribl_control_plane/models/updatepipelinebyidop.py +19 -1
- cribl_control_plane/models/updateroutesbyidop.py +19 -1
- cribl_control_plane/nodes.py +16 -8
- cribl_control_plane/packs.py +32 -20
- cribl_control_plane/pipelines.py +30 -20
- cribl_control_plane/routes_sdk.py +28 -16
- cribl_control_plane/samples.py +12 -8
- cribl_control_plane/sources.py +30 -20
- cribl_control_plane/statuses.py +6 -4
- cribl_control_plane/summaries.py +6 -8
- cribl_control_plane/teams.py +6 -4
- cribl_control_plane/tokens.py +2 -0
- cribl_control_plane/utils/forms.py +21 -10
- cribl_control_plane/utils/queryparams.py +14 -2
- cribl_control_plane/utils/retries.py +69 -5
- cribl_control_plane/utils/unmarshal_json_response.py +15 -1
- cribl_control_plane/versions_configs.py +6 -4
- {cribl_control_plane-0.2.1rc11.dist-info → cribl_control_plane-0.4.0a12.dist-info}/METADATA +28 -36
- {cribl_control_plane-0.2.1rc11.dist-info → cribl_control_plane-0.4.0a12.dist-info}/RECORD +117 -131
- cribl_control_plane-0.4.0a12.dist-info/licenses/LICENSE +201 -0
- cribl_control_plane/models/countedlistconfiggroup.py +0 -20
- cribl_control_plane/models/countedlistcribllakedataset.py +0 -20
- cribl_control_plane/models/countedlistdistributedsummary.py +0 -20
- cribl_control_plane/models/countedlistgitcommitsummary.py +0 -20
- cribl_control_plane/models/countedlistgitcountresult.py +0 -20
- cribl_control_plane/models/countedlistgitfilesresponse.py +0 -20
- cribl_control_plane/models/countedlistgitlogresult.py +0 -20
- cribl_control_plane/models/countedlistgitrevertresult.py +0 -20
- cribl_control_plane/models/countedlistgitstatusresult.py +0 -20
- cribl_control_plane/models/countedlistinputsplunkhec.py +0 -20
- cribl_control_plane/models/countedlistjobinfo.py +0 -20
- cribl_control_plane/models/countedlistmasterworkerentry.py +0 -20
- cribl_control_plane/models/countedlistnumber.py +0 -19
- cribl_control_plane/models/countedlistobject.py +0 -19
- cribl_control_plane/models/countedlistoutputsamplesresponse.py +0 -20
- cribl_control_plane/models/countedlistoutputtestresponse.py +0 -20
- cribl_control_plane/models/countedlistpackinfo.py +0 -20
- cribl_control_plane/models/countedlistteamaccesscontrollist.py +0 -20
- cribl_control_plane/models/countedlistuseraccesscontrollist.py +0 -20
- {cribl_control_plane-0.2.1rc11.dist-info → cribl_control_plane-0.4.0a12.dist-info}/WHEEL +0 -0
|
@@ -234,6 +234,8 @@ class OutputS3TypedDict(TypedDict):
|
|
|
234
234
|
r"""If a file fails to move to its final destination after the maximum number of retries, move it to a designated directory to prevent further errors"""
|
|
235
235
|
on_disk_full_backpressure: NotRequired[OutputS3DiskSpaceProtection]
|
|
236
236
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
237
|
+
force_close_on_shutdown: NotRequired[bool]
|
|
238
|
+
r"""Force all staged files to close during an orderly Node shutdown. This triggers immediate upload of in-progress data — regardless of idle time, file age, or size thresholds — to minimize data loss."""
|
|
237
239
|
max_file_open_time_sec: NotRequired[float]
|
|
238
240
|
r"""Maximum amount of time to write to a file. Files open for longer than this will be closed and moved to final output location."""
|
|
239
241
|
max_file_idle_time_sec: NotRequired[float]
|
|
@@ -277,6 +279,8 @@ class OutputS3TypedDict(TypedDict):
|
|
|
277
279
|
r"""Parquet tools can use the checksum of a Parquet page to verify data integrity"""
|
|
278
280
|
empty_dir_cleanup_sec: NotRequired[float]
|
|
279
281
|
r"""How frequently, in seconds, to clean up empty directories"""
|
|
282
|
+
directory_batch_size: NotRequired[float]
|
|
283
|
+
r"""Number of directories to process in each batch during cleanup of empty directories. Minimum is 10, maximum is 10000. Higher values may require more memory."""
|
|
280
284
|
deadletter_path: NotRequired[str]
|
|
281
285
|
r"""Storage location for files that fail to reach their final destination after maximum retries are exceeded"""
|
|
282
286
|
max_retry_num: NotRequired[float]
|
|
@@ -474,6 +478,11 @@ class OutputS3(BaseModel):
|
|
|
474
478
|
] = OutputS3DiskSpaceProtection.BLOCK
|
|
475
479
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
476
480
|
|
|
481
|
+
force_close_on_shutdown: Annotated[
|
|
482
|
+
Optional[bool], pydantic.Field(alias="forceCloseOnShutdown")
|
|
483
|
+
] = False
|
|
484
|
+
r"""Force all staged files to close during an orderly Node shutdown. This triggers immediate upload of in-progress data — regardless of idle time, file age, or size thresholds — to minimize data loss."""
|
|
485
|
+
|
|
477
486
|
max_file_open_time_sec: Annotated[
|
|
478
487
|
Optional[float], pydantic.Field(alias="maxFileOpenTimeSec")
|
|
479
488
|
] = 300
|
|
@@ -588,6 +597,11 @@ class OutputS3(BaseModel):
|
|
|
588
597
|
] = 300
|
|
589
598
|
r"""How frequently, in seconds, to clean up empty directories"""
|
|
590
599
|
|
|
600
|
+
directory_batch_size: Annotated[
|
|
601
|
+
Optional[float], pydantic.Field(alias="directoryBatchSize")
|
|
602
|
+
] = 1000
|
|
603
|
+
r"""Number of directories to process in each batch during cleanup of empty directories. Minimum is 10, maximum is 10000. Higher values may require more memory."""
|
|
604
|
+
|
|
591
605
|
deadletter_path: Annotated[
|
|
592
606
|
Optional[str], pydantic.Field(alias="deadletterPath")
|
|
593
607
|
] = "$CRIBL_HOME/state/outputs/dead-letter"
|
|
@@ -202,6 +202,8 @@ class OutputSecurityLakeTypedDict(TypedDict):
|
|
|
202
202
|
r"""If a file fails to move to its final destination after the maximum number of retries, move it to a designated directory to prevent further errors"""
|
|
203
203
|
on_disk_full_backpressure: NotRequired[OutputSecurityLakeDiskSpaceProtection]
|
|
204
204
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
205
|
+
force_close_on_shutdown: NotRequired[bool]
|
|
206
|
+
r"""Force all staged files to close during an orderly Node shutdown. This triggers immediate upload of in-progress data — regardless of idle time, file age, or size thresholds — to minimize data loss."""
|
|
205
207
|
max_file_open_time_sec: NotRequired[float]
|
|
206
208
|
r"""Maximum amount of time to write to a file. Files open for longer than this will be closed and moved to final output location."""
|
|
207
209
|
max_file_idle_time_sec: NotRequired[float]
|
|
@@ -239,6 +241,8 @@ class OutputSecurityLakeTypedDict(TypedDict):
|
|
|
239
241
|
r"""Select or create a stored secret that references your access key and secret key"""
|
|
240
242
|
empty_dir_cleanup_sec: NotRequired[float]
|
|
241
243
|
r"""How frequently, in seconds, to clean up empty directories"""
|
|
244
|
+
directory_batch_size: NotRequired[float]
|
|
245
|
+
r"""Number of directories to process in each batch during cleanup of empty directories. Minimum is 10, maximum is 10000. Higher values may require more memory."""
|
|
242
246
|
parquet_schema: NotRequired[str]
|
|
243
247
|
r"""To add a new schema, navigate to Processing > Knowledge > Parquet Schemas"""
|
|
244
248
|
deadletter_path: NotRequired[str]
|
|
@@ -422,6 +426,11 @@ class OutputSecurityLake(BaseModel):
|
|
|
422
426
|
] = OutputSecurityLakeDiskSpaceProtection.BLOCK
|
|
423
427
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
424
428
|
|
|
429
|
+
force_close_on_shutdown: Annotated[
|
|
430
|
+
Optional[bool], pydantic.Field(alias="forceCloseOnShutdown")
|
|
431
|
+
] = False
|
|
432
|
+
r"""Force all staged files to close during an orderly Node shutdown. This triggers immediate upload of in-progress data — regardless of idle time, file age, or size thresholds — to minimize data loss."""
|
|
433
|
+
|
|
425
434
|
max_file_open_time_sec: Annotated[
|
|
426
435
|
Optional[float], pydantic.Field(alias="maxFileOpenTimeSec")
|
|
427
436
|
] = 300
|
|
@@ -519,6 +528,11 @@ class OutputSecurityLake(BaseModel):
|
|
|
519
528
|
] = 300
|
|
520
529
|
r"""How frequently, in seconds, to clean up empty directories"""
|
|
521
530
|
|
|
531
|
+
directory_batch_size: Annotated[
|
|
532
|
+
Optional[float], pydantic.Field(alias="directoryBatchSize")
|
|
533
|
+
] = 1000
|
|
534
|
+
r"""Number of directories to process in each batch during cleanup of empty directories. Minimum is 10, maximum is 10000. Higher values may require more memory."""
|
|
535
|
+
|
|
522
536
|
parquet_schema: Annotated[Optional[str], pydantic.Field(alias="parquetSchema")] = (
|
|
523
537
|
None
|
|
524
538
|
)
|
|
@@ -346,6 +346,8 @@ class OutputSyslogTypedDict(TypedDict):
|
|
|
346
346
|
r"""Maximum size of syslog messages. Make sure this value is less than or equal to the MTU to avoid UDP packet fragmentation."""
|
|
347
347
|
udp_dns_resolve_period_sec: NotRequired[float]
|
|
348
348
|
r"""How often to resolve the destination hostname to an IP address. Ignored if the destination is an IP address. A value of 0 means every message sent will incur a DNS lookup."""
|
|
349
|
+
enable_ip_spoofing: NotRequired[bool]
|
|
350
|
+
r"""Send Syslog traffic using the original event's Source IP and port. To enable this, you must install the external `udp-sender` helper binary at `/usr/bin/udp-sender` on all Worker Nodes and grant it the `CAP_NET_RAW` capability."""
|
|
349
351
|
pq_strict_ordering: NotRequired[bool]
|
|
350
352
|
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
351
353
|
pq_rate_per_sec: NotRequired[float]
|
|
@@ -502,6 +504,11 @@ class OutputSyslog(BaseModel):
|
|
|
502
504
|
] = 0
|
|
503
505
|
r"""How often to resolve the destination hostname to an IP address. Ignored if the destination is an IP address. A value of 0 means every message sent will incur a DNS lookup."""
|
|
504
506
|
|
|
507
|
+
enable_ip_spoofing: Annotated[
|
|
508
|
+
Optional[bool], pydantic.Field(alias="enableIpSpoofing")
|
|
509
|
+
] = False
|
|
510
|
+
r"""Send Syslog traffic using the original event's Source IP and port. To enable this, you must install the external `udp-sender` helper binary at `/usr/bin/udp-sender` on all Worker Nodes and grant it the `CAP_NET_RAW` capability."""
|
|
511
|
+
|
|
505
512
|
pq_strict_ordering: Annotated[
|
|
506
513
|
Optional[bool], pydantic.Field(alias="pqStrictOrdering")
|
|
507
514
|
] = True
|
|
@@ -73,7 +73,6 @@ class RunnableJobCollectionRunSettingsTypedDict(TypedDict):
|
|
|
73
73
|
|
|
74
74
|
|
|
75
75
|
|
|
76
|
-
|
|
77
76
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
78
77
|
"""
|
|
79
78
|
max_task_size: NotRequired[str]
|
|
@@ -88,7 +87,6 @@ class RunnableJobCollectionRunSettingsTypedDict(TypedDict):
|
|
|
88
87
|
|
|
89
88
|
|
|
90
89
|
|
|
91
|
-
|
|
92
90
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
93
91
|
"""
|
|
94
92
|
|
|
@@ -152,7 +150,6 @@ class RunnableJobCollectionRunSettings(BaseModel):
|
|
|
152
150
|
|
|
153
151
|
|
|
154
152
|
|
|
155
|
-
|
|
156
153
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
157
154
|
"""
|
|
158
155
|
|
|
@@ -170,7 +167,6 @@ class RunnableJobCollectionRunSettings(BaseModel):
|
|
|
170
167
|
|
|
171
168
|
|
|
172
169
|
|
|
173
|
-
|
|
174
170
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
175
171
|
"""
|
|
176
172
|
|
|
@@ -452,7 +448,6 @@ class RunnableJobCollectionRunTypedDict(TypedDict):
|
|
|
452
448
|
|
|
453
449
|
|
|
454
450
|
|
|
455
|
-
|
|
456
451
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
457
452
|
"""
|
|
458
453
|
max_task_size: NotRequired[str]
|
|
@@ -466,7 +461,6 @@ class RunnableJobCollectionRunTypedDict(TypedDict):
|
|
|
466
461
|
|
|
467
462
|
|
|
468
463
|
|
|
469
|
-
|
|
470
464
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
471
465
|
"""
|
|
472
466
|
discover_to_routes: NotRequired[bool]
|
|
@@ -536,7 +530,6 @@ class RunnableJobCollectionRun(BaseModel):
|
|
|
536
530
|
|
|
537
531
|
|
|
538
532
|
|
|
539
|
-
|
|
540
533
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
541
534
|
"""
|
|
542
535
|
|
|
@@ -553,7 +546,6 @@ class RunnableJobCollectionRun(BaseModel):
|
|
|
553
546
|
|
|
554
547
|
|
|
555
548
|
|
|
556
|
-
|
|
557
549
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
558
550
|
"""
|
|
559
551
|
|
|
@@ -73,7 +73,6 @@ class RunnableJobExecutorRunSettingsTypedDict(TypedDict):
|
|
|
73
73
|
|
|
74
74
|
|
|
75
75
|
|
|
76
|
-
|
|
77
76
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
78
77
|
"""
|
|
79
78
|
max_task_size: NotRequired[str]
|
|
@@ -88,7 +87,6 @@ class RunnableJobExecutorRunSettingsTypedDict(TypedDict):
|
|
|
88
87
|
|
|
89
88
|
|
|
90
89
|
|
|
91
|
-
|
|
92
90
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
93
91
|
"""
|
|
94
92
|
|
|
@@ -150,7 +148,6 @@ class RunnableJobExecutorRunSettings(BaseModel):
|
|
|
150
148
|
|
|
151
149
|
|
|
152
150
|
|
|
153
|
-
|
|
154
151
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
155
152
|
"""
|
|
156
153
|
|
|
@@ -168,7 +165,6 @@ class RunnableJobExecutorRunSettings(BaseModel):
|
|
|
168
165
|
|
|
169
166
|
|
|
170
167
|
|
|
171
|
-
|
|
172
168
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
173
169
|
"""
|
|
174
170
|
|
|
@@ -73,7 +73,6 @@ class RunnableJobScheduledSearchRunSettingsTypedDict(TypedDict):
|
|
|
73
73
|
|
|
74
74
|
|
|
75
75
|
|
|
76
|
-
|
|
77
76
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
78
77
|
"""
|
|
79
78
|
max_task_size: NotRequired[str]
|
|
@@ -88,7 +87,6 @@ class RunnableJobScheduledSearchRunSettingsTypedDict(TypedDict):
|
|
|
88
87
|
|
|
89
88
|
|
|
90
89
|
|
|
91
|
-
|
|
92
90
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
93
91
|
"""
|
|
94
92
|
|
|
@@ -151,7 +149,6 @@ class RunnableJobScheduledSearchRunSettings(BaseModel):
|
|
|
151
149
|
|
|
152
150
|
|
|
153
151
|
|
|
154
|
-
|
|
155
152
|
if your lower bundle size is 1MB, you can bundle up to five 200KB files into one task.
|
|
156
153
|
"""
|
|
157
154
|
|
|
@@ -169,7 +166,6 @@ class RunnableJobScheduledSearchRunSettings(BaseModel):
|
|
|
169
166
|
|
|
170
167
|
|
|
171
168
|
|
|
172
|
-
|
|
173
169
|
you can bundle up to five 2MB files into one task. Files greater than this size will be assigned to individual tasks.
|
|
174
170
|
"""
|
|
175
171
|
|
|
@@ -14,7 +14,8 @@ from cribl_control_plane.utils import (
|
|
|
14
14
|
import pydantic
|
|
15
15
|
from pydantic import field_serializer
|
|
16
16
|
from pydantic.functional_validators import PlainValidator
|
|
17
|
-
from
|
|
17
|
+
from typing import List, Optional
|
|
18
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
18
19
|
|
|
19
20
|
|
|
20
21
|
class UpdateConfigGroupByProductAndIDRequestTypedDict(TypedDict):
|
|
@@ -54,3 +55,20 @@ class UpdateConfigGroupByProductAndIDRequest(BaseModel):
|
|
|
54
55
|
except ValueError:
|
|
55
56
|
return value
|
|
56
57
|
return value
|
|
58
|
+
|
|
59
|
+
|
|
60
|
+
class UpdateConfigGroupByProductAndIDResponseTypedDict(TypedDict):
|
|
61
|
+
r"""a list of ConfigGroup objects"""
|
|
62
|
+
|
|
63
|
+
count: NotRequired[int]
|
|
64
|
+
r"""number of items present in the items array"""
|
|
65
|
+
items: NotRequired[List[ConfigGroupTypedDict]]
|
|
66
|
+
|
|
67
|
+
|
|
68
|
+
class UpdateConfigGroupByProductAndIDResponse(BaseModel):
|
|
69
|
+
r"""a list of ConfigGroup objects"""
|
|
70
|
+
|
|
71
|
+
count: Optional[int] = None
|
|
72
|
+
r"""number of items present in the items array"""
|
|
73
|
+
|
|
74
|
+
items: Optional[List[ConfigGroup]] = None
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from .configgroup import ConfigGroup, ConfigGroupTypedDict
|
|
4
5
|
from .deployrequest import DeployRequest, DeployRequestTypedDict
|
|
5
6
|
from .productscore import ProductsCore
|
|
6
7
|
from cribl_control_plane import models
|
|
@@ -13,7 +14,8 @@ from cribl_control_plane.utils import (
|
|
|
13
14
|
)
|
|
14
15
|
from pydantic import field_serializer
|
|
15
16
|
from pydantic.functional_validators import PlainValidator
|
|
16
|
-
from
|
|
17
|
+
from typing import List, Optional
|
|
18
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
17
19
|
|
|
18
20
|
|
|
19
21
|
class UpdateConfigGroupDeployByProductAndIDRequestTypedDict(TypedDict):
|
|
@@ -51,3 +53,20 @@ class UpdateConfigGroupDeployByProductAndIDRequest(BaseModel):
|
|
|
51
53
|
except ValueError:
|
|
52
54
|
return value
|
|
53
55
|
return value
|
|
56
|
+
|
|
57
|
+
|
|
58
|
+
class UpdateConfigGroupDeployByProductAndIDResponseTypedDict(TypedDict):
|
|
59
|
+
r"""a list of ConfigGroup objects"""
|
|
60
|
+
|
|
61
|
+
count: NotRequired[int]
|
|
62
|
+
r"""number of items present in the items array"""
|
|
63
|
+
items: NotRequired[List[ConfigGroupTypedDict]]
|
|
64
|
+
|
|
65
|
+
|
|
66
|
+
class UpdateConfigGroupDeployByProductAndIDResponse(BaseModel):
|
|
67
|
+
r"""a list of ConfigGroup objects"""
|
|
68
|
+
|
|
69
|
+
count: Optional[int] = None
|
|
70
|
+
r"""number of items present in the items array"""
|
|
71
|
+
|
|
72
|
+
items: Optional[List[ConfigGroup]] = None
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from .cribllakedataset import CriblLakeDataset, CriblLakeDatasetTypedDict
|
|
4
5
|
from .cribllakedatasetupdate import (
|
|
5
6
|
CriblLakeDatasetUpdate,
|
|
6
7
|
CriblLakeDatasetUpdateTypedDict,
|
|
@@ -8,7 +9,8 @@ from .cribllakedatasetupdate import (
|
|
|
8
9
|
from cribl_control_plane.types import BaseModel
|
|
9
10
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
10
11
|
import pydantic
|
|
11
|
-
from
|
|
12
|
+
from typing import List, Optional
|
|
13
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
14
|
|
|
13
15
|
|
|
14
16
|
class UpdateCriblLakeDatasetByLakeIDAndIDRequestTypedDict(TypedDict):
|
|
@@ -40,3 +42,20 @@ class UpdateCriblLakeDatasetByLakeIDAndIDRequest(BaseModel):
|
|
|
40
42
|
FieldMetadata(request=RequestMetadata(media_type="application/json")),
|
|
41
43
|
]
|
|
42
44
|
r"""CriblLakeDatasetUpdate object"""
|
|
45
|
+
|
|
46
|
+
|
|
47
|
+
class UpdateCriblLakeDatasetByLakeIDAndIDResponseTypedDict(TypedDict):
|
|
48
|
+
r"""a list of CriblLakeDataset objects"""
|
|
49
|
+
|
|
50
|
+
count: NotRequired[int]
|
|
51
|
+
r"""number of items present in the items array"""
|
|
52
|
+
items: NotRequired[List[CriblLakeDatasetTypedDict]]
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
class UpdateCriblLakeDatasetByLakeIDAndIDResponse(BaseModel):
|
|
56
|
+
r"""a list of CriblLakeDataset objects"""
|
|
57
|
+
|
|
58
|
+
count: Optional[int] = None
|
|
59
|
+
r"""number of items present in the items array"""
|
|
60
|
+
|
|
61
|
+
items: Optional[List[CriblLakeDataset]] = None
|
|
@@ -2,8 +2,9 @@
|
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
4
|
from cribl_control_plane.types import BaseModel
|
|
5
|
+
import pydantic
|
|
5
6
|
from typing import List, Optional
|
|
6
|
-
from typing_extensions import NotRequired, TypedDict
|
|
7
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
7
8
|
|
|
8
9
|
|
|
9
10
|
class UpdateHecTokenRequestMetadatumTypedDict(TypedDict):
|
|
@@ -18,12 +19,17 @@ class UpdateHecTokenRequestMetadatum(BaseModel):
|
|
|
18
19
|
|
|
19
20
|
|
|
20
21
|
class UpdateHecTokenRequestTypedDict(TypedDict):
|
|
22
|
+
allowed_indexes_at_token: NotRequired[List[str]]
|
|
21
23
|
description: NotRequired[str]
|
|
22
24
|
enabled: NotRequired[bool]
|
|
23
25
|
metadata: NotRequired[List[UpdateHecTokenRequestMetadatumTypedDict]]
|
|
24
26
|
|
|
25
27
|
|
|
26
28
|
class UpdateHecTokenRequest(BaseModel):
|
|
29
|
+
allowed_indexes_at_token: Annotated[
|
|
30
|
+
Optional[List[str]], pydantic.Field(alias="allowedIndexesAtToken")
|
|
31
|
+
] = None
|
|
32
|
+
|
|
27
33
|
description: Optional[str] = None
|
|
28
34
|
|
|
29
35
|
enabled: Optional[bool] = None
|
|
@@ -4,7 +4,8 @@ from __future__ import annotations
|
|
|
4
4
|
from .input import Input, InputTypedDict
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
7
|
-
from
|
|
7
|
+
from typing import List, Optional
|
|
8
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
8
9
|
|
|
9
10
|
|
|
10
11
|
class UpdateInputByIDRequestTypedDict(TypedDict):
|
|
@@ -24,3 +25,20 @@ class UpdateInputByIDRequest(BaseModel):
|
|
|
24
25
|
Input, FieldMetadata(request=RequestMetadata(media_type="application/json"))
|
|
25
26
|
]
|
|
26
27
|
r"""Input object"""
|
|
28
|
+
|
|
29
|
+
|
|
30
|
+
class UpdateInputByIDResponseTypedDict(TypedDict):
|
|
31
|
+
r"""a list of Source objects"""
|
|
32
|
+
|
|
33
|
+
count: NotRequired[int]
|
|
34
|
+
r"""number of items present in the items array"""
|
|
35
|
+
items: NotRequired[List[InputTypedDict]]
|
|
36
|
+
|
|
37
|
+
|
|
38
|
+
class UpdateInputByIDResponse(BaseModel):
|
|
39
|
+
r"""a list of Source objects"""
|
|
40
|
+
|
|
41
|
+
count: Optional[int] = None
|
|
42
|
+
r"""number of items present in the items array"""
|
|
43
|
+
|
|
44
|
+
items: Optional[List[Input]] = None
|
|
@@ -1,10 +1,12 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from .inputsplunkhec import InputSplunkHec, InputSplunkHecTypedDict
|
|
4
5
|
from .updatehectokenrequest import UpdateHecTokenRequest, UpdateHecTokenRequestTypedDict
|
|
5
6
|
from cribl_control_plane.types import BaseModel
|
|
6
7
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
7
|
-
from
|
|
8
|
+
from typing import List, Optional
|
|
9
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
8
10
|
|
|
9
11
|
|
|
10
12
|
class UpdateInputHecTokenByIDAndTokenRequestTypedDict(TypedDict):
|
|
@@ -32,3 +34,20 @@ class UpdateInputHecTokenByIDAndTokenRequest(BaseModel):
|
|
|
32
34
|
FieldMetadata(request=RequestMetadata(media_type="application/json")),
|
|
33
35
|
]
|
|
34
36
|
r"""UpdateHecTokenRequest object"""
|
|
37
|
+
|
|
38
|
+
|
|
39
|
+
class UpdateInputHecTokenByIDAndTokenResponseTypedDict(TypedDict):
|
|
40
|
+
r"""a list of InputSplunkHec objects"""
|
|
41
|
+
|
|
42
|
+
count: NotRequired[int]
|
|
43
|
+
r"""number of items present in the items array"""
|
|
44
|
+
items: NotRequired[List[InputSplunkHecTypedDict]]
|
|
45
|
+
|
|
46
|
+
|
|
47
|
+
class UpdateInputHecTokenByIDAndTokenResponse(BaseModel):
|
|
48
|
+
r"""a list of InputSplunkHec objects"""
|
|
49
|
+
|
|
50
|
+
count: Optional[int] = None
|
|
51
|
+
r"""number of items present in the items array"""
|
|
52
|
+
|
|
53
|
+
items: Optional[List[InputSplunkHec]] = None
|
|
@@ -4,7 +4,8 @@ from __future__ import annotations
|
|
|
4
4
|
from .output import Output, OutputTypedDict
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
7
|
-
from
|
|
7
|
+
from typing import List, Optional
|
|
8
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
8
9
|
|
|
9
10
|
|
|
10
11
|
class UpdateOutputByIDRequestTypedDict(TypedDict):
|
|
@@ -24,3 +25,20 @@ class UpdateOutputByIDRequest(BaseModel):
|
|
|
24
25
|
Output, FieldMetadata(request=RequestMetadata(media_type="application/json"))
|
|
25
26
|
]
|
|
26
27
|
r"""Output object"""
|
|
28
|
+
|
|
29
|
+
|
|
30
|
+
class UpdateOutputByIDResponseTypedDict(TypedDict):
|
|
31
|
+
r"""a list of Destination objects"""
|
|
32
|
+
|
|
33
|
+
count: NotRequired[int]
|
|
34
|
+
r"""number of items present in the items array"""
|
|
35
|
+
items: NotRequired[List[OutputTypedDict]]
|
|
36
|
+
|
|
37
|
+
|
|
38
|
+
class UpdateOutputByIDResponse(BaseModel):
|
|
39
|
+
r"""a list of Destination objects"""
|
|
40
|
+
|
|
41
|
+
count: Optional[int] = None
|
|
42
|
+
r"""number of items present in the items array"""
|
|
43
|
+
|
|
44
|
+
items: Optional[List[Output]] = None
|
|
@@ -1,10 +1,12 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from .packinfo import PackInfo, PackInfoTypedDict
|
|
4
5
|
from .packupgraderequest import PackUpgradeRequest, PackUpgradeRequestTypedDict
|
|
5
6
|
from cribl_control_plane.types import BaseModel
|
|
6
7
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
7
|
-
from
|
|
8
|
+
from typing import List, Optional
|
|
9
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
8
10
|
|
|
9
11
|
|
|
10
12
|
class UpdatePacksByIDRequestTypedDict(TypedDict):
|
|
@@ -25,3 +27,20 @@ class UpdatePacksByIDRequest(BaseModel):
|
|
|
25
27
|
FieldMetadata(request=RequestMetadata(media_type="application/json")),
|
|
26
28
|
]
|
|
27
29
|
r"""PackUpgradeRequest object"""
|
|
30
|
+
|
|
31
|
+
|
|
32
|
+
class UpdatePacksByIDResponseTypedDict(TypedDict):
|
|
33
|
+
r"""a list of PackInfo objects"""
|
|
34
|
+
|
|
35
|
+
count: NotRequired[int]
|
|
36
|
+
r"""number of items present in the items array"""
|
|
37
|
+
items: NotRequired[List[PackInfoTypedDict]]
|
|
38
|
+
|
|
39
|
+
|
|
40
|
+
class UpdatePacksByIDResponse(BaseModel):
|
|
41
|
+
r"""a list of PackInfo objects"""
|
|
42
|
+
|
|
43
|
+
count: Optional[int] = None
|
|
44
|
+
r"""number of items present in the items array"""
|
|
45
|
+
|
|
46
|
+
items: Optional[List[PackInfo]] = None
|
|
@@ -5,7 +5,8 @@ from .pipeline import Pipeline, PipelineTypedDict
|
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
7
7
|
import pydantic
|
|
8
|
-
from
|
|
8
|
+
from typing import List, Optional
|
|
9
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
10
|
|
|
10
11
|
|
|
11
12
|
class UpdatePipelineByIDRequestTypedDict(TypedDict):
|
|
@@ -27,3 +28,20 @@ class UpdatePipelineByIDRequest(BaseModel):
|
|
|
27
28
|
Pipeline, FieldMetadata(request=RequestMetadata(media_type="application/json"))
|
|
28
29
|
]
|
|
29
30
|
r"""Pipeline object"""
|
|
31
|
+
|
|
32
|
+
|
|
33
|
+
class UpdatePipelineByIDResponseTypedDict(TypedDict):
|
|
34
|
+
r"""a list of Pipeline objects"""
|
|
35
|
+
|
|
36
|
+
count: NotRequired[int]
|
|
37
|
+
r"""number of items present in the items array"""
|
|
38
|
+
items: NotRequired[List[PipelineTypedDict]]
|
|
39
|
+
|
|
40
|
+
|
|
41
|
+
class UpdatePipelineByIDResponse(BaseModel):
|
|
42
|
+
r"""a list of Pipeline objects"""
|
|
43
|
+
|
|
44
|
+
count: Optional[int] = None
|
|
45
|
+
r"""number of items present in the items array"""
|
|
46
|
+
|
|
47
|
+
items: Optional[List[Pipeline]] = None
|
|
@@ -5,7 +5,8 @@ from .routes import Routes, RoutesTypedDict
|
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata, RequestMetadata
|
|
7
7
|
import pydantic
|
|
8
|
-
from
|
|
8
|
+
from typing import List, Optional
|
|
9
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
10
|
|
|
10
11
|
|
|
11
12
|
class UpdateRoutesByIDRequestTypedDict(TypedDict):
|
|
@@ -27,3 +28,20 @@ class UpdateRoutesByIDRequest(BaseModel):
|
|
|
27
28
|
Routes, FieldMetadata(request=RequestMetadata(media_type="application/json"))
|
|
28
29
|
]
|
|
29
30
|
r"""Routes object"""
|
|
31
|
+
|
|
32
|
+
|
|
33
|
+
class UpdateRoutesByIDResponseTypedDict(TypedDict):
|
|
34
|
+
r"""a list of Routes objects"""
|
|
35
|
+
|
|
36
|
+
count: NotRequired[int]
|
|
37
|
+
r"""number of items present in the items array"""
|
|
38
|
+
items: NotRequired[List[RoutesTypedDict]]
|
|
39
|
+
|
|
40
|
+
|
|
41
|
+
class UpdateRoutesByIDResponse(BaseModel):
|
|
42
|
+
r"""a list of Routes objects"""
|
|
43
|
+
|
|
44
|
+
count: Optional[int] = None
|
|
45
|
+
r"""number of items present in the items array"""
|
|
46
|
+
|
|
47
|
+
items: Optional[List[Routes]] = None
|
cribl_control_plane/nodes.py
CHANGED
|
@@ -37,7 +37,7 @@ class Nodes(BaseSDK):
|
|
|
37
37
|
server_url: Optional[str] = None,
|
|
38
38
|
timeout_ms: Optional[int] = None,
|
|
39
39
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
40
|
-
) -> models.
|
|
40
|
+
) -> models.ListMasterWorkerEntryResponse:
|
|
41
41
|
r"""Get detailed metadata for Worker and Edge Nodes
|
|
42
42
|
|
|
43
43
|
Get detailed metadata for Worker and Edge Nodes.
|
|
@@ -85,6 +85,7 @@ class Nodes(BaseSDK):
|
|
|
85
85
|
accept_header_value="application/json",
|
|
86
86
|
http_headers=http_headers,
|
|
87
87
|
security=self.sdk_configuration.security,
|
|
88
|
+
allow_empty_value=None,
|
|
88
89
|
timeout_ms=timeout_ms,
|
|
89
90
|
)
|
|
90
91
|
|
|
@@ -114,7 +115,7 @@ class Nodes(BaseSDK):
|
|
|
114
115
|
response_data: Any = None
|
|
115
116
|
if utils.match_response(http_res, "200", "application/json"):
|
|
116
117
|
return unmarshal_json_response(
|
|
117
|
-
models.
|
|
118
|
+
models.ListMasterWorkerEntryResponse, http_res
|
|
118
119
|
)
|
|
119
120
|
if utils.match_response(http_res, "500", "application/json"):
|
|
120
121
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
@@ -141,7 +142,7 @@ class Nodes(BaseSDK):
|
|
|
141
142
|
server_url: Optional[str] = None,
|
|
142
143
|
timeout_ms: Optional[int] = None,
|
|
143
144
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
144
|
-
) -> models.
|
|
145
|
+
) -> models.ListMasterWorkerEntryResponse:
|
|
145
146
|
r"""Get detailed metadata for Worker and Edge Nodes
|
|
146
147
|
|
|
147
148
|
Get detailed metadata for Worker and Edge Nodes.
|
|
@@ -189,6 +190,7 @@ class Nodes(BaseSDK):
|
|
|
189
190
|
accept_header_value="application/json",
|
|
190
191
|
http_headers=http_headers,
|
|
191
192
|
security=self.sdk_configuration.security,
|
|
193
|
+
allow_empty_value=None,
|
|
192
194
|
timeout_ms=timeout_ms,
|
|
193
195
|
)
|
|
194
196
|
|
|
@@ -218,7 +220,7 @@ class Nodes(BaseSDK):
|
|
|
218
220
|
response_data: Any = None
|
|
219
221
|
if utils.match_response(http_res, "200", "application/json"):
|
|
220
222
|
return unmarshal_json_response(
|
|
221
|
-
models.
|
|
223
|
+
models.ListMasterWorkerEntryResponse, http_res
|
|
222
224
|
)
|
|
223
225
|
if utils.match_response(http_res, "500", "application/json"):
|
|
224
226
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
@@ -240,7 +242,7 @@ class Nodes(BaseSDK):
|
|
|
240
242
|
server_url: Optional[str] = None,
|
|
241
243
|
timeout_ms: Optional[int] = None,
|
|
242
244
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
243
|
-
) -> models.
|
|
245
|
+
) -> models.GetMasterWorkerEntryResponse:
|
|
244
246
|
r"""Get a count of Worker and Edge Nodes
|
|
245
247
|
|
|
246
248
|
Get a count of all Worker and Edge Nodes.
|
|
@@ -278,6 +280,7 @@ class Nodes(BaseSDK):
|
|
|
278
280
|
accept_header_value="application/json",
|
|
279
281
|
http_headers=http_headers,
|
|
280
282
|
security=self.sdk_configuration.security,
|
|
283
|
+
allow_empty_value=None,
|
|
281
284
|
timeout_ms=timeout_ms,
|
|
282
285
|
)
|
|
283
286
|
|
|
@@ -306,7 +309,9 @@ class Nodes(BaseSDK):
|
|
|
306
309
|
|
|
307
310
|
response_data: Any = None
|
|
308
311
|
if utils.match_response(http_res, "200", "application/json"):
|
|
309
|
-
return unmarshal_json_response(
|
|
312
|
+
return unmarshal_json_response(
|
|
313
|
+
models.GetMasterWorkerEntryResponse, http_res
|
|
314
|
+
)
|
|
310
315
|
if utils.match_response(http_res, "500", "application/json"):
|
|
311
316
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
312
317
|
raise errors.Error(response_data, http_res)
|
|
@@ -327,7 +332,7 @@ class Nodes(BaseSDK):
|
|
|
327
332
|
server_url: Optional[str] = None,
|
|
328
333
|
timeout_ms: Optional[int] = None,
|
|
329
334
|
http_headers: Optional[Mapping[str, str]] = None,
|
|
330
|
-
) -> models.
|
|
335
|
+
) -> models.GetMasterWorkerEntryResponse:
|
|
331
336
|
r"""Get a count of Worker and Edge Nodes
|
|
332
337
|
|
|
333
338
|
Get a count of all Worker and Edge Nodes.
|
|
@@ -365,6 +370,7 @@ class Nodes(BaseSDK):
|
|
|
365
370
|
accept_header_value="application/json",
|
|
366
371
|
http_headers=http_headers,
|
|
367
372
|
security=self.sdk_configuration.security,
|
|
373
|
+
allow_empty_value=None,
|
|
368
374
|
timeout_ms=timeout_ms,
|
|
369
375
|
)
|
|
370
376
|
|
|
@@ -393,7 +399,9 @@ class Nodes(BaseSDK):
|
|
|
393
399
|
|
|
394
400
|
response_data: Any = None
|
|
395
401
|
if utils.match_response(http_res, "200", "application/json"):
|
|
396
|
-
return unmarshal_json_response(
|
|
402
|
+
return unmarshal_json_response(
|
|
403
|
+
models.GetMasterWorkerEntryResponse, http_res
|
|
404
|
+
)
|
|
397
405
|
if utils.match_response(http_res, "500", "application/json"):
|
|
398
406
|
response_data = unmarshal_json_response(errors.ErrorData, http_res)
|
|
399
407
|
raise errors.Error(response_data, http_res)
|