cribl-control-plane 0.2.1rc7__py3-none-any.whl → 0.3.0a1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +4 -4
- cribl_control_plane/errors/__init__.py +5 -8
- cribl_control_plane/errors/{healthserverstatus_error.py → healthstatus_error.py} +9 -10
- cribl_control_plane/groups_sdk.py +28 -52
- cribl_control_plane/health.py +16 -22
- cribl_control_plane/models/__init__.py +54 -217
- cribl_control_plane/models/appmode.py +14 -0
- cribl_control_plane/models/authtoken.py +1 -5
- cribl_control_plane/models/cacheconnection.py +0 -20
- cribl_control_plane/models/configgroup.py +7 -55
- cribl_control_plane/models/configgroupcloud.py +1 -11
- cribl_control_plane/models/createconfiggroupbyproductop.py +5 -17
- cribl_control_plane/models/createroutesappendbyidop.py +2 -2
- cribl_control_plane/models/createversionundoop.py +3 -3
- cribl_control_plane/models/cribllakedataset.py +1 -11
- cribl_control_plane/models/cribllakedatasetupdate.py +1 -11
- cribl_control_plane/models/datasetmetadata.py +1 -11
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +0 -11
- cribl_control_plane/models/deleteoutputpqbyidop.py +2 -2
- cribl_control_plane/models/distributedsummary.py +0 -6
- cribl_control_plane/models/error.py +16 -0
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +0 -20
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +0 -20
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +0 -11
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +0 -11
- cribl_control_plane/models/gethealthinfoop.py +17 -0
- cribl_control_plane/models/getsummaryop.py +0 -11
- cribl_control_plane/models/hbcriblinfo.py +3 -24
- cribl_control_plane/models/{healthserverstatus.py → healthstatus.py} +8 -27
- cribl_control_plane/models/heartbeatmetadata.py +0 -3
- cribl_control_plane/models/input.py +78 -80
- cribl_control_plane/models/inputappscope.py +17 -80
- cribl_control_plane/models/inputazureblob.py +1 -33
- cribl_control_plane/models/inputcollection.py +1 -24
- cribl_control_plane/models/inputconfluentcloud.py +18 -195
- cribl_control_plane/models/inputcribl.py +1 -24
- cribl_control_plane/models/inputcriblhttp.py +17 -62
- cribl_control_plane/models/inputcribllakehttp.py +17 -62
- cribl_control_plane/models/inputcriblmetrics.py +1 -24
- cribl_control_plane/models/inputcribltcp.py +17 -62
- cribl_control_plane/models/inputcrowdstrike.py +1 -54
- cribl_control_plane/models/inputdatadogagent.py +17 -62
- cribl_control_plane/models/inputdatagen.py +1 -24
- cribl_control_plane/models/inputedgeprometheus.py +34 -147
- cribl_control_plane/models/inputelastic.py +27 -119
- cribl_control_plane/models/inputeventhub.py +1 -182
- cribl_control_plane/models/inputexec.py +1 -33
- cribl_control_plane/models/inputfile.py +3 -42
- cribl_control_plane/models/inputfirehose.py +17 -62
- cribl_control_plane/models/inputgooglepubsub.py +1 -36
- cribl_control_plane/models/inputgrafana.py +32 -157
- cribl_control_plane/models/inputhttp.py +17 -62
- cribl_control_plane/models/inputhttpraw.py +17 -62
- cribl_control_plane/models/inputjournalfiles.py +1 -24
- cribl_control_plane/models/inputkafka.py +17 -189
- cribl_control_plane/models/inputkinesis.py +1 -80
- cribl_control_plane/models/inputkubeevents.py +1 -24
- cribl_control_plane/models/inputkubelogs.py +1 -33
- cribl_control_plane/models/inputkubemetrics.py +1 -33
- cribl_control_plane/models/inputloki.py +17 -71
- cribl_control_plane/models/inputmetrics.py +17 -62
- cribl_control_plane/models/inputmodeldriventelemetry.py +17 -62
- cribl_control_plane/models/inputmsk.py +18 -81
- cribl_control_plane/models/inputnetflow.py +1 -24
- cribl_control_plane/models/inputoffice365mgmt.py +1 -67
- cribl_control_plane/models/inputoffice365msgtrace.py +1 -67
- cribl_control_plane/models/inputoffice365service.py +1 -67
- cribl_control_plane/models/inputopentelemetry.py +16 -92
- cribl_control_plane/models/inputprometheus.py +34 -138
- cribl_control_plane/models/inputprometheusrw.py +17 -71
- cribl_control_plane/models/inputrawudp.py +1 -24
- cribl_control_plane/models/inputs3.py +1 -45
- cribl_control_plane/models/inputs3inventory.py +1 -54
- cribl_control_plane/models/inputsecuritylake.py +1 -54
- cribl_control_plane/models/inputsnmp.py +1 -40
- cribl_control_plane/models/inputsplunk.py +17 -85
- cribl_control_plane/models/inputsplunkhec.py +16 -70
- cribl_control_plane/models/inputsplunksearch.py +1 -63
- cribl_control_plane/models/inputsqs.py +1 -56
- cribl_control_plane/models/inputsyslog.py +32 -121
- cribl_control_plane/models/inputsystemmetrics.py +9 -142
- cribl_control_plane/models/inputsystemstate.py +1 -33
- cribl_control_plane/models/inputtcp.py +17 -81
- cribl_control_plane/models/inputtcpjson.py +17 -71
- cribl_control_plane/models/inputwef.py +1 -71
- cribl_control_plane/models/inputwindowsmetrics.py +9 -129
- cribl_control_plane/models/inputwineventlogs.py +1 -60
- cribl_control_plane/models/inputwiz.py +1 -45
- cribl_control_plane/models/inputwizwebhook.py +17 -62
- cribl_control_plane/models/inputzscalerhec.py +16 -70
- cribl_control_plane/models/jobinfo.py +1 -4
- cribl_control_plane/models/jobstatus.py +3 -34
- cribl_control_plane/models/listconfiggroupbyproductop.py +0 -11
- cribl_control_plane/models/logininfo.py +3 -3
- cribl_control_plane/models/masterworkerentry.py +1 -11
- cribl_control_plane/models/nodeprovidedinfo.py +1 -11
- cribl_control_plane/models/nodeupgradestatus.py +0 -38
- cribl_control_plane/models/output.py +88 -93
- cribl_control_plane/models/outputazureblob.py +1 -110
- cribl_control_plane/models/outputazuredataexplorer.py +87 -452
- cribl_control_plane/models/outputazureeventhub.py +19 -281
- cribl_control_plane/models/outputazurelogs.py +19 -115
- cribl_control_plane/models/outputchronicle.py +19 -115
- cribl_control_plane/models/outputclickhouse.py +19 -155
- cribl_control_plane/models/outputcloudwatch.py +19 -106
- cribl_control_plane/models/outputconfluentcloud.py +38 -311
- cribl_control_plane/models/outputcriblhttp.py +19 -135
- cribl_control_plane/models/outputcribllake.py +1 -97
- cribl_control_plane/models/outputcribltcp.py +19 -132
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +20 -129
- cribl_control_plane/models/outputdatadog.py +19 -159
- cribl_control_plane/models/outputdataset.py +19 -143
- cribl_control_plane/models/outputdiskspool.py +1 -11
- cribl_control_plane/models/outputdls3.py +1 -152
- cribl_control_plane/models/outputdynatracehttp.py +19 -160
- cribl_control_plane/models/outputdynatraceotlp.py +19 -160
- cribl_control_plane/models/outputelastic.py +19 -163
- cribl_control_plane/models/outputelasticcloud.py +19 -140
- cribl_control_plane/models/outputexabeam.py +1 -61
- cribl_control_plane/models/outputfilesystem.py +1 -87
- cribl_control_plane/models/outputgooglechronicle.py +20 -166
- cribl_control_plane/models/outputgooglecloudlogging.py +20 -131
- cribl_control_plane/models/outputgooglecloudstorage.py +1 -136
- cribl_control_plane/models/outputgooglepubsub.py +19 -106
- cribl_control_plane/models/outputgrafanacloud.py +37 -288
- cribl_control_plane/models/outputgraphite.py +19 -105
- cribl_control_plane/models/outputhoneycomb.py +19 -115
- cribl_control_plane/models/outputhumiohec.py +19 -126
- cribl_control_plane/models/outputinfluxdb.py +19 -130
- cribl_control_plane/models/outputkafka.py +34 -302
- cribl_control_plane/models/outputkinesis.py +19 -133
- cribl_control_plane/models/outputloki.py +17 -129
- cribl_control_plane/models/outputminio.py +1 -145
- cribl_control_plane/models/outputmsk.py +34 -193
- cribl_control_plane/models/outputnewrelic.py +19 -136
- cribl_control_plane/models/outputnewrelicevents.py +20 -128
- cribl_control_plane/models/outputopentelemetry.py +19 -178
- cribl_control_plane/models/outputprometheus.py +19 -115
- cribl_control_plane/models/outputring.py +1 -31
- cribl_control_plane/models/outputs3.py +1 -152
- cribl_control_plane/models/outputsecuritylake.py +1 -114
- cribl_control_plane/models/outputsentinel.py +19 -135
- cribl_control_plane/models/outputsentineloneaisiem.py +20 -134
- cribl_control_plane/models/outputservicenow.py +19 -168
- cribl_control_plane/models/outputsignalfx.py +19 -115
- cribl_control_plane/models/outputsns.py +17 -113
- cribl_control_plane/models/outputsplunk.py +19 -153
- cribl_control_plane/models/outputsplunkhec.py +19 -208
- cribl_control_plane/models/outputsplunklb.py +19 -182
- cribl_control_plane/models/outputsqs.py +17 -124
- cribl_control_plane/models/outputstatsd.py +19 -105
- cribl_control_plane/models/outputstatsdext.py +19 -105
- cribl_control_plane/models/outputsumologic.py +19 -117
- cribl_control_plane/models/outputsyslog.py +96 -259
- cribl_control_plane/models/outputtcpjson.py +19 -141
- cribl_control_plane/models/outputwavefront.py +19 -115
- cribl_control_plane/models/outputwebhook.py +19 -161
- cribl_control_plane/models/outputxsiam.py +17 -113
- cribl_control_plane/models/packinfo.py +5 -8
- cribl_control_plane/models/packinstallinfo.py +5 -8
- cribl_control_plane/models/resourcepolicy.py +0 -11
- cribl_control_plane/models/{uploadpackresponse.py → routecloneconf.py} +4 -4
- cribl_control_plane/models/routeconf.py +4 -3
- cribl_control_plane/models/runnablejobcollection.py +9 -72
- cribl_control_plane/models/runnablejobexecutor.py +9 -32
- cribl_control_plane/models/runnablejobscheduledsearch.py +9 -23
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +0 -11
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +0 -11
- cribl_control_plane/packs.py +7 -202
- cribl_control_plane/routes_sdk.py +6 -6
- cribl_control_plane/tokens.py +15 -23
- {cribl_control_plane-0.2.1rc7.dist-info → cribl_control_plane-0.3.0a1.dist-info}/METADATA +9 -50
- cribl_control_plane-0.3.0a1.dist-info/RECORD +330 -0
- cribl_control_plane/models/groupcreaterequest.py +0 -171
- cribl_control_plane/models/outpostnodeinfo.py +0 -16
- cribl_control_plane/models/outputdatabricks.py +0 -482
- cribl_control_plane/models/updatepacksop.py +0 -25
- cribl_control_plane-0.2.1rc7.dist-info/RECORD +0 -331
- {cribl_control_plane-0.2.1rc7.dist-info → cribl_control_plane-0.3.0a1.dist-info}/WHEEL +0 -0
|
@@ -1,12 +1,11 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import
|
|
4
|
+
from cribl_control_plane import utils
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import validate_open_enum
|
|
7
7
|
from enum import Enum
|
|
8
8
|
import pydantic
|
|
9
|
-
from pydantic import field_serializer
|
|
10
9
|
from pydantic.functional_validators import PlainValidator
|
|
11
10
|
from typing import List, Optional
|
|
12
11
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
@@ -19,11 +18,8 @@ class OutputS3Type(str, Enum):
|
|
|
19
18
|
class OutputS3AuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
20
19
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
21
20
|
|
|
22
|
-
# Auto
|
|
23
21
|
AUTO = "auto"
|
|
24
|
-
# Manual
|
|
25
22
|
MANUAL = "manual"
|
|
26
|
-
# Secret Key pair
|
|
27
23
|
SECRET = "secret"
|
|
28
24
|
|
|
29
25
|
|
|
@@ -37,78 +33,54 @@ class OutputS3SignatureVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
37
33
|
class OutputS3ObjectACL(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
38
34
|
r"""Object ACL to assign to uploaded objects"""
|
|
39
35
|
|
|
40
|
-
# Private
|
|
41
36
|
PRIVATE = "private"
|
|
42
|
-
# Public Read Only
|
|
43
37
|
PUBLIC_READ = "public-read"
|
|
44
|
-
# Public Read/Write
|
|
45
38
|
PUBLIC_READ_WRITE = "public-read-write"
|
|
46
|
-
# Authenticated Read Only
|
|
47
39
|
AUTHENTICATED_READ = "authenticated-read"
|
|
48
|
-
# AWS EC2 AMI Read Only
|
|
49
40
|
AWS_EXEC_READ = "aws-exec-read"
|
|
50
|
-
# Bucket Owner Read Only
|
|
51
41
|
BUCKET_OWNER_READ = "bucket-owner-read"
|
|
52
|
-
# Bucket Owner Full Control
|
|
53
42
|
BUCKET_OWNER_FULL_CONTROL = "bucket-owner-full-control"
|
|
54
43
|
|
|
55
44
|
|
|
56
45
|
class OutputS3StorageClass(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
57
46
|
r"""Storage class to select for uploaded objects"""
|
|
58
47
|
|
|
59
|
-
# Standard
|
|
60
48
|
STANDARD = "STANDARD"
|
|
61
|
-
# Reduced Redundancy Storage
|
|
62
49
|
REDUCED_REDUNDANCY = "REDUCED_REDUNDANCY"
|
|
63
|
-
# Standard, Infrequent Access
|
|
64
50
|
STANDARD_IA = "STANDARD_IA"
|
|
65
|
-
# One Zone, Infrequent Access
|
|
66
51
|
ONEZONE_IA = "ONEZONE_IA"
|
|
67
|
-
# Intelligent Tiering
|
|
68
52
|
INTELLIGENT_TIERING = "INTELLIGENT_TIERING"
|
|
69
|
-
# Glacier Flexible Retrieval
|
|
70
53
|
GLACIER = "GLACIER"
|
|
71
|
-
# Glacier Instant Retrieval
|
|
72
54
|
GLACIER_IR = "GLACIER_IR"
|
|
73
|
-
# Glacier Deep Archive
|
|
74
55
|
DEEP_ARCHIVE = "DEEP_ARCHIVE"
|
|
75
56
|
|
|
76
57
|
|
|
77
58
|
class OutputS3ServerSideEncryptionForUploadedObjects(
|
|
78
59
|
str, Enum, metaclass=utils.OpenEnumMeta
|
|
79
60
|
):
|
|
80
|
-
# Amazon S3 Managed Key
|
|
81
61
|
AES256 = "AES256"
|
|
82
|
-
# AWS KMS Managed Key
|
|
83
62
|
AWS_KMS = "aws:kms"
|
|
84
63
|
|
|
85
64
|
|
|
86
65
|
class OutputS3DataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
87
66
|
r"""Format of the output data"""
|
|
88
67
|
|
|
89
|
-
# JSON
|
|
90
68
|
JSON = "json"
|
|
91
|
-
# Raw
|
|
92
69
|
RAW = "raw"
|
|
93
|
-
# Parquet
|
|
94
70
|
PARQUET = "parquet"
|
|
95
71
|
|
|
96
72
|
|
|
97
73
|
class OutputS3BackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
98
74
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
99
75
|
|
|
100
|
-
# Block
|
|
101
76
|
BLOCK = "block"
|
|
102
|
-
# Drop
|
|
103
77
|
DROP = "drop"
|
|
104
78
|
|
|
105
79
|
|
|
106
80
|
class OutputS3DiskSpaceProtection(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
107
81
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
108
82
|
|
|
109
|
-
# Block
|
|
110
83
|
BLOCK = "block"
|
|
111
|
-
# Drop
|
|
112
84
|
DROP = "drop"
|
|
113
85
|
|
|
114
86
|
|
|
@@ -122,31 +94,23 @@ class OutputS3Compression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
122
94
|
class OutputS3CompressionLevel(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
123
95
|
r"""Compression level to apply before moving files to final destination"""
|
|
124
96
|
|
|
125
|
-
# Best Speed
|
|
126
97
|
BEST_SPEED = "best_speed"
|
|
127
|
-
# Normal
|
|
128
98
|
NORMAL = "normal"
|
|
129
|
-
# Best Compression
|
|
130
99
|
BEST_COMPRESSION = "best_compression"
|
|
131
100
|
|
|
132
101
|
|
|
133
102
|
class OutputS3ParquetVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
134
103
|
r"""Determines which data types are supported and how they are represented"""
|
|
135
104
|
|
|
136
|
-
# 1.0
|
|
137
105
|
PARQUET_1_0 = "PARQUET_1_0"
|
|
138
|
-
# 2.4
|
|
139
106
|
PARQUET_2_4 = "PARQUET_2_4"
|
|
140
|
-
# 2.6
|
|
141
107
|
PARQUET_2_6 = "PARQUET_2_6"
|
|
142
108
|
|
|
143
109
|
|
|
144
110
|
class OutputS3DataPageVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
145
111
|
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
146
112
|
|
|
147
|
-
# V1
|
|
148
113
|
DATA_PAGE_V1 = "DATA_PAGE_V1"
|
|
149
|
-
# V2
|
|
150
114
|
DATA_PAGE_V2 = "DATA_PAGE_V2"
|
|
151
115
|
|
|
152
116
|
|
|
@@ -255,8 +219,6 @@ class OutputS3TypedDict(TypedDict):
|
|
|
255
219
|
r"""Compression level to apply before moving files to final destination"""
|
|
256
220
|
automatic_schema: NotRequired[bool]
|
|
257
221
|
r"""Automatically calculate the schema based on the events of each Parquet file generated"""
|
|
258
|
-
parquet_schema: NotRequired[str]
|
|
259
|
-
r"""To add a new schema, navigate to Processing > Knowledge > Parquet Schemas"""
|
|
260
222
|
parquet_version: NotRequired[OutputS3ParquetVersion]
|
|
261
223
|
r"""Determines which data types are supported and how they are represented"""
|
|
262
224
|
parquet_data_page_version: NotRequired[OutputS3DataPageVersion]
|
|
@@ -526,11 +488,6 @@ class OutputS3(BaseModel):
|
|
|
526
488
|
] = False
|
|
527
489
|
r"""Automatically calculate the schema based on the events of each Parquet file generated"""
|
|
528
490
|
|
|
529
|
-
parquet_schema: Annotated[Optional[str], pydantic.Field(alias="parquetSchema")] = (
|
|
530
|
-
None
|
|
531
|
-
)
|
|
532
|
-
r"""To add a new schema, navigate to Processing > Knowledge > Parquet Schemas"""
|
|
533
|
-
|
|
534
491
|
parquet_version: Annotated[
|
|
535
492
|
Annotated[
|
|
536
493
|
Optional[OutputS3ParquetVersion], PlainValidator(validate_open_enum(False))
|
|
@@ -595,111 +552,3 @@ class OutputS3(BaseModel):
|
|
|
595
552
|
|
|
596
553
|
max_retry_num: Annotated[Optional[float], pydantic.Field(alias="maxRetryNum")] = 20
|
|
597
554
|
r"""The maximum number of times a file will attempt to move to its final destination before being dead-lettered"""
|
|
598
|
-
|
|
599
|
-
@field_serializer("aws_authentication_method")
|
|
600
|
-
def serialize_aws_authentication_method(self, value):
|
|
601
|
-
if isinstance(value, str):
|
|
602
|
-
try:
|
|
603
|
-
return models.OutputS3AuthenticationMethod(value)
|
|
604
|
-
except ValueError:
|
|
605
|
-
return value
|
|
606
|
-
return value
|
|
607
|
-
|
|
608
|
-
@field_serializer("signature_version")
|
|
609
|
-
def serialize_signature_version(self, value):
|
|
610
|
-
if isinstance(value, str):
|
|
611
|
-
try:
|
|
612
|
-
return models.OutputS3SignatureVersion(value)
|
|
613
|
-
except ValueError:
|
|
614
|
-
return value
|
|
615
|
-
return value
|
|
616
|
-
|
|
617
|
-
@field_serializer("object_acl")
|
|
618
|
-
def serialize_object_acl(self, value):
|
|
619
|
-
if isinstance(value, str):
|
|
620
|
-
try:
|
|
621
|
-
return models.OutputS3ObjectACL(value)
|
|
622
|
-
except ValueError:
|
|
623
|
-
return value
|
|
624
|
-
return value
|
|
625
|
-
|
|
626
|
-
@field_serializer("storage_class")
|
|
627
|
-
def serialize_storage_class(self, value):
|
|
628
|
-
if isinstance(value, str):
|
|
629
|
-
try:
|
|
630
|
-
return models.OutputS3StorageClass(value)
|
|
631
|
-
except ValueError:
|
|
632
|
-
return value
|
|
633
|
-
return value
|
|
634
|
-
|
|
635
|
-
@field_serializer("server_side_encryption")
|
|
636
|
-
def serialize_server_side_encryption(self, value):
|
|
637
|
-
if isinstance(value, str):
|
|
638
|
-
try:
|
|
639
|
-
return models.OutputS3ServerSideEncryptionForUploadedObjects(value)
|
|
640
|
-
except ValueError:
|
|
641
|
-
return value
|
|
642
|
-
return value
|
|
643
|
-
|
|
644
|
-
@field_serializer("format_")
|
|
645
|
-
def serialize_format_(self, value):
|
|
646
|
-
if isinstance(value, str):
|
|
647
|
-
try:
|
|
648
|
-
return models.OutputS3DataFormat(value)
|
|
649
|
-
except ValueError:
|
|
650
|
-
return value
|
|
651
|
-
return value
|
|
652
|
-
|
|
653
|
-
@field_serializer("on_backpressure")
|
|
654
|
-
def serialize_on_backpressure(self, value):
|
|
655
|
-
if isinstance(value, str):
|
|
656
|
-
try:
|
|
657
|
-
return models.OutputS3BackpressureBehavior(value)
|
|
658
|
-
except ValueError:
|
|
659
|
-
return value
|
|
660
|
-
return value
|
|
661
|
-
|
|
662
|
-
@field_serializer("on_disk_full_backpressure")
|
|
663
|
-
def serialize_on_disk_full_backpressure(self, value):
|
|
664
|
-
if isinstance(value, str):
|
|
665
|
-
try:
|
|
666
|
-
return models.OutputS3DiskSpaceProtection(value)
|
|
667
|
-
except ValueError:
|
|
668
|
-
return value
|
|
669
|
-
return value
|
|
670
|
-
|
|
671
|
-
@field_serializer("compress")
|
|
672
|
-
def serialize_compress(self, value):
|
|
673
|
-
if isinstance(value, str):
|
|
674
|
-
try:
|
|
675
|
-
return models.OutputS3Compression(value)
|
|
676
|
-
except ValueError:
|
|
677
|
-
return value
|
|
678
|
-
return value
|
|
679
|
-
|
|
680
|
-
@field_serializer("compression_level")
|
|
681
|
-
def serialize_compression_level(self, value):
|
|
682
|
-
if isinstance(value, str):
|
|
683
|
-
try:
|
|
684
|
-
return models.OutputS3CompressionLevel(value)
|
|
685
|
-
except ValueError:
|
|
686
|
-
return value
|
|
687
|
-
return value
|
|
688
|
-
|
|
689
|
-
@field_serializer("parquet_version")
|
|
690
|
-
def serialize_parquet_version(self, value):
|
|
691
|
-
if isinstance(value, str):
|
|
692
|
-
try:
|
|
693
|
-
return models.OutputS3ParquetVersion(value)
|
|
694
|
-
except ValueError:
|
|
695
|
-
return value
|
|
696
|
-
return value
|
|
697
|
-
|
|
698
|
-
@field_serializer("parquet_data_page_version")
|
|
699
|
-
def serialize_parquet_data_page_version(self, value):
|
|
700
|
-
if isinstance(value, str):
|
|
701
|
-
try:
|
|
702
|
-
return models.OutputS3DataPageVersion(value)
|
|
703
|
-
except ValueError:
|
|
704
|
-
return value
|
|
705
|
-
return value
|
|
@@ -1,12 +1,11 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import
|
|
4
|
+
from cribl_control_plane import utils
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import validate_open_enum
|
|
7
7
|
from enum import Enum
|
|
8
8
|
import pydantic
|
|
9
|
-
from pydantic import field_serializer
|
|
10
9
|
from pydantic.functional_validators import PlainValidator
|
|
11
10
|
from typing import List, Optional
|
|
12
11
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
@@ -19,11 +18,8 @@ class OutputSecurityLakeType(str, Enum):
|
|
|
19
18
|
class OutputSecurityLakeAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
20
19
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
21
20
|
|
|
22
|
-
# Auto
|
|
23
21
|
AUTO = "auto"
|
|
24
|
-
# Manual
|
|
25
22
|
MANUAL = "manual"
|
|
26
|
-
# Secret Key pair
|
|
27
23
|
SECRET = "secret"
|
|
28
24
|
|
|
29
25
|
|
|
@@ -37,87 +33,61 @@ class OutputSecurityLakeSignatureVersion(str, Enum, metaclass=utils.OpenEnumMeta
|
|
|
37
33
|
class OutputSecurityLakeObjectACL(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
38
34
|
r"""Object ACL to assign to uploaded objects"""
|
|
39
35
|
|
|
40
|
-
# Private
|
|
41
36
|
PRIVATE = "private"
|
|
42
|
-
# Public Read Only
|
|
43
37
|
PUBLIC_READ = "public-read"
|
|
44
|
-
# Public Read/Write
|
|
45
38
|
PUBLIC_READ_WRITE = "public-read-write"
|
|
46
|
-
# Authenticated Read Only
|
|
47
39
|
AUTHENTICATED_READ = "authenticated-read"
|
|
48
|
-
# AWS EC2 AMI Read Only
|
|
49
40
|
AWS_EXEC_READ = "aws-exec-read"
|
|
50
|
-
# Bucket Owner Read Only
|
|
51
41
|
BUCKET_OWNER_READ = "bucket-owner-read"
|
|
52
|
-
# Bucket Owner Full Control
|
|
53
42
|
BUCKET_OWNER_FULL_CONTROL = "bucket-owner-full-control"
|
|
54
43
|
|
|
55
44
|
|
|
56
45
|
class OutputSecurityLakeStorageClass(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
57
46
|
r"""Storage class to select for uploaded objects"""
|
|
58
47
|
|
|
59
|
-
# Standard
|
|
60
48
|
STANDARD = "STANDARD"
|
|
61
|
-
# Reduced Redundancy Storage
|
|
62
49
|
REDUCED_REDUNDANCY = "REDUCED_REDUNDANCY"
|
|
63
|
-
# Standard, Infrequent Access
|
|
64
50
|
STANDARD_IA = "STANDARD_IA"
|
|
65
|
-
# One Zone, Infrequent Access
|
|
66
51
|
ONEZONE_IA = "ONEZONE_IA"
|
|
67
|
-
# Intelligent Tiering
|
|
68
52
|
INTELLIGENT_TIERING = "INTELLIGENT_TIERING"
|
|
69
|
-
# Glacier Flexible Retrieval
|
|
70
53
|
GLACIER = "GLACIER"
|
|
71
|
-
# Glacier Instant Retrieval
|
|
72
54
|
GLACIER_IR = "GLACIER_IR"
|
|
73
|
-
# Glacier Deep Archive
|
|
74
55
|
DEEP_ARCHIVE = "DEEP_ARCHIVE"
|
|
75
56
|
|
|
76
57
|
|
|
77
58
|
class OutputSecurityLakeServerSideEncryptionForUploadedObjects(
|
|
78
59
|
str, Enum, metaclass=utils.OpenEnumMeta
|
|
79
60
|
):
|
|
80
|
-
# Amazon S3 Managed Key
|
|
81
61
|
AES256 = "AES256"
|
|
82
|
-
# AWS KMS Managed Key
|
|
83
62
|
AWS_KMS = "aws:kms"
|
|
84
63
|
|
|
85
64
|
|
|
86
65
|
class OutputSecurityLakeBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
87
66
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
88
67
|
|
|
89
|
-
# Block
|
|
90
68
|
BLOCK = "block"
|
|
91
|
-
# Drop
|
|
92
69
|
DROP = "drop"
|
|
93
70
|
|
|
94
71
|
|
|
95
72
|
class OutputSecurityLakeDiskSpaceProtection(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
96
73
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
97
74
|
|
|
98
|
-
# Block
|
|
99
75
|
BLOCK = "block"
|
|
100
|
-
# Drop
|
|
101
76
|
DROP = "drop"
|
|
102
77
|
|
|
103
78
|
|
|
104
79
|
class OutputSecurityLakeParquetVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
105
80
|
r"""Determines which data types are supported and how they are represented"""
|
|
106
81
|
|
|
107
|
-
# 1.0
|
|
108
82
|
PARQUET_1_0 = "PARQUET_1_0"
|
|
109
|
-
# 2.4
|
|
110
83
|
PARQUET_2_4 = "PARQUET_2_4"
|
|
111
|
-
# 2.6
|
|
112
84
|
PARQUET_2_6 = "PARQUET_2_6"
|
|
113
85
|
|
|
114
86
|
|
|
115
87
|
class OutputSecurityLakeDataPageVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
116
88
|
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
117
89
|
|
|
118
|
-
# V1
|
|
119
90
|
DATA_PAGE_V1 = "DATA_PAGE_V1"
|
|
120
|
-
# V2
|
|
121
91
|
DATA_PAGE_V2 = "DATA_PAGE_V2"
|
|
122
92
|
|
|
123
93
|
|
|
@@ -531,86 +501,3 @@ class OutputSecurityLake(BaseModel):
|
|
|
531
501
|
|
|
532
502
|
max_retry_num: Annotated[Optional[float], pydantic.Field(alias="maxRetryNum")] = 20
|
|
533
503
|
r"""The maximum number of times a file will attempt to move to its final destination before being dead-lettered"""
|
|
534
|
-
|
|
535
|
-
@field_serializer("aws_authentication_method")
|
|
536
|
-
def serialize_aws_authentication_method(self, value):
|
|
537
|
-
if isinstance(value, str):
|
|
538
|
-
try:
|
|
539
|
-
return models.OutputSecurityLakeAuthenticationMethod(value)
|
|
540
|
-
except ValueError:
|
|
541
|
-
return value
|
|
542
|
-
return value
|
|
543
|
-
|
|
544
|
-
@field_serializer("signature_version")
|
|
545
|
-
def serialize_signature_version(self, value):
|
|
546
|
-
if isinstance(value, str):
|
|
547
|
-
try:
|
|
548
|
-
return models.OutputSecurityLakeSignatureVersion(value)
|
|
549
|
-
except ValueError:
|
|
550
|
-
return value
|
|
551
|
-
return value
|
|
552
|
-
|
|
553
|
-
@field_serializer("object_acl")
|
|
554
|
-
def serialize_object_acl(self, value):
|
|
555
|
-
if isinstance(value, str):
|
|
556
|
-
try:
|
|
557
|
-
return models.OutputSecurityLakeObjectACL(value)
|
|
558
|
-
except ValueError:
|
|
559
|
-
return value
|
|
560
|
-
return value
|
|
561
|
-
|
|
562
|
-
@field_serializer("storage_class")
|
|
563
|
-
def serialize_storage_class(self, value):
|
|
564
|
-
if isinstance(value, str):
|
|
565
|
-
try:
|
|
566
|
-
return models.OutputSecurityLakeStorageClass(value)
|
|
567
|
-
except ValueError:
|
|
568
|
-
return value
|
|
569
|
-
return value
|
|
570
|
-
|
|
571
|
-
@field_serializer("server_side_encryption")
|
|
572
|
-
def serialize_server_side_encryption(self, value):
|
|
573
|
-
if isinstance(value, str):
|
|
574
|
-
try:
|
|
575
|
-
return models.OutputSecurityLakeServerSideEncryptionForUploadedObjects(
|
|
576
|
-
value
|
|
577
|
-
)
|
|
578
|
-
except ValueError:
|
|
579
|
-
return value
|
|
580
|
-
return value
|
|
581
|
-
|
|
582
|
-
@field_serializer("on_backpressure")
|
|
583
|
-
def serialize_on_backpressure(self, value):
|
|
584
|
-
if isinstance(value, str):
|
|
585
|
-
try:
|
|
586
|
-
return models.OutputSecurityLakeBackpressureBehavior(value)
|
|
587
|
-
except ValueError:
|
|
588
|
-
return value
|
|
589
|
-
return value
|
|
590
|
-
|
|
591
|
-
@field_serializer("on_disk_full_backpressure")
|
|
592
|
-
def serialize_on_disk_full_backpressure(self, value):
|
|
593
|
-
if isinstance(value, str):
|
|
594
|
-
try:
|
|
595
|
-
return models.OutputSecurityLakeDiskSpaceProtection(value)
|
|
596
|
-
except ValueError:
|
|
597
|
-
return value
|
|
598
|
-
return value
|
|
599
|
-
|
|
600
|
-
@field_serializer("parquet_version")
|
|
601
|
-
def serialize_parquet_version(self, value):
|
|
602
|
-
if isinstance(value, str):
|
|
603
|
-
try:
|
|
604
|
-
return models.OutputSecurityLakeParquetVersion(value)
|
|
605
|
-
except ValueError:
|
|
606
|
-
return value
|
|
607
|
-
return value
|
|
608
|
-
|
|
609
|
-
@field_serializer("parquet_data_page_version")
|
|
610
|
-
def serialize_parquet_data_page_version(self, value):
|
|
611
|
-
if isinstance(value, str):
|
|
612
|
-
try:
|
|
613
|
-
return models.OutputSecurityLakeDataPageVersion(value)
|
|
614
|
-
except ValueError:
|
|
615
|
-
return value
|
|
616
|
-
return value
|