cribl-control-plane 0.0.44a2__py3-none-any.whl → 0.0.46__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +5 -3
- cribl_control_plane/errors/healthstatus_error.py +2 -8
- cribl_control_plane/models/__init__.py +3 -3
- cribl_control_plane/models/appmode.py +1 -2
- cribl_control_plane/models/cacheconnection.py +2 -10
- cribl_control_plane/models/cacheconnectionbackfillstatus.py +1 -2
- cribl_control_plane/models/cloudprovider.py +1 -2
- cribl_control_plane/models/configgroup.py +2 -7
- cribl_control_plane/models/configgroupcloud.py +2 -6
- cribl_control_plane/models/createconfiggroupbyproductop.py +2 -8
- cribl_control_plane/models/cribllakedataset.py +2 -8
- cribl_control_plane/models/datasetmetadata.py +2 -8
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +2 -7
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -3
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +2 -7
- cribl_control_plane/models/getsummaryop.py +2 -7
- cribl_control_plane/models/hbcriblinfo.py +6 -6
- cribl_control_plane/models/healthstatus.py +4 -7
- cribl_control_plane/models/inputappscope.py +14 -34
- cribl_control_plane/models/inputazureblob.py +6 -17
- cribl_control_plane/models/inputcollection.py +4 -11
- cribl_control_plane/models/inputconfluentcloud.py +20 -47
- cribl_control_plane/models/inputcribl.py +4 -11
- cribl_control_plane/models/inputcriblhttp.py +8 -23
- cribl_control_plane/models/inputcribllakehttp.py +10 -22
- cribl_control_plane/models/inputcriblmetrics.py +4 -12
- cribl_control_plane/models/inputcribltcp.py +8 -23
- cribl_control_plane/models/inputcrowdstrike.py +10 -26
- cribl_control_plane/models/inputdatadogagent.py +8 -24
- cribl_control_plane/models/inputdatagen.py +4 -11
- cribl_control_plane/models/inputedgeprometheus.py +24 -58
- cribl_control_plane/models/inputelastic.py +14 -40
- cribl_control_plane/models/inputeventhub.py +6 -15
- cribl_control_plane/models/inputexec.py +6 -14
- cribl_control_plane/models/inputfile.py +6 -15
- cribl_control_plane/models/inputfirehose.py +8 -23
- cribl_control_plane/models/inputgooglepubsub.py +6 -19
- cribl_control_plane/models/inputgrafana.py +24 -67
- cribl_control_plane/models/inputhttp.py +8 -23
- cribl_control_plane/models/inputhttpraw.py +8 -23
- cribl_control_plane/models/inputjournalfiles.py +4 -12
- cribl_control_plane/models/inputkafka.py +16 -46
- cribl_control_plane/models/inputkinesis.py +14 -38
- cribl_control_plane/models/inputkubeevents.py +4 -11
- cribl_control_plane/models/inputkubelogs.py +8 -16
- cribl_control_plane/models/inputkubemetrics.py +8 -16
- cribl_control_plane/models/inputloki.py +10 -29
- cribl_control_plane/models/inputmetrics.py +8 -23
- cribl_control_plane/models/inputmodeldriventelemetry.py +10 -27
- cribl_control_plane/models/inputmsk.py +18 -53
- cribl_control_plane/models/inputnetflow.py +4 -11
- cribl_control_plane/models/inputoffice365mgmt.py +14 -33
- cribl_control_plane/models/inputoffice365msgtrace.py +16 -35
- cribl_control_plane/models/inputoffice365service.py +16 -35
- cribl_control_plane/models/inputopentelemetry.py +16 -38
- cribl_control_plane/models/inputprometheus.py +18 -50
- cribl_control_plane/models/inputprometheusrw.py +10 -30
- cribl_control_plane/models/inputrawudp.py +4 -11
- cribl_control_plane/models/inputs3.py +8 -21
- cribl_control_plane/models/inputs3inventory.py +10 -26
- cribl_control_plane/models/inputsecuritylake.py +10 -27
- cribl_control_plane/models/inputsnmp.py +6 -16
- cribl_control_plane/models/inputsplunk.py +12 -33
- cribl_control_plane/models/inputsplunkhec.py +10 -29
- cribl_control_plane/models/inputsplunksearch.py +14 -33
- cribl_control_plane/models/inputsqs.py +10 -27
- cribl_control_plane/models/inputsyslog.py +16 -43
- cribl_control_plane/models/inputsystemmetrics.py +24 -48
- cribl_control_plane/models/inputsystemstate.py +8 -16
- cribl_control_plane/models/inputtcp.py +10 -29
- cribl_control_plane/models/inputtcpjson.py +10 -29
- cribl_control_plane/models/inputwef.py +14 -37
- cribl_control_plane/models/inputwindowsmetrics.py +24 -44
- cribl_control_plane/models/inputwineventlogs.py +10 -20
- cribl_control_plane/models/inputwiz.py +8 -21
- cribl_control_plane/models/inputwizwebhook.py +8 -23
- cribl_control_plane/models/inputzscalerhec.py +10 -29
- cribl_control_plane/models/lakehouseconnectiontype.py +1 -2
- cribl_control_plane/models/listconfiggroupbyproductop.py +1 -3
- cribl_control_plane/models/masterworkerentry.py +2 -7
- cribl_control_plane/models/nodeactiveupgradestatus.py +1 -2
- cribl_control_plane/models/nodefailedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeskippedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeupgradestate.py +1 -2
- cribl_control_plane/models/nodeupgradestatus.py +5 -13
- cribl_control_plane/models/outputazureblob.py +18 -48
- cribl_control_plane/models/outputazuredataexplorer.py +28 -73
- cribl_control_plane/models/outputazureeventhub.py +18 -40
- cribl_control_plane/models/outputazurelogs.py +12 -35
- cribl_control_plane/models/outputclickhouse.py +20 -55
- cribl_control_plane/models/outputcloudwatch.py +10 -29
- cribl_control_plane/models/outputconfluentcloud.py +32 -77
- cribl_control_plane/models/outputcriblhttp.py +16 -44
- cribl_control_plane/models/outputcribllake.py +16 -46
- cribl_control_plane/models/outputcribltcp.py +18 -45
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +14 -49
- cribl_control_plane/models/outputdatadog.py +20 -48
- cribl_control_plane/models/outputdataset.py +18 -46
- cribl_control_plane/models/outputdiskspool.py +2 -7
- cribl_control_plane/models/outputdls3.py +24 -68
- cribl_control_plane/models/outputdynatracehttp.py +20 -53
- cribl_control_plane/models/outputdynatraceotlp.py +22 -55
- cribl_control_plane/models/outputelastic.py +18 -43
- cribl_control_plane/models/outputelasticcloud.py +12 -36
- cribl_control_plane/models/outputexabeam.py +10 -29
- cribl_control_plane/models/outputfilesystem.py +14 -39
- cribl_control_plane/models/outputgooglechronicle.py +16 -50
- cribl_control_plane/models/outputgooglecloudlogging.py +14 -41
- cribl_control_plane/models/outputgooglecloudstorage.py +24 -66
- cribl_control_plane/models/outputgooglepubsub.py +10 -31
- cribl_control_plane/models/outputgrafanacloud.py +32 -97
- cribl_control_plane/models/outputgraphite.py +14 -31
- cribl_control_plane/models/outputhoneycomb.py +12 -35
- cribl_control_plane/models/outputhumiohec.py +16 -43
- cribl_control_plane/models/outputinfluxdb.py +16 -42
- cribl_control_plane/models/outputkafka.py +28 -74
- cribl_control_plane/models/outputkinesis.py +16 -40
- cribl_control_plane/models/outputloki.py +16 -41
- cribl_control_plane/models/outputminio.py +24 -65
- cribl_control_plane/models/outputmsk.py +30 -82
- cribl_control_plane/models/outputnewrelic.py +18 -43
- cribl_control_plane/models/outputnewrelicevents.py +14 -41
- cribl_control_plane/models/outputopentelemetry.py +26 -67
- cribl_control_plane/models/outputprometheus.py +12 -35
- cribl_control_plane/models/outputring.py +8 -19
- cribl_control_plane/models/outputs3.py +26 -68
- cribl_control_plane/models/outputsecuritylake.py +18 -52
- cribl_control_plane/models/outputsentinel.py +18 -45
- cribl_control_plane/models/outputsentineloneaisiem.py +18 -50
- cribl_control_plane/models/outputservicenow.py +24 -60
- cribl_control_plane/models/outputsignalfx.py +14 -37
- cribl_control_plane/models/outputsns.py +14 -36
- cribl_control_plane/models/outputsplunk.py +24 -60
- cribl_control_plane/models/outputsplunkhec.py +12 -35
- cribl_control_plane/models/outputsplunklb.py +30 -77
- cribl_control_plane/models/outputsqs.py +16 -41
- cribl_control_plane/models/outputstatsd.py +14 -30
- cribl_control_plane/models/outputstatsdext.py +12 -29
- cribl_control_plane/models/outputsumologic.py +12 -35
- cribl_control_plane/models/outputsyslog.py +24 -58
- cribl_control_plane/models/outputtcpjson.py +20 -52
- cribl_control_plane/models/outputwavefront.py +12 -35
- cribl_control_plane/models/outputwebhook.py +22 -58
- cribl_control_plane/models/outputxsiam.py +14 -35
- cribl_control_plane/models/productscore.py +1 -2
- cribl_control_plane/models/rbacresource.py +1 -2
- cribl_control_plane/models/resourcepolicy.py +2 -4
- cribl_control_plane/models/runnablejobcollection.py +13 -30
- cribl_control_plane/models/runnablejobexecutor.py +4 -13
- cribl_control_plane/models/runnablejobscheduledsearch.py +2 -7
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +2 -8
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +2 -8
- cribl_control_plane/models/workertypes.py +1 -2
- {cribl_control_plane-0.0.44a2.dist-info → cribl_control_plane-0.0.46.dist-info}/METADATA +3 -2
- {cribl_control_plane-0.0.44a2.dist-info → cribl_control_plane-0.0.46.dist-info}/RECORD +158 -158
- {cribl_control_plane-0.0.44a2.dist-info → cribl_control_plane-0.0.46.dist-info}/WHEEL +1 -1
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import Any, List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputHTTPConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputHTTPMode(str, Enum
|
|
26
|
+
class InputHTTPMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputHTTPCompression(str, Enum
|
|
33
|
+
class InputHTTPCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputHTTPPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputHTTPPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputHTTPMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputHTTPMode.ALWAYS
|
|
67
|
+
mode: Optional[InputHTTPMode] = InputHTTPMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputHTTPPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputHTTPCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputHTTPCompression.NONE
|
|
91
|
+
compress: Optional[InputHTTPCompression] = InputHTTPCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,14 +96,14 @@ class InputHTTPPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class InputHTTPMinimumTLSVersion(str, Enum
|
|
99
|
+
class InputHTTPMinimumTLSVersion(str, Enum):
|
|
107
100
|
TL_SV1 = "TLSv1"
|
|
108
101
|
TL_SV1_1 = "TLSv1.1"
|
|
109
102
|
TL_SV1_2 = "TLSv1.2"
|
|
110
103
|
TL_SV1_3 = "TLSv1.3"
|
|
111
104
|
|
|
112
105
|
|
|
113
|
-
class InputHTTPMaximumTLSVersion(str, Enum
|
|
106
|
+
class InputHTTPMaximumTLSVersion(str, Enum):
|
|
114
107
|
TL_SV1 = "TLSv1"
|
|
115
108
|
TL_SV1_1 = "TLSv1.1"
|
|
116
109
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -169,19 +162,11 @@ class InputHTTPTLSSettingsServerSide(BaseModel):
|
|
|
169
162
|
] = None
|
|
170
163
|
|
|
171
164
|
min_version: Annotated[
|
|
172
|
-
|
|
173
|
-
Optional[InputHTTPMinimumTLSVersion],
|
|
174
|
-
PlainValidator(validate_open_enum(False)),
|
|
175
|
-
],
|
|
176
|
-
pydantic.Field(alias="minVersion"),
|
|
165
|
+
Optional[InputHTTPMinimumTLSVersion], pydantic.Field(alias="minVersion")
|
|
177
166
|
] = None
|
|
178
167
|
|
|
179
168
|
max_version: Annotated[
|
|
180
|
-
|
|
181
|
-
Optional[InputHTTPMaximumTLSVersion],
|
|
182
|
-
PlainValidator(validate_open_enum(False)),
|
|
183
|
-
],
|
|
184
|
-
pydantic.Field(alias="maxVersion"),
|
|
169
|
+
Optional[InputHTTPMaximumTLSVersion], pydantic.Field(alias="maxVersion")
|
|
185
170
|
] = None
|
|
186
171
|
|
|
187
172
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import Any, List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputHTTPRawConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputHTTPRawMode(str, Enum
|
|
26
|
+
class InputHTTPRawMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputHTTPRawCompression(str, Enum
|
|
33
|
+
class InputHTTPRawCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputHTTPRawPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputHTTPRawPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputHTTPRawMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputHTTPRawMode.ALWAYS
|
|
67
|
+
mode: Optional[InputHTTPRawMode] = InputHTTPRawMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputHTTPRawPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputHTTPRawCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputHTTPRawCompression.NONE
|
|
91
|
+
compress: Optional[InputHTTPRawCompression] = InputHTTPRawCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,14 +96,14 @@ class InputHTTPRawPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class InputHTTPRawMinimumTLSVersion(str, Enum
|
|
99
|
+
class InputHTTPRawMinimumTLSVersion(str, Enum):
|
|
107
100
|
TL_SV1 = "TLSv1"
|
|
108
101
|
TL_SV1_1 = "TLSv1.1"
|
|
109
102
|
TL_SV1_2 = "TLSv1.2"
|
|
110
103
|
TL_SV1_3 = "TLSv1.3"
|
|
111
104
|
|
|
112
105
|
|
|
113
|
-
class InputHTTPRawMaximumTLSVersion(str, Enum
|
|
106
|
+
class InputHTTPRawMaximumTLSVersion(str, Enum):
|
|
114
107
|
TL_SV1 = "TLSv1"
|
|
115
108
|
TL_SV1_1 = "TLSv1.1"
|
|
116
109
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -169,19 +162,11 @@ class InputHTTPRawTLSSettingsServerSide(BaseModel):
|
|
|
169
162
|
] = None
|
|
170
163
|
|
|
171
164
|
min_version: Annotated[
|
|
172
|
-
|
|
173
|
-
Optional[InputHTTPRawMinimumTLSVersion],
|
|
174
|
-
PlainValidator(validate_open_enum(False)),
|
|
175
|
-
],
|
|
176
|
-
pydantic.Field(alias="minVersion"),
|
|
165
|
+
Optional[InputHTTPRawMinimumTLSVersion], pydantic.Field(alias="minVersion")
|
|
177
166
|
] = None
|
|
178
167
|
|
|
179
168
|
max_version: Annotated[
|
|
180
|
-
|
|
181
|
-
Optional[InputHTTPRawMaximumTLSVersion],
|
|
182
|
-
PlainValidator(validate_open_enum(False)),
|
|
183
|
-
],
|
|
184
|
-
pydantic.Field(alias="maxVersion"),
|
|
169
|
+
Optional[InputHTTPRawMaximumTLSVersion], pydantic.Field(alias="maxVersion")
|
|
185
170
|
] = None
|
|
186
171
|
|
|
187
172
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputJournalFilesConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputJournalFilesMode(str, Enum
|
|
26
|
+
class InputJournalFilesMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputJournalFilesCompression(str, Enum
|
|
33
|
+
class InputJournalFilesCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputJournalFilesPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputJournalFilesPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputJournalFilesMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputJournalFilesMode.ALWAYS
|
|
67
|
+
mode: Optional[InputJournalFilesMode] = InputJournalFilesMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,10 +88,7 @@ class InputJournalFilesPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputJournalFilesCompression],
|
|
98
|
-
PlainValidator(validate_open_enum(False)),
|
|
99
|
-
] = InputJournalFilesCompression.NONE
|
|
91
|
+
compress: Optional[InputJournalFilesCompression] = InputJournalFilesCompression.NONE
|
|
100
92
|
r"""Codec to use to compress the persisted data"""
|
|
101
93
|
|
|
102
94
|
pq_controls: Annotated[
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputKafkaConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputKafkaMode(str, Enum
|
|
26
|
+
class InputKafkaMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputKafkaCompression(str, Enum
|
|
33
|
+
class InputKafkaCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputKafkaPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputKafkaPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputKafkaMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputKafkaMode.ALWAYS
|
|
67
|
+
mode: Optional[InputKafkaMode] = InputKafkaMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputKafkaPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputKafkaCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputKafkaCompression.NONE
|
|
91
|
+
compress: Optional[InputKafkaCompression] = InputKafkaCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,7 +96,7 @@ class InputKafkaPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class InputKafkaSchemaType(str, Enum
|
|
99
|
+
class InputKafkaSchemaType(str, Enum):
|
|
107
100
|
r"""The schema format used to encode and decode event data"""
|
|
108
101
|
|
|
109
102
|
AVRO = "avro"
|
|
@@ -129,18 +122,14 @@ class InputKafkaAuth(BaseModel):
|
|
|
129
122
|
r"""Select or create a secret that references your credentials"""
|
|
130
123
|
|
|
131
124
|
|
|
132
|
-
class InputKafkaKafkaSchemaRegistryMinimumTLSVersion(
|
|
133
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
134
|
-
):
|
|
125
|
+
class InputKafkaKafkaSchemaRegistryMinimumTLSVersion(str, Enum):
|
|
135
126
|
TL_SV1 = "TLSv1"
|
|
136
127
|
TL_SV1_1 = "TLSv1.1"
|
|
137
128
|
TL_SV1_2 = "TLSv1.2"
|
|
138
129
|
TL_SV1_3 = "TLSv1.3"
|
|
139
130
|
|
|
140
131
|
|
|
141
|
-
class InputKafkaKafkaSchemaRegistryMaximumTLSVersion(
|
|
142
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
143
|
-
):
|
|
132
|
+
class InputKafkaKafkaSchemaRegistryMaximumTLSVersion(str, Enum):
|
|
144
133
|
TL_SV1 = "TLSv1"
|
|
145
134
|
TL_SV1_1 = "TLSv1.1"
|
|
146
135
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -200,18 +189,12 @@ class InputKafkaKafkaSchemaRegistryTLSSettingsClientSide(BaseModel):
|
|
|
200
189
|
r"""Passphrase to use to decrypt private key"""
|
|
201
190
|
|
|
202
191
|
min_version: Annotated[
|
|
203
|
-
|
|
204
|
-
Optional[InputKafkaKafkaSchemaRegistryMinimumTLSVersion],
|
|
205
|
-
PlainValidator(validate_open_enum(False)),
|
|
206
|
-
],
|
|
192
|
+
Optional[InputKafkaKafkaSchemaRegistryMinimumTLSVersion],
|
|
207
193
|
pydantic.Field(alias="minVersion"),
|
|
208
194
|
] = None
|
|
209
195
|
|
|
210
196
|
max_version: Annotated[
|
|
211
|
-
|
|
212
|
-
Optional[InputKafkaKafkaSchemaRegistryMaximumTLSVersion],
|
|
213
|
-
PlainValidator(validate_open_enum(False)),
|
|
214
|
-
],
|
|
197
|
+
Optional[InputKafkaKafkaSchemaRegistryMaximumTLSVersion],
|
|
215
198
|
pydantic.Field(alias="maxVersion"),
|
|
216
199
|
] = None
|
|
217
200
|
|
|
@@ -242,10 +225,7 @@ class InputKafkaKafkaSchemaRegistryAuthentication(BaseModel):
|
|
|
242
225
|
r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
|
|
243
226
|
|
|
244
227
|
schema_type: Annotated[
|
|
245
|
-
|
|
246
|
-
Optional[InputKafkaSchemaType], PlainValidator(validate_open_enum(False))
|
|
247
|
-
],
|
|
248
|
-
pydantic.Field(alias="schemaType"),
|
|
228
|
+
Optional[InputKafkaSchemaType], pydantic.Field(alias="schemaType")
|
|
249
229
|
] = InputKafkaSchemaType.AVRO
|
|
250
230
|
r"""The schema format used to encode and decode event data"""
|
|
251
231
|
|
|
@@ -268,7 +248,7 @@ class InputKafkaKafkaSchemaRegistryAuthentication(BaseModel):
|
|
|
268
248
|
tls: Optional[InputKafkaKafkaSchemaRegistryTLSSettingsClientSide] = None
|
|
269
249
|
|
|
270
250
|
|
|
271
|
-
class InputKafkaSASLMechanism(str, Enum
|
|
251
|
+
class InputKafkaSASLMechanism(str, Enum):
|
|
272
252
|
PLAIN = "plain"
|
|
273
253
|
SCRAM_SHA_256 = "scram-sha-256"
|
|
274
254
|
SCRAM_SHA_512 = "scram-sha-512"
|
|
@@ -289,9 +269,7 @@ class InputKafkaAuthentication(BaseModel):
|
|
|
289
269
|
|
|
290
270
|
disabled: Optional[bool] = True
|
|
291
271
|
|
|
292
|
-
mechanism:
|
|
293
|
-
Optional[InputKafkaSASLMechanism], PlainValidator(validate_open_enum(False))
|
|
294
|
-
] = InputKafkaSASLMechanism.PLAIN
|
|
272
|
+
mechanism: Optional[InputKafkaSASLMechanism] = InputKafkaSASLMechanism.PLAIN
|
|
295
273
|
|
|
296
274
|
oauth_enabled: Annotated[Optional[bool], pydantic.Field(alias="oauthEnabled")] = (
|
|
297
275
|
False
|
|
@@ -299,14 +277,14 @@ class InputKafkaAuthentication(BaseModel):
|
|
|
299
277
|
r"""Enable OAuth authentication"""
|
|
300
278
|
|
|
301
279
|
|
|
302
|
-
class InputKafkaMinimumTLSVersion(str, Enum
|
|
280
|
+
class InputKafkaMinimumTLSVersion(str, Enum):
|
|
303
281
|
TL_SV1 = "TLSv1"
|
|
304
282
|
TL_SV1_1 = "TLSv1.1"
|
|
305
283
|
TL_SV1_2 = "TLSv1.2"
|
|
306
284
|
TL_SV1_3 = "TLSv1.3"
|
|
307
285
|
|
|
308
286
|
|
|
309
|
-
class InputKafkaMaximumTLSVersion(str, Enum
|
|
287
|
+
class InputKafkaMaximumTLSVersion(str, Enum):
|
|
310
288
|
TL_SV1 = "TLSv1"
|
|
311
289
|
TL_SV1_1 = "TLSv1.1"
|
|
312
290
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -366,19 +344,11 @@ class InputKafkaTLSSettingsClientSide(BaseModel):
|
|
|
366
344
|
r"""Passphrase to use to decrypt private key"""
|
|
367
345
|
|
|
368
346
|
min_version: Annotated[
|
|
369
|
-
|
|
370
|
-
Optional[InputKafkaMinimumTLSVersion],
|
|
371
|
-
PlainValidator(validate_open_enum(False)),
|
|
372
|
-
],
|
|
373
|
-
pydantic.Field(alias="minVersion"),
|
|
347
|
+
Optional[InputKafkaMinimumTLSVersion], pydantic.Field(alias="minVersion")
|
|
374
348
|
] = None
|
|
375
349
|
|
|
376
350
|
max_version: Annotated[
|
|
377
|
-
|
|
378
|
-
Optional[InputKafkaMaximumTLSVersion],
|
|
379
|
-
PlainValidator(validate_open_enum(False)),
|
|
380
|
-
],
|
|
381
|
-
pydantic.Field(alias="maxVersion"),
|
|
351
|
+
Optional[InputKafkaMaximumTLSVersion], pydantic.Field(alias="maxVersion")
|
|
382
352
|
] = None
|
|
383
353
|
|
|
384
354
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputKinesisConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputKinesisMode(str, Enum
|
|
26
|
+
class InputKinesisMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputKinesisCompression(str, Enum
|
|
33
|
+
class InputKinesisCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputKinesisPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputKinesisPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputKinesisMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputKinesisMode.ALWAYS
|
|
67
|
+
mode: Optional[InputKinesisMode] = InputKinesisMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputKinesisPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputKinesisCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputKinesisCompression.NONE
|
|
91
|
+
compress: Optional[InputKinesisCompression] = InputKinesisCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,14 +96,14 @@ class InputKinesisPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class ShardIteratorStart(str, Enum
|
|
99
|
+
class ShardIteratorStart(str, Enum):
|
|
107
100
|
r"""Location at which to start reading a shard for the first time"""
|
|
108
101
|
|
|
109
102
|
TRIM_HORIZON = "TRIM_HORIZON"
|
|
110
103
|
LATEST = "LATEST"
|
|
111
104
|
|
|
112
105
|
|
|
113
|
-
class InputKinesisRecordDataFormat(str, Enum
|
|
106
|
+
class InputKinesisRecordDataFormat(str, Enum):
|
|
114
107
|
r"""Format of data inside the Kinesis Stream records. Gzip compression is automatically detected."""
|
|
115
108
|
|
|
116
109
|
CRIBL = "cribl"
|
|
@@ -119,14 +112,14 @@ class InputKinesisRecordDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
119
112
|
LINE = "line"
|
|
120
113
|
|
|
121
114
|
|
|
122
|
-
class ShardLoadBalancing(str, Enum
|
|
115
|
+
class ShardLoadBalancing(str, Enum):
|
|
123
116
|
r"""The load-balancing algorithm to use for spreading out shards across Workers and Worker Processes"""
|
|
124
117
|
|
|
125
118
|
CONSISTENT_HASHING = "ConsistentHashing"
|
|
126
119
|
ROUND_ROBIN = "RoundRobin"
|
|
127
120
|
|
|
128
121
|
|
|
129
|
-
class InputKinesisAuthenticationMethod(str, Enum
|
|
122
|
+
class InputKinesisAuthenticationMethod(str, Enum):
|
|
130
123
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
131
124
|
|
|
132
125
|
AUTO = "auto"
|
|
@@ -134,7 +127,7 @@ class InputKinesisAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
134
127
|
SECRET = "secret"
|
|
135
128
|
|
|
136
129
|
|
|
137
|
-
class InputKinesisSignatureVersion(str, Enum
|
|
130
|
+
class InputKinesisSignatureVersion(str, Enum):
|
|
138
131
|
r"""Signature version to use for signing Kinesis stream requests"""
|
|
139
132
|
|
|
140
133
|
V2 = "v2"
|
|
@@ -266,19 +259,12 @@ class InputKinesis(BaseModel):
|
|
|
266
259
|
r"""A JavaScript expression to be called with each shardId for the stream. If the expression evaluates to a truthy value, the shard will be processed."""
|
|
267
260
|
|
|
268
261
|
shard_iterator_type: Annotated[
|
|
269
|
-
|
|
270
|
-
Optional[ShardIteratorStart], PlainValidator(validate_open_enum(False))
|
|
271
|
-
],
|
|
272
|
-
pydantic.Field(alias="shardIteratorType"),
|
|
262
|
+
Optional[ShardIteratorStart], pydantic.Field(alias="shardIteratorType")
|
|
273
263
|
] = ShardIteratorStart.TRIM_HORIZON
|
|
274
264
|
r"""Location at which to start reading a shard for the first time"""
|
|
275
265
|
|
|
276
266
|
payload_format: Annotated[
|
|
277
|
-
|
|
278
|
-
Optional[InputKinesisRecordDataFormat],
|
|
279
|
-
PlainValidator(validate_open_enum(False)),
|
|
280
|
-
],
|
|
281
|
-
pydantic.Field(alias="payloadFormat"),
|
|
267
|
+
Optional[InputKinesisRecordDataFormat], pydantic.Field(alias="payloadFormat")
|
|
282
268
|
] = InputKinesisRecordDataFormat.CRIBL
|
|
283
269
|
r"""Format of data inside the Kinesis Stream records. Gzip compression is automatically detected."""
|
|
284
270
|
|
|
@@ -293,18 +279,12 @@ class InputKinesis(BaseModel):
|
|
|
293
279
|
r"""Maximum number of records, across all shards, to pull down at once per Worker Process"""
|
|
294
280
|
|
|
295
281
|
load_balancing_algorithm: Annotated[
|
|
296
|
-
|
|
297
|
-
Optional[ShardLoadBalancing], PlainValidator(validate_open_enum(False))
|
|
298
|
-
],
|
|
299
|
-
pydantic.Field(alias="loadBalancingAlgorithm"),
|
|
282
|
+
Optional[ShardLoadBalancing], pydantic.Field(alias="loadBalancingAlgorithm")
|
|
300
283
|
] = ShardLoadBalancing.CONSISTENT_HASHING
|
|
301
284
|
r"""The load-balancing algorithm to use for spreading out shards across Workers and Worker Processes"""
|
|
302
285
|
|
|
303
286
|
aws_authentication_method: Annotated[
|
|
304
|
-
|
|
305
|
-
Optional[InputKinesisAuthenticationMethod],
|
|
306
|
-
PlainValidator(validate_open_enum(False)),
|
|
307
|
-
],
|
|
287
|
+
Optional[InputKinesisAuthenticationMethod],
|
|
308
288
|
pydantic.Field(alias="awsAuthenticationMethod"),
|
|
309
289
|
] = InputKinesisAuthenticationMethod.AUTO
|
|
310
290
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
@@ -317,11 +297,7 @@ class InputKinesis(BaseModel):
|
|
|
317
297
|
r"""Kinesis stream service endpoint. If empty, defaults to the AWS Region-specific endpoint. Otherwise, it must point to Kinesis stream-compatible endpoint."""
|
|
318
298
|
|
|
319
299
|
signature_version: Annotated[
|
|
320
|
-
|
|
321
|
-
Optional[InputKinesisSignatureVersion],
|
|
322
|
-
PlainValidator(validate_open_enum(False)),
|
|
323
|
-
],
|
|
324
|
-
pydantic.Field(alias="signatureVersion"),
|
|
300
|
+
Optional[InputKinesisSignatureVersion], pydantic.Field(alias="signatureVersion")
|
|
325
301
|
] = InputKinesisSignatureVersion.V4
|
|
326
302
|
r"""Signature version to use for signing Kinesis stream requests"""
|
|
327
303
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputKubeEventsConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputKubeEventsMode(str, Enum
|
|
26
|
+
class InputKubeEventsMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputKubeEventsCompression(str, Enum
|
|
33
|
+
class InputKubeEventsCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputKubeEventsPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputKubeEventsPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputKubeEventsMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputKubeEventsMode.ALWAYS
|
|
67
|
+
mode: Optional[InputKubeEventsMode] = InputKubeEventsMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputKubeEventsPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputKubeEventsCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputKubeEventsCompression.NONE
|
|
91
|
+
compress: Optional[InputKubeEventsCompression] = InputKubeEventsCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputKubeLogsConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputKubeLogsMode(str, Enum
|
|
26
|
+
class InputKubeLogsMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputKubeLogsPqCompression(str, Enum
|
|
33
|
+
class InputKubeLogsPqCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputKubeLogsPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputKubeLogsPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputKubeLogsMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputKubeLogsMode.ALWAYS
|
|
67
|
+
mode: Optional[InputKubeLogsMode] = InputKubeLogsMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputKubeLogsPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputKubeLogsPqCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputKubeLogsPqCompression.NONE
|
|
91
|
+
compress: Optional[InputKubeLogsPqCompression] = InputKubeLogsPqCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -131,7 +124,7 @@ class InputKubeLogsMetadatum(BaseModel):
|
|
|
131
124
|
r"""JavaScript expression to compute field's value, enclosed in quotes or backticks. (Can evaluate to a constant.)"""
|
|
132
125
|
|
|
133
126
|
|
|
134
|
-
class InputKubeLogsPersistenceCompression(str, Enum
|
|
127
|
+
class InputKubeLogsPersistenceCompression(str, Enum):
|
|
135
128
|
r"""Data compression format. Default is gzip."""
|
|
136
129
|
|
|
137
130
|
NONE = "none"
|
|
@@ -164,10 +157,9 @@ class InputKubeLogsDiskSpooling(BaseModel):
|
|
|
164
157
|
max_data_time: Annotated[Optional[str], pydantic.Field(alias="maxDataTime")] = "24h"
|
|
165
158
|
r"""Maximum amount of time to retain data before older buckets are deleted. Examples: 2h, 4d. Default is 24h."""
|
|
166
159
|
|
|
167
|
-
compress:
|
|
168
|
-
|
|
169
|
-
|
|
170
|
-
] = InputKubeLogsPersistenceCompression.GZIP
|
|
160
|
+
compress: Optional[InputKubeLogsPersistenceCompression] = (
|
|
161
|
+
InputKubeLogsPersistenceCompression.GZIP
|
|
162
|
+
)
|
|
171
163
|
r"""Data compression format. Default is gzip."""
|
|
172
164
|
|
|
173
165
|
|