cribl-control-plane 0.0.50rc1__py3-none-any.whl → 0.0.51__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +6 -4
- cribl_control_plane/errors/healthstatus_error.py +2 -8
- cribl_control_plane/health.py +2 -6
- cribl_control_plane/httpclient.py +0 -1
- cribl_control_plane/models/__init__.py +4 -21
- cribl_control_plane/models/appmode.py +1 -2
- cribl_control_plane/models/cacheconnection.py +2 -10
- cribl_control_plane/models/cacheconnectionbackfillstatus.py +1 -2
- cribl_control_plane/models/cloudprovider.py +1 -2
- cribl_control_plane/models/configgroup.py +2 -7
- cribl_control_plane/models/configgroupcloud.py +2 -6
- cribl_control_plane/models/createconfiggroupbyproductop.py +2 -8
- cribl_control_plane/models/createinputhectokenbyidop.py +5 -6
- cribl_control_plane/models/createversionpushop.py +5 -5
- cribl_control_plane/models/createversionundoop.py +3 -3
- cribl_control_plane/models/cribllakedataset.py +2 -8
- cribl_control_plane/models/datasetmetadata.py +2 -8
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +2 -7
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -3
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +2 -7
- cribl_control_plane/models/getsummaryop.py +2 -7
- cribl_control_plane/models/getversionshowop.py +5 -6
- cribl_control_plane/models/gitinfo.py +3 -14
- cribl_control_plane/models/hbcriblinfo.py +1 -11
- cribl_control_plane/models/healthstatus.py +4 -7
- cribl_control_plane/models/inputappscope.py +14 -34
- cribl_control_plane/models/inputazureblob.py +6 -17
- cribl_control_plane/models/inputcollection.py +4 -11
- cribl_control_plane/models/inputconfluentcloud.py +20 -47
- cribl_control_plane/models/inputcribl.py +4 -11
- cribl_control_plane/models/inputcriblhttp.py +8 -23
- cribl_control_plane/models/inputcribllakehttp.py +10 -22
- cribl_control_plane/models/inputcriblmetrics.py +4 -12
- cribl_control_plane/models/inputcribltcp.py +8 -23
- cribl_control_plane/models/inputcrowdstrike.py +10 -26
- cribl_control_plane/models/inputdatadogagent.py +8 -24
- cribl_control_plane/models/inputdatagen.py +4 -11
- cribl_control_plane/models/inputedgeprometheus.py +24 -58
- cribl_control_plane/models/inputelastic.py +14 -40
- cribl_control_plane/models/inputeventhub.py +6 -15
- cribl_control_plane/models/inputexec.py +6 -14
- cribl_control_plane/models/inputfile.py +6 -15
- cribl_control_plane/models/inputfirehose.py +8 -23
- cribl_control_plane/models/inputgooglepubsub.py +6 -19
- cribl_control_plane/models/inputgrafana.py +24 -67
- cribl_control_plane/models/inputhttp.py +8 -23
- cribl_control_plane/models/inputhttpraw.py +8 -23
- cribl_control_plane/models/inputjournalfiles.py +4 -12
- cribl_control_plane/models/inputkafka.py +16 -46
- cribl_control_plane/models/inputkinesis.py +14 -38
- cribl_control_plane/models/inputkubeevents.py +4 -11
- cribl_control_plane/models/inputkubelogs.py +8 -16
- cribl_control_plane/models/inputkubemetrics.py +8 -16
- cribl_control_plane/models/inputloki.py +10 -29
- cribl_control_plane/models/inputmetrics.py +8 -23
- cribl_control_plane/models/inputmodeldriventelemetry.py +10 -32
- cribl_control_plane/models/inputmsk.py +18 -53
- cribl_control_plane/models/inputnetflow.py +4 -11
- cribl_control_plane/models/inputoffice365mgmt.py +14 -33
- cribl_control_plane/models/inputoffice365msgtrace.py +16 -35
- cribl_control_plane/models/inputoffice365service.py +16 -35
- cribl_control_plane/models/inputopentelemetry.py +16 -38
- cribl_control_plane/models/inputprometheus.py +18 -50
- cribl_control_plane/models/inputprometheusrw.py +10 -30
- cribl_control_plane/models/inputrawudp.py +4 -11
- cribl_control_plane/models/inputs3.py +8 -21
- cribl_control_plane/models/inputs3inventory.py +10 -26
- cribl_control_plane/models/inputsecuritylake.py +10 -27
- cribl_control_plane/models/inputsnmp.py +6 -16
- cribl_control_plane/models/inputsplunk.py +12 -33
- cribl_control_plane/models/inputsplunkhec.py +10 -29
- cribl_control_plane/models/inputsplunksearch.py +14 -33
- cribl_control_plane/models/inputsqs.py +10 -27
- cribl_control_plane/models/inputsyslog.py +16 -43
- cribl_control_plane/models/inputsystemmetrics.py +24 -48
- cribl_control_plane/models/inputsystemstate.py +8 -16
- cribl_control_plane/models/inputtcp.py +10 -29
- cribl_control_plane/models/inputtcpjson.py +10 -29
- cribl_control_plane/models/inputwef.py +14 -37
- cribl_control_plane/models/inputwindowsmetrics.py +24 -44
- cribl_control_plane/models/inputwineventlogs.py +10 -20
- cribl_control_plane/models/inputwiz.py +8 -21
- cribl_control_plane/models/inputwizwebhook.py +8 -23
- cribl_control_plane/models/inputzscalerhec.py +10 -29
- cribl_control_plane/models/jobinfo.py +1 -4
- cribl_control_plane/models/lakehouseconnectiontype.py +1 -2
- cribl_control_plane/models/listconfiggroupbyproductop.py +1 -3
- cribl_control_plane/models/masterworkerentry.py +2 -7
- cribl_control_plane/models/nodeactiveupgradestatus.py +1 -2
- cribl_control_plane/models/nodefailedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeprovidedinfo.py +1 -4
- cribl_control_plane/models/nodeskippedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeupgradestate.py +1 -2
- cribl_control_plane/models/nodeupgradestatus.py +5 -13
- cribl_control_plane/models/outputazureblob.py +18 -48
- cribl_control_plane/models/outputazuredataexplorer.py +28 -73
- cribl_control_plane/models/outputazureeventhub.py +18 -40
- cribl_control_plane/models/outputazurelogs.py +12 -35
- cribl_control_plane/models/outputclickhouse.py +20 -55
- cribl_control_plane/models/outputcloudwatch.py +10 -29
- cribl_control_plane/models/outputconfluentcloud.py +32 -77
- cribl_control_plane/models/outputcriblhttp.py +16 -44
- cribl_control_plane/models/outputcribllake.py +16 -46
- cribl_control_plane/models/outputcribltcp.py +18 -45
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +14 -49
- cribl_control_plane/models/outputdatadog.py +20 -48
- cribl_control_plane/models/outputdataset.py +18 -46
- cribl_control_plane/models/outputdiskspool.py +2 -7
- cribl_control_plane/models/outputdls3.py +24 -68
- cribl_control_plane/models/outputdynatracehttp.py +20 -53
- cribl_control_plane/models/outputdynatraceotlp.py +22 -55
- cribl_control_plane/models/outputelastic.py +18 -43
- cribl_control_plane/models/outputelasticcloud.py +12 -36
- cribl_control_plane/models/outputexabeam.py +10 -29
- cribl_control_plane/models/outputfilesystem.py +14 -39
- cribl_control_plane/models/outputgooglechronicle.py +16 -50
- cribl_control_plane/models/outputgooglecloudlogging.py +18 -50
- cribl_control_plane/models/outputgooglecloudstorage.py +24 -66
- cribl_control_plane/models/outputgooglepubsub.py +10 -31
- cribl_control_plane/models/outputgrafanacloud.py +32 -97
- cribl_control_plane/models/outputgraphite.py +14 -31
- cribl_control_plane/models/outputhoneycomb.py +12 -35
- cribl_control_plane/models/outputhumiohec.py +16 -43
- cribl_control_plane/models/outputinfluxdb.py +16 -42
- cribl_control_plane/models/outputkafka.py +28 -74
- cribl_control_plane/models/outputkinesis.py +16 -40
- cribl_control_plane/models/outputloki.py +16 -41
- cribl_control_plane/models/outputminio.py +24 -65
- cribl_control_plane/models/outputmsk.py +30 -82
- cribl_control_plane/models/outputnewrelic.py +18 -43
- cribl_control_plane/models/outputnewrelicevents.py +14 -41
- cribl_control_plane/models/outputopentelemetry.py +26 -67
- cribl_control_plane/models/outputprometheus.py +12 -35
- cribl_control_plane/models/outputring.py +8 -19
- cribl_control_plane/models/outputs3.py +26 -68
- cribl_control_plane/models/outputsecuritylake.py +18 -52
- cribl_control_plane/models/outputsentinel.py +18 -45
- cribl_control_plane/models/outputsentineloneaisiem.py +18 -50
- cribl_control_plane/models/outputservicenow.py +24 -60
- cribl_control_plane/models/outputsignalfx.py +14 -37
- cribl_control_plane/models/outputsns.py +14 -36
- cribl_control_plane/models/outputsplunk.py +24 -60
- cribl_control_plane/models/outputsplunkhec.py +12 -35
- cribl_control_plane/models/outputsplunklb.py +30 -77
- cribl_control_plane/models/outputsqs.py +16 -41
- cribl_control_plane/models/outputstatsd.py +14 -30
- cribl_control_plane/models/outputstatsdext.py +12 -29
- cribl_control_plane/models/outputsumologic.py +12 -35
- cribl_control_plane/models/outputsyslog.py +24 -58
- cribl_control_plane/models/outputtcpjson.py +20 -52
- cribl_control_plane/models/outputwavefront.py +12 -35
- cribl_control_plane/models/outputwebhook.py +22 -58
- cribl_control_plane/models/outputxsiam.py +14 -35
- cribl_control_plane/models/packinfo.py +0 -3
- cribl_control_plane/models/packinstallinfo.py +0 -3
- cribl_control_plane/models/productscore.py +1 -2
- cribl_control_plane/models/rbacresource.py +1 -2
- cribl_control_plane/models/resourcepolicy.py +2 -4
- cribl_control_plane/models/runnablejobcollection.py +13 -30
- cribl_control_plane/models/runnablejobexecutor.py +4 -13
- cribl_control_plane/models/runnablejobscheduledsearch.py +2 -7
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +2 -8
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +2 -8
- cribl_control_plane/models/updateinputhectokenbyidandtokenop.py +5 -6
- cribl_control_plane/models/workertypes.py +1 -2
- {cribl_control_plane-0.0.50rc1.dist-info → cribl_control_plane-0.0.51.dist-info}/METADATA +14 -5
- cribl_control_plane-0.0.51.dist-info/RECORD +325 -0
- cribl_control_plane/models/error.py +0 -16
- cribl_control_plane/models/gethealthinfoop.py +0 -17
- cribl_control_plane/models/gitshowresult.py +0 -19
- cribl_control_plane-0.0.50rc1.dist-info/RECORD +0 -328
- {cribl_control_plane-0.0.50rc1.dist-info → cribl_control_plane-0.0.51.dist-info}/WHEEL +0 -0
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -15,7 +12,7 @@ class OutputFilesystemType(str, Enum):
|
|
|
15
12
|
FILESYSTEM = "filesystem"
|
|
16
13
|
|
|
17
14
|
|
|
18
|
-
class OutputFilesystemDataFormat(str, Enum
|
|
15
|
+
class OutputFilesystemDataFormat(str, Enum):
|
|
19
16
|
r"""Format of the output data"""
|
|
20
17
|
|
|
21
18
|
JSON = "json"
|
|
@@ -23,28 +20,28 @@ class OutputFilesystemDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
23
20
|
PARQUET = "parquet"
|
|
24
21
|
|
|
25
22
|
|
|
26
|
-
class OutputFilesystemBackpressureBehavior(str, Enum
|
|
23
|
+
class OutputFilesystemBackpressureBehavior(str, Enum):
|
|
27
24
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
28
25
|
|
|
29
26
|
BLOCK = "block"
|
|
30
27
|
DROP = "drop"
|
|
31
28
|
|
|
32
29
|
|
|
33
|
-
class OutputFilesystemDiskSpaceProtection(str, Enum
|
|
30
|
+
class OutputFilesystemDiskSpaceProtection(str, Enum):
|
|
34
31
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
35
32
|
|
|
36
33
|
BLOCK = "block"
|
|
37
34
|
DROP = "drop"
|
|
38
35
|
|
|
39
36
|
|
|
40
|
-
class OutputFilesystemCompression(str, Enum
|
|
37
|
+
class OutputFilesystemCompression(str, Enum):
|
|
41
38
|
r"""Data compression format to apply to HTTP content before it is delivered"""
|
|
42
39
|
|
|
43
40
|
NONE = "none"
|
|
44
41
|
GZIP = "gzip"
|
|
45
42
|
|
|
46
43
|
|
|
47
|
-
class OutputFilesystemCompressionLevel(str, Enum
|
|
44
|
+
class OutputFilesystemCompressionLevel(str, Enum):
|
|
48
45
|
r"""Compression level to apply before moving files to final destination"""
|
|
49
46
|
|
|
50
47
|
BEST_SPEED = "best_speed"
|
|
@@ -52,7 +49,7 @@ class OutputFilesystemCompressionLevel(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
52
49
|
BEST_COMPRESSION = "best_compression"
|
|
53
50
|
|
|
54
51
|
|
|
55
|
-
class OutputFilesystemParquetVersion(str, Enum
|
|
52
|
+
class OutputFilesystemParquetVersion(str, Enum):
|
|
56
53
|
r"""Determines which data types are supported and how they are represented"""
|
|
57
54
|
|
|
58
55
|
PARQUET_1_0 = "PARQUET_1_0"
|
|
@@ -60,7 +57,7 @@ class OutputFilesystemParquetVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
60
57
|
PARQUET_2_6 = "PARQUET_2_6"
|
|
61
58
|
|
|
62
59
|
|
|
63
|
-
class OutputFilesystemDataPageVersion(str, Enum
|
|
60
|
+
class OutputFilesystemDataPageVersion(str, Enum):
|
|
64
61
|
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
65
62
|
|
|
66
63
|
DATA_PAGE_V1 = "DATA_PAGE_V1"
|
|
@@ -199,11 +196,7 @@ class OutputFilesystem(BaseModel):
|
|
|
199
196
|
r"""JavaScript expression defining how files are partitioned and organized. Default is date-based. If blank, Stream will fall back to the event's __partition field value – if present – otherwise to each location's root directory."""
|
|
200
197
|
|
|
201
198
|
format_: Annotated[
|
|
202
|
-
|
|
203
|
-
Optional[OutputFilesystemDataFormat],
|
|
204
|
-
PlainValidator(validate_open_enum(False)),
|
|
205
|
-
],
|
|
206
|
-
pydantic.Field(alias="format"),
|
|
199
|
+
Optional[OutputFilesystemDataFormat], pydantic.Field(alias="format")
|
|
207
200
|
] = OutputFilesystemDataFormat.JSON
|
|
208
201
|
r"""Format of the output data"""
|
|
209
202
|
|
|
@@ -246,10 +239,7 @@ class OutputFilesystem(BaseModel):
|
|
|
246
239
|
r"""Buffer size used to write to a file"""
|
|
247
240
|
|
|
248
241
|
on_backpressure: Annotated[
|
|
249
|
-
|
|
250
|
-
Optional[OutputFilesystemBackpressureBehavior],
|
|
251
|
-
PlainValidator(validate_open_enum(False)),
|
|
252
|
-
],
|
|
242
|
+
Optional[OutputFilesystemBackpressureBehavior],
|
|
253
243
|
pydantic.Field(alias="onBackpressure"),
|
|
254
244
|
] = OutputFilesystemBackpressureBehavior.BLOCK
|
|
255
245
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -260,26 +250,18 @@ class OutputFilesystem(BaseModel):
|
|
|
260
250
|
r"""If a file fails to move to its final destination after the maximum number of retries, move it to a designated directory to prevent further errors"""
|
|
261
251
|
|
|
262
252
|
on_disk_full_backpressure: Annotated[
|
|
263
|
-
|
|
264
|
-
Optional[OutputFilesystemDiskSpaceProtection],
|
|
265
|
-
PlainValidator(validate_open_enum(False)),
|
|
266
|
-
],
|
|
253
|
+
Optional[OutputFilesystemDiskSpaceProtection],
|
|
267
254
|
pydantic.Field(alias="onDiskFullBackpressure"),
|
|
268
255
|
] = OutputFilesystemDiskSpaceProtection.BLOCK
|
|
269
256
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
270
257
|
|
|
271
258
|
description: Optional[str] = None
|
|
272
259
|
|
|
273
|
-
compress:
|
|
274
|
-
Optional[OutputFilesystemCompression], PlainValidator(validate_open_enum(False))
|
|
275
|
-
] = OutputFilesystemCompression.GZIP
|
|
260
|
+
compress: Optional[OutputFilesystemCompression] = OutputFilesystemCompression.GZIP
|
|
276
261
|
r"""Data compression format to apply to HTTP content before it is delivered"""
|
|
277
262
|
|
|
278
263
|
compression_level: Annotated[
|
|
279
|
-
|
|
280
|
-
Optional[OutputFilesystemCompressionLevel],
|
|
281
|
-
PlainValidator(validate_open_enum(False)),
|
|
282
|
-
],
|
|
264
|
+
Optional[OutputFilesystemCompressionLevel],
|
|
283
265
|
pydantic.Field(alias="compressionLevel"),
|
|
284
266
|
] = OutputFilesystemCompressionLevel.BEST_SPEED
|
|
285
267
|
r"""Compression level to apply before moving files to final destination"""
|
|
@@ -290,19 +272,12 @@ class OutputFilesystem(BaseModel):
|
|
|
290
272
|
r"""Automatically calculate the schema based on the events of each Parquet file generated"""
|
|
291
273
|
|
|
292
274
|
parquet_version: Annotated[
|
|
293
|
-
|
|
294
|
-
Optional[OutputFilesystemParquetVersion],
|
|
295
|
-
PlainValidator(validate_open_enum(False)),
|
|
296
|
-
],
|
|
297
|
-
pydantic.Field(alias="parquetVersion"),
|
|
275
|
+
Optional[OutputFilesystemParquetVersion], pydantic.Field(alias="parquetVersion")
|
|
298
276
|
] = OutputFilesystemParquetVersion.PARQUET_2_6
|
|
299
277
|
r"""Determines which data types are supported and how they are represented"""
|
|
300
278
|
|
|
301
279
|
parquet_data_page_version: Annotated[
|
|
302
|
-
|
|
303
|
-
Optional[OutputFilesystemDataPageVersion],
|
|
304
|
-
PlainValidator(validate_open_enum(False)),
|
|
305
|
-
],
|
|
280
|
+
Optional[OutputFilesystemDataPageVersion],
|
|
306
281
|
pydantic.Field(alias="parquetDataPageVersion"),
|
|
307
282
|
] = OutputFilesystemDataPageVersion.DATA_PAGE_V2
|
|
308
283
|
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -15,14 +12,12 @@ class OutputGoogleChronicleType(str, Enum):
|
|
|
15
12
|
GOOGLE_CHRONICLE = "google_chronicle"
|
|
16
13
|
|
|
17
14
|
|
|
18
|
-
class OutputGoogleChronicleAPIVersion(str, Enum
|
|
15
|
+
class OutputGoogleChronicleAPIVersion(str, Enum):
|
|
19
16
|
V1 = "v1"
|
|
20
17
|
V2 = "v2"
|
|
21
18
|
|
|
22
19
|
|
|
23
|
-
class OutputGoogleChronicleAuthenticationMethod(
|
|
24
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
25
|
-
):
|
|
20
|
+
class OutputGoogleChronicleAuthenticationMethod(str, Enum):
|
|
26
21
|
MANUAL = "manual"
|
|
27
22
|
SECRET = "secret"
|
|
28
23
|
SERVICE_ACCOUNT = "serviceAccount"
|
|
@@ -83,7 +78,7 @@ class OutputGoogleChronicleTimeoutRetrySettings(BaseModel):
|
|
|
83
78
|
r"""The maximum backoff interval, in milliseconds, Cribl Stream should apply. Default (and minimum) is 10,000 ms (10 seconds); maximum is 180,000 ms (180 seconds)."""
|
|
84
79
|
|
|
85
80
|
|
|
86
|
-
class SendEventsAs(str, Enum
|
|
81
|
+
class SendEventsAs(str, Enum):
|
|
87
82
|
UNSTRUCTURED = "unstructured"
|
|
88
83
|
UDM = "udm"
|
|
89
84
|
|
|
@@ -99,9 +94,7 @@ class OutputGoogleChronicleExtraHTTPHeader(BaseModel):
|
|
|
99
94
|
name: Optional[str] = None
|
|
100
95
|
|
|
101
96
|
|
|
102
|
-
class OutputGoogleChronicleFailedRequestLoggingMode(
|
|
103
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
104
|
-
):
|
|
97
|
+
class OutputGoogleChronicleFailedRequestLoggingMode(str, Enum):
|
|
105
98
|
r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
|
|
106
99
|
|
|
107
100
|
PAYLOAD = "payload"
|
|
@@ -109,9 +102,7 @@ class OutputGoogleChronicleFailedRequestLoggingMode(
|
|
|
109
102
|
NONE = "none"
|
|
110
103
|
|
|
111
104
|
|
|
112
|
-
class OutputGoogleChronicleBackpressureBehavior(
|
|
113
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
114
|
-
):
|
|
105
|
+
class OutputGoogleChronicleBackpressureBehavior(str, Enum):
|
|
115
106
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
116
107
|
|
|
117
108
|
BLOCK = "block"
|
|
@@ -141,21 +132,21 @@ class CustomLabel(BaseModel):
|
|
|
141
132
|
value: str
|
|
142
133
|
|
|
143
134
|
|
|
144
|
-
class OutputGoogleChronicleCompression(str, Enum
|
|
135
|
+
class OutputGoogleChronicleCompression(str, Enum):
|
|
145
136
|
r"""Codec to use to compress the persisted data"""
|
|
146
137
|
|
|
147
138
|
NONE = "none"
|
|
148
139
|
GZIP = "gzip"
|
|
149
140
|
|
|
150
141
|
|
|
151
|
-
class OutputGoogleChronicleQueueFullBehavior(str, Enum
|
|
142
|
+
class OutputGoogleChronicleQueueFullBehavior(str, Enum):
|
|
152
143
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
153
144
|
|
|
154
145
|
BLOCK = "block"
|
|
155
146
|
DROP = "drop"
|
|
156
147
|
|
|
157
148
|
|
|
158
|
-
class OutputGoogleChronicleMode(str, Enum
|
|
149
|
+
class OutputGoogleChronicleMode(str, Enum):
|
|
159
150
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
160
151
|
|
|
161
152
|
ERROR = "error"
|
|
@@ -285,18 +276,11 @@ class OutputGoogleChronicle(BaseModel):
|
|
|
285
276
|
r"""Tags for filtering and grouping in @{product}"""
|
|
286
277
|
|
|
287
278
|
api_version: Annotated[
|
|
288
|
-
|
|
289
|
-
Optional[OutputGoogleChronicleAPIVersion],
|
|
290
|
-
PlainValidator(validate_open_enum(False)),
|
|
291
|
-
],
|
|
292
|
-
pydantic.Field(alias="apiVersion"),
|
|
279
|
+
Optional[OutputGoogleChronicleAPIVersion], pydantic.Field(alias="apiVersion")
|
|
293
280
|
] = OutputGoogleChronicleAPIVersion.V1
|
|
294
281
|
|
|
295
282
|
authentication_method: Annotated[
|
|
296
|
-
|
|
297
|
-
Optional[OutputGoogleChronicleAuthenticationMethod],
|
|
298
|
-
PlainValidator(validate_open_enum(False)),
|
|
299
|
-
],
|
|
283
|
+
Optional[OutputGoogleChronicleAuthenticationMethod],
|
|
300
284
|
pydantic.Field(alias="authenticationMethod"),
|
|
301
285
|
] = OutputGoogleChronicleAuthenticationMethod.SERVICE_ACCOUNT
|
|
302
286
|
|
|
@@ -317,8 +301,7 @@ class OutputGoogleChronicle(BaseModel):
|
|
|
317
301
|
r"""Honor any Retry-After header that specifies a delay (in seconds) no longer than 180 seconds after the retry request. @{product} limits the delay to 180 seconds, even if the Retry-After header specifies a longer delay. When enabled, takes precedence over user-configured retry options. When disabled, all Retry-After headers are ignored."""
|
|
318
302
|
|
|
319
303
|
log_format_type: Annotated[
|
|
320
|
-
|
|
321
|
-
pydantic.Field(alias="logFormatType"),
|
|
304
|
+
Optional[SendEventsAs], pydantic.Field(alias="logFormatType")
|
|
322
305
|
] = SendEventsAs.UNSTRUCTURED
|
|
323
306
|
|
|
324
307
|
region: Optional[str] = None
|
|
@@ -363,10 +346,7 @@ class OutputGoogleChronicle(BaseModel):
|
|
|
363
346
|
r"""Headers to add to all events"""
|
|
364
347
|
|
|
365
348
|
failed_request_logging_mode: Annotated[
|
|
366
|
-
|
|
367
|
-
Optional[OutputGoogleChronicleFailedRequestLoggingMode],
|
|
368
|
-
PlainValidator(validate_open_enum(False)),
|
|
369
|
-
],
|
|
349
|
+
Optional[OutputGoogleChronicleFailedRequestLoggingMode],
|
|
370
350
|
pydantic.Field(alias="failedRequestLoggingMode"),
|
|
371
351
|
] = OutputGoogleChronicleFailedRequestLoggingMode.NONE
|
|
372
352
|
r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
|
|
@@ -382,10 +362,7 @@ class OutputGoogleChronicle(BaseModel):
|
|
|
382
362
|
r"""Enable round-robin DNS lookup. When a DNS server returns multiple addresses, @{product} will cycle through them in the order returned."""
|
|
383
363
|
|
|
384
364
|
on_backpressure: Annotated[
|
|
385
|
-
|
|
386
|
-
Optional[OutputGoogleChronicleBackpressureBehavior],
|
|
387
|
-
PlainValidator(validate_open_enum(False)),
|
|
388
|
-
],
|
|
365
|
+
Optional[OutputGoogleChronicleBackpressureBehavior],
|
|
389
366
|
pydantic.Field(alias="onBackpressure"),
|
|
390
367
|
] = OutputGoogleChronicleBackpressureBehavior.BLOCK
|
|
391
368
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -453,29 +430,18 @@ class OutputGoogleChronicle(BaseModel):
|
|
|
453
430
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
454
431
|
|
|
455
432
|
pq_compress: Annotated[
|
|
456
|
-
|
|
457
|
-
Optional[OutputGoogleChronicleCompression],
|
|
458
|
-
PlainValidator(validate_open_enum(False)),
|
|
459
|
-
],
|
|
460
|
-
pydantic.Field(alias="pqCompress"),
|
|
433
|
+
Optional[OutputGoogleChronicleCompression], pydantic.Field(alias="pqCompress")
|
|
461
434
|
] = OutputGoogleChronicleCompression.NONE
|
|
462
435
|
r"""Codec to use to compress the persisted data"""
|
|
463
436
|
|
|
464
437
|
pq_on_backpressure: Annotated[
|
|
465
|
-
|
|
466
|
-
Optional[OutputGoogleChronicleQueueFullBehavior],
|
|
467
|
-
PlainValidator(validate_open_enum(False)),
|
|
468
|
-
],
|
|
438
|
+
Optional[OutputGoogleChronicleQueueFullBehavior],
|
|
469
439
|
pydantic.Field(alias="pqOnBackpressure"),
|
|
470
440
|
] = OutputGoogleChronicleQueueFullBehavior.BLOCK
|
|
471
441
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
472
442
|
|
|
473
443
|
pq_mode: Annotated[
|
|
474
|
-
|
|
475
|
-
Optional[OutputGoogleChronicleMode],
|
|
476
|
-
PlainValidator(validate_open_enum(False)),
|
|
477
|
-
],
|
|
478
|
-
pydantic.Field(alias="pqMode"),
|
|
444
|
+
Optional[OutputGoogleChronicleMode], pydantic.Field(alias="pqMode")
|
|
479
445
|
] = OutputGoogleChronicleMode.ERROR
|
|
480
446
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
481
447
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -15,14 +12,14 @@ class OutputGoogleCloudLoggingType(str, Enum):
|
|
|
15
12
|
GOOGLE_CLOUD_LOGGING = "google_cloud_logging"
|
|
16
13
|
|
|
17
14
|
|
|
18
|
-
class LogLocationType(str, Enum
|
|
15
|
+
class LogLocationType(str, Enum):
|
|
19
16
|
PROJECT = "project"
|
|
20
17
|
ORGANIZATION = "organization"
|
|
21
18
|
BILLING_ACCOUNT = "billingAccount"
|
|
22
19
|
FOLDER = "folder"
|
|
23
20
|
|
|
24
21
|
|
|
25
|
-
class PayloadFormat(str, Enum
|
|
22
|
+
class PayloadFormat(str, Enum):
|
|
26
23
|
r"""Format to use when sending payload. Defaults to Text."""
|
|
27
24
|
|
|
28
25
|
TEXT = "text"
|
|
@@ -59,9 +56,7 @@ class ResourceTypeLabel(BaseModel):
|
|
|
59
56
|
r"""JavaScript expression to compute the label's value."""
|
|
60
57
|
|
|
61
58
|
|
|
62
|
-
class OutputGoogleCloudLoggingGoogleAuthenticationMethod(
|
|
63
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
64
|
-
):
|
|
59
|
+
class OutputGoogleCloudLoggingGoogleAuthenticationMethod(str, Enum):
|
|
65
60
|
r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
|
|
66
61
|
|
|
67
62
|
AUTO = "auto"
|
|
@@ -69,9 +64,7 @@ class OutputGoogleCloudLoggingGoogleAuthenticationMethod(
|
|
|
69
64
|
SECRET = "secret"
|
|
70
65
|
|
|
71
66
|
|
|
72
|
-
class OutputGoogleCloudLoggingBackpressureBehavior(
|
|
73
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
74
|
-
):
|
|
67
|
+
class OutputGoogleCloudLoggingBackpressureBehavior(str, Enum):
|
|
75
68
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
76
69
|
|
|
77
70
|
BLOCK = "block"
|
|
@@ -79,23 +72,21 @@ class OutputGoogleCloudLoggingBackpressureBehavior(
|
|
|
79
72
|
QUEUE = "queue"
|
|
80
73
|
|
|
81
74
|
|
|
82
|
-
class OutputGoogleCloudLoggingCompression(str, Enum
|
|
75
|
+
class OutputGoogleCloudLoggingCompression(str, Enum):
|
|
83
76
|
r"""Codec to use to compress the persisted data"""
|
|
84
77
|
|
|
85
78
|
NONE = "none"
|
|
86
79
|
GZIP = "gzip"
|
|
87
80
|
|
|
88
81
|
|
|
89
|
-
class OutputGoogleCloudLoggingQueueFullBehavior(
|
|
90
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
91
|
-
):
|
|
82
|
+
class OutputGoogleCloudLoggingQueueFullBehavior(str, Enum):
|
|
92
83
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
93
84
|
|
|
94
85
|
BLOCK = "block"
|
|
95
86
|
DROP = "drop"
|
|
96
87
|
|
|
97
88
|
|
|
98
|
-
class OutputGoogleCloudLoggingMode(str, Enum
|
|
89
|
+
class OutputGoogleCloudLoggingMode(str, Enum):
|
|
99
90
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
100
91
|
|
|
101
92
|
ERROR = "error"
|
|
@@ -115,9 +106,9 @@ class OutputGoogleCloudLoggingTypedDict(TypedDict):
|
|
|
115
106
|
type: OutputGoogleCloudLoggingType
|
|
116
107
|
log_location_type: LogLocationType
|
|
117
108
|
log_name_expression: str
|
|
118
|
-
r"""JavaScript expression to compute the value of the log name.
|
|
109
|
+
r"""JavaScript expression to compute the value of the log name."""
|
|
119
110
|
log_location_expression: str
|
|
120
|
-
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated.
|
|
111
|
+
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated."""
|
|
121
112
|
id: NotRequired[str]
|
|
122
113
|
r"""Unique ID for this output"""
|
|
123
114
|
pipeline: NotRequired[str]
|
|
@@ -128,7 +119,6 @@ class OutputGoogleCloudLoggingTypedDict(TypedDict):
|
|
|
128
119
|
r"""Optionally, enable this config only on a specified Git branch. If empty, will be enabled everywhere."""
|
|
129
120
|
streamtags: NotRequired[List[str]]
|
|
130
121
|
r"""Tags for filtering and grouping in @{product}"""
|
|
131
|
-
sanitize_log_names: NotRequired[bool]
|
|
132
122
|
payload_format: NotRequired[PayloadFormat]
|
|
133
123
|
r"""Format to use when sending payload. Defaults to Text."""
|
|
134
124
|
log_labels: NotRequired[List[LogLabelTypedDict]]
|
|
@@ -243,17 +233,16 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
243
233
|
type: OutputGoogleCloudLoggingType
|
|
244
234
|
|
|
245
235
|
log_location_type: Annotated[
|
|
246
|
-
|
|
247
|
-
pydantic.Field(alias="logLocationType"),
|
|
236
|
+
LogLocationType, pydantic.Field(alias="logLocationType")
|
|
248
237
|
]
|
|
249
238
|
|
|
250
239
|
log_name_expression: Annotated[str, pydantic.Field(alias="logNameExpression")]
|
|
251
|
-
r"""JavaScript expression to compute the value of the log name.
|
|
240
|
+
r"""JavaScript expression to compute the value of the log name."""
|
|
252
241
|
|
|
253
242
|
log_location_expression: Annotated[
|
|
254
243
|
str, pydantic.Field(alias="logLocationExpression")
|
|
255
244
|
]
|
|
256
|
-
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated.
|
|
245
|
+
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated."""
|
|
257
246
|
|
|
258
247
|
id: Optional[str] = None
|
|
259
248
|
r"""Unique ID for this output"""
|
|
@@ -272,13 +261,8 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
272
261
|
streamtags: Optional[List[str]] = None
|
|
273
262
|
r"""Tags for filtering and grouping in @{product}"""
|
|
274
263
|
|
|
275
|
-
sanitize_log_names: Annotated[
|
|
276
|
-
Optional[bool], pydantic.Field(alias="sanitizeLogNames")
|
|
277
|
-
] = False
|
|
278
|
-
|
|
279
264
|
payload_format: Annotated[
|
|
280
|
-
|
|
281
|
-
pydantic.Field(alias="payloadFormat"),
|
|
265
|
+
Optional[PayloadFormat], pydantic.Field(alias="payloadFormat")
|
|
282
266
|
] = PayloadFormat.TEXT
|
|
283
267
|
r"""Format to use when sending payload. Defaults to Text."""
|
|
284
268
|
|
|
@@ -308,10 +292,7 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
308
292
|
r"""JavaScript expression to compute the value of the insert ID field."""
|
|
309
293
|
|
|
310
294
|
google_auth_method: Annotated[
|
|
311
|
-
|
|
312
|
-
Optional[OutputGoogleCloudLoggingGoogleAuthenticationMethod],
|
|
313
|
-
PlainValidator(validate_open_enum(False)),
|
|
314
|
-
],
|
|
295
|
+
Optional[OutputGoogleCloudLoggingGoogleAuthenticationMethod],
|
|
315
296
|
pydantic.Field(alias="googleAuthMethod"),
|
|
316
297
|
] = OutputGoogleCloudLoggingGoogleAuthenticationMethod.MANUAL
|
|
317
298
|
r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
|
|
@@ -494,10 +475,7 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
494
475
|
r"""A JavaScript expression that evaluates to the the sampling decision of the span associated with the log entry. See the [documentation](https://cloud.google.com/logging/docs/reference/v2/rest/v2/LogEntry) for details."""
|
|
495
476
|
|
|
496
477
|
on_backpressure: Annotated[
|
|
497
|
-
|
|
498
|
-
Optional[OutputGoogleCloudLoggingBackpressureBehavior],
|
|
499
|
-
PlainValidator(validate_open_enum(False)),
|
|
500
|
-
],
|
|
478
|
+
Optional[OutputGoogleCloudLoggingBackpressureBehavior],
|
|
501
479
|
pydantic.Field(alias="onBackpressure"),
|
|
502
480
|
] = OutputGoogleCloudLoggingBackpressureBehavior.BLOCK
|
|
503
481
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -528,29 +506,19 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
528
506
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
529
507
|
|
|
530
508
|
pq_compress: Annotated[
|
|
531
|
-
|
|
532
|
-
Optional[OutputGoogleCloudLoggingCompression],
|
|
533
|
-
PlainValidator(validate_open_enum(False)),
|
|
534
|
-
],
|
|
509
|
+
Optional[OutputGoogleCloudLoggingCompression],
|
|
535
510
|
pydantic.Field(alias="pqCompress"),
|
|
536
511
|
] = OutputGoogleCloudLoggingCompression.NONE
|
|
537
512
|
r"""Codec to use to compress the persisted data"""
|
|
538
513
|
|
|
539
514
|
pq_on_backpressure: Annotated[
|
|
540
|
-
|
|
541
|
-
Optional[OutputGoogleCloudLoggingQueueFullBehavior],
|
|
542
|
-
PlainValidator(validate_open_enum(False)),
|
|
543
|
-
],
|
|
515
|
+
Optional[OutputGoogleCloudLoggingQueueFullBehavior],
|
|
544
516
|
pydantic.Field(alias="pqOnBackpressure"),
|
|
545
517
|
] = OutputGoogleCloudLoggingQueueFullBehavior.BLOCK
|
|
546
518
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
547
519
|
|
|
548
520
|
pq_mode: Annotated[
|
|
549
|
-
|
|
550
|
-
Optional[OutputGoogleCloudLoggingMode],
|
|
551
|
-
PlainValidator(validate_open_enum(False)),
|
|
552
|
-
],
|
|
553
|
-
pydantic.Field(alias="pqMode"),
|
|
521
|
+
Optional[OutputGoogleCloudLoggingMode], pydantic.Field(alias="pqMode")
|
|
554
522
|
] = OutputGoogleCloudLoggingMode.ERROR
|
|
555
523
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
556
524
|
|