cribl-control-plane 0.0.50rc2__py3-none-any.whl → 0.0.51__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +5 -3
- cribl_control_plane/errors/healthstatus_error.py +2 -8
- cribl_control_plane/groups_sdk.py +4 -4
- cribl_control_plane/health.py +2 -6
- cribl_control_plane/models/__init__.py +31 -56
- cribl_control_plane/models/appmode.py +13 -0
- cribl_control_plane/models/cacheconnection.py +2 -10
- cribl_control_plane/models/cacheconnectionbackfillstatus.py +1 -2
- cribl_control_plane/models/cloudprovider.py +1 -2
- cribl_control_plane/models/configgroup.py +4 -24
- cribl_control_plane/models/configgroupcloud.py +2 -6
- cribl_control_plane/models/createconfiggroupbyproductop.py +2 -8
- cribl_control_plane/models/createinputhectokenbyidop.py +5 -6
- cribl_control_plane/models/createversionpushop.py +5 -5
- cribl_control_plane/models/cribllakedataset.py +2 -8
- cribl_control_plane/models/datasetmetadata.py +2 -8
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +2 -7
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -3
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +2 -7
- cribl_control_plane/models/getsummaryop.py +2 -7
- cribl_control_plane/models/getversionshowop.py +5 -6
- cribl_control_plane/models/gitinfo.py +3 -14
- cribl_control_plane/models/hbcriblinfo.py +3 -24
- cribl_control_plane/models/healthstatus.py +4 -7
- cribl_control_plane/models/heartbeatmetadata.py +0 -3
- cribl_control_plane/models/input.py +63 -65
- cribl_control_plane/models/inputappscope.py +14 -34
- cribl_control_plane/models/inputazureblob.py +6 -17
- cribl_control_plane/models/inputcollection.py +4 -11
- cribl_control_plane/models/inputconfluentcloud.py +32 -41
- cribl_control_plane/models/inputcribl.py +4 -11
- cribl_control_plane/models/inputcriblhttp.py +8 -23
- cribl_control_plane/models/inputcribllakehttp.py +10 -22
- cribl_control_plane/models/inputcriblmetrics.py +4 -12
- cribl_control_plane/models/inputcribltcp.py +8 -23
- cribl_control_plane/models/inputcrowdstrike.py +10 -26
- cribl_control_plane/models/inputdatadogagent.py +8 -24
- cribl_control_plane/models/inputdatagen.py +4 -11
- cribl_control_plane/models/inputedgeprometheus.py +24 -58
- cribl_control_plane/models/inputelastic.py +14 -40
- cribl_control_plane/models/inputeventhub.py +6 -15
- cribl_control_plane/models/inputexec.py +6 -14
- cribl_control_plane/models/inputfile.py +6 -15
- cribl_control_plane/models/inputfirehose.py +8 -23
- cribl_control_plane/models/inputgooglepubsub.py +6 -19
- cribl_control_plane/models/inputgrafana.py +24 -67
- cribl_control_plane/models/inputhttp.py +8 -23
- cribl_control_plane/models/inputhttpraw.py +8 -23
- cribl_control_plane/models/inputjournalfiles.py +4 -12
- cribl_control_plane/models/inputkafka.py +28 -41
- cribl_control_plane/models/inputkinesis.py +14 -38
- cribl_control_plane/models/inputkubeevents.py +4 -11
- cribl_control_plane/models/inputkubelogs.py +8 -16
- cribl_control_plane/models/inputkubemetrics.py +8 -16
- cribl_control_plane/models/inputloki.py +10 -29
- cribl_control_plane/models/inputmetrics.py +8 -23
- cribl_control_plane/models/inputmodeldriventelemetry.py +10 -32
- cribl_control_plane/models/inputmsk.py +30 -48
- cribl_control_plane/models/inputnetflow.py +4 -11
- cribl_control_plane/models/inputoffice365mgmt.py +14 -33
- cribl_control_plane/models/inputoffice365msgtrace.py +16 -35
- cribl_control_plane/models/inputoffice365service.py +16 -35
- cribl_control_plane/models/inputopentelemetry.py +16 -38
- cribl_control_plane/models/inputprometheus.py +18 -50
- cribl_control_plane/models/inputprometheusrw.py +10 -30
- cribl_control_plane/models/inputrawudp.py +4 -11
- cribl_control_plane/models/inputs3.py +8 -21
- cribl_control_plane/models/inputs3inventory.py +10 -26
- cribl_control_plane/models/inputsecuritylake.py +10 -27
- cribl_control_plane/models/inputsnmp.py +6 -16
- cribl_control_plane/models/inputsplunk.py +12 -33
- cribl_control_plane/models/inputsplunkhec.py +10 -29
- cribl_control_plane/models/inputsplunksearch.py +14 -33
- cribl_control_plane/models/inputsqs.py +10 -27
- cribl_control_plane/models/inputsyslog.py +16 -43
- cribl_control_plane/models/inputsystemmetrics.py +24 -48
- cribl_control_plane/models/inputsystemstate.py +8 -16
- cribl_control_plane/models/inputtcp.py +10 -29
- cribl_control_plane/models/inputtcpjson.py +10 -29
- cribl_control_plane/models/inputwef.py +14 -37
- cribl_control_plane/models/inputwindowsmetrics.py +24 -44
- cribl_control_plane/models/inputwineventlogs.py +10 -20
- cribl_control_plane/models/inputwiz.py +8 -21
- cribl_control_plane/models/inputwizwebhook.py +8 -23
- cribl_control_plane/models/inputzscalerhec.py +10 -29
- cribl_control_plane/models/lakehouseconnectiontype.py +1 -2
- cribl_control_plane/models/listconfiggroupbyproductop.py +1 -3
- cribl_control_plane/models/masterworkerentry.py +2 -7
- cribl_control_plane/models/nodeactiveupgradestatus.py +1 -2
- cribl_control_plane/models/nodefailedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeprovidedinfo.py +0 -3
- cribl_control_plane/models/nodeskippedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeupgradestate.py +1 -2
- cribl_control_plane/models/nodeupgradestatus.py +5 -13
- cribl_control_plane/models/output.py +79 -84
- cribl_control_plane/models/outputazureblob.py +18 -48
- cribl_control_plane/models/outputazuredataexplorer.py +28 -73
- cribl_control_plane/models/outputazureeventhub.py +18 -40
- cribl_control_plane/models/outputazurelogs.py +12 -35
- cribl_control_plane/models/outputclickhouse.py +20 -55
- cribl_control_plane/models/outputcloudwatch.py +10 -29
- cribl_control_plane/models/outputconfluentcloud.py +44 -71
- cribl_control_plane/models/outputcriblhttp.py +16 -44
- cribl_control_plane/models/outputcribllake.py +16 -46
- cribl_control_plane/models/outputcribltcp.py +18 -45
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +14 -49
- cribl_control_plane/models/outputdatadog.py +20 -48
- cribl_control_plane/models/outputdataset.py +18 -46
- cribl_control_plane/models/outputdiskspool.py +2 -7
- cribl_control_plane/models/outputdls3.py +24 -68
- cribl_control_plane/models/outputdynatracehttp.py +20 -53
- cribl_control_plane/models/outputdynatraceotlp.py +22 -55
- cribl_control_plane/models/outputelastic.py +18 -43
- cribl_control_plane/models/outputelasticcloud.py +12 -36
- cribl_control_plane/models/outputexabeam.py +10 -29
- cribl_control_plane/models/outputfilesystem.py +14 -39
- cribl_control_plane/models/outputgooglechronicle.py +16 -50
- cribl_control_plane/models/outputgooglecloudlogging.py +18 -50
- cribl_control_plane/models/outputgooglecloudstorage.py +24 -66
- cribl_control_plane/models/outputgooglepubsub.py +10 -31
- cribl_control_plane/models/outputgrafanacloud.py +32 -97
- cribl_control_plane/models/outputgraphite.py +14 -31
- cribl_control_plane/models/outputhoneycomb.py +12 -35
- cribl_control_plane/models/outputhumiohec.py +16 -43
- cribl_control_plane/models/outputinfluxdb.py +16 -42
- cribl_control_plane/models/outputkafka.py +40 -69
- cribl_control_plane/models/outputkinesis.py +16 -40
- cribl_control_plane/models/outputloki.py +16 -41
- cribl_control_plane/models/outputminio.py +24 -65
- cribl_control_plane/models/outputmsk.py +42 -77
- cribl_control_plane/models/outputnewrelic.py +18 -43
- cribl_control_plane/models/outputnewrelicevents.py +14 -41
- cribl_control_plane/models/outputopentelemetry.py +26 -67
- cribl_control_plane/models/outputprometheus.py +12 -35
- cribl_control_plane/models/outputring.py +8 -19
- cribl_control_plane/models/outputs3.py +26 -68
- cribl_control_plane/models/outputsecuritylake.py +18 -52
- cribl_control_plane/models/outputsentinel.py +18 -45
- cribl_control_plane/models/outputsentineloneaisiem.py +18 -50
- cribl_control_plane/models/outputservicenow.py +24 -60
- cribl_control_plane/models/outputsignalfx.py +14 -37
- cribl_control_plane/models/outputsns.py +14 -36
- cribl_control_plane/models/outputsplunk.py +24 -60
- cribl_control_plane/models/outputsplunkhec.py +12 -35
- cribl_control_plane/models/outputsplunklb.py +30 -77
- cribl_control_plane/models/outputsqs.py +16 -41
- cribl_control_plane/models/outputstatsd.py +14 -30
- cribl_control_plane/models/outputstatsdext.py +12 -29
- cribl_control_plane/models/outputsumologic.py +12 -35
- cribl_control_plane/models/outputsyslog.py +24 -58
- cribl_control_plane/models/outputtcpjson.py +20 -52
- cribl_control_plane/models/outputwavefront.py +12 -35
- cribl_control_plane/models/outputwebhook.py +22 -58
- cribl_control_plane/models/outputxsiam.py +14 -35
- cribl_control_plane/models/productscore.py +1 -2
- cribl_control_plane/models/rbacresource.py +1 -2
- cribl_control_plane/models/resourcepolicy.py +2 -4
- cribl_control_plane/models/routecloneconf.py +13 -0
- cribl_control_plane/models/routeconf.py +4 -3
- cribl_control_plane/models/runnablejobcollection.py +13 -30
- cribl_control_plane/models/runnablejobexecutor.py +4 -13
- cribl_control_plane/models/runnablejobscheduledsearch.py +2 -7
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +2 -8
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +2 -8
- cribl_control_plane/models/updateinputhectokenbyidandtokenop.py +5 -6
- cribl_control_plane/models/workertypes.py +1 -2
- {cribl_control_plane-0.0.50rc2.dist-info → cribl_control_plane-0.0.51.dist-info}/METADATA +14 -5
- cribl_control_plane-0.0.51.dist-info/RECORD +325 -0
- cribl_control_plane/models/error.py +0 -16
- cribl_control_plane/models/gethealthinfoop.py +0 -17
- cribl_control_plane/models/gitshowresult.py +0 -19
- cribl_control_plane/models/outputdatabricks.py +0 -282
- cribl_control_plane-0.0.50rc2.dist-info/RECORD +0 -327
- {cribl_control_plane-0.0.50rc2.dist-info → cribl_control_plane-0.0.51.dist-info}/WHEEL +0 -0
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -15,14 +12,14 @@ class OutputGoogleCloudLoggingType(str, Enum):
|
|
|
15
12
|
GOOGLE_CLOUD_LOGGING = "google_cloud_logging"
|
|
16
13
|
|
|
17
14
|
|
|
18
|
-
class LogLocationType(str, Enum
|
|
15
|
+
class LogLocationType(str, Enum):
|
|
19
16
|
PROJECT = "project"
|
|
20
17
|
ORGANIZATION = "organization"
|
|
21
18
|
BILLING_ACCOUNT = "billingAccount"
|
|
22
19
|
FOLDER = "folder"
|
|
23
20
|
|
|
24
21
|
|
|
25
|
-
class PayloadFormat(str, Enum
|
|
22
|
+
class PayloadFormat(str, Enum):
|
|
26
23
|
r"""Format to use when sending payload. Defaults to Text."""
|
|
27
24
|
|
|
28
25
|
TEXT = "text"
|
|
@@ -59,9 +56,7 @@ class ResourceTypeLabel(BaseModel):
|
|
|
59
56
|
r"""JavaScript expression to compute the label's value."""
|
|
60
57
|
|
|
61
58
|
|
|
62
|
-
class OutputGoogleCloudLoggingGoogleAuthenticationMethod(
|
|
63
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
64
|
-
):
|
|
59
|
+
class OutputGoogleCloudLoggingGoogleAuthenticationMethod(str, Enum):
|
|
65
60
|
r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
|
|
66
61
|
|
|
67
62
|
AUTO = "auto"
|
|
@@ -69,9 +64,7 @@ class OutputGoogleCloudLoggingGoogleAuthenticationMethod(
|
|
|
69
64
|
SECRET = "secret"
|
|
70
65
|
|
|
71
66
|
|
|
72
|
-
class OutputGoogleCloudLoggingBackpressureBehavior(
|
|
73
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
74
|
-
):
|
|
67
|
+
class OutputGoogleCloudLoggingBackpressureBehavior(str, Enum):
|
|
75
68
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
76
69
|
|
|
77
70
|
BLOCK = "block"
|
|
@@ -79,23 +72,21 @@ class OutputGoogleCloudLoggingBackpressureBehavior(
|
|
|
79
72
|
QUEUE = "queue"
|
|
80
73
|
|
|
81
74
|
|
|
82
|
-
class OutputGoogleCloudLoggingCompression(str, Enum
|
|
75
|
+
class OutputGoogleCloudLoggingCompression(str, Enum):
|
|
83
76
|
r"""Codec to use to compress the persisted data"""
|
|
84
77
|
|
|
85
78
|
NONE = "none"
|
|
86
79
|
GZIP = "gzip"
|
|
87
80
|
|
|
88
81
|
|
|
89
|
-
class OutputGoogleCloudLoggingQueueFullBehavior(
|
|
90
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
91
|
-
):
|
|
82
|
+
class OutputGoogleCloudLoggingQueueFullBehavior(str, Enum):
|
|
92
83
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
93
84
|
|
|
94
85
|
BLOCK = "block"
|
|
95
86
|
DROP = "drop"
|
|
96
87
|
|
|
97
88
|
|
|
98
|
-
class OutputGoogleCloudLoggingMode(str, Enum
|
|
89
|
+
class OutputGoogleCloudLoggingMode(str, Enum):
|
|
99
90
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
100
91
|
|
|
101
92
|
ERROR = "error"
|
|
@@ -115,9 +106,9 @@ class OutputGoogleCloudLoggingTypedDict(TypedDict):
|
|
|
115
106
|
type: OutputGoogleCloudLoggingType
|
|
116
107
|
log_location_type: LogLocationType
|
|
117
108
|
log_name_expression: str
|
|
118
|
-
r"""JavaScript expression to compute the value of the log name.
|
|
109
|
+
r"""JavaScript expression to compute the value of the log name."""
|
|
119
110
|
log_location_expression: str
|
|
120
|
-
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated.
|
|
111
|
+
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated."""
|
|
121
112
|
id: NotRequired[str]
|
|
122
113
|
r"""Unique ID for this output"""
|
|
123
114
|
pipeline: NotRequired[str]
|
|
@@ -128,7 +119,6 @@ class OutputGoogleCloudLoggingTypedDict(TypedDict):
|
|
|
128
119
|
r"""Optionally, enable this config only on a specified Git branch. If empty, will be enabled everywhere."""
|
|
129
120
|
streamtags: NotRequired[List[str]]
|
|
130
121
|
r"""Tags for filtering and grouping in @{product}"""
|
|
131
|
-
sanitize_log_names: NotRequired[bool]
|
|
132
122
|
payload_format: NotRequired[PayloadFormat]
|
|
133
123
|
r"""Format to use when sending payload. Defaults to Text."""
|
|
134
124
|
log_labels: NotRequired[List[LogLabelTypedDict]]
|
|
@@ -243,17 +233,16 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
243
233
|
type: OutputGoogleCloudLoggingType
|
|
244
234
|
|
|
245
235
|
log_location_type: Annotated[
|
|
246
|
-
|
|
247
|
-
pydantic.Field(alias="logLocationType"),
|
|
236
|
+
LogLocationType, pydantic.Field(alias="logLocationType")
|
|
248
237
|
]
|
|
249
238
|
|
|
250
239
|
log_name_expression: Annotated[str, pydantic.Field(alias="logNameExpression")]
|
|
251
|
-
r"""JavaScript expression to compute the value of the log name.
|
|
240
|
+
r"""JavaScript expression to compute the value of the log name."""
|
|
252
241
|
|
|
253
242
|
log_location_expression: Annotated[
|
|
254
243
|
str, pydantic.Field(alias="logLocationExpression")
|
|
255
244
|
]
|
|
256
|
-
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated.
|
|
245
|
+
r"""JavaScript expression to compute the value of the folder ID with which log entries should be associated."""
|
|
257
246
|
|
|
258
247
|
id: Optional[str] = None
|
|
259
248
|
r"""Unique ID for this output"""
|
|
@@ -272,13 +261,8 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
272
261
|
streamtags: Optional[List[str]] = None
|
|
273
262
|
r"""Tags for filtering and grouping in @{product}"""
|
|
274
263
|
|
|
275
|
-
sanitize_log_names: Annotated[
|
|
276
|
-
Optional[bool], pydantic.Field(alias="sanitizeLogNames")
|
|
277
|
-
] = False
|
|
278
|
-
|
|
279
264
|
payload_format: Annotated[
|
|
280
|
-
|
|
281
|
-
pydantic.Field(alias="payloadFormat"),
|
|
265
|
+
Optional[PayloadFormat], pydantic.Field(alias="payloadFormat")
|
|
282
266
|
] = PayloadFormat.TEXT
|
|
283
267
|
r"""Format to use when sending payload. Defaults to Text."""
|
|
284
268
|
|
|
@@ -308,10 +292,7 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
308
292
|
r"""JavaScript expression to compute the value of the insert ID field."""
|
|
309
293
|
|
|
310
294
|
google_auth_method: Annotated[
|
|
311
|
-
|
|
312
|
-
Optional[OutputGoogleCloudLoggingGoogleAuthenticationMethod],
|
|
313
|
-
PlainValidator(validate_open_enum(False)),
|
|
314
|
-
],
|
|
295
|
+
Optional[OutputGoogleCloudLoggingGoogleAuthenticationMethod],
|
|
315
296
|
pydantic.Field(alias="googleAuthMethod"),
|
|
316
297
|
] = OutputGoogleCloudLoggingGoogleAuthenticationMethod.MANUAL
|
|
317
298
|
r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
|
|
@@ -494,10 +475,7 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
494
475
|
r"""A JavaScript expression that evaluates to the the sampling decision of the span associated with the log entry. See the [documentation](https://cloud.google.com/logging/docs/reference/v2/rest/v2/LogEntry) for details."""
|
|
495
476
|
|
|
496
477
|
on_backpressure: Annotated[
|
|
497
|
-
|
|
498
|
-
Optional[OutputGoogleCloudLoggingBackpressureBehavior],
|
|
499
|
-
PlainValidator(validate_open_enum(False)),
|
|
500
|
-
],
|
|
478
|
+
Optional[OutputGoogleCloudLoggingBackpressureBehavior],
|
|
501
479
|
pydantic.Field(alias="onBackpressure"),
|
|
502
480
|
] = OutputGoogleCloudLoggingBackpressureBehavior.BLOCK
|
|
503
481
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -528,29 +506,19 @@ class OutputGoogleCloudLogging(BaseModel):
|
|
|
528
506
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
529
507
|
|
|
530
508
|
pq_compress: Annotated[
|
|
531
|
-
|
|
532
|
-
Optional[OutputGoogleCloudLoggingCompression],
|
|
533
|
-
PlainValidator(validate_open_enum(False)),
|
|
534
|
-
],
|
|
509
|
+
Optional[OutputGoogleCloudLoggingCompression],
|
|
535
510
|
pydantic.Field(alias="pqCompress"),
|
|
536
511
|
] = OutputGoogleCloudLoggingCompression.NONE
|
|
537
512
|
r"""Codec to use to compress the persisted data"""
|
|
538
513
|
|
|
539
514
|
pq_on_backpressure: Annotated[
|
|
540
|
-
|
|
541
|
-
Optional[OutputGoogleCloudLoggingQueueFullBehavior],
|
|
542
|
-
PlainValidator(validate_open_enum(False)),
|
|
543
|
-
],
|
|
515
|
+
Optional[OutputGoogleCloudLoggingQueueFullBehavior],
|
|
544
516
|
pydantic.Field(alias="pqOnBackpressure"),
|
|
545
517
|
] = OutputGoogleCloudLoggingQueueFullBehavior.BLOCK
|
|
546
518
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
547
519
|
|
|
548
520
|
pq_mode: Annotated[
|
|
549
|
-
|
|
550
|
-
Optional[OutputGoogleCloudLoggingMode],
|
|
551
|
-
PlainValidator(validate_open_enum(False)),
|
|
552
|
-
],
|
|
553
|
-
pydantic.Field(alias="pqMode"),
|
|
521
|
+
Optional[OutputGoogleCloudLoggingMode], pydantic.Field(alias="pqMode")
|
|
554
522
|
] = OutputGoogleCloudLoggingMode.ERROR
|
|
555
523
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
556
524
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -15,22 +12,20 @@ class OutputGoogleCloudStorageType(str, Enum):
|
|
|
15
12
|
GOOGLE_CLOUD_STORAGE = "google_cloud_storage"
|
|
16
13
|
|
|
17
14
|
|
|
18
|
-
class OutputGoogleCloudStorageSignatureVersion(str, Enum
|
|
15
|
+
class OutputGoogleCloudStorageSignatureVersion(str, Enum):
|
|
19
16
|
r"""Signature version to use for signing Google Cloud Storage requests"""
|
|
20
17
|
|
|
21
18
|
V2 = "v2"
|
|
22
19
|
V4 = "v4"
|
|
23
20
|
|
|
24
21
|
|
|
25
|
-
class OutputGoogleCloudStorageAuthenticationMethod(
|
|
26
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
27
|
-
):
|
|
22
|
+
class OutputGoogleCloudStorageAuthenticationMethod(str, Enum):
|
|
28
23
|
AUTO = "auto"
|
|
29
24
|
MANUAL = "manual"
|
|
30
25
|
SECRET = "secret"
|
|
31
26
|
|
|
32
27
|
|
|
33
|
-
class OutputGoogleCloudStorageObjectACL(str, Enum
|
|
28
|
+
class OutputGoogleCloudStorageObjectACL(str, Enum):
|
|
34
29
|
r"""Object ACL to assign to uploaded objects"""
|
|
35
30
|
|
|
36
31
|
PRIVATE = "private"
|
|
@@ -41,7 +36,7 @@ class OutputGoogleCloudStorageObjectACL(str, Enum, metaclass=utils.OpenEnumMeta)
|
|
|
41
36
|
PUBLIC_READ = "public-read"
|
|
42
37
|
|
|
43
38
|
|
|
44
|
-
class OutputGoogleCloudStorageStorageClass(str, Enum
|
|
39
|
+
class OutputGoogleCloudStorageStorageClass(str, Enum):
|
|
45
40
|
r"""Storage class to select for uploaded objects"""
|
|
46
41
|
|
|
47
42
|
STANDARD = "STANDARD"
|
|
@@ -50,7 +45,7 @@ class OutputGoogleCloudStorageStorageClass(str, Enum, metaclass=utils.OpenEnumMe
|
|
|
50
45
|
ARCHIVE = "ARCHIVE"
|
|
51
46
|
|
|
52
47
|
|
|
53
|
-
class OutputGoogleCloudStorageDataFormat(str, Enum
|
|
48
|
+
class OutputGoogleCloudStorageDataFormat(str, Enum):
|
|
54
49
|
r"""Format of the output data"""
|
|
55
50
|
|
|
56
51
|
JSON = "json"
|
|
@@ -58,32 +53,28 @@ class OutputGoogleCloudStorageDataFormat(str, Enum, metaclass=utils.OpenEnumMeta
|
|
|
58
53
|
PARQUET = "parquet"
|
|
59
54
|
|
|
60
55
|
|
|
61
|
-
class OutputGoogleCloudStorageBackpressureBehavior(
|
|
62
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
63
|
-
):
|
|
56
|
+
class OutputGoogleCloudStorageBackpressureBehavior(str, Enum):
|
|
64
57
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
65
58
|
|
|
66
59
|
BLOCK = "block"
|
|
67
60
|
DROP = "drop"
|
|
68
61
|
|
|
69
62
|
|
|
70
|
-
class OutputGoogleCloudStorageDiskSpaceProtection(
|
|
71
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
72
|
-
):
|
|
63
|
+
class OutputGoogleCloudStorageDiskSpaceProtection(str, Enum):
|
|
73
64
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
74
65
|
|
|
75
66
|
BLOCK = "block"
|
|
76
67
|
DROP = "drop"
|
|
77
68
|
|
|
78
69
|
|
|
79
|
-
class OutputGoogleCloudStorageCompression(str, Enum
|
|
70
|
+
class OutputGoogleCloudStorageCompression(str, Enum):
|
|
80
71
|
r"""Data compression format to apply to HTTP content before it is delivered"""
|
|
81
72
|
|
|
82
73
|
NONE = "none"
|
|
83
74
|
GZIP = "gzip"
|
|
84
75
|
|
|
85
76
|
|
|
86
|
-
class OutputGoogleCloudStorageCompressionLevel(str, Enum
|
|
77
|
+
class OutputGoogleCloudStorageCompressionLevel(str, Enum):
|
|
87
78
|
r"""Compression level to apply before moving files to final destination"""
|
|
88
79
|
|
|
89
80
|
BEST_SPEED = "best_speed"
|
|
@@ -91,7 +82,7 @@ class OutputGoogleCloudStorageCompressionLevel(str, Enum, metaclass=utils.OpenEn
|
|
|
91
82
|
BEST_COMPRESSION = "best_compression"
|
|
92
83
|
|
|
93
84
|
|
|
94
|
-
class OutputGoogleCloudStorageParquetVersion(str, Enum
|
|
85
|
+
class OutputGoogleCloudStorageParquetVersion(str, Enum):
|
|
95
86
|
r"""Determines which data types are supported and how they are represented"""
|
|
96
87
|
|
|
97
88
|
PARQUET_1_0 = "PARQUET_1_0"
|
|
@@ -99,7 +90,7 @@ class OutputGoogleCloudStorageParquetVersion(str, Enum, metaclass=utils.OpenEnum
|
|
|
99
90
|
PARQUET_2_6 = "PARQUET_2_6"
|
|
100
91
|
|
|
101
92
|
|
|
102
|
-
class OutputGoogleCloudStorageDataPageVersion(str, Enum
|
|
93
|
+
class OutputGoogleCloudStorageDataPageVersion(str, Enum):
|
|
103
94
|
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
104
95
|
|
|
105
96
|
DATA_PAGE_V1 = "DATA_PAGE_V1"
|
|
@@ -253,19 +244,13 @@ class OutputGoogleCloudStorage(BaseModel):
|
|
|
253
244
|
r"""Google Cloud Storage service endpoint"""
|
|
254
245
|
|
|
255
246
|
signature_version: Annotated[
|
|
256
|
-
|
|
257
|
-
Optional[OutputGoogleCloudStorageSignatureVersion],
|
|
258
|
-
PlainValidator(validate_open_enum(False)),
|
|
259
|
-
],
|
|
247
|
+
Optional[OutputGoogleCloudStorageSignatureVersion],
|
|
260
248
|
pydantic.Field(alias="signatureVersion"),
|
|
261
249
|
] = OutputGoogleCloudStorageSignatureVersion.V4
|
|
262
250
|
r"""Signature version to use for signing Google Cloud Storage requests"""
|
|
263
251
|
|
|
264
252
|
aws_authentication_method: Annotated[
|
|
265
|
-
|
|
266
|
-
Optional[OutputGoogleCloudStorageAuthenticationMethod],
|
|
267
|
-
PlainValidator(validate_open_enum(False)),
|
|
268
|
-
],
|
|
253
|
+
Optional[OutputGoogleCloudStorageAuthenticationMethod],
|
|
269
254
|
pydantic.Field(alias="awsAuthenticationMethod"),
|
|
270
255
|
] = OutputGoogleCloudStorageAuthenticationMethod.MANUAL
|
|
271
256
|
|
|
@@ -283,19 +268,12 @@ class OutputGoogleCloudStorage(BaseModel):
|
|
|
283
268
|
r"""Disable if you can access files within the bucket but not the bucket itself"""
|
|
284
269
|
|
|
285
270
|
object_acl: Annotated[
|
|
286
|
-
|
|
287
|
-
Optional[OutputGoogleCloudStorageObjectACL],
|
|
288
|
-
PlainValidator(validate_open_enum(False)),
|
|
289
|
-
],
|
|
290
|
-
pydantic.Field(alias="objectACL"),
|
|
271
|
+
Optional[OutputGoogleCloudStorageObjectACL], pydantic.Field(alias="objectACL")
|
|
291
272
|
] = OutputGoogleCloudStorageObjectACL.PRIVATE
|
|
292
273
|
r"""Object ACL to assign to uploaded objects"""
|
|
293
274
|
|
|
294
275
|
storage_class: Annotated[
|
|
295
|
-
|
|
296
|
-
Optional[OutputGoogleCloudStorageStorageClass],
|
|
297
|
-
PlainValidator(validate_open_enum(False)),
|
|
298
|
-
],
|
|
276
|
+
Optional[OutputGoogleCloudStorageStorageClass],
|
|
299
277
|
pydantic.Field(alias="storageClass"),
|
|
300
278
|
] = None
|
|
301
279
|
r"""Storage class to select for uploaded objects"""
|
|
@@ -326,11 +304,7 @@ class OutputGoogleCloudStorage(BaseModel):
|
|
|
326
304
|
r"""JavaScript expression defining how files are partitioned and organized. Default is date-based. If blank, Stream will fall back to the event's __partition field value – if present – otherwise to each location's root directory."""
|
|
327
305
|
|
|
328
306
|
format_: Annotated[
|
|
329
|
-
|
|
330
|
-
Optional[OutputGoogleCloudStorageDataFormat],
|
|
331
|
-
PlainValidator(validate_open_enum(False)),
|
|
332
|
-
],
|
|
333
|
-
pydantic.Field(alias="format"),
|
|
307
|
+
Optional[OutputGoogleCloudStorageDataFormat], pydantic.Field(alias="format")
|
|
334
308
|
] = OutputGoogleCloudStorageDataFormat.JSON
|
|
335
309
|
r"""Format of the output data"""
|
|
336
310
|
|
|
@@ -373,10 +347,7 @@ class OutputGoogleCloudStorage(BaseModel):
|
|
|
373
347
|
r"""Buffer size used to write to a file"""
|
|
374
348
|
|
|
375
349
|
on_backpressure: Annotated[
|
|
376
|
-
|
|
377
|
-
Optional[OutputGoogleCloudStorageBackpressureBehavior],
|
|
378
|
-
PlainValidator(validate_open_enum(False)),
|
|
379
|
-
],
|
|
350
|
+
Optional[OutputGoogleCloudStorageBackpressureBehavior],
|
|
380
351
|
pydantic.Field(alias="onBackpressure"),
|
|
381
352
|
] = OutputGoogleCloudStorageBackpressureBehavior.BLOCK
|
|
382
353
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -387,27 +358,20 @@ class OutputGoogleCloudStorage(BaseModel):
|
|
|
387
358
|
r"""If a file fails to move to its final destination after the maximum number of retries, move it to a designated directory to prevent further errors"""
|
|
388
359
|
|
|
389
360
|
on_disk_full_backpressure: Annotated[
|
|
390
|
-
|
|
391
|
-
Optional[OutputGoogleCloudStorageDiskSpaceProtection],
|
|
392
|
-
PlainValidator(validate_open_enum(False)),
|
|
393
|
-
],
|
|
361
|
+
Optional[OutputGoogleCloudStorageDiskSpaceProtection],
|
|
394
362
|
pydantic.Field(alias="onDiskFullBackpressure"),
|
|
395
363
|
] = OutputGoogleCloudStorageDiskSpaceProtection.BLOCK
|
|
396
364
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
397
365
|
|
|
398
366
|
description: Optional[str] = None
|
|
399
367
|
|
|
400
|
-
compress:
|
|
401
|
-
|
|
402
|
-
|
|
403
|
-
] = OutputGoogleCloudStorageCompression.GZIP
|
|
368
|
+
compress: Optional[OutputGoogleCloudStorageCompression] = (
|
|
369
|
+
OutputGoogleCloudStorageCompression.GZIP
|
|
370
|
+
)
|
|
404
371
|
r"""Data compression format to apply to HTTP content before it is delivered"""
|
|
405
372
|
|
|
406
373
|
compression_level: Annotated[
|
|
407
|
-
|
|
408
|
-
Optional[OutputGoogleCloudStorageCompressionLevel],
|
|
409
|
-
PlainValidator(validate_open_enum(False)),
|
|
410
|
-
],
|
|
374
|
+
Optional[OutputGoogleCloudStorageCompressionLevel],
|
|
411
375
|
pydantic.Field(alias="compressionLevel"),
|
|
412
376
|
] = OutputGoogleCloudStorageCompressionLevel.BEST_SPEED
|
|
413
377
|
r"""Compression level to apply before moving files to final destination"""
|
|
@@ -418,19 +382,13 @@ class OutputGoogleCloudStorage(BaseModel):
|
|
|
418
382
|
r"""Automatically calculate the schema based on the events of each Parquet file generated"""
|
|
419
383
|
|
|
420
384
|
parquet_version: Annotated[
|
|
421
|
-
|
|
422
|
-
Optional[OutputGoogleCloudStorageParquetVersion],
|
|
423
|
-
PlainValidator(validate_open_enum(False)),
|
|
424
|
-
],
|
|
385
|
+
Optional[OutputGoogleCloudStorageParquetVersion],
|
|
425
386
|
pydantic.Field(alias="parquetVersion"),
|
|
426
387
|
] = OutputGoogleCloudStorageParquetVersion.PARQUET_2_6
|
|
427
388
|
r"""Determines which data types are supported and how they are represented"""
|
|
428
389
|
|
|
429
390
|
parquet_data_page_version: Annotated[
|
|
430
|
-
|
|
431
|
-
Optional[OutputGoogleCloudStorageDataPageVersion],
|
|
432
|
-
PlainValidator(validate_open_enum(False)),
|
|
433
|
-
],
|
|
391
|
+
Optional[OutputGoogleCloudStorageDataPageVersion],
|
|
434
392
|
pydantic.Field(alias="parquetDataPageVersion"),
|
|
435
393
|
] = OutputGoogleCloudStorageDataPageVersion.DATA_PAGE_V2
|
|
436
394
|
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -15,9 +12,7 @@ class OutputGooglePubsubType(str, Enum):
|
|
|
15
12
|
GOOGLE_PUBSUB = "google_pubsub"
|
|
16
13
|
|
|
17
14
|
|
|
18
|
-
class OutputGooglePubsubGoogleAuthenticationMethod(
|
|
19
|
-
str, Enum, metaclass=utils.OpenEnumMeta
|
|
20
|
-
):
|
|
15
|
+
class OutputGooglePubsubGoogleAuthenticationMethod(str, Enum):
|
|
21
16
|
r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
|
|
22
17
|
|
|
23
18
|
AUTO = "auto"
|
|
@@ -25,7 +20,7 @@ class OutputGooglePubsubGoogleAuthenticationMethod(
|
|
|
25
20
|
SECRET = "secret"
|
|
26
21
|
|
|
27
22
|
|
|
28
|
-
class OutputGooglePubsubBackpressureBehavior(str, Enum
|
|
23
|
+
class OutputGooglePubsubBackpressureBehavior(str, Enum):
|
|
29
24
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
30
25
|
|
|
31
26
|
BLOCK = "block"
|
|
@@ -33,21 +28,21 @@ class OutputGooglePubsubBackpressureBehavior(str, Enum, metaclass=utils.OpenEnum
|
|
|
33
28
|
QUEUE = "queue"
|
|
34
29
|
|
|
35
30
|
|
|
36
|
-
class OutputGooglePubsubCompression(str, Enum
|
|
31
|
+
class OutputGooglePubsubCompression(str, Enum):
|
|
37
32
|
r"""Codec to use to compress the persisted data"""
|
|
38
33
|
|
|
39
34
|
NONE = "none"
|
|
40
35
|
GZIP = "gzip"
|
|
41
36
|
|
|
42
37
|
|
|
43
|
-
class OutputGooglePubsubQueueFullBehavior(str, Enum
|
|
38
|
+
class OutputGooglePubsubQueueFullBehavior(str, Enum):
|
|
44
39
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
45
40
|
|
|
46
41
|
BLOCK = "block"
|
|
47
42
|
DROP = "drop"
|
|
48
43
|
|
|
49
44
|
|
|
50
|
-
class OutputGooglePubsubMode(str, Enum
|
|
45
|
+
class OutputGooglePubsubMode(str, Enum):
|
|
51
46
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
52
47
|
|
|
53
48
|
ERROR = "error"
|
|
@@ -154,10 +149,7 @@ class OutputGooglePubsub(BaseModel):
|
|
|
154
149
|
r"""Region to publish messages to. Select 'default' to allow Google to auto-select the nearest region. When using ordered delivery, the selected region must be allowed by message storage policy."""
|
|
155
150
|
|
|
156
151
|
google_auth_method: Annotated[
|
|
157
|
-
|
|
158
|
-
Optional[OutputGooglePubsubGoogleAuthenticationMethod],
|
|
159
|
-
PlainValidator(validate_open_enum(False)),
|
|
160
|
-
],
|
|
152
|
+
Optional[OutputGooglePubsubGoogleAuthenticationMethod],
|
|
161
153
|
pydantic.Field(alias="googleAuthMethod"),
|
|
162
154
|
] = OutputGooglePubsubGoogleAuthenticationMethod.MANUAL
|
|
163
155
|
r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
|
|
@@ -197,10 +189,7 @@ class OutputGooglePubsub(BaseModel):
|
|
|
197
189
|
r"""The maximum number of in-progress API requests before backpressure is applied."""
|
|
198
190
|
|
|
199
191
|
on_backpressure: Annotated[
|
|
200
|
-
|
|
201
|
-
Optional[OutputGooglePubsubBackpressureBehavior],
|
|
202
|
-
PlainValidator(validate_open_enum(False)),
|
|
203
|
-
],
|
|
192
|
+
Optional[OutputGooglePubsubBackpressureBehavior],
|
|
204
193
|
pydantic.Field(alias="onBackpressure"),
|
|
205
194
|
] = OutputGooglePubsubBackpressureBehavior.BLOCK
|
|
206
195
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -221,28 +210,18 @@ class OutputGooglePubsub(BaseModel):
|
|
|
221
210
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
222
211
|
|
|
223
212
|
pq_compress: Annotated[
|
|
224
|
-
|
|
225
|
-
Optional[OutputGooglePubsubCompression],
|
|
226
|
-
PlainValidator(validate_open_enum(False)),
|
|
227
|
-
],
|
|
228
|
-
pydantic.Field(alias="pqCompress"),
|
|
213
|
+
Optional[OutputGooglePubsubCompression], pydantic.Field(alias="pqCompress")
|
|
229
214
|
] = OutputGooglePubsubCompression.NONE
|
|
230
215
|
r"""Codec to use to compress the persisted data"""
|
|
231
216
|
|
|
232
217
|
pq_on_backpressure: Annotated[
|
|
233
|
-
|
|
234
|
-
Optional[OutputGooglePubsubQueueFullBehavior],
|
|
235
|
-
PlainValidator(validate_open_enum(False)),
|
|
236
|
-
],
|
|
218
|
+
Optional[OutputGooglePubsubQueueFullBehavior],
|
|
237
219
|
pydantic.Field(alias="pqOnBackpressure"),
|
|
238
220
|
] = OutputGooglePubsubQueueFullBehavior.BLOCK
|
|
239
221
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
240
222
|
|
|
241
223
|
pq_mode: Annotated[
|
|
242
|
-
|
|
243
|
-
Optional[OutputGooglePubsubMode], PlainValidator(validate_open_enum(False))
|
|
244
|
-
],
|
|
245
|
-
pydantic.Field(alias="pqMode"),
|
|
224
|
+
Optional[OutputGooglePubsubMode], pydantic.Field(alias="pqMode")
|
|
246
225
|
] = OutputGooglePubsubMode.ERROR
|
|
247
226
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
248
227
|
|