cribl-control-plane 0.2.1rc4__py3-none-any.whl → 0.2.1rc6__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +3 -3
- cribl_control_plane/groups_sdk.py +2 -2
- cribl_control_plane/models/__init__.py +3 -99
- cribl_control_plane/models/createconfiggroupbyproductop.py +2 -2
- cribl_control_plane/models/input.py +14 -14
- cribl_control_plane/models/inputappscope.py +16 -20
- cribl_control_plane/models/inputconfluentcloud.py +0 -110
- cribl_control_plane/models/inputcriblhttp.py +16 -20
- cribl_control_plane/models/inputcribllakehttp.py +16 -20
- cribl_control_plane/models/inputcribltcp.py +16 -20
- cribl_control_plane/models/inputdatadogagent.py +16 -20
- cribl_control_plane/models/inputedgeprometheus.py +36 -44
- cribl_control_plane/models/inputelastic.py +27 -44
- cribl_control_plane/models/inputeventhub.py +0 -118
- cribl_control_plane/models/inputfirehose.py +16 -20
- cribl_control_plane/models/inputgrafana.py +31 -39
- cribl_control_plane/models/inputhttp.py +16 -20
- cribl_control_plane/models/inputhttpraw.py +16 -20
- cribl_control_plane/models/inputkafka.py +0 -108
- cribl_control_plane/models/inputloki.py +16 -20
- cribl_control_plane/models/inputmetrics.py +16 -20
- cribl_control_plane/models/inputmodeldriventelemetry.py +16 -20
- cribl_control_plane/models/inputopentelemetry.py +15 -19
- cribl_control_plane/models/inputprometheus.py +36 -44
- cribl_control_plane/models/inputprometheusrw.py +16 -20
- cribl_control_plane/models/inputsplunk.py +16 -20
- cribl_control_plane/models/inputsplunkhec.py +15 -19
- cribl_control_plane/models/inputsyslog.py +31 -39
- cribl_control_plane/models/inputsystemmetrics.py +10 -20
- cribl_control_plane/models/inputtcp.py +16 -30
- cribl_control_plane/models/inputtcpjson.py +16 -20
- cribl_control_plane/models/inputwindowsmetrics.py +10 -20
- cribl_control_plane/models/inputwineventlogs.py +0 -14
- cribl_control_plane/models/inputwizwebhook.py +16 -20
- cribl_control_plane/models/inputzscalerhec.py +15 -19
- cribl_control_plane/models/output.py +22 -22
- cribl_control_plane/models/outputazureblob.py +0 -7
- cribl_control_plane/models/outputazuredataexplorer.py +93 -283
- cribl_control_plane/models/outputazureeventhub.py +21 -169
- cribl_control_plane/models/outputazurelogs.py +21 -49
- cribl_control_plane/models/outputchronicle.py +21 -49
- cribl_control_plane/models/outputclickhouse.py +21 -49
- cribl_control_plane/models/outputcloudwatch.py +21 -49
- cribl_control_plane/models/outputconfluentcloud.py +22 -167
- cribl_control_plane/models/outputcriblhttp.py +21 -49
- cribl_control_plane/models/outputcribltcp.py +21 -49
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +22 -50
- cribl_control_plane/models/outputdatabricks.py +0 -7
- cribl_control_plane/models/outputdatadog.py +21 -49
- cribl_control_plane/models/outputdataset.py +21 -49
- cribl_control_plane/models/outputdls3.py +0 -7
- cribl_control_plane/models/outputdynatracehttp.py +21 -49
- cribl_control_plane/models/outputdynatraceotlp.py +21 -49
- cribl_control_plane/models/outputelastic.py +21 -74
- cribl_control_plane/models/outputelasticcloud.py +21 -74
- cribl_control_plane/models/outputfilesystem.py +0 -7
- cribl_control_plane/models/outputgooglechronicle.py +22 -65
- cribl_control_plane/models/outputgooglecloudlogging.py +22 -50
- cribl_control_plane/models/outputgooglecloudstorage.py +0 -7
- cribl_control_plane/models/outputgooglepubsub.py +21 -49
- cribl_control_plane/models/outputgrafanacloud.py +42 -98
- cribl_control_plane/models/outputgraphite.py +21 -49
- cribl_control_plane/models/outputhoneycomb.py +21 -49
- cribl_control_plane/models/outputhumiohec.py +21 -49
- cribl_control_plane/models/outputinfluxdb.py +21 -49
- cribl_control_plane/models/outputkafka.py +19 -162
- cribl_control_plane/models/outputkinesis.py +21 -56
- cribl_control_plane/models/outputloki.py +19 -47
- cribl_control_plane/models/outputminio.py +0 -7
- cribl_control_plane/models/outputmsk.py +19 -54
- cribl_control_plane/models/outputnewrelic.py +21 -49
- cribl_control_plane/models/outputnewrelicevents.py +22 -50
- cribl_control_plane/models/outputopentelemetry.py +21 -49
- cribl_control_plane/models/outputprometheus.py +21 -49
- cribl_control_plane/models/outputs3.py +0 -7
- cribl_control_plane/models/outputsentinel.py +21 -49
- cribl_control_plane/models/outputsentineloneaisiem.py +22 -50
- cribl_control_plane/models/outputservicenow.py +21 -49
- cribl_control_plane/models/outputsignalfx.py +21 -49
- cribl_control_plane/models/outputsns.py +19 -47
- cribl_control_plane/models/outputsplunk.py +21 -49
- cribl_control_plane/models/outputsplunkhec.py +21 -49
- cribl_control_plane/models/outputsplunklb.py +21 -49
- cribl_control_plane/models/outputsqs.py +19 -47
- cribl_control_plane/models/outputstatsd.py +21 -49
- cribl_control_plane/models/outputstatsdext.py +21 -49
- cribl_control_plane/models/outputsumologic.py +21 -49
- cribl_control_plane/models/outputsyslog.py +99 -129
- cribl_control_plane/models/outputtcpjson.py +21 -49
- cribl_control_plane/models/outputwavefront.py +21 -49
- cribl_control_plane/models/outputwebhook.py +21 -49
- cribl_control_plane/models/outputxsiam.py +19 -47
- cribl_control_plane/models/runnablejobcollection.py +8 -12
- cribl_control_plane/models/runnablejobexecutor.py +8 -12
- cribl_control_plane/models/runnablejobscheduledsearch.py +8 -12
- {cribl_control_plane-0.2.1rc4.dist-info → cribl_control_plane-0.2.1rc6.dist-info}/METADATA +1 -1
- {cribl_control_plane-0.2.1rc4.dist-info → cribl_control_plane-0.2.1rc6.dist-info}/RECORD +98 -98
- {cribl_control_plane-0.2.1rc4.dist-info → cribl_control_plane-0.2.1rc6.dist-info}/WHEEL +0 -0
|
@@ -37,17 +37,6 @@ class OutputCloudwatchBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMe
|
|
|
37
37
|
QUEUE = "queue"
|
|
38
38
|
|
|
39
39
|
|
|
40
|
-
class OutputCloudwatchMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
41
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
42
|
-
|
|
43
|
-
# Error
|
|
44
|
-
ERROR = "error"
|
|
45
|
-
# Backpressure
|
|
46
|
-
ALWAYS = "always"
|
|
47
|
-
# Always On
|
|
48
|
-
BACKPRESSURE = "backpressure"
|
|
49
|
-
|
|
50
|
-
|
|
51
40
|
class OutputCloudwatchCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
52
41
|
r"""Codec to use to compress the persisted data"""
|
|
53
42
|
|
|
@@ -66,6 +55,17 @@ class OutputCloudwatchQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta)
|
|
|
66
55
|
DROP = "drop"
|
|
67
56
|
|
|
68
57
|
|
|
58
|
+
class OutputCloudwatchMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
59
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
60
|
+
|
|
61
|
+
# Error
|
|
62
|
+
ERROR = "error"
|
|
63
|
+
# Backpressure
|
|
64
|
+
BACKPRESSURE = "backpressure"
|
|
65
|
+
# Always On
|
|
66
|
+
ALWAYS = "always"
|
|
67
|
+
|
|
68
|
+
|
|
69
69
|
class OutputCloudwatchPqControlsTypedDict(TypedDict):
|
|
70
70
|
pass
|
|
71
71
|
|
|
@@ -121,16 +121,6 @@ class OutputCloudwatchTypedDict(TypedDict):
|
|
|
121
121
|
aws_api_key: NotRequired[str]
|
|
122
122
|
aws_secret: NotRequired[str]
|
|
123
123
|
r"""Select or create a stored secret that references your access key and secret key"""
|
|
124
|
-
pq_strict_ordering: NotRequired[bool]
|
|
125
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
126
|
-
pq_rate_per_sec: NotRequired[float]
|
|
127
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
128
|
-
pq_mode: NotRequired[OutputCloudwatchMode]
|
|
129
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
130
|
-
pq_max_buffer_size: NotRequired[float]
|
|
131
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
132
|
-
pq_max_backpressure_sec: NotRequired[float]
|
|
133
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
134
124
|
pq_max_file_size: NotRequired[str]
|
|
135
125
|
r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
|
|
136
126
|
pq_max_size: NotRequired[str]
|
|
@@ -141,6 +131,8 @@ class OutputCloudwatchTypedDict(TypedDict):
|
|
|
141
131
|
r"""Codec to use to compress the persisted data"""
|
|
142
132
|
pq_on_backpressure: NotRequired[OutputCloudwatchQueueFullBehavior]
|
|
143
133
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
134
|
+
pq_mode: NotRequired[OutputCloudwatchMode]
|
|
135
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
144
136
|
pq_controls: NotRequired[OutputCloudwatchPqControlsTypedDict]
|
|
145
137
|
|
|
146
138
|
|
|
@@ -248,34 +240,6 @@ class OutputCloudwatch(BaseModel):
|
|
|
248
240
|
aws_secret: Annotated[Optional[str], pydantic.Field(alias="awsSecret")] = None
|
|
249
241
|
r"""Select or create a stored secret that references your access key and secret key"""
|
|
250
242
|
|
|
251
|
-
pq_strict_ordering: Annotated[
|
|
252
|
-
Optional[bool], pydantic.Field(alias="pqStrictOrdering")
|
|
253
|
-
] = True
|
|
254
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
255
|
-
|
|
256
|
-
pq_rate_per_sec: Annotated[
|
|
257
|
-
Optional[float], pydantic.Field(alias="pqRatePerSec")
|
|
258
|
-
] = 0
|
|
259
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
260
|
-
|
|
261
|
-
pq_mode: Annotated[
|
|
262
|
-
Annotated[
|
|
263
|
-
Optional[OutputCloudwatchMode], PlainValidator(validate_open_enum(False))
|
|
264
|
-
],
|
|
265
|
-
pydantic.Field(alias="pqMode"),
|
|
266
|
-
] = OutputCloudwatchMode.ERROR
|
|
267
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
268
|
-
|
|
269
|
-
pq_max_buffer_size: Annotated[
|
|
270
|
-
Optional[float], pydantic.Field(alias="pqMaxBufferSize")
|
|
271
|
-
] = 42
|
|
272
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
273
|
-
|
|
274
|
-
pq_max_backpressure_sec: Annotated[
|
|
275
|
-
Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
|
|
276
|
-
] = 30
|
|
277
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
278
|
-
|
|
279
243
|
pq_max_file_size: Annotated[
|
|
280
244
|
Optional[str], pydantic.Field(alias="pqMaxFileSize")
|
|
281
245
|
] = "1 MB"
|
|
@@ -307,6 +271,14 @@ class OutputCloudwatch(BaseModel):
|
|
|
307
271
|
] = OutputCloudwatchQueueFullBehavior.BLOCK
|
|
308
272
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
309
273
|
|
|
274
|
+
pq_mode: Annotated[
|
|
275
|
+
Annotated[
|
|
276
|
+
Optional[OutputCloudwatchMode], PlainValidator(validate_open_enum(False))
|
|
277
|
+
],
|
|
278
|
+
pydantic.Field(alias="pqMode"),
|
|
279
|
+
] = OutputCloudwatchMode.ERROR
|
|
280
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
281
|
+
|
|
310
282
|
pq_controls: Annotated[
|
|
311
283
|
Optional[OutputCloudwatchPqControls], pydantic.Field(alias="pqControls")
|
|
312
284
|
] = None
|
|
@@ -297,13 +297,6 @@ class OutputConfluentCloudKafkaSchemaRegistryAuthentication(BaseModel):
|
|
|
297
297
|
r"""Used when __valueSchemaIdOut is not present, to transform _raw, leave blank if value transformation is not required by default."""
|
|
298
298
|
|
|
299
299
|
|
|
300
|
-
class OutputConfluentCloudAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
301
|
-
r"""Enter credentials directly, or select a stored secret"""
|
|
302
|
-
|
|
303
|
-
MANUAL = "manual"
|
|
304
|
-
SECRET = "secret"
|
|
305
|
-
|
|
306
|
-
|
|
307
300
|
class OutputConfluentCloudSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
308
301
|
# PLAIN
|
|
309
302
|
PLAIN = "plain"
|
|
@@ -315,58 +308,13 @@ class OutputConfluentCloudSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta)
|
|
|
315
308
|
KERBEROS = "kerberos"
|
|
316
309
|
|
|
317
310
|
|
|
318
|
-
class OutputConfluentCloudOauthParamTypedDict(TypedDict):
|
|
319
|
-
name: str
|
|
320
|
-
value: str
|
|
321
|
-
|
|
322
|
-
|
|
323
|
-
class OutputConfluentCloudOauthParam(BaseModel):
|
|
324
|
-
name: str
|
|
325
|
-
|
|
326
|
-
value: str
|
|
327
|
-
|
|
328
|
-
|
|
329
|
-
class OutputConfluentCloudSaslExtensionTypedDict(TypedDict):
|
|
330
|
-
name: str
|
|
331
|
-
value: str
|
|
332
|
-
|
|
333
|
-
|
|
334
|
-
class OutputConfluentCloudSaslExtension(BaseModel):
|
|
335
|
-
name: str
|
|
336
|
-
|
|
337
|
-
value: str
|
|
338
|
-
|
|
339
|
-
|
|
340
311
|
class OutputConfluentCloudAuthenticationTypedDict(TypedDict):
|
|
341
312
|
r"""Authentication parameters to use when connecting to brokers. Using TLS is highly recommended."""
|
|
342
313
|
|
|
343
314
|
disabled: NotRequired[bool]
|
|
344
|
-
username: NotRequired[str]
|
|
345
|
-
password: NotRequired[str]
|
|
346
|
-
auth_type: NotRequired[OutputConfluentCloudAuthenticationMethod]
|
|
347
|
-
r"""Enter credentials directly, or select a stored secret"""
|
|
348
|
-
credentials_secret: NotRequired[str]
|
|
349
|
-
r"""Select or create a secret that references your credentials"""
|
|
350
315
|
mechanism: NotRequired[OutputConfluentCloudSASLMechanism]
|
|
351
|
-
keytab_location: NotRequired[str]
|
|
352
|
-
r"""Location of keytab file for authentication principal"""
|
|
353
|
-
principal: NotRequired[str]
|
|
354
|
-
r"""Authentication principal, such as `kafka_user@example.com`"""
|
|
355
|
-
broker_service_class: NotRequired[str]
|
|
356
|
-
r"""Kerberos service class for Kafka brokers, such as `kafka`"""
|
|
357
316
|
oauth_enabled: NotRequired[bool]
|
|
358
317
|
r"""Enable OAuth authentication"""
|
|
359
|
-
token_url: NotRequired[str]
|
|
360
|
-
r"""URL of the token endpoint to use for OAuth authentication"""
|
|
361
|
-
client_id: NotRequired[str]
|
|
362
|
-
r"""Client ID to use for OAuth authentication"""
|
|
363
|
-
oauth_secret_type: NotRequired[str]
|
|
364
|
-
client_text_secret: NotRequired[str]
|
|
365
|
-
r"""Select or create a stored text secret"""
|
|
366
|
-
oauth_params: NotRequired[List[OutputConfluentCloudOauthParamTypedDict]]
|
|
367
|
-
r"""Additional fields to send to the token endpoint, such as scope or audience"""
|
|
368
|
-
sasl_extensions: NotRequired[List[OutputConfluentCloudSaslExtensionTypedDict]]
|
|
369
|
-
r"""Additional SASL extension fields, such as Confluent's logicalCluster or identityPoolId"""
|
|
370
318
|
|
|
371
319
|
|
|
372
320
|
class OutputConfluentCloudAuthentication(BaseModel):
|
|
@@ -374,74 +322,16 @@ class OutputConfluentCloudAuthentication(BaseModel):
|
|
|
374
322
|
|
|
375
323
|
disabled: Optional[bool] = True
|
|
376
324
|
|
|
377
|
-
username: Optional[str] = None
|
|
378
|
-
|
|
379
|
-
password: Optional[str] = None
|
|
380
|
-
|
|
381
|
-
auth_type: Annotated[
|
|
382
|
-
Annotated[
|
|
383
|
-
Optional[OutputConfluentCloudAuthenticationMethod],
|
|
384
|
-
PlainValidator(validate_open_enum(False)),
|
|
385
|
-
],
|
|
386
|
-
pydantic.Field(alias="authType"),
|
|
387
|
-
] = OutputConfluentCloudAuthenticationMethod.MANUAL
|
|
388
|
-
r"""Enter credentials directly, or select a stored secret"""
|
|
389
|
-
|
|
390
|
-
credentials_secret: Annotated[
|
|
391
|
-
Optional[str], pydantic.Field(alias="credentialsSecret")
|
|
392
|
-
] = None
|
|
393
|
-
r"""Select or create a secret that references your credentials"""
|
|
394
|
-
|
|
395
325
|
mechanism: Annotated[
|
|
396
326
|
Optional[OutputConfluentCloudSASLMechanism],
|
|
397
327
|
PlainValidator(validate_open_enum(False)),
|
|
398
328
|
] = OutputConfluentCloudSASLMechanism.PLAIN
|
|
399
329
|
|
|
400
|
-
keytab_location: Annotated[
|
|
401
|
-
Optional[str], pydantic.Field(alias="keytabLocation")
|
|
402
|
-
] = None
|
|
403
|
-
r"""Location of keytab file for authentication principal"""
|
|
404
|
-
|
|
405
|
-
principal: Optional[str] = None
|
|
406
|
-
r"""Authentication principal, such as `kafka_user@example.com`"""
|
|
407
|
-
|
|
408
|
-
broker_service_class: Annotated[
|
|
409
|
-
Optional[str], pydantic.Field(alias="brokerServiceClass")
|
|
410
|
-
] = None
|
|
411
|
-
r"""Kerberos service class for Kafka brokers, such as `kafka`"""
|
|
412
|
-
|
|
413
330
|
oauth_enabled: Annotated[Optional[bool], pydantic.Field(alias="oauthEnabled")] = (
|
|
414
331
|
False
|
|
415
332
|
)
|
|
416
333
|
r"""Enable OAuth authentication"""
|
|
417
334
|
|
|
418
|
-
token_url: Annotated[Optional[str], pydantic.Field(alias="tokenUrl")] = None
|
|
419
|
-
r"""URL of the token endpoint to use for OAuth authentication"""
|
|
420
|
-
|
|
421
|
-
client_id: Annotated[Optional[str], pydantic.Field(alias="clientId")] = None
|
|
422
|
-
r"""Client ID to use for OAuth authentication"""
|
|
423
|
-
|
|
424
|
-
oauth_secret_type: Annotated[
|
|
425
|
-
Optional[str], pydantic.Field(alias="oauthSecretType")
|
|
426
|
-
] = "secret"
|
|
427
|
-
|
|
428
|
-
client_text_secret: Annotated[
|
|
429
|
-
Optional[str], pydantic.Field(alias="clientTextSecret")
|
|
430
|
-
] = None
|
|
431
|
-
r"""Select or create a stored text secret"""
|
|
432
|
-
|
|
433
|
-
oauth_params: Annotated[
|
|
434
|
-
Optional[List[OutputConfluentCloudOauthParam]],
|
|
435
|
-
pydantic.Field(alias="oauthParams"),
|
|
436
|
-
] = None
|
|
437
|
-
r"""Additional fields to send to the token endpoint, such as scope or audience"""
|
|
438
|
-
|
|
439
|
-
sasl_extensions: Annotated[
|
|
440
|
-
Optional[List[OutputConfluentCloudSaslExtension]],
|
|
441
|
-
pydantic.Field(alias="saslExtensions"),
|
|
442
|
-
] = None
|
|
443
|
-
r"""Additional SASL extension fields, such as Confluent's logicalCluster or identityPoolId"""
|
|
444
|
-
|
|
445
335
|
|
|
446
336
|
class OutputConfluentCloudBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
447
337
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
@@ -454,17 +344,6 @@ class OutputConfluentCloudBackpressureBehavior(str, Enum, metaclass=utils.OpenEn
|
|
|
454
344
|
QUEUE = "queue"
|
|
455
345
|
|
|
456
346
|
|
|
457
|
-
class OutputConfluentCloudMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
458
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
459
|
-
|
|
460
|
-
# Error
|
|
461
|
-
ERROR = "error"
|
|
462
|
-
# Backpressure
|
|
463
|
-
ALWAYS = "always"
|
|
464
|
-
# Always On
|
|
465
|
-
BACKPRESSURE = "backpressure"
|
|
466
|
-
|
|
467
|
-
|
|
468
347
|
class OutputConfluentCloudPqCompressCompression(
|
|
469
348
|
str, Enum, metaclass=utils.OpenEnumMeta
|
|
470
349
|
):
|
|
@@ -485,6 +364,17 @@ class OutputConfluentCloudQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumM
|
|
|
485
364
|
DROP = "drop"
|
|
486
365
|
|
|
487
366
|
|
|
367
|
+
class OutputConfluentCloudMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
368
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
369
|
+
|
|
370
|
+
# Error
|
|
371
|
+
ERROR = "error"
|
|
372
|
+
# Backpressure
|
|
373
|
+
BACKPRESSURE = "backpressure"
|
|
374
|
+
# Always On
|
|
375
|
+
ALWAYS = "always"
|
|
376
|
+
|
|
377
|
+
|
|
488
378
|
class OutputConfluentCloudPqControlsTypedDict(TypedDict):
|
|
489
379
|
pass
|
|
490
380
|
|
|
@@ -548,18 +438,6 @@ class OutputConfluentCloudTypedDict(TypedDict):
|
|
|
548
438
|
description: NotRequired[str]
|
|
549
439
|
protobuf_library_id: NotRequired[str]
|
|
550
440
|
r"""Select a set of Protobuf definitions for the events you want to send"""
|
|
551
|
-
protobuf_encoding_id: NotRequired[str]
|
|
552
|
-
r"""Select the type of object you want the Protobuf definitions to use for event encoding"""
|
|
553
|
-
pq_strict_ordering: NotRequired[bool]
|
|
554
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
555
|
-
pq_rate_per_sec: NotRequired[float]
|
|
556
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
557
|
-
pq_mode: NotRequired[OutputConfluentCloudMode]
|
|
558
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
559
|
-
pq_max_buffer_size: NotRequired[float]
|
|
560
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
561
|
-
pq_max_backpressure_sec: NotRequired[float]
|
|
562
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
563
441
|
pq_max_file_size: NotRequired[str]
|
|
564
442
|
r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
|
|
565
443
|
pq_max_size: NotRequired[str]
|
|
@@ -570,6 +448,8 @@ class OutputConfluentCloudTypedDict(TypedDict):
|
|
|
570
448
|
r"""Codec to use to compress the persisted data"""
|
|
571
449
|
pq_on_backpressure: NotRequired[OutputConfluentCloudQueueFullBehavior]
|
|
572
450
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
451
|
+
pq_mode: NotRequired[OutputConfluentCloudMode]
|
|
452
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
573
453
|
pq_controls: NotRequired[OutputConfluentCloudPqControlsTypedDict]
|
|
574
454
|
|
|
575
455
|
|
|
@@ -695,40 +575,6 @@ class OutputConfluentCloud(BaseModel):
|
|
|
695
575
|
] = None
|
|
696
576
|
r"""Select a set of Protobuf definitions for the events you want to send"""
|
|
697
577
|
|
|
698
|
-
protobuf_encoding_id: Annotated[
|
|
699
|
-
Optional[str], pydantic.Field(alias="protobufEncodingId")
|
|
700
|
-
] = None
|
|
701
|
-
r"""Select the type of object you want the Protobuf definitions to use for event encoding"""
|
|
702
|
-
|
|
703
|
-
pq_strict_ordering: Annotated[
|
|
704
|
-
Optional[bool], pydantic.Field(alias="pqStrictOrdering")
|
|
705
|
-
] = True
|
|
706
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
707
|
-
|
|
708
|
-
pq_rate_per_sec: Annotated[
|
|
709
|
-
Optional[float], pydantic.Field(alias="pqRatePerSec")
|
|
710
|
-
] = 0
|
|
711
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
712
|
-
|
|
713
|
-
pq_mode: Annotated[
|
|
714
|
-
Annotated[
|
|
715
|
-
Optional[OutputConfluentCloudMode],
|
|
716
|
-
PlainValidator(validate_open_enum(False)),
|
|
717
|
-
],
|
|
718
|
-
pydantic.Field(alias="pqMode"),
|
|
719
|
-
] = OutputConfluentCloudMode.ERROR
|
|
720
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
721
|
-
|
|
722
|
-
pq_max_buffer_size: Annotated[
|
|
723
|
-
Optional[float], pydantic.Field(alias="pqMaxBufferSize")
|
|
724
|
-
] = 42
|
|
725
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
726
|
-
|
|
727
|
-
pq_max_backpressure_sec: Annotated[
|
|
728
|
-
Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
|
|
729
|
-
] = 30
|
|
730
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
731
|
-
|
|
732
578
|
pq_max_file_size: Annotated[
|
|
733
579
|
Optional[str], pydantic.Field(alias="pqMaxFileSize")
|
|
734
580
|
] = "1 MB"
|
|
@@ -760,6 +606,15 @@ class OutputConfluentCloud(BaseModel):
|
|
|
760
606
|
] = OutputConfluentCloudQueueFullBehavior.BLOCK
|
|
761
607
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
762
608
|
|
|
609
|
+
pq_mode: Annotated[
|
|
610
|
+
Annotated[
|
|
611
|
+
Optional[OutputConfluentCloudMode],
|
|
612
|
+
PlainValidator(validate_open_enum(False)),
|
|
613
|
+
],
|
|
614
|
+
pydantic.Field(alias="pqMode"),
|
|
615
|
+
] = OutputConfluentCloudMode.ERROR
|
|
616
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
617
|
+
|
|
763
618
|
pq_controls: Annotated[
|
|
764
619
|
Optional[OutputConfluentCloudPqControls], pydantic.Field(alias="pqControls")
|
|
765
620
|
] = None
|
|
@@ -209,17 +209,6 @@ class OutputCriblHTTPURL(BaseModel):
|
|
|
209
209
|
r"""Assign a weight (>0) to each endpoint to indicate its traffic-handling capability"""
|
|
210
210
|
|
|
211
211
|
|
|
212
|
-
class OutputCriblHTTPMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
213
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
214
|
-
|
|
215
|
-
# Error
|
|
216
|
-
ERROR = "error"
|
|
217
|
-
# Backpressure
|
|
218
|
-
ALWAYS = "always"
|
|
219
|
-
# Always On
|
|
220
|
-
BACKPRESSURE = "backpressure"
|
|
221
|
-
|
|
222
|
-
|
|
223
212
|
class OutputCriblHTTPPqCompressCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
224
213
|
r"""Codec to use to compress the persisted data"""
|
|
225
214
|
|
|
@@ -238,6 +227,17 @@ class OutputCriblHTTPQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
238
227
|
DROP = "drop"
|
|
239
228
|
|
|
240
229
|
|
|
230
|
+
class OutputCriblHTTPMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
231
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
232
|
+
|
|
233
|
+
# Error
|
|
234
|
+
ERROR = "error"
|
|
235
|
+
# Backpressure
|
|
236
|
+
BACKPRESSURE = "backpressure"
|
|
237
|
+
# Always On
|
|
238
|
+
ALWAYS = "always"
|
|
239
|
+
|
|
240
|
+
|
|
241
241
|
class OutputCriblHTTPPqControlsTypedDict(TypedDict):
|
|
242
242
|
pass
|
|
243
243
|
|
|
@@ -309,16 +309,6 @@ class OutputCriblHTTPTypedDict(TypedDict):
|
|
|
309
309
|
r"""The interval in which to re-resolve any hostnames and pick up destinations from A records"""
|
|
310
310
|
load_balance_stats_period_sec: NotRequired[float]
|
|
311
311
|
r"""How far back in time to keep traffic stats for load balancing purposes"""
|
|
312
|
-
pq_strict_ordering: NotRequired[bool]
|
|
313
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
314
|
-
pq_rate_per_sec: NotRequired[float]
|
|
315
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
316
|
-
pq_mode: NotRequired[OutputCriblHTTPMode]
|
|
317
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
318
|
-
pq_max_buffer_size: NotRequired[float]
|
|
319
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
320
|
-
pq_max_backpressure_sec: NotRequired[float]
|
|
321
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
322
312
|
pq_max_file_size: NotRequired[str]
|
|
323
313
|
r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
|
|
324
314
|
pq_max_size: NotRequired[str]
|
|
@@ -329,6 +319,8 @@ class OutputCriblHTTPTypedDict(TypedDict):
|
|
|
329
319
|
r"""Codec to use to compress the persisted data"""
|
|
330
320
|
pq_on_backpressure: NotRequired[OutputCriblHTTPQueueFullBehavior]
|
|
331
321
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
322
|
+
pq_mode: NotRequired[OutputCriblHTTPMode]
|
|
323
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
332
324
|
pq_controls: NotRequired[OutputCriblHTTPPqControlsTypedDict]
|
|
333
325
|
|
|
334
326
|
|
|
@@ -473,34 +465,6 @@ class OutputCriblHTTP(BaseModel):
|
|
|
473
465
|
] = 300
|
|
474
466
|
r"""How far back in time to keep traffic stats for load balancing purposes"""
|
|
475
467
|
|
|
476
|
-
pq_strict_ordering: Annotated[
|
|
477
|
-
Optional[bool], pydantic.Field(alias="pqStrictOrdering")
|
|
478
|
-
] = True
|
|
479
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
480
|
-
|
|
481
|
-
pq_rate_per_sec: Annotated[
|
|
482
|
-
Optional[float], pydantic.Field(alias="pqRatePerSec")
|
|
483
|
-
] = 0
|
|
484
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
485
|
-
|
|
486
|
-
pq_mode: Annotated[
|
|
487
|
-
Annotated[
|
|
488
|
-
Optional[OutputCriblHTTPMode], PlainValidator(validate_open_enum(False))
|
|
489
|
-
],
|
|
490
|
-
pydantic.Field(alias="pqMode"),
|
|
491
|
-
] = OutputCriblHTTPMode.ERROR
|
|
492
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
493
|
-
|
|
494
|
-
pq_max_buffer_size: Annotated[
|
|
495
|
-
Optional[float], pydantic.Field(alias="pqMaxBufferSize")
|
|
496
|
-
] = 42
|
|
497
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
498
|
-
|
|
499
|
-
pq_max_backpressure_sec: Annotated[
|
|
500
|
-
Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
|
|
501
|
-
] = 30
|
|
502
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
503
|
-
|
|
504
468
|
pq_max_file_size: Annotated[
|
|
505
469
|
Optional[str], pydantic.Field(alias="pqMaxFileSize")
|
|
506
470
|
] = "1 MB"
|
|
@@ -532,6 +496,14 @@ class OutputCriblHTTP(BaseModel):
|
|
|
532
496
|
] = OutputCriblHTTPQueueFullBehavior.BLOCK
|
|
533
497
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
534
498
|
|
|
499
|
+
pq_mode: Annotated[
|
|
500
|
+
Annotated[
|
|
501
|
+
Optional[OutputCriblHTTPMode], PlainValidator(validate_open_enum(False))
|
|
502
|
+
],
|
|
503
|
+
pydantic.Field(alias="pqMode"),
|
|
504
|
+
] = OutputCriblHTTPMode.ERROR
|
|
505
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
506
|
+
|
|
535
507
|
pq_controls: Annotated[
|
|
536
508
|
Optional[OutputCriblHTTPPqControls], pydantic.Field(alias="pqControls")
|
|
537
509
|
] = None
|
|
@@ -157,17 +157,6 @@ class OutputCriblTCPHost(BaseModel):
|
|
|
157
157
|
r"""Assign a weight (>0) to each endpoint to indicate its traffic-handling capability"""
|
|
158
158
|
|
|
159
159
|
|
|
160
|
-
class OutputCriblTCPMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
161
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
162
|
-
|
|
163
|
-
# Error
|
|
164
|
-
ERROR = "error"
|
|
165
|
-
# Backpressure
|
|
166
|
-
ALWAYS = "always"
|
|
167
|
-
# Always On
|
|
168
|
-
BACKPRESSURE = "backpressure"
|
|
169
|
-
|
|
170
|
-
|
|
171
160
|
class OutputCriblTCPPqCompressCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
172
161
|
r"""Codec to use to compress the persisted data"""
|
|
173
162
|
|
|
@@ -186,6 +175,17 @@ class OutputCriblTCPQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
|
186
175
|
DROP = "drop"
|
|
187
176
|
|
|
188
177
|
|
|
178
|
+
class OutputCriblTCPMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
179
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
180
|
+
|
|
181
|
+
# Error
|
|
182
|
+
ERROR = "error"
|
|
183
|
+
# Backpressure
|
|
184
|
+
BACKPRESSURE = "backpressure"
|
|
185
|
+
# Always On
|
|
186
|
+
ALWAYS = "always"
|
|
187
|
+
|
|
188
|
+
|
|
189
189
|
class OutputCriblTCPPqControlsTypedDict(TypedDict):
|
|
190
190
|
pass
|
|
191
191
|
|
|
@@ -240,16 +240,6 @@ class OutputCriblTCPTypedDict(TypedDict):
|
|
|
240
240
|
r"""How far back in time to keep traffic stats for load balancing purposes"""
|
|
241
241
|
max_concurrent_senders: NotRequired[float]
|
|
242
242
|
r"""Maximum number of concurrent connections (per Worker Process). A random set of IPs will be picked on every DNS resolution period. Use 0 for unlimited."""
|
|
243
|
-
pq_strict_ordering: NotRequired[bool]
|
|
244
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
245
|
-
pq_rate_per_sec: NotRequired[float]
|
|
246
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
247
|
-
pq_mode: NotRequired[OutputCriblTCPMode]
|
|
248
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
249
|
-
pq_max_buffer_size: NotRequired[float]
|
|
250
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
251
|
-
pq_max_backpressure_sec: NotRequired[float]
|
|
252
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
253
243
|
pq_max_file_size: NotRequired[str]
|
|
254
244
|
r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
|
|
255
245
|
pq_max_size: NotRequired[str]
|
|
@@ -260,6 +250,8 @@ class OutputCriblTCPTypedDict(TypedDict):
|
|
|
260
250
|
r"""Codec to use to compress the persisted data"""
|
|
261
251
|
pq_on_backpressure: NotRequired[OutputCriblTCPQueueFullBehavior]
|
|
262
252
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
253
|
+
pq_mode: NotRequired[OutputCriblTCPMode]
|
|
254
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
263
255
|
pq_controls: NotRequired[OutputCriblTCPPqControlsTypedDict]
|
|
264
256
|
|
|
265
257
|
|
|
@@ -363,34 +355,6 @@ class OutputCriblTCP(BaseModel):
|
|
|
363
355
|
] = 0
|
|
364
356
|
r"""Maximum number of concurrent connections (per Worker Process). A random set of IPs will be picked on every DNS resolution period. Use 0 for unlimited."""
|
|
365
357
|
|
|
366
|
-
pq_strict_ordering: Annotated[
|
|
367
|
-
Optional[bool], pydantic.Field(alias="pqStrictOrdering")
|
|
368
|
-
] = True
|
|
369
|
-
r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
|
|
370
|
-
|
|
371
|
-
pq_rate_per_sec: Annotated[
|
|
372
|
-
Optional[float], pydantic.Field(alias="pqRatePerSec")
|
|
373
|
-
] = 0
|
|
374
|
-
r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
|
|
375
|
-
|
|
376
|
-
pq_mode: Annotated[
|
|
377
|
-
Annotated[
|
|
378
|
-
Optional[OutputCriblTCPMode], PlainValidator(validate_open_enum(False))
|
|
379
|
-
],
|
|
380
|
-
pydantic.Field(alias="pqMode"),
|
|
381
|
-
] = OutputCriblTCPMode.ERROR
|
|
382
|
-
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
383
|
-
|
|
384
|
-
pq_max_buffer_size: Annotated[
|
|
385
|
-
Optional[float], pydantic.Field(alias="pqMaxBufferSize")
|
|
386
|
-
] = 42
|
|
387
|
-
r"""The maximum number of events to hold in memory before writing the events to disk"""
|
|
388
|
-
|
|
389
|
-
pq_max_backpressure_sec: Annotated[
|
|
390
|
-
Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
|
|
391
|
-
] = 30
|
|
392
|
-
r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
|
|
393
|
-
|
|
394
358
|
pq_max_file_size: Annotated[
|
|
395
359
|
Optional[str], pydantic.Field(alias="pqMaxFileSize")
|
|
396
360
|
] = "1 MB"
|
|
@@ -422,6 +386,14 @@ class OutputCriblTCP(BaseModel):
|
|
|
422
386
|
] = OutputCriblTCPQueueFullBehavior.BLOCK
|
|
423
387
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
424
388
|
|
|
389
|
+
pq_mode: Annotated[
|
|
390
|
+
Annotated[
|
|
391
|
+
Optional[OutputCriblTCPMode], PlainValidator(validate_open_enum(False))
|
|
392
|
+
],
|
|
393
|
+
pydantic.Field(alias="pqMode"),
|
|
394
|
+
] = OutputCriblTCPMode.ERROR
|
|
395
|
+
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
396
|
+
|
|
425
397
|
pq_controls: Annotated[
|
|
426
398
|
Optional[OutputCriblTCPPqControls], pydantic.Field(alias="pqControls")
|
|
427
399
|
] = None
|