cribl-control-plane 0.2.1rc4__py3-none-any.whl → 0.2.1rc5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (110) hide show
  1. cribl_control_plane/_version.py +3 -3
  2. cribl_control_plane/groups_sdk.py +3 -0
  3. cribl_control_plane/mappings.py +1185 -0
  4. cribl_control_plane/models/__init__.py +149 -105
  5. cribl_control_plane/models/createadminproductsmappingsactivatebyproductop.py +52 -0
  6. cribl_control_plane/models/createadminproductsmappingsbyproductop.py +53 -0
  7. cribl_control_plane/models/deleteadminproductsmappingsbyproductandidop.py +51 -0
  8. cribl_control_plane/models/getadminproductsmappingsbyproductandidop.py +51 -0
  9. cribl_control_plane/models/getadminproductsmappingsbyproductop.py +44 -0
  10. cribl_control_plane/models/input.py +14 -14
  11. cribl_control_plane/models/inputappscope.py +16 -20
  12. cribl_control_plane/models/inputconfluentcloud.py +0 -110
  13. cribl_control_plane/models/inputcriblhttp.py +16 -20
  14. cribl_control_plane/models/inputcribllakehttp.py +16 -20
  15. cribl_control_plane/models/inputcribltcp.py +16 -20
  16. cribl_control_plane/models/inputdatadogagent.py +16 -20
  17. cribl_control_plane/models/inputedgeprometheus.py +36 -44
  18. cribl_control_plane/models/inputelastic.py +27 -44
  19. cribl_control_plane/models/inputeventhub.py +0 -118
  20. cribl_control_plane/models/inputfirehose.py +16 -20
  21. cribl_control_plane/models/inputgrafana.py +31 -39
  22. cribl_control_plane/models/inputhttp.py +16 -20
  23. cribl_control_plane/models/inputhttpraw.py +16 -20
  24. cribl_control_plane/models/inputkafka.py +0 -108
  25. cribl_control_plane/models/inputloki.py +16 -20
  26. cribl_control_plane/models/inputmetrics.py +16 -20
  27. cribl_control_plane/models/inputmodeldriventelemetry.py +16 -20
  28. cribl_control_plane/models/inputopentelemetry.py +15 -19
  29. cribl_control_plane/models/inputprometheus.py +36 -44
  30. cribl_control_plane/models/inputprometheusrw.py +16 -20
  31. cribl_control_plane/models/inputsplunk.py +16 -20
  32. cribl_control_plane/models/inputsplunkhec.py +15 -19
  33. cribl_control_plane/models/inputsyslog.py +31 -39
  34. cribl_control_plane/models/inputsystemmetrics.py +10 -20
  35. cribl_control_plane/models/inputtcp.py +16 -30
  36. cribl_control_plane/models/inputtcpjson.py +16 -20
  37. cribl_control_plane/models/inputwindowsmetrics.py +10 -20
  38. cribl_control_plane/models/inputwineventlogs.py +0 -14
  39. cribl_control_plane/models/inputwizwebhook.py +16 -20
  40. cribl_control_plane/models/inputzscalerhec.py +15 -19
  41. cribl_control_plane/models/mappingruleset.py +53 -0
  42. cribl_control_plane/models/mappingrulesetevalmappingfunction.py +71 -0
  43. cribl_control_plane/models/mappingrulesetgenericmappingfunction.py +29 -0
  44. cribl_control_plane/models/output.py +22 -22
  45. cribl_control_plane/models/outputazureblob.py +0 -7
  46. cribl_control_plane/models/outputazuredataexplorer.py +93 -283
  47. cribl_control_plane/models/outputazureeventhub.py +21 -169
  48. cribl_control_plane/models/outputazurelogs.py +21 -49
  49. cribl_control_plane/models/outputchronicle.py +21 -49
  50. cribl_control_plane/models/outputclickhouse.py +21 -49
  51. cribl_control_plane/models/outputcloudwatch.py +21 -49
  52. cribl_control_plane/models/outputconfluentcloud.py +22 -167
  53. cribl_control_plane/models/outputcriblhttp.py +21 -49
  54. cribl_control_plane/models/outputcribltcp.py +21 -49
  55. cribl_control_plane/models/outputcrowdstrikenextgensiem.py +22 -50
  56. cribl_control_plane/models/outputdatabricks.py +0 -7
  57. cribl_control_plane/models/outputdatadog.py +21 -49
  58. cribl_control_plane/models/outputdataset.py +21 -49
  59. cribl_control_plane/models/outputdls3.py +0 -7
  60. cribl_control_plane/models/outputdynatracehttp.py +21 -49
  61. cribl_control_plane/models/outputdynatraceotlp.py +21 -49
  62. cribl_control_plane/models/outputelastic.py +21 -74
  63. cribl_control_plane/models/outputelasticcloud.py +21 -74
  64. cribl_control_plane/models/outputfilesystem.py +0 -7
  65. cribl_control_plane/models/outputgooglechronicle.py +22 -65
  66. cribl_control_plane/models/outputgooglecloudlogging.py +22 -50
  67. cribl_control_plane/models/outputgooglecloudstorage.py +0 -7
  68. cribl_control_plane/models/outputgooglepubsub.py +21 -49
  69. cribl_control_plane/models/outputgrafanacloud.py +42 -98
  70. cribl_control_plane/models/outputgraphite.py +21 -49
  71. cribl_control_plane/models/outputhoneycomb.py +21 -49
  72. cribl_control_plane/models/outputhumiohec.py +21 -49
  73. cribl_control_plane/models/outputinfluxdb.py +21 -49
  74. cribl_control_plane/models/outputkafka.py +19 -162
  75. cribl_control_plane/models/outputkinesis.py +21 -56
  76. cribl_control_plane/models/outputloki.py +19 -47
  77. cribl_control_plane/models/outputminio.py +0 -7
  78. cribl_control_plane/models/outputmsk.py +19 -54
  79. cribl_control_plane/models/outputnewrelic.py +21 -49
  80. cribl_control_plane/models/outputnewrelicevents.py +22 -50
  81. cribl_control_plane/models/outputopentelemetry.py +21 -49
  82. cribl_control_plane/models/outputprometheus.py +21 -49
  83. cribl_control_plane/models/outputs3.py +0 -7
  84. cribl_control_plane/models/outputsentinel.py +21 -49
  85. cribl_control_plane/models/outputsentineloneaisiem.py +22 -50
  86. cribl_control_plane/models/outputservicenow.py +21 -49
  87. cribl_control_plane/models/outputsignalfx.py +21 -49
  88. cribl_control_plane/models/outputsns.py +19 -47
  89. cribl_control_plane/models/outputsplunk.py +21 -49
  90. cribl_control_plane/models/outputsplunkhec.py +21 -49
  91. cribl_control_plane/models/outputsplunklb.py +21 -49
  92. cribl_control_plane/models/outputsqs.py +19 -47
  93. cribl_control_plane/models/outputstatsd.py +21 -49
  94. cribl_control_plane/models/outputstatsdext.py +21 -49
  95. cribl_control_plane/models/outputsumologic.py +21 -49
  96. cribl_control_plane/models/outputsyslog.py +99 -129
  97. cribl_control_plane/models/outputtcpjson.py +21 -49
  98. cribl_control_plane/models/outputwavefront.py +21 -49
  99. cribl_control_plane/models/outputwebhook.py +21 -49
  100. cribl_control_plane/models/outputxsiam.py +19 -47
  101. cribl_control_plane/models/pipeline.py +4 -4
  102. cribl_control_plane/models/rulesetid.py +13 -0
  103. cribl_control_plane/models/runnablejobcollection.py +8 -12
  104. cribl_control_plane/models/runnablejobexecutor.py +8 -12
  105. cribl_control_plane/models/runnablejobscheduledsearch.py +8 -12
  106. cribl_control_plane/models/updateadminproductsmappingsbyproductandidop.py +63 -0
  107. cribl_control_plane/pipelines.py +8 -8
  108. {cribl_control_plane-0.2.1rc4.dist-info → cribl_control_plane-0.2.1rc5.dist-info}/METADATA +11 -2
  109. {cribl_control_plane-0.2.1rc4.dist-info → cribl_control_plane-0.2.1rc5.dist-info}/RECORD +110 -99
  110. {cribl_control_plane-0.2.1rc4.dist-info → cribl_control_plane-0.2.1rc5.dist-info}/WHEEL +0 -0
@@ -37,17 +37,6 @@ class OutputCloudwatchBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMe
37
37
  QUEUE = "queue"
38
38
 
39
39
 
40
- class OutputCloudwatchMode(str, Enum, metaclass=utils.OpenEnumMeta):
41
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
42
-
43
- # Error
44
- ERROR = "error"
45
- # Backpressure
46
- ALWAYS = "always"
47
- # Always On
48
- BACKPRESSURE = "backpressure"
49
-
50
-
51
40
  class OutputCloudwatchCompression(str, Enum, metaclass=utils.OpenEnumMeta):
52
41
  r"""Codec to use to compress the persisted data"""
53
42
 
@@ -66,6 +55,17 @@ class OutputCloudwatchQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta)
66
55
  DROP = "drop"
67
56
 
68
57
 
58
+ class OutputCloudwatchMode(str, Enum, metaclass=utils.OpenEnumMeta):
59
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
60
+
61
+ # Error
62
+ ERROR = "error"
63
+ # Backpressure
64
+ BACKPRESSURE = "backpressure"
65
+ # Always On
66
+ ALWAYS = "always"
67
+
68
+
69
69
  class OutputCloudwatchPqControlsTypedDict(TypedDict):
70
70
  pass
71
71
 
@@ -121,16 +121,6 @@ class OutputCloudwatchTypedDict(TypedDict):
121
121
  aws_api_key: NotRequired[str]
122
122
  aws_secret: NotRequired[str]
123
123
  r"""Select or create a stored secret that references your access key and secret key"""
124
- pq_strict_ordering: NotRequired[bool]
125
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
126
- pq_rate_per_sec: NotRequired[float]
127
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
128
- pq_mode: NotRequired[OutputCloudwatchMode]
129
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
130
- pq_max_buffer_size: NotRequired[float]
131
- r"""The maximum number of events to hold in memory before writing the events to disk"""
132
- pq_max_backpressure_sec: NotRequired[float]
133
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
134
124
  pq_max_file_size: NotRequired[str]
135
125
  r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
136
126
  pq_max_size: NotRequired[str]
@@ -141,6 +131,8 @@ class OutputCloudwatchTypedDict(TypedDict):
141
131
  r"""Codec to use to compress the persisted data"""
142
132
  pq_on_backpressure: NotRequired[OutputCloudwatchQueueFullBehavior]
143
133
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
134
+ pq_mode: NotRequired[OutputCloudwatchMode]
135
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
144
136
  pq_controls: NotRequired[OutputCloudwatchPqControlsTypedDict]
145
137
 
146
138
 
@@ -248,34 +240,6 @@ class OutputCloudwatch(BaseModel):
248
240
  aws_secret: Annotated[Optional[str], pydantic.Field(alias="awsSecret")] = None
249
241
  r"""Select or create a stored secret that references your access key and secret key"""
250
242
 
251
- pq_strict_ordering: Annotated[
252
- Optional[bool], pydantic.Field(alias="pqStrictOrdering")
253
- ] = True
254
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
255
-
256
- pq_rate_per_sec: Annotated[
257
- Optional[float], pydantic.Field(alias="pqRatePerSec")
258
- ] = 0
259
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
260
-
261
- pq_mode: Annotated[
262
- Annotated[
263
- Optional[OutputCloudwatchMode], PlainValidator(validate_open_enum(False))
264
- ],
265
- pydantic.Field(alias="pqMode"),
266
- ] = OutputCloudwatchMode.ERROR
267
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
268
-
269
- pq_max_buffer_size: Annotated[
270
- Optional[float], pydantic.Field(alias="pqMaxBufferSize")
271
- ] = 42
272
- r"""The maximum number of events to hold in memory before writing the events to disk"""
273
-
274
- pq_max_backpressure_sec: Annotated[
275
- Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
276
- ] = 30
277
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
278
-
279
243
  pq_max_file_size: Annotated[
280
244
  Optional[str], pydantic.Field(alias="pqMaxFileSize")
281
245
  ] = "1 MB"
@@ -307,6 +271,14 @@ class OutputCloudwatch(BaseModel):
307
271
  ] = OutputCloudwatchQueueFullBehavior.BLOCK
308
272
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
309
273
 
274
+ pq_mode: Annotated[
275
+ Annotated[
276
+ Optional[OutputCloudwatchMode], PlainValidator(validate_open_enum(False))
277
+ ],
278
+ pydantic.Field(alias="pqMode"),
279
+ ] = OutputCloudwatchMode.ERROR
280
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
281
+
310
282
  pq_controls: Annotated[
311
283
  Optional[OutputCloudwatchPqControls], pydantic.Field(alias="pqControls")
312
284
  ] = None
@@ -297,13 +297,6 @@ class OutputConfluentCloudKafkaSchemaRegistryAuthentication(BaseModel):
297
297
  r"""Used when __valueSchemaIdOut is not present, to transform _raw, leave blank if value transformation is not required by default."""
298
298
 
299
299
 
300
- class OutputConfluentCloudAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
301
- r"""Enter credentials directly, or select a stored secret"""
302
-
303
- MANUAL = "manual"
304
- SECRET = "secret"
305
-
306
-
307
300
  class OutputConfluentCloudSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta):
308
301
  # PLAIN
309
302
  PLAIN = "plain"
@@ -315,58 +308,13 @@ class OutputConfluentCloudSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta)
315
308
  KERBEROS = "kerberos"
316
309
 
317
310
 
318
- class OutputConfluentCloudOauthParamTypedDict(TypedDict):
319
- name: str
320
- value: str
321
-
322
-
323
- class OutputConfluentCloudOauthParam(BaseModel):
324
- name: str
325
-
326
- value: str
327
-
328
-
329
- class OutputConfluentCloudSaslExtensionTypedDict(TypedDict):
330
- name: str
331
- value: str
332
-
333
-
334
- class OutputConfluentCloudSaslExtension(BaseModel):
335
- name: str
336
-
337
- value: str
338
-
339
-
340
311
  class OutputConfluentCloudAuthenticationTypedDict(TypedDict):
341
312
  r"""Authentication parameters to use when connecting to brokers. Using TLS is highly recommended."""
342
313
 
343
314
  disabled: NotRequired[bool]
344
- username: NotRequired[str]
345
- password: NotRequired[str]
346
- auth_type: NotRequired[OutputConfluentCloudAuthenticationMethod]
347
- r"""Enter credentials directly, or select a stored secret"""
348
- credentials_secret: NotRequired[str]
349
- r"""Select or create a secret that references your credentials"""
350
315
  mechanism: NotRequired[OutputConfluentCloudSASLMechanism]
351
- keytab_location: NotRequired[str]
352
- r"""Location of keytab file for authentication principal"""
353
- principal: NotRequired[str]
354
- r"""Authentication principal, such as `kafka_user@example.com`"""
355
- broker_service_class: NotRequired[str]
356
- r"""Kerberos service class for Kafka brokers, such as `kafka`"""
357
316
  oauth_enabled: NotRequired[bool]
358
317
  r"""Enable OAuth authentication"""
359
- token_url: NotRequired[str]
360
- r"""URL of the token endpoint to use for OAuth authentication"""
361
- client_id: NotRequired[str]
362
- r"""Client ID to use for OAuth authentication"""
363
- oauth_secret_type: NotRequired[str]
364
- client_text_secret: NotRequired[str]
365
- r"""Select or create a stored text secret"""
366
- oauth_params: NotRequired[List[OutputConfluentCloudOauthParamTypedDict]]
367
- r"""Additional fields to send to the token endpoint, such as scope or audience"""
368
- sasl_extensions: NotRequired[List[OutputConfluentCloudSaslExtensionTypedDict]]
369
- r"""Additional SASL extension fields, such as Confluent's logicalCluster or identityPoolId"""
370
318
 
371
319
 
372
320
  class OutputConfluentCloudAuthentication(BaseModel):
@@ -374,74 +322,16 @@ class OutputConfluentCloudAuthentication(BaseModel):
374
322
 
375
323
  disabled: Optional[bool] = True
376
324
 
377
- username: Optional[str] = None
378
-
379
- password: Optional[str] = None
380
-
381
- auth_type: Annotated[
382
- Annotated[
383
- Optional[OutputConfluentCloudAuthenticationMethod],
384
- PlainValidator(validate_open_enum(False)),
385
- ],
386
- pydantic.Field(alias="authType"),
387
- ] = OutputConfluentCloudAuthenticationMethod.MANUAL
388
- r"""Enter credentials directly, or select a stored secret"""
389
-
390
- credentials_secret: Annotated[
391
- Optional[str], pydantic.Field(alias="credentialsSecret")
392
- ] = None
393
- r"""Select or create a secret that references your credentials"""
394
-
395
325
  mechanism: Annotated[
396
326
  Optional[OutputConfluentCloudSASLMechanism],
397
327
  PlainValidator(validate_open_enum(False)),
398
328
  ] = OutputConfluentCloudSASLMechanism.PLAIN
399
329
 
400
- keytab_location: Annotated[
401
- Optional[str], pydantic.Field(alias="keytabLocation")
402
- ] = None
403
- r"""Location of keytab file for authentication principal"""
404
-
405
- principal: Optional[str] = None
406
- r"""Authentication principal, such as `kafka_user@example.com`"""
407
-
408
- broker_service_class: Annotated[
409
- Optional[str], pydantic.Field(alias="brokerServiceClass")
410
- ] = None
411
- r"""Kerberos service class for Kafka brokers, such as `kafka`"""
412
-
413
330
  oauth_enabled: Annotated[Optional[bool], pydantic.Field(alias="oauthEnabled")] = (
414
331
  False
415
332
  )
416
333
  r"""Enable OAuth authentication"""
417
334
 
418
- token_url: Annotated[Optional[str], pydantic.Field(alias="tokenUrl")] = None
419
- r"""URL of the token endpoint to use for OAuth authentication"""
420
-
421
- client_id: Annotated[Optional[str], pydantic.Field(alias="clientId")] = None
422
- r"""Client ID to use for OAuth authentication"""
423
-
424
- oauth_secret_type: Annotated[
425
- Optional[str], pydantic.Field(alias="oauthSecretType")
426
- ] = "secret"
427
-
428
- client_text_secret: Annotated[
429
- Optional[str], pydantic.Field(alias="clientTextSecret")
430
- ] = None
431
- r"""Select or create a stored text secret"""
432
-
433
- oauth_params: Annotated[
434
- Optional[List[OutputConfluentCloudOauthParam]],
435
- pydantic.Field(alias="oauthParams"),
436
- ] = None
437
- r"""Additional fields to send to the token endpoint, such as scope or audience"""
438
-
439
- sasl_extensions: Annotated[
440
- Optional[List[OutputConfluentCloudSaslExtension]],
441
- pydantic.Field(alias="saslExtensions"),
442
- ] = None
443
- r"""Additional SASL extension fields, such as Confluent's logicalCluster or identityPoolId"""
444
-
445
335
 
446
336
  class OutputConfluentCloudBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
447
337
  r"""How to handle events when all receivers are exerting backpressure"""
@@ -454,17 +344,6 @@ class OutputConfluentCloudBackpressureBehavior(str, Enum, metaclass=utils.OpenEn
454
344
  QUEUE = "queue"
455
345
 
456
346
 
457
- class OutputConfluentCloudMode(str, Enum, metaclass=utils.OpenEnumMeta):
458
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
459
-
460
- # Error
461
- ERROR = "error"
462
- # Backpressure
463
- ALWAYS = "always"
464
- # Always On
465
- BACKPRESSURE = "backpressure"
466
-
467
-
468
347
  class OutputConfluentCloudPqCompressCompression(
469
348
  str, Enum, metaclass=utils.OpenEnumMeta
470
349
  ):
@@ -485,6 +364,17 @@ class OutputConfluentCloudQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumM
485
364
  DROP = "drop"
486
365
 
487
366
 
367
+ class OutputConfluentCloudMode(str, Enum, metaclass=utils.OpenEnumMeta):
368
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
369
+
370
+ # Error
371
+ ERROR = "error"
372
+ # Backpressure
373
+ BACKPRESSURE = "backpressure"
374
+ # Always On
375
+ ALWAYS = "always"
376
+
377
+
488
378
  class OutputConfluentCloudPqControlsTypedDict(TypedDict):
489
379
  pass
490
380
 
@@ -548,18 +438,6 @@ class OutputConfluentCloudTypedDict(TypedDict):
548
438
  description: NotRequired[str]
549
439
  protobuf_library_id: NotRequired[str]
550
440
  r"""Select a set of Protobuf definitions for the events you want to send"""
551
- protobuf_encoding_id: NotRequired[str]
552
- r"""Select the type of object you want the Protobuf definitions to use for event encoding"""
553
- pq_strict_ordering: NotRequired[bool]
554
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
555
- pq_rate_per_sec: NotRequired[float]
556
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
557
- pq_mode: NotRequired[OutputConfluentCloudMode]
558
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
559
- pq_max_buffer_size: NotRequired[float]
560
- r"""The maximum number of events to hold in memory before writing the events to disk"""
561
- pq_max_backpressure_sec: NotRequired[float]
562
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
563
441
  pq_max_file_size: NotRequired[str]
564
442
  r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
565
443
  pq_max_size: NotRequired[str]
@@ -570,6 +448,8 @@ class OutputConfluentCloudTypedDict(TypedDict):
570
448
  r"""Codec to use to compress the persisted data"""
571
449
  pq_on_backpressure: NotRequired[OutputConfluentCloudQueueFullBehavior]
572
450
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
451
+ pq_mode: NotRequired[OutputConfluentCloudMode]
452
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
573
453
  pq_controls: NotRequired[OutputConfluentCloudPqControlsTypedDict]
574
454
 
575
455
 
@@ -695,40 +575,6 @@ class OutputConfluentCloud(BaseModel):
695
575
  ] = None
696
576
  r"""Select a set of Protobuf definitions for the events you want to send"""
697
577
 
698
- protobuf_encoding_id: Annotated[
699
- Optional[str], pydantic.Field(alias="protobufEncodingId")
700
- ] = None
701
- r"""Select the type of object you want the Protobuf definitions to use for event encoding"""
702
-
703
- pq_strict_ordering: Annotated[
704
- Optional[bool], pydantic.Field(alias="pqStrictOrdering")
705
- ] = True
706
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
707
-
708
- pq_rate_per_sec: Annotated[
709
- Optional[float], pydantic.Field(alias="pqRatePerSec")
710
- ] = 0
711
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
712
-
713
- pq_mode: Annotated[
714
- Annotated[
715
- Optional[OutputConfluentCloudMode],
716
- PlainValidator(validate_open_enum(False)),
717
- ],
718
- pydantic.Field(alias="pqMode"),
719
- ] = OutputConfluentCloudMode.ERROR
720
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
721
-
722
- pq_max_buffer_size: Annotated[
723
- Optional[float], pydantic.Field(alias="pqMaxBufferSize")
724
- ] = 42
725
- r"""The maximum number of events to hold in memory before writing the events to disk"""
726
-
727
- pq_max_backpressure_sec: Annotated[
728
- Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
729
- ] = 30
730
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
731
-
732
578
  pq_max_file_size: Annotated[
733
579
  Optional[str], pydantic.Field(alias="pqMaxFileSize")
734
580
  ] = "1 MB"
@@ -760,6 +606,15 @@ class OutputConfluentCloud(BaseModel):
760
606
  ] = OutputConfluentCloudQueueFullBehavior.BLOCK
761
607
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
762
608
 
609
+ pq_mode: Annotated[
610
+ Annotated[
611
+ Optional[OutputConfluentCloudMode],
612
+ PlainValidator(validate_open_enum(False)),
613
+ ],
614
+ pydantic.Field(alias="pqMode"),
615
+ ] = OutputConfluentCloudMode.ERROR
616
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
617
+
763
618
  pq_controls: Annotated[
764
619
  Optional[OutputConfluentCloudPqControls], pydantic.Field(alias="pqControls")
765
620
  ] = None
@@ -209,17 +209,6 @@ class OutputCriblHTTPURL(BaseModel):
209
209
  r"""Assign a weight (>0) to each endpoint to indicate its traffic-handling capability"""
210
210
 
211
211
 
212
- class OutputCriblHTTPMode(str, Enum, metaclass=utils.OpenEnumMeta):
213
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
214
-
215
- # Error
216
- ERROR = "error"
217
- # Backpressure
218
- ALWAYS = "always"
219
- # Always On
220
- BACKPRESSURE = "backpressure"
221
-
222
-
223
212
  class OutputCriblHTTPPqCompressCompression(str, Enum, metaclass=utils.OpenEnumMeta):
224
213
  r"""Codec to use to compress the persisted data"""
225
214
 
@@ -238,6 +227,17 @@ class OutputCriblHTTPQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
238
227
  DROP = "drop"
239
228
 
240
229
 
230
+ class OutputCriblHTTPMode(str, Enum, metaclass=utils.OpenEnumMeta):
231
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
232
+
233
+ # Error
234
+ ERROR = "error"
235
+ # Backpressure
236
+ BACKPRESSURE = "backpressure"
237
+ # Always On
238
+ ALWAYS = "always"
239
+
240
+
241
241
  class OutputCriblHTTPPqControlsTypedDict(TypedDict):
242
242
  pass
243
243
 
@@ -309,16 +309,6 @@ class OutputCriblHTTPTypedDict(TypedDict):
309
309
  r"""The interval in which to re-resolve any hostnames and pick up destinations from A records"""
310
310
  load_balance_stats_period_sec: NotRequired[float]
311
311
  r"""How far back in time to keep traffic stats for load balancing purposes"""
312
- pq_strict_ordering: NotRequired[bool]
313
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
314
- pq_rate_per_sec: NotRequired[float]
315
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
316
- pq_mode: NotRequired[OutputCriblHTTPMode]
317
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
318
- pq_max_buffer_size: NotRequired[float]
319
- r"""The maximum number of events to hold in memory before writing the events to disk"""
320
- pq_max_backpressure_sec: NotRequired[float]
321
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
322
312
  pq_max_file_size: NotRequired[str]
323
313
  r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
324
314
  pq_max_size: NotRequired[str]
@@ -329,6 +319,8 @@ class OutputCriblHTTPTypedDict(TypedDict):
329
319
  r"""Codec to use to compress the persisted data"""
330
320
  pq_on_backpressure: NotRequired[OutputCriblHTTPQueueFullBehavior]
331
321
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
322
+ pq_mode: NotRequired[OutputCriblHTTPMode]
323
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
332
324
  pq_controls: NotRequired[OutputCriblHTTPPqControlsTypedDict]
333
325
 
334
326
 
@@ -473,34 +465,6 @@ class OutputCriblHTTP(BaseModel):
473
465
  ] = 300
474
466
  r"""How far back in time to keep traffic stats for load balancing purposes"""
475
467
 
476
- pq_strict_ordering: Annotated[
477
- Optional[bool], pydantic.Field(alias="pqStrictOrdering")
478
- ] = True
479
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
480
-
481
- pq_rate_per_sec: Annotated[
482
- Optional[float], pydantic.Field(alias="pqRatePerSec")
483
- ] = 0
484
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
485
-
486
- pq_mode: Annotated[
487
- Annotated[
488
- Optional[OutputCriblHTTPMode], PlainValidator(validate_open_enum(False))
489
- ],
490
- pydantic.Field(alias="pqMode"),
491
- ] = OutputCriblHTTPMode.ERROR
492
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
493
-
494
- pq_max_buffer_size: Annotated[
495
- Optional[float], pydantic.Field(alias="pqMaxBufferSize")
496
- ] = 42
497
- r"""The maximum number of events to hold in memory before writing the events to disk"""
498
-
499
- pq_max_backpressure_sec: Annotated[
500
- Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
501
- ] = 30
502
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
503
-
504
468
  pq_max_file_size: Annotated[
505
469
  Optional[str], pydantic.Field(alias="pqMaxFileSize")
506
470
  ] = "1 MB"
@@ -532,6 +496,14 @@ class OutputCriblHTTP(BaseModel):
532
496
  ] = OutputCriblHTTPQueueFullBehavior.BLOCK
533
497
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
534
498
 
499
+ pq_mode: Annotated[
500
+ Annotated[
501
+ Optional[OutputCriblHTTPMode], PlainValidator(validate_open_enum(False))
502
+ ],
503
+ pydantic.Field(alias="pqMode"),
504
+ ] = OutputCriblHTTPMode.ERROR
505
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
506
+
535
507
  pq_controls: Annotated[
536
508
  Optional[OutputCriblHTTPPqControls], pydantic.Field(alias="pqControls")
537
509
  ] = None
@@ -157,17 +157,6 @@ class OutputCriblTCPHost(BaseModel):
157
157
  r"""Assign a weight (>0) to each endpoint to indicate its traffic-handling capability"""
158
158
 
159
159
 
160
- class OutputCriblTCPMode(str, Enum, metaclass=utils.OpenEnumMeta):
161
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
162
-
163
- # Error
164
- ERROR = "error"
165
- # Backpressure
166
- ALWAYS = "always"
167
- # Always On
168
- BACKPRESSURE = "backpressure"
169
-
170
-
171
160
  class OutputCriblTCPPqCompressCompression(str, Enum, metaclass=utils.OpenEnumMeta):
172
161
  r"""Codec to use to compress the persisted data"""
173
162
 
@@ -186,6 +175,17 @@ class OutputCriblTCPQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
186
175
  DROP = "drop"
187
176
 
188
177
 
178
+ class OutputCriblTCPMode(str, Enum, metaclass=utils.OpenEnumMeta):
179
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
180
+
181
+ # Error
182
+ ERROR = "error"
183
+ # Backpressure
184
+ BACKPRESSURE = "backpressure"
185
+ # Always On
186
+ ALWAYS = "always"
187
+
188
+
189
189
  class OutputCriblTCPPqControlsTypedDict(TypedDict):
190
190
  pass
191
191
 
@@ -240,16 +240,6 @@ class OutputCriblTCPTypedDict(TypedDict):
240
240
  r"""How far back in time to keep traffic stats for load balancing purposes"""
241
241
  max_concurrent_senders: NotRequired[float]
242
242
  r"""Maximum number of concurrent connections (per Worker Process). A random set of IPs will be picked on every DNS resolution period. Use 0 for unlimited."""
243
- pq_strict_ordering: NotRequired[bool]
244
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
245
- pq_rate_per_sec: NotRequired[float]
246
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
247
- pq_mode: NotRequired[OutputCriblTCPMode]
248
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
249
- pq_max_buffer_size: NotRequired[float]
250
- r"""The maximum number of events to hold in memory before writing the events to disk"""
251
- pq_max_backpressure_sec: NotRequired[float]
252
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
253
243
  pq_max_file_size: NotRequired[str]
254
244
  r"""The maximum size to store in each queue file before closing and optionally compressing (KB, MB, etc.)"""
255
245
  pq_max_size: NotRequired[str]
@@ -260,6 +250,8 @@ class OutputCriblTCPTypedDict(TypedDict):
260
250
  r"""Codec to use to compress the persisted data"""
261
251
  pq_on_backpressure: NotRequired[OutputCriblTCPQueueFullBehavior]
262
252
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
253
+ pq_mode: NotRequired[OutputCriblTCPMode]
254
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
263
255
  pq_controls: NotRequired[OutputCriblTCPPqControlsTypedDict]
264
256
 
265
257
 
@@ -363,34 +355,6 @@ class OutputCriblTCP(BaseModel):
363
355
  ] = 0
364
356
  r"""Maximum number of concurrent connections (per Worker Process). A random set of IPs will be picked on every DNS resolution period. Use 0 for unlimited."""
365
357
 
366
- pq_strict_ordering: Annotated[
367
- Optional[bool], pydantic.Field(alias="pqStrictOrdering")
368
- ] = True
369
- r"""Use FIFO (first in, first out) processing. Disable to forward new events to receivers before queue is flushed."""
370
-
371
- pq_rate_per_sec: Annotated[
372
- Optional[float], pydantic.Field(alias="pqRatePerSec")
373
- ] = 0
374
- r"""Throttling rate (in events per second) to impose while writing to Destinations from PQ. Defaults to 0, which disables throttling."""
375
-
376
- pq_mode: Annotated[
377
- Annotated[
378
- Optional[OutputCriblTCPMode], PlainValidator(validate_open_enum(False))
379
- ],
380
- pydantic.Field(alias="pqMode"),
381
- ] = OutputCriblTCPMode.ERROR
382
- r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
383
-
384
- pq_max_buffer_size: Annotated[
385
- Optional[float], pydantic.Field(alias="pqMaxBufferSize")
386
- ] = 42
387
- r"""The maximum number of events to hold in memory before writing the events to disk"""
388
-
389
- pq_max_backpressure_sec: Annotated[
390
- Optional[float], pydantic.Field(alias="pqMaxBackpressureSec")
391
- ] = 30
392
- r"""How long (in seconds) to wait for backpressure to resolve before engaging the queue"""
393
-
394
358
  pq_max_file_size: Annotated[
395
359
  Optional[str], pydantic.Field(alias="pqMaxFileSize")
396
360
  ] = "1 MB"
@@ -422,6 +386,14 @@ class OutputCriblTCP(BaseModel):
422
386
  ] = OutputCriblTCPQueueFullBehavior.BLOCK
423
387
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
424
388
 
389
+ pq_mode: Annotated[
390
+ Annotated[
391
+ Optional[OutputCriblTCPMode], PlainValidator(validate_open_enum(False))
392
+ ],
393
+ pydantic.Field(alias="pqMode"),
394
+ ] = OutputCriblTCPMode.ERROR
395
+ r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
396
+
425
397
  pq_controls: Annotated[
426
398
  Optional[OutputCriblTCPPqControls], pydantic.Field(alias="pqControls")
427
399
  ] = None