cribl-control-plane 0.0.44__py3-none-any.whl → 0.0.44a2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +3 -3
- cribl_control_plane/errors/healthstatus_error.py +8 -2
- cribl_control_plane/models/__init__.py +3 -3
- cribl_control_plane/models/appmode.py +2 -1
- cribl_control_plane/models/cacheconnection.py +10 -2
- cribl_control_plane/models/cacheconnectionbackfillstatus.py +2 -1
- cribl_control_plane/models/cloudprovider.py +2 -1
- cribl_control_plane/models/configgroup.py +7 -2
- cribl_control_plane/models/configgroupcloud.py +6 -2
- cribl_control_plane/models/createconfiggroupbyproductop.py +8 -2
- cribl_control_plane/models/cribllakedataset.py +8 -2
- cribl_control_plane/models/datasetmetadata.py +8 -2
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +7 -2
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +4 -2
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +4 -2
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +3 -1
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +7 -2
- cribl_control_plane/models/getsummaryop.py +7 -2
- cribl_control_plane/models/hbcriblinfo.py +6 -1
- cribl_control_plane/models/healthstatus.py +7 -4
- cribl_control_plane/models/inputappscope.py +34 -14
- cribl_control_plane/models/inputazureblob.py +17 -6
- cribl_control_plane/models/inputcollection.py +11 -4
- cribl_control_plane/models/inputconfluentcloud.py +47 -20
- cribl_control_plane/models/inputcribl.py +11 -4
- cribl_control_plane/models/inputcriblhttp.py +23 -8
- cribl_control_plane/models/inputcribllakehttp.py +22 -10
- cribl_control_plane/models/inputcriblmetrics.py +12 -4
- cribl_control_plane/models/inputcribltcp.py +23 -8
- cribl_control_plane/models/inputcrowdstrike.py +26 -10
- cribl_control_plane/models/inputdatadogagent.py +24 -8
- cribl_control_plane/models/inputdatagen.py +11 -4
- cribl_control_plane/models/inputedgeprometheus.py +58 -24
- cribl_control_plane/models/inputelastic.py +40 -14
- cribl_control_plane/models/inputeventhub.py +15 -6
- cribl_control_plane/models/inputexec.py +14 -6
- cribl_control_plane/models/inputfile.py +15 -6
- cribl_control_plane/models/inputfirehose.py +23 -8
- cribl_control_plane/models/inputgooglepubsub.py +19 -6
- cribl_control_plane/models/inputgrafana.py +67 -24
- cribl_control_plane/models/inputhttp.py +23 -8
- cribl_control_plane/models/inputhttpraw.py +23 -8
- cribl_control_plane/models/inputjournalfiles.py +12 -4
- cribl_control_plane/models/inputkafka.py +46 -16
- cribl_control_plane/models/inputkinesis.py +38 -14
- cribl_control_plane/models/inputkubeevents.py +11 -4
- cribl_control_plane/models/inputkubelogs.py +16 -8
- cribl_control_plane/models/inputkubemetrics.py +16 -8
- cribl_control_plane/models/inputloki.py +29 -10
- cribl_control_plane/models/inputmetrics.py +23 -8
- cribl_control_plane/models/inputmodeldriventelemetry.py +27 -10
- cribl_control_plane/models/inputmsk.py +53 -18
- cribl_control_plane/models/inputnetflow.py +11 -4
- cribl_control_plane/models/inputoffice365mgmt.py +33 -14
- cribl_control_plane/models/inputoffice365msgtrace.py +35 -16
- cribl_control_plane/models/inputoffice365service.py +35 -16
- cribl_control_plane/models/inputopentelemetry.py +38 -16
- cribl_control_plane/models/inputprometheus.py +50 -18
- cribl_control_plane/models/inputprometheusrw.py +30 -10
- cribl_control_plane/models/inputrawudp.py +11 -4
- cribl_control_plane/models/inputs3.py +21 -8
- cribl_control_plane/models/inputs3inventory.py +26 -10
- cribl_control_plane/models/inputsecuritylake.py +27 -10
- cribl_control_plane/models/inputsnmp.py +16 -6
- cribl_control_plane/models/inputsplunk.py +33 -12
- cribl_control_plane/models/inputsplunkhec.py +29 -10
- cribl_control_plane/models/inputsplunksearch.py +33 -14
- cribl_control_plane/models/inputsqs.py +27 -10
- cribl_control_plane/models/inputsyslog.py +43 -16
- cribl_control_plane/models/inputsystemmetrics.py +48 -24
- cribl_control_plane/models/inputsystemstate.py +16 -8
- cribl_control_plane/models/inputtcp.py +29 -10
- cribl_control_plane/models/inputtcpjson.py +29 -10
- cribl_control_plane/models/inputwef.py +37 -14
- cribl_control_plane/models/inputwindowsmetrics.py +44 -24
- cribl_control_plane/models/inputwineventlogs.py +20 -10
- cribl_control_plane/models/inputwiz.py +21 -8
- cribl_control_plane/models/inputwizwebhook.py +23 -8
- cribl_control_plane/models/inputzscalerhec.py +29 -10
- cribl_control_plane/models/lakehouseconnectiontype.py +2 -1
- cribl_control_plane/models/listconfiggroupbyproductop.py +3 -1
- cribl_control_plane/models/masterworkerentry.py +7 -2
- cribl_control_plane/models/nodeactiveupgradestatus.py +2 -1
- cribl_control_plane/models/nodefailedupgradestatus.py +2 -1
- cribl_control_plane/models/nodeskippedupgradestatus.py +2 -1
- cribl_control_plane/models/nodeupgradestate.py +2 -1
- cribl_control_plane/models/nodeupgradestatus.py +13 -5
- cribl_control_plane/models/outputazureblob.py +48 -18
- cribl_control_plane/models/outputazuredataexplorer.py +74 -29
- cribl_control_plane/models/outputazureeventhub.py +40 -18
- cribl_control_plane/models/outputazurelogs.py +36 -13
- cribl_control_plane/models/outputclickhouse.py +56 -21
- cribl_control_plane/models/outputcloudwatch.py +29 -10
- cribl_control_plane/models/outputconfluentcloud.py +77 -32
- cribl_control_plane/models/outputcriblhttp.py +46 -18
- cribl_control_plane/models/outputcribllake.py +46 -16
- cribl_control_plane/models/outputcribltcp.py +45 -18
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +50 -15
- cribl_control_plane/models/outputdatadog.py +48 -20
- cribl_control_plane/models/outputdataset.py +46 -18
- cribl_control_plane/models/outputdiskspool.py +7 -2
- cribl_control_plane/models/outputdls3.py +68 -24
- cribl_control_plane/models/outputdynatracehttp.py +54 -21
- cribl_control_plane/models/outputdynatraceotlp.py +56 -23
- cribl_control_plane/models/outputelastic.py +44 -19
- cribl_control_plane/models/outputelasticcloud.py +37 -13
- cribl_control_plane/models/outputexabeam.py +29 -10
- cribl_control_plane/models/outputfilesystem.py +39 -14
- cribl_control_plane/models/outputgooglechronicle.py +50 -16
- cribl_control_plane/models/outputgooglecloudlogging.py +41 -14
- cribl_control_plane/models/outputgooglecloudstorage.py +66 -24
- cribl_control_plane/models/outputgooglepubsub.py +31 -10
- cribl_control_plane/models/outputgrafanacloud.py +99 -34
- cribl_control_plane/models/outputgraphite.py +31 -14
- cribl_control_plane/models/outputhoneycomb.py +36 -13
- cribl_control_plane/models/outputhumiohec.py +44 -17
- cribl_control_plane/models/outputinfluxdb.py +43 -17
- cribl_control_plane/models/outputkafka.py +74 -28
- cribl_control_plane/models/outputkinesis.py +40 -16
- cribl_control_plane/models/outputloki.py +41 -16
- cribl_control_plane/models/outputminio.py +65 -24
- cribl_control_plane/models/outputmsk.py +82 -30
- cribl_control_plane/models/outputnewrelic.py +43 -18
- cribl_control_plane/models/outputnewrelicevents.py +42 -15
- cribl_control_plane/models/outputopentelemetry.py +68 -27
- cribl_control_plane/models/outputprometheus.py +36 -13
- cribl_control_plane/models/outputring.py +19 -8
- cribl_control_plane/models/outputs3.py +68 -26
- cribl_control_plane/models/outputsecuritylake.py +52 -18
- cribl_control_plane/models/outputsentinel.py +45 -18
- cribl_control_plane/models/outputsentineloneaisiem.py +51 -19
- cribl_control_plane/models/outputservicenow.py +61 -25
- cribl_control_plane/models/outputsignalfx.py +38 -15
- cribl_control_plane/models/outputsns.py +36 -14
- cribl_control_plane/models/outputsplunk.py +60 -24
- cribl_control_plane/models/outputsplunkhec.py +36 -13
- cribl_control_plane/models/outputsplunklb.py +77 -30
- cribl_control_plane/models/outputsqs.py +41 -16
- cribl_control_plane/models/outputstatsd.py +30 -14
- cribl_control_plane/models/outputstatsdext.py +29 -12
- cribl_control_plane/models/outputsumologic.py +35 -12
- cribl_control_plane/models/outputsyslog.py +58 -24
- cribl_control_plane/models/outputtcpjson.py +52 -20
- cribl_control_plane/models/outputwavefront.py +36 -13
- cribl_control_plane/models/outputwebhook.py +58 -22
- cribl_control_plane/models/outputxsiam.py +36 -15
- cribl_control_plane/models/productscore.py +2 -1
- cribl_control_plane/models/rbacresource.py +2 -1
- cribl_control_plane/models/resourcepolicy.py +4 -2
- cribl_control_plane/models/runnablejobcollection.py +30 -13
- cribl_control_plane/models/runnablejobexecutor.py +13 -4
- cribl_control_plane/models/runnablejobscheduledsearch.py +7 -2
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +8 -2
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +8 -2
- cribl_control_plane/models/workertypes.py +2 -1
- {cribl_control_plane-0.0.44.dist-info → cribl_control_plane-0.0.44a2.dist-info}/METADATA +1 -1
- {cribl_control_plane-0.0.44.dist-info → cribl_control_plane-0.0.44a2.dist-info}/RECORD +158 -158
- {cribl_control_plane-0.0.44.dist-info → cribl_control_plane-0.0.44a2.dist-info}/WHEEL +0 -0
|
@@ -1,9 +1,12 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from cribl_control_plane import utils
|
|
4
5
|
from cribl_control_plane.types import BaseModel
|
|
6
|
+
from cribl_control_plane.utils import validate_open_enum
|
|
5
7
|
from enum import Enum
|
|
6
8
|
import pydantic
|
|
9
|
+
from pydantic.functional_validators import PlainValidator
|
|
7
10
|
from typing import List, Optional
|
|
8
11
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
12
|
|
|
@@ -12,7 +15,7 @@ class OutputMskType(str, Enum):
|
|
|
12
15
|
MSK = "msk"
|
|
13
16
|
|
|
14
17
|
|
|
15
|
-
class OutputMskAcknowledgments(int, Enum):
|
|
18
|
+
class OutputMskAcknowledgments(int, Enum, metaclass=utils.OpenEnumMeta):
|
|
16
19
|
r"""Control the number of required acknowledgments."""
|
|
17
20
|
|
|
18
21
|
ONE = 1
|
|
@@ -20,7 +23,7 @@ class OutputMskAcknowledgments(int, Enum):
|
|
|
20
23
|
MINUS_1 = -1
|
|
21
24
|
|
|
22
25
|
|
|
23
|
-
class OutputMskRecordDataFormat(str, Enum):
|
|
26
|
+
class OutputMskRecordDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
24
27
|
r"""Format to use to serialize events before writing to Kafka."""
|
|
25
28
|
|
|
26
29
|
JSON = "json"
|
|
@@ -28,7 +31,7 @@ class OutputMskRecordDataFormat(str, Enum):
|
|
|
28
31
|
PROTOBUF = "protobuf"
|
|
29
32
|
|
|
30
33
|
|
|
31
|
-
class OutputMskCompression(str, Enum):
|
|
34
|
+
class OutputMskCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
32
35
|
r"""Codec to use to compress the data before sending to Kafka"""
|
|
33
36
|
|
|
34
37
|
NONE = "none"
|
|
@@ -37,7 +40,7 @@ class OutputMskCompression(str, Enum):
|
|
|
37
40
|
LZ4 = "lz4"
|
|
38
41
|
|
|
39
42
|
|
|
40
|
-
class OutputMskSchemaType(str, Enum):
|
|
43
|
+
class OutputMskSchemaType(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
41
44
|
r"""The schema format used to encode and decode event data"""
|
|
42
45
|
|
|
43
46
|
AVRO = "avro"
|
|
@@ -63,14 +66,18 @@ class OutputMskAuth(BaseModel):
|
|
|
63
66
|
r"""Select or create a secret that references your credentials"""
|
|
64
67
|
|
|
65
68
|
|
|
66
|
-
class OutputMskKafkaSchemaRegistryMinimumTLSVersion(
|
|
69
|
+
class OutputMskKafkaSchemaRegistryMinimumTLSVersion(
|
|
70
|
+
str, Enum, metaclass=utils.OpenEnumMeta
|
|
71
|
+
):
|
|
67
72
|
TL_SV1 = "TLSv1"
|
|
68
73
|
TL_SV1_1 = "TLSv1.1"
|
|
69
74
|
TL_SV1_2 = "TLSv1.2"
|
|
70
75
|
TL_SV1_3 = "TLSv1.3"
|
|
71
76
|
|
|
72
77
|
|
|
73
|
-
class OutputMskKafkaSchemaRegistryMaximumTLSVersion(
|
|
78
|
+
class OutputMskKafkaSchemaRegistryMaximumTLSVersion(
|
|
79
|
+
str, Enum, metaclass=utils.OpenEnumMeta
|
|
80
|
+
):
|
|
74
81
|
TL_SV1 = "TLSv1"
|
|
75
82
|
TL_SV1_1 = "TLSv1.1"
|
|
76
83
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -130,12 +137,18 @@ class OutputMskKafkaSchemaRegistryTLSSettingsClientSide(BaseModel):
|
|
|
130
137
|
r"""Passphrase to use to decrypt private key"""
|
|
131
138
|
|
|
132
139
|
min_version: Annotated[
|
|
133
|
-
|
|
140
|
+
Annotated[
|
|
141
|
+
Optional[OutputMskKafkaSchemaRegistryMinimumTLSVersion],
|
|
142
|
+
PlainValidator(validate_open_enum(False)),
|
|
143
|
+
],
|
|
134
144
|
pydantic.Field(alias="minVersion"),
|
|
135
145
|
] = None
|
|
136
146
|
|
|
137
147
|
max_version: Annotated[
|
|
138
|
-
|
|
148
|
+
Annotated[
|
|
149
|
+
Optional[OutputMskKafkaSchemaRegistryMaximumTLSVersion],
|
|
150
|
+
PlainValidator(validate_open_enum(False)),
|
|
151
|
+
],
|
|
139
152
|
pydantic.Field(alias="maxVersion"),
|
|
140
153
|
] = None
|
|
141
154
|
|
|
@@ -170,7 +183,10 @@ class OutputMskKafkaSchemaRegistryAuthentication(BaseModel):
|
|
|
170
183
|
r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
|
|
171
184
|
|
|
172
185
|
schema_type: Annotated[
|
|
173
|
-
|
|
186
|
+
Annotated[
|
|
187
|
+
Optional[OutputMskSchemaType], PlainValidator(validate_open_enum(False))
|
|
188
|
+
],
|
|
189
|
+
pydantic.Field(alias="schemaType"),
|
|
174
190
|
] = OutputMskSchemaType.AVRO
|
|
175
191
|
r"""The schema format used to encode and decode event data"""
|
|
176
192
|
|
|
@@ -203,7 +219,7 @@ class OutputMskKafkaSchemaRegistryAuthentication(BaseModel):
|
|
|
203
219
|
r"""Used when __valueSchemaIdOut is not present, to transform _raw, leave blank if value transformation is not required by default."""
|
|
204
220
|
|
|
205
221
|
|
|
206
|
-
class OutputMskAuthenticationMethod(str, Enum):
|
|
222
|
+
class OutputMskAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
207
223
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
208
224
|
|
|
209
225
|
AUTO = "auto"
|
|
@@ -211,21 +227,21 @@ class OutputMskAuthenticationMethod(str, Enum):
|
|
|
211
227
|
SECRET = "secret"
|
|
212
228
|
|
|
213
229
|
|
|
214
|
-
class OutputMskSignatureVersion(str, Enum):
|
|
230
|
+
class OutputMskSignatureVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
215
231
|
r"""Signature version to use for signing MSK cluster requests"""
|
|
216
232
|
|
|
217
233
|
V2 = "v2"
|
|
218
234
|
V4 = "v4"
|
|
219
235
|
|
|
220
236
|
|
|
221
|
-
class OutputMskMinimumTLSVersion(str, Enum):
|
|
237
|
+
class OutputMskMinimumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
222
238
|
TL_SV1 = "TLSv1"
|
|
223
239
|
TL_SV1_1 = "TLSv1.1"
|
|
224
240
|
TL_SV1_2 = "TLSv1.2"
|
|
225
241
|
TL_SV1_3 = "TLSv1.3"
|
|
226
242
|
|
|
227
243
|
|
|
228
|
-
class OutputMskMaximumTLSVersion(str, Enum):
|
|
244
|
+
class OutputMskMaximumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
229
245
|
TL_SV1 = "TLSv1"
|
|
230
246
|
TL_SV1_1 = "TLSv1.1"
|
|
231
247
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -285,15 +301,23 @@ class OutputMskTLSSettingsClientSide(BaseModel):
|
|
|
285
301
|
r"""Passphrase to use to decrypt private key"""
|
|
286
302
|
|
|
287
303
|
min_version: Annotated[
|
|
288
|
-
|
|
304
|
+
Annotated[
|
|
305
|
+
Optional[OutputMskMinimumTLSVersion],
|
|
306
|
+
PlainValidator(validate_open_enum(False)),
|
|
307
|
+
],
|
|
308
|
+
pydantic.Field(alias="minVersion"),
|
|
289
309
|
] = None
|
|
290
310
|
|
|
291
311
|
max_version: Annotated[
|
|
292
|
-
|
|
312
|
+
Annotated[
|
|
313
|
+
Optional[OutputMskMaximumTLSVersion],
|
|
314
|
+
PlainValidator(validate_open_enum(False)),
|
|
315
|
+
],
|
|
316
|
+
pydantic.Field(alias="maxVersion"),
|
|
293
317
|
] = None
|
|
294
318
|
|
|
295
319
|
|
|
296
|
-
class OutputMskBackpressureBehavior(str, Enum):
|
|
320
|
+
class OutputMskBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
297
321
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
298
322
|
|
|
299
323
|
BLOCK = "block"
|
|
@@ -301,21 +325,21 @@ class OutputMskBackpressureBehavior(str, Enum):
|
|
|
301
325
|
QUEUE = "queue"
|
|
302
326
|
|
|
303
327
|
|
|
304
|
-
class OutputMskPqCompressCompression(str, Enum):
|
|
328
|
+
class OutputMskPqCompressCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
305
329
|
r"""Codec to use to compress the persisted data"""
|
|
306
330
|
|
|
307
331
|
NONE = "none"
|
|
308
332
|
GZIP = "gzip"
|
|
309
333
|
|
|
310
334
|
|
|
311
|
-
class OutputMskQueueFullBehavior(str, Enum):
|
|
335
|
+
class OutputMskQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
312
336
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
313
337
|
|
|
314
338
|
BLOCK = "block"
|
|
315
339
|
DROP = "drop"
|
|
316
340
|
|
|
317
341
|
|
|
318
|
-
class OutputMskMode(str, Enum):
|
|
342
|
+
class OutputMskMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
319
343
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
320
344
|
|
|
321
345
|
ERROR = "error"
|
|
@@ -452,15 +476,23 @@ class OutputMsk(BaseModel):
|
|
|
452
476
|
streamtags: Optional[List[str]] = None
|
|
453
477
|
r"""Tags for filtering and grouping in @{product}"""
|
|
454
478
|
|
|
455
|
-
ack:
|
|
479
|
+
ack: Annotated[
|
|
480
|
+
Optional[OutputMskAcknowledgments], PlainValidator(validate_open_enum(True))
|
|
481
|
+
] = OutputMskAcknowledgments.ONE
|
|
456
482
|
r"""Control the number of required acknowledgments."""
|
|
457
483
|
|
|
458
484
|
format_: Annotated[
|
|
459
|
-
|
|
485
|
+
Annotated[
|
|
486
|
+
Optional[OutputMskRecordDataFormat],
|
|
487
|
+
PlainValidator(validate_open_enum(False)),
|
|
488
|
+
],
|
|
489
|
+
pydantic.Field(alias="format"),
|
|
460
490
|
] = OutputMskRecordDataFormat.JSON
|
|
461
491
|
r"""Format to use to serialize events before writing to Kafka."""
|
|
462
492
|
|
|
463
|
-
compression:
|
|
493
|
+
compression: Annotated[
|
|
494
|
+
Optional[OutputMskCompression], PlainValidator(validate_open_enum(False))
|
|
495
|
+
] = OutputMskCompression.GZIP
|
|
464
496
|
r"""Codec to use to compress the data before sending to Kafka"""
|
|
465
497
|
|
|
466
498
|
max_record_size_kb: Annotated[
|
|
@@ -518,7 +550,10 @@ class OutputMsk(BaseModel):
|
|
|
518
550
|
r"""Specifies a time window during which @{product} can reauthenticate if needed. Creates the window measuring backward from the moment when credentials are set to expire."""
|
|
519
551
|
|
|
520
552
|
aws_authentication_method: Annotated[
|
|
521
|
-
|
|
553
|
+
Annotated[
|
|
554
|
+
Optional[OutputMskAuthenticationMethod],
|
|
555
|
+
PlainValidator(validate_open_enum(False)),
|
|
556
|
+
],
|
|
522
557
|
pydantic.Field(alias="awsAuthenticationMethod"),
|
|
523
558
|
] = OutputMskAuthenticationMethod.AUTO
|
|
524
559
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
@@ -531,7 +566,11 @@ class OutputMsk(BaseModel):
|
|
|
531
566
|
r"""MSK cluster service endpoint. If empty, defaults to the AWS Region-specific endpoint. Otherwise, it must point to MSK cluster-compatible endpoint."""
|
|
532
567
|
|
|
533
568
|
signature_version: Annotated[
|
|
534
|
-
|
|
569
|
+
Annotated[
|
|
570
|
+
Optional[OutputMskSignatureVersion],
|
|
571
|
+
PlainValidator(validate_open_enum(False)),
|
|
572
|
+
],
|
|
573
|
+
pydantic.Field(alias="signatureVersion"),
|
|
535
574
|
] = OutputMskSignatureVersion.V4
|
|
536
575
|
r"""Signature version to use for signing MSK cluster requests"""
|
|
537
576
|
|
|
@@ -568,7 +607,11 @@ class OutputMsk(BaseModel):
|
|
|
568
607
|
tls: Optional[OutputMskTLSSettingsClientSide] = None
|
|
569
608
|
|
|
570
609
|
on_backpressure: Annotated[
|
|
571
|
-
|
|
610
|
+
Annotated[
|
|
611
|
+
Optional[OutputMskBackpressureBehavior],
|
|
612
|
+
PlainValidator(validate_open_enum(False)),
|
|
613
|
+
],
|
|
614
|
+
pydantic.Field(alias="onBackpressure"),
|
|
572
615
|
] = OutputMskBackpressureBehavior.BLOCK
|
|
573
616
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
574
617
|
|
|
@@ -598,18 +641,27 @@ class OutputMsk(BaseModel):
|
|
|
598
641
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
599
642
|
|
|
600
643
|
pq_compress: Annotated[
|
|
601
|
-
|
|
644
|
+
Annotated[
|
|
645
|
+
Optional[OutputMskPqCompressCompression],
|
|
646
|
+
PlainValidator(validate_open_enum(False)),
|
|
647
|
+
],
|
|
648
|
+
pydantic.Field(alias="pqCompress"),
|
|
602
649
|
] = OutputMskPqCompressCompression.NONE
|
|
603
650
|
r"""Codec to use to compress the persisted data"""
|
|
604
651
|
|
|
605
652
|
pq_on_backpressure: Annotated[
|
|
606
|
-
|
|
653
|
+
Annotated[
|
|
654
|
+
Optional[OutputMskQueueFullBehavior],
|
|
655
|
+
PlainValidator(validate_open_enum(False)),
|
|
656
|
+
],
|
|
657
|
+
pydantic.Field(alias="pqOnBackpressure"),
|
|
607
658
|
] = OutputMskQueueFullBehavior.BLOCK
|
|
608
659
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
609
660
|
|
|
610
|
-
pq_mode: Annotated[
|
|
611
|
-
OutputMskMode
|
|
612
|
-
|
|
661
|
+
pq_mode: Annotated[
|
|
662
|
+
Annotated[Optional[OutputMskMode], PlainValidator(validate_open_enum(False))],
|
|
663
|
+
pydantic.Field(alias="pqMode"),
|
|
664
|
+
] = OutputMskMode.ERROR
|
|
613
665
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
614
666
|
|
|
615
667
|
pq_controls: Annotated[
|
|
@@ -1,9 +1,12 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from cribl_control_plane import utils
|
|
4
5
|
from cribl_control_plane.types import BaseModel
|
|
6
|
+
from cribl_control_plane.utils import validate_open_enum
|
|
5
7
|
from enum import Enum
|
|
6
8
|
import pydantic
|
|
9
|
+
from pydantic.functional_validators import PlainValidator
|
|
7
10
|
from typing import List, Optional
|
|
8
11
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
12
|
|
|
@@ -12,7 +15,7 @@ class OutputNewrelicType(str, Enum):
|
|
|
12
15
|
NEWRELIC = "newrelic"
|
|
13
16
|
|
|
14
17
|
|
|
15
|
-
class OutputNewrelicRegion(str, Enum):
|
|
18
|
+
class OutputNewrelicRegion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
16
19
|
r"""Which New Relic region endpoint to use."""
|
|
17
20
|
|
|
18
21
|
US = "US"
|
|
@@ -20,7 +23,7 @@ class OutputNewrelicRegion(str, Enum):
|
|
|
20
23
|
CUSTOM = "Custom"
|
|
21
24
|
|
|
22
25
|
|
|
23
|
-
class FieldName(str, Enum):
|
|
26
|
+
class FieldName(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
24
27
|
SERVICE = "service"
|
|
25
28
|
HOSTNAME = "hostname"
|
|
26
29
|
TIMESTAMP = "timestamp"
|
|
@@ -34,7 +37,7 @@ class OutputNewrelicMetadatumTypedDict(TypedDict):
|
|
|
34
37
|
|
|
35
38
|
|
|
36
39
|
class OutputNewrelicMetadatum(BaseModel):
|
|
37
|
-
name: FieldName
|
|
40
|
+
name: Annotated[FieldName, PlainValidator(validate_open_enum(False))]
|
|
38
41
|
|
|
39
42
|
value: str
|
|
40
43
|
r"""JavaScript expression to compute field's value, enclosed in quotes or backticks. (Can evaluate to a constant.)"""
|
|
@@ -51,7 +54,7 @@ class OutputNewrelicExtraHTTPHeader(BaseModel):
|
|
|
51
54
|
name: Optional[str] = None
|
|
52
55
|
|
|
53
56
|
|
|
54
|
-
class OutputNewrelicFailedRequestLoggingMode(str, Enum):
|
|
57
|
+
class OutputNewrelicFailedRequestLoggingMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
55
58
|
r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
|
|
56
59
|
|
|
57
60
|
PAYLOAD = "payload"
|
|
@@ -113,7 +116,7 @@ class OutputNewrelicTimeoutRetrySettings(BaseModel):
|
|
|
113
116
|
r"""The maximum backoff interval, in milliseconds, Cribl Stream should apply. Default (and minimum) is 10,000 ms (10 seconds); maximum is 180,000 ms (180 seconds)."""
|
|
114
117
|
|
|
115
118
|
|
|
116
|
-
class OutputNewrelicBackpressureBehavior(str, Enum):
|
|
119
|
+
class OutputNewrelicBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
117
120
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
118
121
|
|
|
119
122
|
BLOCK = "block"
|
|
@@ -121,28 +124,28 @@ class OutputNewrelicBackpressureBehavior(str, Enum):
|
|
|
121
124
|
QUEUE = "queue"
|
|
122
125
|
|
|
123
126
|
|
|
124
|
-
class OutputNewrelicAuthenticationMethod(str, Enum):
|
|
127
|
+
class OutputNewrelicAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
125
128
|
r"""Enter API key directly, or select a stored secret"""
|
|
126
129
|
|
|
127
130
|
MANUAL = "manual"
|
|
128
131
|
SECRET = "secret"
|
|
129
132
|
|
|
130
133
|
|
|
131
|
-
class OutputNewrelicCompression(str, Enum):
|
|
134
|
+
class OutputNewrelicCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
132
135
|
r"""Codec to use to compress the persisted data"""
|
|
133
136
|
|
|
134
137
|
NONE = "none"
|
|
135
138
|
GZIP = "gzip"
|
|
136
139
|
|
|
137
140
|
|
|
138
|
-
class OutputNewrelicQueueFullBehavior(str, Enum):
|
|
141
|
+
class OutputNewrelicQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
139
142
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
140
143
|
|
|
141
144
|
BLOCK = "block"
|
|
142
145
|
DROP = "drop"
|
|
143
146
|
|
|
144
147
|
|
|
145
|
-
class OutputNewrelicMode(str, Enum):
|
|
148
|
+
class OutputNewrelicMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
146
149
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
147
150
|
|
|
148
151
|
ERROR = "error"
|
|
@@ -257,7 +260,9 @@ class OutputNewrelic(BaseModel):
|
|
|
257
260
|
streamtags: Optional[List[str]] = None
|
|
258
261
|
r"""Tags for filtering and grouping in @{product}"""
|
|
259
262
|
|
|
260
|
-
region:
|
|
263
|
+
region: Annotated[
|
|
264
|
+
Optional[OutputNewrelicRegion], PlainValidator(validate_open_enum(False))
|
|
265
|
+
] = OutputNewrelicRegion.US
|
|
261
266
|
r"""Which New Relic region endpoint to use."""
|
|
262
267
|
|
|
263
268
|
log_type: Annotated[Optional[str], pydantic.Field(alias="logType")] = ""
|
|
@@ -313,7 +318,10 @@ class OutputNewrelic(BaseModel):
|
|
|
313
318
|
r"""Enable round-robin DNS lookup. When a DNS server returns multiple addresses, @{product} will cycle through them in the order returned. For optimal performance, consider enabling this setting for non-load balanced destinations."""
|
|
314
319
|
|
|
315
320
|
failed_request_logging_mode: Annotated[
|
|
316
|
-
|
|
321
|
+
Annotated[
|
|
322
|
+
Optional[OutputNewrelicFailedRequestLoggingMode],
|
|
323
|
+
PlainValidator(validate_open_enum(False)),
|
|
324
|
+
],
|
|
317
325
|
pydantic.Field(alias="failedRequestLoggingMode"),
|
|
318
326
|
] = OutputNewrelicFailedRequestLoggingMode.NONE
|
|
319
327
|
r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
|
|
@@ -340,13 +348,20 @@ class OutputNewrelic(BaseModel):
|
|
|
340
348
|
r"""Honor any Retry-After header that specifies a delay (in seconds) no longer than 180 seconds after the retry request. @{product} limits the delay to 180 seconds, even if the Retry-After header specifies a longer delay. When enabled, takes precedence over user-configured retry options. When disabled, all Retry-After headers are ignored."""
|
|
341
349
|
|
|
342
350
|
on_backpressure: Annotated[
|
|
343
|
-
|
|
351
|
+
Annotated[
|
|
352
|
+
Optional[OutputNewrelicBackpressureBehavior],
|
|
353
|
+
PlainValidator(validate_open_enum(False)),
|
|
354
|
+
],
|
|
344
355
|
pydantic.Field(alias="onBackpressure"),
|
|
345
356
|
] = OutputNewrelicBackpressureBehavior.BLOCK
|
|
346
357
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
347
358
|
|
|
348
359
|
auth_type: Annotated[
|
|
349
|
-
|
|
360
|
+
Annotated[
|
|
361
|
+
Optional[OutputNewrelicAuthenticationMethod],
|
|
362
|
+
PlainValidator(validate_open_enum(False)),
|
|
363
|
+
],
|
|
364
|
+
pydantic.Field(alias="authType"),
|
|
350
365
|
] = OutputNewrelicAuthenticationMethod.MANUAL
|
|
351
366
|
r"""Enter API key directly, or select a stored secret"""
|
|
352
367
|
|
|
@@ -373,19 +388,29 @@ class OutputNewrelic(BaseModel):
|
|
|
373
388
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
374
389
|
|
|
375
390
|
pq_compress: Annotated[
|
|
376
|
-
|
|
391
|
+
Annotated[
|
|
392
|
+
Optional[OutputNewrelicCompression],
|
|
393
|
+
PlainValidator(validate_open_enum(False)),
|
|
394
|
+
],
|
|
395
|
+
pydantic.Field(alias="pqCompress"),
|
|
377
396
|
] = OutputNewrelicCompression.NONE
|
|
378
397
|
r"""Codec to use to compress the persisted data"""
|
|
379
398
|
|
|
380
399
|
pq_on_backpressure: Annotated[
|
|
381
|
-
|
|
400
|
+
Annotated[
|
|
401
|
+
Optional[OutputNewrelicQueueFullBehavior],
|
|
402
|
+
PlainValidator(validate_open_enum(False)),
|
|
403
|
+
],
|
|
382
404
|
pydantic.Field(alias="pqOnBackpressure"),
|
|
383
405
|
] = OutputNewrelicQueueFullBehavior.BLOCK
|
|
384
406
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
385
407
|
|
|
386
|
-
pq_mode: Annotated[
|
|
387
|
-
|
|
388
|
-
|
|
408
|
+
pq_mode: Annotated[
|
|
409
|
+
Annotated[
|
|
410
|
+
Optional[OutputNewrelicMode], PlainValidator(validate_open_enum(False))
|
|
411
|
+
],
|
|
412
|
+
pydantic.Field(alias="pqMode"),
|
|
413
|
+
] = OutputNewrelicMode.ERROR
|
|
389
414
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
390
415
|
|
|
391
416
|
pq_controls: Annotated[
|
|
@@ -1,9 +1,12 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from cribl_control_plane import utils
|
|
4
5
|
from cribl_control_plane.types import BaseModel
|
|
6
|
+
from cribl_control_plane.utils import validate_open_enum
|
|
5
7
|
from enum import Enum
|
|
6
8
|
import pydantic
|
|
9
|
+
from pydantic.functional_validators import PlainValidator
|
|
7
10
|
from typing import List, Optional
|
|
8
11
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
12
|
|
|
@@ -12,7 +15,7 @@ class OutputNewrelicEventsType(str, Enum):
|
|
|
12
15
|
NEWRELIC_EVENTS = "newrelic_events"
|
|
13
16
|
|
|
14
17
|
|
|
15
|
-
class OutputNewrelicEventsRegion(str, Enum):
|
|
18
|
+
class OutputNewrelicEventsRegion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
16
19
|
r"""Which New Relic region endpoint to use."""
|
|
17
20
|
|
|
18
21
|
US = "US"
|
|
@@ -31,7 +34,9 @@ class OutputNewrelicEventsExtraHTTPHeader(BaseModel):
|
|
|
31
34
|
name: Optional[str] = None
|
|
32
35
|
|
|
33
36
|
|
|
34
|
-
class OutputNewrelicEventsFailedRequestLoggingMode(
|
|
37
|
+
class OutputNewrelicEventsFailedRequestLoggingMode(
|
|
38
|
+
str, Enum, metaclass=utils.OpenEnumMeta
|
|
39
|
+
):
|
|
35
40
|
r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
|
|
36
41
|
|
|
37
42
|
PAYLOAD = "payload"
|
|
@@ -93,7 +98,7 @@ class OutputNewrelicEventsTimeoutRetrySettings(BaseModel):
|
|
|
93
98
|
r"""The maximum backoff interval, in milliseconds, Cribl Stream should apply. Default (and minimum) is 10,000 ms (10 seconds); maximum is 180,000 ms (180 seconds)."""
|
|
94
99
|
|
|
95
100
|
|
|
96
|
-
class OutputNewrelicEventsBackpressureBehavior(str, Enum):
|
|
101
|
+
class OutputNewrelicEventsBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
97
102
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
98
103
|
|
|
99
104
|
BLOCK = "block"
|
|
@@ -101,28 +106,28 @@ class OutputNewrelicEventsBackpressureBehavior(str, Enum):
|
|
|
101
106
|
QUEUE = "queue"
|
|
102
107
|
|
|
103
108
|
|
|
104
|
-
class OutputNewrelicEventsAuthenticationMethod(str, Enum):
|
|
109
|
+
class OutputNewrelicEventsAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
105
110
|
r"""Enter API key directly, or select a stored secret"""
|
|
106
111
|
|
|
107
112
|
MANUAL = "manual"
|
|
108
113
|
SECRET = "secret"
|
|
109
114
|
|
|
110
115
|
|
|
111
|
-
class OutputNewrelicEventsCompression(str, Enum):
|
|
116
|
+
class OutputNewrelicEventsCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
112
117
|
r"""Codec to use to compress the persisted data"""
|
|
113
118
|
|
|
114
119
|
NONE = "none"
|
|
115
120
|
GZIP = "gzip"
|
|
116
121
|
|
|
117
122
|
|
|
118
|
-
class OutputNewrelicEventsQueueFullBehavior(str, Enum):
|
|
123
|
+
class OutputNewrelicEventsQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
119
124
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
120
125
|
|
|
121
126
|
BLOCK = "block"
|
|
122
127
|
DROP = "drop"
|
|
123
128
|
|
|
124
129
|
|
|
125
|
-
class OutputNewrelicEventsMode(str, Enum):
|
|
130
|
+
class OutputNewrelicEventsMode(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
126
131
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
127
132
|
|
|
128
133
|
ERROR = "error"
|
|
@@ -243,7 +248,9 @@ class OutputNewrelicEvents(BaseModel):
|
|
|
243
248
|
streamtags: Optional[List[str]] = None
|
|
244
249
|
r"""Tags for filtering and grouping in @{product}"""
|
|
245
250
|
|
|
246
|
-
region:
|
|
251
|
+
region: Annotated[
|
|
252
|
+
Optional[OutputNewrelicEventsRegion], PlainValidator(validate_open_enum(False))
|
|
253
|
+
] = OutputNewrelicEventsRegion.US
|
|
247
254
|
r"""Which New Relic region endpoint to use."""
|
|
248
255
|
|
|
249
256
|
concurrency: Optional[float] = 5
|
|
@@ -290,7 +297,10 @@ class OutputNewrelicEvents(BaseModel):
|
|
|
290
297
|
r"""Enable round-robin DNS lookup. When a DNS server returns multiple addresses, @{product} will cycle through them in the order returned. For optimal performance, consider enabling this setting for non-load balanced destinations."""
|
|
291
298
|
|
|
292
299
|
failed_request_logging_mode: Annotated[
|
|
293
|
-
|
|
300
|
+
Annotated[
|
|
301
|
+
Optional[OutputNewrelicEventsFailedRequestLoggingMode],
|
|
302
|
+
PlainValidator(validate_open_enum(False)),
|
|
303
|
+
],
|
|
294
304
|
pydantic.Field(alias="failedRequestLoggingMode"),
|
|
295
305
|
] = OutputNewrelicEventsFailedRequestLoggingMode.NONE
|
|
296
306
|
r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
|
|
@@ -313,17 +323,23 @@ class OutputNewrelicEvents(BaseModel):
|
|
|
313
323
|
|
|
314
324
|
response_honor_retry_after_header: Annotated[
|
|
315
325
|
Optional[bool], pydantic.Field(alias="responseHonorRetryAfterHeader")
|
|
316
|
-
] =
|
|
326
|
+
] = True
|
|
317
327
|
r"""Honor any Retry-After header that specifies a delay (in seconds) no longer than 180 seconds after the retry request. @{product} limits the delay to 180 seconds, even if the Retry-After header specifies a longer delay. When enabled, takes precedence over user-configured retry options. When disabled, all Retry-After headers are ignored."""
|
|
318
328
|
|
|
319
329
|
on_backpressure: Annotated[
|
|
320
|
-
|
|
330
|
+
Annotated[
|
|
331
|
+
Optional[OutputNewrelicEventsBackpressureBehavior],
|
|
332
|
+
PlainValidator(validate_open_enum(False)),
|
|
333
|
+
],
|
|
321
334
|
pydantic.Field(alias="onBackpressure"),
|
|
322
335
|
] = OutputNewrelicEventsBackpressureBehavior.BLOCK
|
|
323
336
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
324
337
|
|
|
325
338
|
auth_type: Annotated[
|
|
326
|
-
|
|
339
|
+
Annotated[
|
|
340
|
+
Optional[OutputNewrelicEventsAuthenticationMethod],
|
|
341
|
+
PlainValidator(validate_open_enum(False)),
|
|
342
|
+
],
|
|
327
343
|
pydantic.Field(alias="authType"),
|
|
328
344
|
] = OutputNewrelicEventsAuthenticationMethod.MANUAL
|
|
329
345
|
r"""Enter API key directly, or select a stored secret"""
|
|
@@ -346,18 +362,29 @@ class OutputNewrelicEvents(BaseModel):
|
|
|
346
362
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/<output-id>."""
|
|
347
363
|
|
|
348
364
|
pq_compress: Annotated[
|
|
349
|
-
|
|
365
|
+
Annotated[
|
|
366
|
+
Optional[OutputNewrelicEventsCompression],
|
|
367
|
+
PlainValidator(validate_open_enum(False)),
|
|
368
|
+
],
|
|
369
|
+
pydantic.Field(alias="pqCompress"),
|
|
350
370
|
] = OutputNewrelicEventsCompression.NONE
|
|
351
371
|
r"""Codec to use to compress the persisted data"""
|
|
352
372
|
|
|
353
373
|
pq_on_backpressure: Annotated[
|
|
354
|
-
|
|
374
|
+
Annotated[
|
|
375
|
+
Optional[OutputNewrelicEventsQueueFullBehavior],
|
|
376
|
+
PlainValidator(validate_open_enum(False)),
|
|
377
|
+
],
|
|
355
378
|
pydantic.Field(alias="pqOnBackpressure"),
|
|
356
379
|
] = OutputNewrelicEventsQueueFullBehavior.BLOCK
|
|
357
380
|
r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
|
|
358
381
|
|
|
359
382
|
pq_mode: Annotated[
|
|
360
|
-
|
|
383
|
+
Annotated[
|
|
384
|
+
Optional[OutputNewrelicEventsMode],
|
|
385
|
+
PlainValidator(validate_open_enum(False)),
|
|
386
|
+
],
|
|
387
|
+
pydantic.Field(alias="pqMode"),
|
|
361
388
|
] = OutputNewrelicEventsMode.ERROR
|
|
362
389
|
r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
|
|
363
390
|
|