cribl-control-plane 0.0.48a1__py3-none-any.whl → 0.0.49__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +6 -4
- cribl_control_plane/errors/healthstatus_error.py +2 -8
- cribl_control_plane/models/__init__.py +12 -12
- cribl_control_plane/models/appmode.py +13 -0
- cribl_control_plane/models/cacheconnection.py +2 -10
- cribl_control_plane/models/cacheconnectionbackfillstatus.py +1 -2
- cribl_control_plane/models/cloudprovider.py +1 -2
- cribl_control_plane/models/configgroup.py +2 -7
- cribl_control_plane/models/configgroupcloud.py +2 -6
- cribl_control_plane/models/createconfiggroupbyproductop.py +2 -8
- cribl_control_plane/models/cribllakedataset.py +2 -8
- cribl_control_plane/models/datasetmetadata.py +2 -8
- cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +2 -7
- cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +2 -4
- cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -3
- cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +2 -7
- cribl_control_plane/models/getsummaryop.py +2 -7
- cribl_control_plane/models/hbcriblinfo.py +3 -19
- cribl_control_plane/models/healthstatus.py +4 -7
- cribl_control_plane/models/heartbeatmetadata.py +0 -3
- cribl_control_plane/models/inputappscope.py +14 -34
- cribl_control_plane/models/inputazureblob.py +6 -17
- cribl_control_plane/models/inputcollection.py +4 -11
- cribl_control_plane/models/inputconfluentcloud.py +20 -47
- cribl_control_plane/models/inputcribl.py +4 -11
- cribl_control_plane/models/inputcriblhttp.py +8 -23
- cribl_control_plane/models/inputcribllakehttp.py +10 -22
- cribl_control_plane/models/inputcriblmetrics.py +4 -12
- cribl_control_plane/models/inputcribltcp.py +8 -23
- cribl_control_plane/models/inputcrowdstrike.py +10 -26
- cribl_control_plane/models/inputdatadogagent.py +8 -24
- cribl_control_plane/models/inputdatagen.py +4 -11
- cribl_control_plane/models/inputedgeprometheus.py +24 -58
- cribl_control_plane/models/inputelastic.py +14 -40
- cribl_control_plane/models/inputeventhub.py +6 -15
- cribl_control_plane/models/inputexec.py +6 -14
- cribl_control_plane/models/inputfile.py +6 -15
- cribl_control_plane/models/inputfirehose.py +8 -23
- cribl_control_plane/models/inputgooglepubsub.py +6 -19
- cribl_control_plane/models/inputgrafana.py +24 -67
- cribl_control_plane/models/inputhttp.py +8 -23
- cribl_control_plane/models/inputhttpraw.py +8 -23
- cribl_control_plane/models/inputjournalfiles.py +4 -12
- cribl_control_plane/models/inputkafka.py +16 -46
- cribl_control_plane/models/inputkinesis.py +14 -38
- cribl_control_plane/models/inputkubeevents.py +4 -11
- cribl_control_plane/models/inputkubelogs.py +8 -16
- cribl_control_plane/models/inputkubemetrics.py +8 -16
- cribl_control_plane/models/inputloki.py +10 -29
- cribl_control_plane/models/inputmetrics.py +8 -23
- cribl_control_plane/models/inputmodeldriventelemetry.py +10 -32
- cribl_control_plane/models/inputmsk.py +18 -53
- cribl_control_plane/models/inputnetflow.py +4 -11
- cribl_control_plane/models/inputoffice365mgmt.py +14 -33
- cribl_control_plane/models/inputoffice365msgtrace.py +16 -35
- cribl_control_plane/models/inputoffice365service.py +16 -35
- cribl_control_plane/models/inputopentelemetry.py +16 -38
- cribl_control_plane/models/inputprometheus.py +18 -50
- cribl_control_plane/models/inputprometheusrw.py +10 -30
- cribl_control_plane/models/inputrawudp.py +4 -11
- cribl_control_plane/models/inputs3.py +8 -21
- cribl_control_plane/models/inputs3inventory.py +10 -26
- cribl_control_plane/models/inputsecuritylake.py +10 -27
- cribl_control_plane/models/inputsnmp.py +6 -16
- cribl_control_plane/models/inputsplunk.py +12 -33
- cribl_control_plane/models/inputsplunkhec.py +10 -29
- cribl_control_plane/models/inputsplunksearch.py +14 -33
- cribl_control_plane/models/inputsqs.py +10 -27
- cribl_control_plane/models/inputsyslog.py +16 -43
- cribl_control_plane/models/inputsystemmetrics.py +24 -48
- cribl_control_plane/models/inputsystemstate.py +8 -16
- cribl_control_plane/models/inputtcp.py +10 -29
- cribl_control_plane/models/inputtcpjson.py +10 -29
- cribl_control_plane/models/inputwef.py +14 -37
- cribl_control_plane/models/inputwindowsmetrics.py +24 -44
- cribl_control_plane/models/inputwineventlogs.py +10 -20
- cribl_control_plane/models/inputwiz.py +8 -21
- cribl_control_plane/models/inputwizwebhook.py +8 -23
- cribl_control_plane/models/inputzscalerhec.py +10 -29
- cribl_control_plane/models/lakehouseconnectiontype.py +1 -2
- cribl_control_plane/models/listconfiggroupbyproductop.py +1 -3
- cribl_control_plane/models/masterworkerentry.py +2 -7
- cribl_control_plane/models/nodeactiveupgradestatus.py +1 -2
- cribl_control_plane/models/nodefailedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeprovidedinfo.py +0 -3
- cribl_control_plane/models/nodeskippedupgradestatus.py +1 -2
- cribl_control_plane/models/nodeupgradestate.py +1 -2
- cribl_control_plane/models/nodeupgradestatus.py +5 -13
- cribl_control_plane/models/outputazureblob.py +18 -48
- cribl_control_plane/models/outputazuredataexplorer.py +28 -73
- cribl_control_plane/models/outputazureeventhub.py +18 -40
- cribl_control_plane/models/outputazurelogs.py +12 -35
- cribl_control_plane/models/outputclickhouse.py +20 -55
- cribl_control_plane/models/outputcloudwatch.py +10 -29
- cribl_control_plane/models/outputconfluentcloud.py +32 -77
- cribl_control_plane/models/outputcriblhttp.py +16 -44
- cribl_control_plane/models/outputcribllake.py +16 -46
- cribl_control_plane/models/outputcribltcp.py +18 -45
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +14 -49
- cribl_control_plane/models/outputdatadog.py +20 -48
- cribl_control_plane/models/outputdataset.py +18 -46
- cribl_control_plane/models/outputdiskspool.py +2 -7
- cribl_control_plane/models/outputdls3.py +24 -68
- cribl_control_plane/models/outputdynatracehttp.py +20 -53
- cribl_control_plane/models/outputdynatraceotlp.py +22 -55
- cribl_control_plane/models/outputelastic.py +18 -43
- cribl_control_plane/models/outputelasticcloud.py +12 -36
- cribl_control_plane/models/outputexabeam.py +10 -29
- cribl_control_plane/models/outputfilesystem.py +14 -39
- cribl_control_plane/models/outputgooglechronicle.py +16 -50
- cribl_control_plane/models/outputgooglecloudlogging.py +14 -41
- cribl_control_plane/models/outputgooglecloudstorage.py +24 -66
- cribl_control_plane/models/outputgooglepubsub.py +10 -31
- cribl_control_plane/models/outputgrafanacloud.py +32 -97
- cribl_control_plane/models/outputgraphite.py +14 -31
- cribl_control_plane/models/outputhoneycomb.py +12 -35
- cribl_control_plane/models/outputhumiohec.py +16 -43
- cribl_control_plane/models/outputinfluxdb.py +16 -42
- cribl_control_plane/models/outputkafka.py +28 -74
- cribl_control_plane/models/outputkinesis.py +16 -40
- cribl_control_plane/models/outputloki.py +16 -41
- cribl_control_plane/models/outputminio.py +24 -65
- cribl_control_plane/models/outputmsk.py +30 -82
- cribl_control_plane/models/outputnewrelic.py +18 -43
- cribl_control_plane/models/outputnewrelicevents.py +14 -41
- cribl_control_plane/models/outputopentelemetry.py +26 -67
- cribl_control_plane/models/outputprometheus.py +12 -35
- cribl_control_plane/models/outputring.py +8 -19
- cribl_control_plane/models/outputs3.py +26 -68
- cribl_control_plane/models/outputsecuritylake.py +18 -52
- cribl_control_plane/models/outputsentinel.py +18 -45
- cribl_control_plane/models/outputsentineloneaisiem.py +18 -50
- cribl_control_plane/models/outputservicenow.py +24 -60
- cribl_control_plane/models/outputsignalfx.py +14 -37
- cribl_control_plane/models/outputsns.py +14 -36
- cribl_control_plane/models/outputsplunk.py +24 -60
- cribl_control_plane/models/outputsplunkhec.py +12 -35
- cribl_control_plane/models/outputsplunklb.py +30 -77
- cribl_control_plane/models/outputsqs.py +16 -41
- cribl_control_plane/models/outputstatsd.py +14 -30
- cribl_control_plane/models/outputstatsdext.py +12 -29
- cribl_control_plane/models/outputsumologic.py +12 -35
- cribl_control_plane/models/outputsyslog.py +24 -58
- cribl_control_plane/models/outputtcpjson.py +20 -52
- cribl_control_plane/models/outputwavefront.py +12 -35
- cribl_control_plane/models/outputwebhook.py +22 -58
- cribl_control_plane/models/outputxsiam.py +14 -35
- cribl_control_plane/models/productscore.py +1 -2
- cribl_control_plane/models/rbacresource.py +1 -2
- cribl_control_plane/models/resourcepolicy.py +2 -4
- cribl_control_plane/models/routecloneconf.py +13 -0
- cribl_control_plane/models/routeconf.py +4 -3
- cribl_control_plane/models/runnablejobcollection.py +13 -30
- cribl_control_plane/models/runnablejobexecutor.py +4 -13
- cribl_control_plane/models/runnablejobscheduledsearch.py +2 -7
- cribl_control_plane/models/updateconfiggroupbyproductandidop.py +2 -8
- cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +2 -8
- cribl_control_plane/models/workertypes.py +1 -2
- cribl_control_plane/sdk.py +2 -2
- cribl_control_plane/utils/annotations.py +32 -8
- {cribl_control_plane-0.0.48a1.dist-info → cribl_control_plane-0.0.49.dist-info}/METADATA +2 -1
- {cribl_control_plane-0.0.48a1.dist-info → cribl_control_plane-0.0.49.dist-info}/RECORD +164 -162
- {cribl_control_plane-0.0.48a1.dist-info → cribl_control_plane-0.0.49.dist-info}/WHEEL +0 -0
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputS3InventoryConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputS3InventoryMode(str, Enum
|
|
26
|
+
class InputS3InventoryMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputS3InventoryCompression(str, Enum
|
|
33
|
+
class InputS3InventoryCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputS3InventoryPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputS3InventoryPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputS3InventoryMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputS3InventoryMode.ALWAYS
|
|
67
|
+
mode: Optional[InputS3InventoryMode] = InputS3InventoryMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputS3InventoryPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputS3InventoryCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputS3InventoryCompression.NONE
|
|
91
|
+
compress: Optional[InputS3InventoryCompression] = InputS3InventoryCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,7 +96,7 @@ class InputS3InventoryPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class InputS3InventoryAuthenticationMethod(str, Enum
|
|
99
|
+
class InputS3InventoryAuthenticationMethod(str, Enum):
|
|
107
100
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
108
101
|
|
|
109
102
|
AUTO = "auto"
|
|
@@ -111,7 +104,7 @@ class InputS3InventoryAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMe
|
|
|
111
104
|
SECRET = "secret"
|
|
112
105
|
|
|
113
106
|
|
|
114
|
-
class InputS3InventorySignatureVersion(str, Enum
|
|
107
|
+
class InputS3InventorySignatureVersion(str, Enum):
|
|
115
108
|
r"""Signature version to use for signing S3 requests"""
|
|
116
109
|
|
|
117
110
|
V2 = "v2"
|
|
@@ -164,7 +157,7 @@ class InputS3InventoryCheckpointing(BaseModel):
|
|
|
164
157
|
r"""The number of times to retry processing when a processing error occurs. If Skip file on error is enabled, this setting is ignored."""
|
|
165
158
|
|
|
166
159
|
|
|
167
|
-
class InputS3InventoryTagAfterProcessing(str, Enum
|
|
160
|
+
class InputS3InventoryTagAfterProcessing(str, Enum):
|
|
168
161
|
FALSE = "false"
|
|
169
162
|
TRUE = "true"
|
|
170
163
|
|
|
@@ -301,10 +294,7 @@ class InputS3Inventory(BaseModel):
|
|
|
301
294
|
r"""SQS queue owner's AWS account ID. Leave empty if SQS queue is in same AWS account."""
|
|
302
295
|
|
|
303
296
|
aws_authentication_method: Annotated[
|
|
304
|
-
|
|
305
|
-
Optional[InputS3InventoryAuthenticationMethod],
|
|
306
|
-
PlainValidator(validate_open_enum(False)),
|
|
307
|
-
],
|
|
297
|
+
Optional[InputS3InventoryAuthenticationMethod],
|
|
308
298
|
pydantic.Field(alias="awsAuthenticationMethod"),
|
|
309
299
|
] = InputS3InventoryAuthenticationMethod.AUTO
|
|
310
300
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
@@ -320,10 +310,7 @@ class InputS3Inventory(BaseModel):
|
|
|
320
310
|
r"""S3 service endpoint. If empty, defaults to the AWS Region-specific endpoint. Otherwise, it must point to S3-compatible endpoint."""
|
|
321
311
|
|
|
322
312
|
signature_version: Annotated[
|
|
323
|
-
|
|
324
|
-
Optional[InputS3InventorySignatureVersion],
|
|
325
|
-
PlainValidator(validate_open_enum(False)),
|
|
326
|
-
],
|
|
313
|
+
Optional[InputS3InventorySignatureVersion],
|
|
327
314
|
pydantic.Field(alias="signatureVersion"),
|
|
328
315
|
] = InputS3InventorySignatureVersion.V4
|
|
329
316
|
r"""Signature version to use for signing S3 requests"""
|
|
@@ -442,10 +429,7 @@ class InputS3Inventory(BaseModel):
|
|
|
442
429
|
r"""Select or create a stored secret that references your access key and secret key"""
|
|
443
430
|
|
|
444
431
|
tag_after_processing: Annotated[
|
|
445
|
-
|
|
446
|
-
Optional[InputS3InventoryTagAfterProcessing],
|
|
447
|
-
PlainValidator(validate_open_enum(False)),
|
|
448
|
-
],
|
|
432
|
+
Optional[InputS3InventoryTagAfterProcessing],
|
|
449
433
|
pydantic.Field(alias="tagAfterProcessing"),
|
|
450
434
|
] = None
|
|
451
435
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputSecurityLakeConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputSecurityLakeMode(str, Enum
|
|
26
|
+
class InputSecurityLakeMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputSecurityLakeCompression(str, Enum
|
|
33
|
+
class InputSecurityLakeCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputSecurityLakePqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputSecurityLakePq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputSecurityLakeMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputSecurityLakeMode.ALWAYS
|
|
67
|
+
mode: Optional[InputSecurityLakeMode] = InputSecurityLakeMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,10 +88,7 @@ class InputSecurityLakePq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputSecurityLakeCompression],
|
|
98
|
-
PlainValidator(validate_open_enum(False)),
|
|
99
|
-
] = InputSecurityLakeCompression.NONE
|
|
91
|
+
compress: Optional[InputSecurityLakeCompression] = InputSecurityLakeCompression.NONE
|
|
100
92
|
r"""Codec to use to compress the persisted data"""
|
|
101
93
|
|
|
102
94
|
pq_controls: Annotated[
|
|
@@ -104,7 +96,7 @@ class InputSecurityLakePq(BaseModel):
|
|
|
104
96
|
] = None
|
|
105
97
|
|
|
106
98
|
|
|
107
|
-
class InputSecurityLakeAuthenticationMethod(str, Enum
|
|
99
|
+
class InputSecurityLakeAuthenticationMethod(str, Enum):
|
|
108
100
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
109
101
|
|
|
110
102
|
AUTO = "auto"
|
|
@@ -112,7 +104,7 @@ class InputSecurityLakeAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumM
|
|
|
112
104
|
SECRET = "secret"
|
|
113
105
|
|
|
114
106
|
|
|
115
|
-
class InputSecurityLakeSignatureVersion(str, Enum
|
|
107
|
+
class InputSecurityLakeSignatureVersion(str, Enum):
|
|
116
108
|
r"""Signature version to use for signing S3 requests"""
|
|
117
109
|
|
|
118
110
|
V2 = "v2"
|
|
@@ -165,7 +157,7 @@ class InputSecurityLakeCheckpointing(BaseModel):
|
|
|
165
157
|
r"""The number of times to retry processing when a processing error occurs. If Skip file on error is enabled, this setting is ignored."""
|
|
166
158
|
|
|
167
159
|
|
|
168
|
-
class InputSecurityLakeTagAfterProcessing(str, Enum
|
|
160
|
+
class InputSecurityLakeTagAfterProcessing(str, Enum):
|
|
169
161
|
FALSE = "false"
|
|
170
162
|
TRUE = "true"
|
|
171
163
|
|
|
@@ -298,10 +290,7 @@ class InputSecurityLake(BaseModel):
|
|
|
298
290
|
r"""SQS queue owner's AWS account ID. Leave empty if SQS queue is in same AWS account."""
|
|
299
291
|
|
|
300
292
|
aws_authentication_method: Annotated[
|
|
301
|
-
|
|
302
|
-
Optional[InputSecurityLakeAuthenticationMethod],
|
|
303
|
-
PlainValidator(validate_open_enum(False)),
|
|
304
|
-
],
|
|
293
|
+
Optional[InputSecurityLakeAuthenticationMethod],
|
|
305
294
|
pydantic.Field(alias="awsAuthenticationMethod"),
|
|
306
295
|
] = InputSecurityLakeAuthenticationMethod.AUTO
|
|
307
296
|
r"""AWS authentication method. Choose Auto to use IAM roles."""
|
|
@@ -317,10 +306,7 @@ class InputSecurityLake(BaseModel):
|
|
|
317
306
|
r"""S3 service endpoint. If empty, defaults to the AWS Region-specific endpoint. Otherwise, it must point to S3-compatible endpoint."""
|
|
318
307
|
|
|
319
308
|
signature_version: Annotated[
|
|
320
|
-
|
|
321
|
-
Optional[InputSecurityLakeSignatureVersion],
|
|
322
|
-
PlainValidator(validate_open_enum(False)),
|
|
323
|
-
],
|
|
309
|
+
Optional[InputSecurityLakeSignatureVersion],
|
|
324
310
|
pydantic.Field(alias="signatureVersion"),
|
|
325
311
|
] = InputSecurityLakeSignatureVersion.V4
|
|
326
312
|
r"""Signature version to use for signing S3 requests"""
|
|
@@ -427,10 +413,7 @@ class InputSecurityLake(BaseModel):
|
|
|
427
413
|
r"""Select or create a stored secret that references your access key and secret key"""
|
|
428
414
|
|
|
429
415
|
tag_after_processing: Annotated[
|
|
430
|
-
|
|
431
|
-
Optional[InputSecurityLakeTagAfterProcessing],
|
|
432
|
-
PlainValidator(validate_open_enum(False)),
|
|
433
|
-
],
|
|
416
|
+
Optional[InputSecurityLakeTagAfterProcessing],
|
|
434
417
|
pydantic.Field(alias="tagAfterProcessing"),
|
|
435
418
|
] = None
|
|
436
419
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import Any, List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputSnmpConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputSnmpMode(str, Enum
|
|
26
|
+
class InputSnmpMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputSnmpCompression(str, Enum
|
|
33
|
+
class InputSnmpCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputSnmpPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputSnmpPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputSnmpMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputSnmpMode.ALWAYS
|
|
67
|
+
mode: Optional[InputSnmpMode] = InputSnmpMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputSnmpPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputSnmpCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputSnmpCompression.NONE
|
|
91
|
+
compress: Optional[InputSnmpCompression] = InputSnmpCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,7 +96,7 @@ class InputSnmpPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class AuthenticationProtocol(str, Enum
|
|
99
|
+
class AuthenticationProtocol(str, Enum):
|
|
107
100
|
NONE = "none"
|
|
108
101
|
MD5 = "md5"
|
|
109
102
|
SHA = "sha"
|
|
@@ -124,10 +117,7 @@ class V3User(BaseModel):
|
|
|
124
117
|
name: str
|
|
125
118
|
|
|
126
119
|
auth_protocol: Annotated[
|
|
127
|
-
|
|
128
|
-
Optional[AuthenticationProtocol], PlainValidator(validate_open_enum(False))
|
|
129
|
-
],
|
|
130
|
-
pydantic.Field(alias="authProtocol"),
|
|
120
|
+
Optional[AuthenticationProtocol], pydantic.Field(alias="authProtocol")
|
|
131
121
|
] = AuthenticationProtocol.NONE
|
|
132
122
|
|
|
133
123
|
auth_key: Annotated[Optional[Any], pydantic.Field(alias="authKey")] = None
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import Any, List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputSplunkConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputSplunkMode(str, Enum
|
|
26
|
+
class InputSplunkMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputSplunkPqCompression(str, Enum
|
|
33
|
+
class InputSplunkPqCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputSplunkPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputSplunkPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputSplunkMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputSplunkMode.ALWAYS
|
|
67
|
+
mode: Optional[InputSplunkMode] = InputSplunkMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputSplunkPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputSplunkPqCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputSplunkPqCompression.NONE
|
|
91
|
+
compress: Optional[InputSplunkPqCompression] = InputSplunkPqCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,14 +96,14 @@ class InputSplunkPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class InputSplunkMinimumTLSVersion(str, Enum
|
|
99
|
+
class InputSplunkMinimumTLSVersion(str, Enum):
|
|
107
100
|
TL_SV1 = "TLSv1"
|
|
108
101
|
TL_SV1_1 = "TLSv1.1"
|
|
109
102
|
TL_SV1_2 = "TLSv1.2"
|
|
110
103
|
TL_SV1_3 = "TLSv1.3"
|
|
111
104
|
|
|
112
105
|
|
|
113
|
-
class InputSplunkMaximumTLSVersion(str, Enum
|
|
106
|
+
class InputSplunkMaximumTLSVersion(str, Enum):
|
|
114
107
|
TL_SV1 = "TLSv1"
|
|
115
108
|
TL_SV1_1 = "TLSv1.1"
|
|
116
109
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -169,19 +162,11 @@ class InputSplunkTLSSettingsServerSide(BaseModel):
|
|
|
169
162
|
] = None
|
|
170
163
|
|
|
171
164
|
min_version: Annotated[
|
|
172
|
-
|
|
173
|
-
Optional[InputSplunkMinimumTLSVersion],
|
|
174
|
-
PlainValidator(validate_open_enum(False)),
|
|
175
|
-
],
|
|
176
|
-
pydantic.Field(alias="minVersion"),
|
|
165
|
+
Optional[InputSplunkMinimumTLSVersion], pydantic.Field(alias="minVersion")
|
|
177
166
|
] = None
|
|
178
167
|
|
|
179
168
|
max_version: Annotated[
|
|
180
|
-
|
|
181
|
-
Optional[InputSplunkMaximumTLSVersion],
|
|
182
|
-
PlainValidator(validate_open_enum(False)),
|
|
183
|
-
],
|
|
184
|
-
pydantic.Field(alias="maxVersion"),
|
|
169
|
+
Optional[InputSplunkMaximumTLSVersion], pydantic.Field(alias="maxVersion")
|
|
185
170
|
] = None
|
|
186
171
|
|
|
187
172
|
|
|
@@ -211,14 +196,14 @@ class InputSplunkAuthToken(BaseModel):
|
|
|
211
196
|
description: Optional[str] = None
|
|
212
197
|
|
|
213
198
|
|
|
214
|
-
class InputSplunkMaxS2SVersion(str, Enum
|
|
199
|
+
class InputSplunkMaxS2SVersion(str, Enum):
|
|
215
200
|
r"""The highest S2S protocol version to advertise during handshake"""
|
|
216
201
|
|
|
217
202
|
V3 = "v3"
|
|
218
203
|
V4 = "v4"
|
|
219
204
|
|
|
220
205
|
|
|
221
|
-
class InputSplunkCompression(str, Enum
|
|
206
|
+
class InputSplunkCompression(str, Enum):
|
|
222
207
|
r"""Controls whether to support reading compressed data from a forwarder. Select 'Automatic' to match the forwarder's configuration, or 'Disabled' to reject compressed connections."""
|
|
223
208
|
|
|
224
209
|
DISABLED = "disabled"
|
|
@@ -369,11 +354,7 @@ class InputSplunk(BaseModel):
|
|
|
369
354
|
r"""Shared secrets to be provided by any Splunk forwarder. If empty, unauthorized access is permitted."""
|
|
370
355
|
|
|
371
356
|
max_s2_sversion: Annotated[
|
|
372
|
-
|
|
373
|
-
Optional[InputSplunkMaxS2SVersion],
|
|
374
|
-
PlainValidator(validate_open_enum(False)),
|
|
375
|
-
],
|
|
376
|
-
pydantic.Field(alias="maxS2Sversion"),
|
|
357
|
+
Optional[InputSplunkMaxS2SVersion], pydantic.Field(alias="maxS2Sversion")
|
|
377
358
|
] = InputSplunkMaxS2SVersion.V3
|
|
378
359
|
r"""The highest S2S protocol version to advertise during handshake"""
|
|
379
360
|
|
|
@@ -394,7 +375,5 @@ class InputSplunk(BaseModel):
|
|
|
394
375
|
] = False
|
|
395
376
|
r"""Extract and process Splunk-generated metrics as Cribl metrics"""
|
|
396
377
|
|
|
397
|
-
compress:
|
|
398
|
-
Optional[InputSplunkCompression], PlainValidator(validate_open_enum(False))
|
|
399
|
-
] = InputSplunkCompression.DISABLED
|
|
378
|
+
compress: Optional[InputSplunkCompression] = InputSplunkCompression.DISABLED
|
|
400
379
|
r"""Controls whether to support reading compressed data from a forwarder. Select 'Automatic' to match the forwarder's configuration, or 'Disabled' to reject compressed connections."""
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import Any, List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputSplunkHecConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputSplunkHecMode(str, Enum
|
|
26
|
+
class InputSplunkHecMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputSplunkHecCompression(str, Enum
|
|
33
|
+
class InputSplunkHecCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputSplunkHecPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputSplunkHecPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputSplunkHecMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputSplunkHecMode.ALWAYS
|
|
67
|
+
mode: Optional[InputSplunkHecMode] = InputSplunkHecMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,9 +88,7 @@ class InputSplunkHecPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputSplunkHecCompression], PlainValidator(validate_open_enum(False))
|
|
98
|
-
] = InputSplunkHecCompression.NONE
|
|
91
|
+
compress: Optional[InputSplunkHecCompression] = InputSplunkHecCompression.NONE
|
|
99
92
|
r"""Codec to use to compress the persisted data"""
|
|
100
93
|
|
|
101
94
|
pq_controls: Annotated[
|
|
@@ -103,7 +96,7 @@ class InputSplunkHecPq(BaseModel):
|
|
|
103
96
|
] = None
|
|
104
97
|
|
|
105
98
|
|
|
106
|
-
class InputSplunkHecAuthenticationMethod(str, Enum
|
|
99
|
+
class InputSplunkHecAuthenticationMethod(str, Enum):
|
|
107
100
|
r"""Select Manual to enter an auth token directly, or select Secret to use a text secret to authenticate"""
|
|
108
101
|
|
|
109
102
|
MANUAL = "manual"
|
|
@@ -141,11 +134,7 @@ class InputSplunkHecAuthToken(BaseModel):
|
|
|
141
134
|
token: Any
|
|
142
135
|
|
|
143
136
|
auth_type: Annotated[
|
|
144
|
-
|
|
145
|
-
Optional[InputSplunkHecAuthenticationMethod],
|
|
146
|
-
PlainValidator(validate_open_enum(False)),
|
|
147
|
-
],
|
|
148
|
-
pydantic.Field(alias="authType"),
|
|
137
|
+
Optional[InputSplunkHecAuthenticationMethod], pydantic.Field(alias="authType")
|
|
149
138
|
] = InputSplunkHecAuthenticationMethod.MANUAL
|
|
150
139
|
r"""Select Manual to enter an auth token directly, or select Secret to use a text secret to authenticate"""
|
|
151
140
|
|
|
@@ -165,14 +154,14 @@ class InputSplunkHecAuthToken(BaseModel):
|
|
|
165
154
|
r"""Fields to add to events referencing this token"""
|
|
166
155
|
|
|
167
156
|
|
|
168
|
-
class InputSplunkHecMinimumTLSVersion(str, Enum
|
|
157
|
+
class InputSplunkHecMinimumTLSVersion(str, Enum):
|
|
169
158
|
TL_SV1 = "TLSv1"
|
|
170
159
|
TL_SV1_1 = "TLSv1.1"
|
|
171
160
|
TL_SV1_2 = "TLSv1.2"
|
|
172
161
|
TL_SV1_3 = "TLSv1.3"
|
|
173
162
|
|
|
174
163
|
|
|
175
|
-
class InputSplunkHecMaximumTLSVersion(str, Enum
|
|
164
|
+
class InputSplunkHecMaximumTLSVersion(str, Enum):
|
|
176
165
|
TL_SV1 = "TLSv1"
|
|
177
166
|
TL_SV1_1 = "TLSv1.1"
|
|
178
167
|
TL_SV1_2 = "TLSv1.2"
|
|
@@ -231,19 +220,11 @@ class InputSplunkHecTLSSettingsServerSide(BaseModel):
|
|
|
231
220
|
] = None
|
|
232
221
|
|
|
233
222
|
min_version: Annotated[
|
|
234
|
-
|
|
235
|
-
Optional[InputSplunkHecMinimumTLSVersion],
|
|
236
|
-
PlainValidator(validate_open_enum(False)),
|
|
237
|
-
],
|
|
238
|
-
pydantic.Field(alias="minVersion"),
|
|
223
|
+
Optional[InputSplunkHecMinimumTLSVersion], pydantic.Field(alias="minVersion")
|
|
239
224
|
] = None
|
|
240
225
|
|
|
241
226
|
max_version: Annotated[
|
|
242
|
-
|
|
243
|
-
Optional[InputSplunkHecMaximumTLSVersion],
|
|
244
|
-
PlainValidator(validate_open_enum(False)),
|
|
245
|
-
],
|
|
246
|
-
pydantic.Field(alias="maxVersion"),
|
|
227
|
+
Optional[InputSplunkHecMaximumTLSVersion], pydantic.Field(alias="maxVersion")
|
|
247
228
|
] = None
|
|
248
229
|
|
|
249
230
|
|
|
@@ -1,12 +1,9 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
-
from cribl_control_plane import utils
|
|
5
4
|
from cribl_control_plane.types import BaseModel
|
|
6
|
-
from cribl_control_plane.utils import validate_open_enum
|
|
7
5
|
from enum import Enum
|
|
8
6
|
import pydantic
|
|
9
|
-
from pydantic.functional_validators import PlainValidator
|
|
10
7
|
from typing import List, Optional
|
|
11
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
12
9
|
|
|
@@ -26,14 +23,14 @@ class InputSplunkSearchConnection(BaseModel):
|
|
|
26
23
|
pipeline: Optional[str] = None
|
|
27
24
|
|
|
28
25
|
|
|
29
|
-
class InputSplunkSearchMode(str, Enum
|
|
26
|
+
class InputSplunkSearchMode(str, Enum):
|
|
30
27
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
31
28
|
|
|
32
29
|
SMART = "smart"
|
|
33
30
|
ALWAYS = "always"
|
|
34
31
|
|
|
35
32
|
|
|
36
|
-
class InputSplunkSearchCompression(str, Enum
|
|
33
|
+
class InputSplunkSearchCompression(str, Enum):
|
|
37
34
|
r"""Codec to use to compress the persisted data"""
|
|
38
35
|
|
|
39
36
|
NONE = "none"
|
|
@@ -67,9 +64,7 @@ class InputSplunkSearchPqTypedDict(TypedDict):
|
|
|
67
64
|
|
|
68
65
|
|
|
69
66
|
class InputSplunkSearchPq(BaseModel):
|
|
70
|
-
mode:
|
|
71
|
-
Optional[InputSplunkSearchMode], PlainValidator(validate_open_enum(False))
|
|
72
|
-
] = InputSplunkSearchMode.ALWAYS
|
|
67
|
+
mode: Optional[InputSplunkSearchMode] = InputSplunkSearchMode.ALWAYS
|
|
73
68
|
r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
|
|
74
69
|
|
|
75
70
|
max_buffer_size: Annotated[
|
|
@@ -93,10 +88,7 @@ class InputSplunkSearchPq(BaseModel):
|
|
|
93
88
|
path: Optional[str] = "$CRIBL_HOME/state/queues"
|
|
94
89
|
r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
|
|
95
90
|
|
|
96
|
-
compress:
|
|
97
|
-
Optional[InputSplunkSearchCompression],
|
|
98
|
-
PlainValidator(validate_open_enum(False)),
|
|
99
|
-
] = InputSplunkSearchCompression.NONE
|
|
91
|
+
compress: Optional[InputSplunkSearchCompression] = InputSplunkSearchCompression.NONE
|
|
100
92
|
r"""Codec to use to compress the persisted data"""
|
|
101
93
|
|
|
102
94
|
pq_controls: Annotated[
|
|
@@ -104,7 +96,7 @@ class InputSplunkSearchPq(BaseModel):
|
|
|
104
96
|
] = None
|
|
105
97
|
|
|
106
98
|
|
|
107
|
-
class OutputMode(str, Enum
|
|
99
|
+
class OutputMode(str, Enum):
|
|
108
100
|
r"""Format of the returned output"""
|
|
109
101
|
|
|
110
102
|
CSV = "csv"
|
|
@@ -137,7 +129,7 @@ class EndpointHeader(BaseModel):
|
|
|
137
129
|
r"""JavaScript expression to compute the header's value, normally enclosed in backticks (e.g., `${earliest}`). If a constant, use single quotes (e.g., 'earliest'). Values without delimiters (e.g., earliest) are evaluated as strings."""
|
|
138
130
|
|
|
139
131
|
|
|
140
|
-
class InputSplunkSearchLogLevel(str, Enum
|
|
132
|
+
class InputSplunkSearchLogLevel(str, Enum):
|
|
141
133
|
r"""Collector runtime log level (verbosity)"""
|
|
142
134
|
|
|
143
135
|
ERROR = "error"
|
|
@@ -159,7 +151,7 @@ class InputSplunkSearchMetadatum(BaseModel):
|
|
|
159
151
|
r"""JavaScript expression to compute field's value, enclosed in quotes or backticks. (Can evaluate to a constant.)"""
|
|
160
152
|
|
|
161
153
|
|
|
162
|
-
class InputSplunkSearchRetryType(str, Enum
|
|
154
|
+
class InputSplunkSearchRetryType(str, Enum):
|
|
163
155
|
r"""The algorithm to use when performing HTTP retries"""
|
|
164
156
|
|
|
165
157
|
NONE = "none"
|
|
@@ -187,9 +179,7 @@ class InputSplunkSearchRetryRulesTypedDict(TypedDict):
|
|
|
187
179
|
|
|
188
180
|
|
|
189
181
|
class InputSplunkSearchRetryRules(BaseModel):
|
|
190
|
-
type:
|
|
191
|
-
Optional[InputSplunkSearchRetryType], PlainValidator(validate_open_enum(False))
|
|
192
|
-
] = InputSplunkSearchRetryType.BACKOFF
|
|
182
|
+
type: Optional[InputSplunkSearchRetryType] = InputSplunkSearchRetryType.BACKOFF
|
|
193
183
|
r"""The algorithm to use when performing HTTP retries"""
|
|
194
184
|
|
|
195
185
|
interval: Optional[float] = 1000
|
|
@@ -220,7 +210,7 @@ class InputSplunkSearchRetryRules(BaseModel):
|
|
|
220
210
|
r"""Retry request when a connection reset (ECONNRESET) error occurs"""
|
|
221
211
|
|
|
222
212
|
|
|
223
|
-
class InputSplunkSearchAuthenticationType(str, Enum
|
|
213
|
+
class InputSplunkSearchAuthenticationType(str, Enum):
|
|
224
214
|
r"""Splunk Search authentication type"""
|
|
225
215
|
|
|
226
216
|
NONE = "none"
|
|
@@ -405,10 +395,9 @@ class InputSplunkSearch(BaseModel):
|
|
|
405
395
|
endpoint: Optional[str] = "/services/search/v2/jobs/export"
|
|
406
396
|
r"""REST API used to create a search"""
|
|
407
397
|
|
|
408
|
-
output_mode: Annotated[
|
|
409
|
-
|
|
410
|
-
|
|
411
|
-
] = OutputMode.JSON
|
|
398
|
+
output_mode: Annotated[Optional[OutputMode], pydantic.Field(alias="outputMode")] = (
|
|
399
|
+
OutputMode.JSON
|
|
400
|
+
)
|
|
412
401
|
r"""Format of the returned output"""
|
|
413
402
|
|
|
414
403
|
endpoint_params: Annotated[
|
|
@@ -422,11 +411,7 @@ class InputSplunkSearch(BaseModel):
|
|
|
422
411
|
r"""Optional request headers to send to the endpoint"""
|
|
423
412
|
|
|
424
413
|
log_level: Annotated[
|
|
425
|
-
|
|
426
|
-
Optional[InputSplunkSearchLogLevel],
|
|
427
|
-
PlainValidator(validate_open_enum(False)),
|
|
428
|
-
],
|
|
429
|
-
pydantic.Field(alias="logLevel"),
|
|
414
|
+
Optional[InputSplunkSearchLogLevel], pydantic.Field(alias="logLevel")
|
|
430
415
|
] = None
|
|
431
416
|
r"""Collector runtime log level (verbosity)"""
|
|
432
417
|
|
|
@@ -487,11 +472,7 @@ class InputSplunkSearch(BaseModel):
|
|
|
487
472
|
r"""How long (in milliseconds) the Event Breaker will wait for new data to be sent to a specific channel before flushing the data stream out, as is, to the Pipelines"""
|
|
488
473
|
|
|
489
474
|
auth_type: Annotated[
|
|
490
|
-
|
|
491
|
-
Optional[InputSplunkSearchAuthenticationType],
|
|
492
|
-
PlainValidator(validate_open_enum(False)),
|
|
493
|
-
],
|
|
494
|
-
pydantic.Field(alias="authType"),
|
|
475
|
+
Optional[InputSplunkSearchAuthenticationType], pydantic.Field(alias="authType")
|
|
495
476
|
] = InputSplunkSearchAuthenticationType.BASIC
|
|
496
477
|
r"""Splunk Search authentication type"""
|
|
497
478
|
|