cribl-control-plane 0.0.14__py3-none-any.whl → 0.0.16__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +3 -3
- cribl_control_plane/{outputs.py → destinations.py} +69 -71
- cribl_control_plane/models/__init__.py +5329 -8
- cribl_control_plane/models/createinputop.py +19253 -2
- cribl_control_plane/models/createoutputop.py +20578 -4
- cribl_control_plane/models/createoutputtestbyidop.py +2 -2
- cribl_control_plane/models/createpipelineop.py +24 -0
- cribl_control_plane/models/deleteoutputbyidop.py +2 -2
- cribl_control_plane/models/deleteoutputpqbyidop.py +2 -2
- cribl_control_plane/models/deletepipelinebyidop.py +37 -0
- cribl_control_plane/models/getoutputbyidop.py +2 -2
- cribl_control_plane/models/getoutputpqbyidop.py +2 -2
- cribl_control_plane/models/getoutputsamplesbyidop.py +2 -2
- cribl_control_plane/models/getpipelinebyidop.py +37 -0
- cribl_control_plane/models/inputconfluentcloud.py +5 -3
- cribl_control_plane/models/inputcriblhttp.py +5 -3
- cribl_control_plane/models/inputcribllakehttp.py +5 -3
- cribl_control_plane/models/inputcribltcp.py +5 -3
- cribl_control_plane/models/inputdatadogagent.py +5 -3
- cribl_control_plane/models/inputedgeprometheus.py +5 -3
- cribl_control_plane/models/inputelastic.py +5 -3
- cribl_control_plane/models/inputeventhub.py +5 -3
- cribl_control_plane/models/inputfirehose.py +5 -3
- cribl_control_plane/models/inputgooglepubsub.py +5 -3
- cribl_control_plane/models/inputgrafana_union.py +10 -6
- cribl_control_plane/models/inputhttp.py +5 -3
- cribl_control_plane/models/inputhttpraw.py +5 -3
- cribl_control_plane/models/inputjournalfiles.py +5 -3
- cribl_control_plane/models/inputkafka.py +5 -3
- cribl_control_plane/models/inputkinesis.py +5 -3
- cribl_control_plane/models/inputloki.py +5 -3
- cribl_control_plane/models/inputmodeldriventelemetry.py +6 -5
- cribl_control_plane/models/inputmsk.py +5 -3
- cribl_control_plane/models/inputnetflow.py +5 -3
- cribl_control_plane/models/inputoffice365mgmt.py +5 -3
- cribl_control_plane/models/inputoffice365msgtrace.py +5 -5
- cribl_control_plane/models/inputoffice365service.py +5 -5
- cribl_control_plane/models/inputopentelemetry.py +5 -3
- cribl_control_plane/models/inputprometheus.py +5 -3
- cribl_control_plane/models/inputprometheusrw.py +5 -3
- cribl_control_plane/models/inputrawudp.py +5 -3
- cribl_control_plane/models/inputsnmp.py +5 -3
- cribl_control_plane/models/inputsplunk.py +5 -3
- cribl_control_plane/models/inputsplunkhec.py +5 -3
- cribl_control_plane/models/inputsplunksearch.py +5 -3
- cribl_control_plane/models/inputsqs.py +5 -3
- cribl_control_plane/models/inputtcp.py +5 -3
- cribl_control_plane/models/inputtcpjson.py +5 -3
- cribl_control_plane/models/inputwef.py +5 -3
- cribl_control_plane/models/inputwiz.py +5 -3
- cribl_control_plane/models/inputzscalerhec.py +5 -3
- cribl_control_plane/models/listoutputop.py +2 -2
- cribl_control_plane/models/listpipelineop.py +24 -0
- cribl_control_plane/models/pipeline.py +70 -0
- cribl_control_plane/models/pipelinefunctionconf.py +53 -0
- cribl_control_plane/models/routes.py +6 -6
- cribl_control_plane/models/updateoutputbyidop.py +4 -4
- cribl_control_plane/models/updatepipelinebyidop.py +47 -0
- cribl_control_plane/pipelines.py +903 -0
- cribl_control_plane/routes_sdk.py +8 -4
- cribl_control_plane/sdk.py +7 -5
- cribl_control_plane/sources.py +8 -10
- {cribl_control_plane-0.0.14.dist-info → cribl_control_plane-0.0.16.dist-info}/METADATA +20 -12
- {cribl_control_plane-0.0.14.dist-info → cribl_control_plane-0.0.16.dist-info}/RECORD +65 -57
- {cribl_control_plane-0.0.14.dist-info → cribl_control_plane-0.0.16.dist-info}/WHEEL +0 -0
|
@@ -11,7 +11,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
11
11
|
|
|
12
12
|
class CreateOutputTestByIDRequestTypedDict(TypedDict):
|
|
13
13
|
id: str
|
|
14
|
-
r"""
|
|
14
|
+
r"""Destination Id"""
|
|
15
15
|
output_test_request: OutputTestRequestTypedDict
|
|
16
16
|
r"""OutputTestRequest object"""
|
|
17
17
|
|
|
@@ -20,7 +20,7 @@ class CreateOutputTestByIDRequest(BaseModel):
|
|
|
20
20
|
id: Annotated[
|
|
21
21
|
str, FieldMetadata(path=PathParamMetadata(style="simple", explode=False))
|
|
22
22
|
]
|
|
23
|
-
r"""
|
|
23
|
+
r"""Destination Id"""
|
|
24
24
|
|
|
25
25
|
output_test_request: Annotated[
|
|
26
26
|
OutputTestRequest,
|
|
@@ -0,0 +1,24 @@
|
|
|
1
|
+
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
from .pipeline import Pipeline, PipelineTypedDict
|
|
5
|
+
from cribl_control_plane.types import BaseModel
|
|
6
|
+
from typing import List, Optional
|
|
7
|
+
from typing_extensions import NotRequired, TypedDict
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
class CreatePipelineResponseTypedDict(TypedDict):
|
|
11
|
+
r"""a list of Pipeline objects"""
|
|
12
|
+
|
|
13
|
+
count: NotRequired[int]
|
|
14
|
+
r"""number of items present in the items array"""
|
|
15
|
+
items: NotRequired[List[PipelineTypedDict]]
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
class CreatePipelineResponse(BaseModel):
|
|
19
|
+
r"""a list of Pipeline objects"""
|
|
20
|
+
|
|
21
|
+
count: Optional[int] = None
|
|
22
|
+
r"""number of items present in the items array"""
|
|
23
|
+
|
|
24
|
+
items: Optional[List[Pipeline]] = None
|
|
@@ -21,7 +21,7 @@ class DeleteOutputByIDRequest(BaseModel):
|
|
|
21
21
|
|
|
22
22
|
|
|
23
23
|
class DeleteOutputByIDResponseTypedDict(TypedDict):
|
|
24
|
-
r"""a list of
|
|
24
|
+
r"""a list of Destination objects"""
|
|
25
25
|
|
|
26
26
|
count: NotRequired[int]
|
|
27
27
|
r"""number of items present in the items array"""
|
|
@@ -29,7 +29,7 @@ class DeleteOutputByIDResponseTypedDict(TypedDict):
|
|
|
29
29
|
|
|
30
30
|
|
|
31
31
|
class DeleteOutputByIDResponse(BaseModel):
|
|
32
|
-
r"""a list of
|
|
32
|
+
r"""a list of Destination objects"""
|
|
33
33
|
|
|
34
34
|
count: Optional[int] = None
|
|
35
35
|
r"""number of items present in the items array"""
|
|
@@ -9,14 +9,14 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
9
9
|
|
|
10
10
|
class DeleteOutputPqByIDRequestTypedDict(TypedDict):
|
|
11
11
|
id: str
|
|
12
|
-
r"""
|
|
12
|
+
r"""Destination Id"""
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
class DeleteOutputPqByIDRequest(BaseModel):
|
|
16
16
|
id: Annotated[
|
|
17
17
|
str, FieldMetadata(path=PathParamMetadata(style="simple", explode=False))
|
|
18
18
|
]
|
|
19
|
-
r"""
|
|
19
|
+
r"""Destination Id"""
|
|
20
20
|
|
|
21
21
|
|
|
22
22
|
class DeleteOutputPqByIDResponseTypedDict(TypedDict):
|
|
@@ -0,0 +1,37 @@
|
|
|
1
|
+
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
from .pipeline import Pipeline, PipelineTypedDict
|
|
5
|
+
from cribl_control_plane.types import BaseModel
|
|
6
|
+
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata
|
|
7
|
+
from typing import List, Optional
|
|
8
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
|
+
|
|
10
|
+
|
|
11
|
+
class DeletePipelineByIDRequestTypedDict(TypedDict):
|
|
12
|
+
id: str
|
|
13
|
+
r"""Unique ID to DELETE"""
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
class DeletePipelineByIDRequest(BaseModel):
|
|
17
|
+
id: Annotated[
|
|
18
|
+
str, FieldMetadata(path=PathParamMetadata(style="simple", explode=False))
|
|
19
|
+
]
|
|
20
|
+
r"""Unique ID to DELETE"""
|
|
21
|
+
|
|
22
|
+
|
|
23
|
+
class DeletePipelineByIDResponseTypedDict(TypedDict):
|
|
24
|
+
r"""a list of Pipeline objects"""
|
|
25
|
+
|
|
26
|
+
count: NotRequired[int]
|
|
27
|
+
r"""number of items present in the items array"""
|
|
28
|
+
items: NotRequired[List[PipelineTypedDict]]
|
|
29
|
+
|
|
30
|
+
|
|
31
|
+
class DeletePipelineByIDResponse(BaseModel):
|
|
32
|
+
r"""a list of Pipeline objects"""
|
|
33
|
+
|
|
34
|
+
count: Optional[int] = None
|
|
35
|
+
r"""number of items present in the items array"""
|
|
36
|
+
|
|
37
|
+
items: Optional[List[Pipeline]] = None
|
|
@@ -21,7 +21,7 @@ class GetOutputByIDRequest(BaseModel):
|
|
|
21
21
|
|
|
22
22
|
|
|
23
23
|
class GetOutputByIDResponseTypedDict(TypedDict):
|
|
24
|
-
r"""a list of
|
|
24
|
+
r"""a list of Destination objects"""
|
|
25
25
|
|
|
26
26
|
count: NotRequired[int]
|
|
27
27
|
r"""number of items present in the items array"""
|
|
@@ -29,7 +29,7 @@ class GetOutputByIDResponseTypedDict(TypedDict):
|
|
|
29
29
|
|
|
30
30
|
|
|
31
31
|
class GetOutputByIDResponse(BaseModel):
|
|
32
|
-
r"""a list of
|
|
32
|
+
r"""a list of Destination objects"""
|
|
33
33
|
|
|
34
34
|
count: Optional[int] = None
|
|
35
35
|
r"""number of items present in the items array"""
|
|
@@ -9,14 +9,14 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
9
9
|
|
|
10
10
|
class GetOutputPqByIDRequestTypedDict(TypedDict):
|
|
11
11
|
id: str
|
|
12
|
-
r"""
|
|
12
|
+
r"""Destination Id"""
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
class GetOutputPqByIDRequest(BaseModel):
|
|
16
16
|
id: Annotated[
|
|
17
17
|
str, FieldMetadata(path=PathParamMetadata(style="simple", explode=False))
|
|
18
18
|
]
|
|
19
|
-
r"""
|
|
19
|
+
r"""Destination Id"""
|
|
20
20
|
|
|
21
21
|
|
|
22
22
|
class GetOutputPqByIDResponseTypedDict(TypedDict):
|
|
@@ -10,14 +10,14 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
10
10
|
|
|
11
11
|
class GetOutputSamplesByIDRequestTypedDict(TypedDict):
|
|
12
12
|
id: str
|
|
13
|
-
r"""
|
|
13
|
+
r"""Destination Id"""
|
|
14
14
|
|
|
15
15
|
|
|
16
16
|
class GetOutputSamplesByIDRequest(BaseModel):
|
|
17
17
|
id: Annotated[
|
|
18
18
|
str, FieldMetadata(path=PathParamMetadata(style="simple", explode=False))
|
|
19
19
|
]
|
|
20
|
-
r"""
|
|
20
|
+
r"""Destination Id"""
|
|
21
21
|
|
|
22
22
|
|
|
23
23
|
class GetOutputSamplesByIDResponseTypedDict(TypedDict):
|
|
@@ -0,0 +1,37 @@
|
|
|
1
|
+
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
from .pipeline import Pipeline, PipelineTypedDict
|
|
5
|
+
from cribl_control_plane.types import BaseModel
|
|
6
|
+
from cribl_control_plane.utils import FieldMetadata, PathParamMetadata
|
|
7
|
+
from typing import List, Optional
|
|
8
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
|
+
|
|
10
|
+
|
|
11
|
+
class GetPipelineByIDRequestTypedDict(TypedDict):
|
|
12
|
+
id: str
|
|
13
|
+
r"""Unique ID to GET"""
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
class GetPipelineByIDRequest(BaseModel):
|
|
17
|
+
id: Annotated[
|
|
18
|
+
str, FieldMetadata(path=PathParamMetadata(style="simple", explode=False))
|
|
19
|
+
]
|
|
20
|
+
r"""Unique ID to GET"""
|
|
21
|
+
|
|
22
|
+
|
|
23
|
+
class GetPipelineByIDResponseTypedDict(TypedDict):
|
|
24
|
+
r"""a list of Pipeline objects"""
|
|
25
|
+
|
|
26
|
+
count: NotRequired[int]
|
|
27
|
+
r"""number of items present in the items array"""
|
|
28
|
+
items: NotRequired[List[PipelineTypedDict]]
|
|
29
|
+
|
|
30
|
+
|
|
31
|
+
class GetPipelineByIDResponse(BaseModel):
|
|
32
|
+
r"""a list of Pipeline objects"""
|
|
33
|
+
|
|
34
|
+
count: Optional[int] = None
|
|
35
|
+
r"""number of items present in the items array"""
|
|
36
|
+
|
|
37
|
+
items: Optional[List[Pipeline]] = None
|
|
@@ -363,13 +363,13 @@ class InputConfluentCloudMetadatum(BaseModel):
|
|
|
363
363
|
|
|
364
364
|
|
|
365
365
|
class InputConfluentCloudTypedDict(TypedDict):
|
|
366
|
-
type: InputConfluentCloudType
|
|
367
366
|
brokers: List[str]
|
|
368
367
|
r"""List of Confluent Cloud bootstrap servers to use, such as yourAccount.confluent.cloud:9092"""
|
|
369
368
|
topics: List[str]
|
|
370
369
|
r"""Topic to subscribe to. Warning: To optimize performance, Cribl suggests subscribing each Kafka Source to a single topic only."""
|
|
371
370
|
id: NotRequired[str]
|
|
372
371
|
r"""Unique ID for this input"""
|
|
372
|
+
type: NotRequired[InputConfluentCloudType]
|
|
373
373
|
disabled: NotRequired[bool]
|
|
374
374
|
pipeline: NotRequired[str]
|
|
375
375
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -443,8 +443,6 @@ class InputConfluentCloudTypedDict(TypedDict):
|
|
|
443
443
|
|
|
444
444
|
|
|
445
445
|
class InputConfluentCloud(BaseModel):
|
|
446
|
-
type: Annotated[InputConfluentCloudType, PlainValidator(validate_open_enum(False))]
|
|
447
|
-
|
|
448
446
|
brokers: List[str]
|
|
449
447
|
r"""List of Confluent Cloud bootstrap servers to use, such as yourAccount.confluent.cloud:9092"""
|
|
450
448
|
|
|
@@ -454,6 +452,10 @@ class InputConfluentCloud(BaseModel):
|
|
|
454
452
|
id: Optional[str] = None
|
|
455
453
|
r"""Unique ID for this input"""
|
|
456
454
|
|
|
455
|
+
type: Annotated[
|
|
456
|
+
Optional[InputConfluentCloudType], PlainValidator(validate_open_enum(False))
|
|
457
|
+
] = None
|
|
458
|
+
|
|
457
459
|
disabled: Optional[bool] = False
|
|
458
460
|
|
|
459
461
|
pipeline: Optional[str] = None
|
|
@@ -186,11 +186,11 @@ class InputCriblHTTPMetadatum(BaseModel):
|
|
|
186
186
|
|
|
187
187
|
|
|
188
188
|
class InputCriblHTTPTypedDict(TypedDict):
|
|
189
|
-
type: InputCriblHTTPType
|
|
190
189
|
port: float
|
|
191
190
|
r"""Port to listen on"""
|
|
192
191
|
id: NotRequired[str]
|
|
193
192
|
r"""Unique ID for this input"""
|
|
193
|
+
type: NotRequired[InputCriblHTTPType]
|
|
194
194
|
disabled: NotRequired[bool]
|
|
195
195
|
pipeline: NotRequired[str]
|
|
196
196
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -238,14 +238,16 @@ class InputCriblHTTPTypedDict(TypedDict):
|
|
|
238
238
|
|
|
239
239
|
|
|
240
240
|
class InputCriblHTTP(BaseModel):
|
|
241
|
-
type: Annotated[InputCriblHTTPType, PlainValidator(validate_open_enum(False))]
|
|
242
|
-
|
|
243
241
|
port: float
|
|
244
242
|
r"""Port to listen on"""
|
|
245
243
|
|
|
246
244
|
id: Optional[str] = None
|
|
247
245
|
r"""Unique ID for this input"""
|
|
248
246
|
|
|
247
|
+
type: Annotated[
|
|
248
|
+
Optional[InputCriblHTTPType], PlainValidator(validate_open_enum(False))
|
|
249
|
+
] = None
|
|
250
|
+
|
|
249
251
|
disabled: Optional[bool] = False
|
|
250
252
|
|
|
251
253
|
pipeline: Optional[str] = None
|
|
@@ -187,11 +187,11 @@ class InputCriblLakeHTTPMetadatum(BaseModel):
|
|
|
187
187
|
|
|
188
188
|
|
|
189
189
|
class InputCriblLakeHTTPTypedDict(TypedDict):
|
|
190
|
-
type: InputCriblLakeHTTPType
|
|
191
190
|
port: float
|
|
192
191
|
r"""Port to listen on"""
|
|
193
192
|
id: NotRequired[str]
|
|
194
193
|
r"""Unique ID for this input"""
|
|
194
|
+
type: NotRequired[InputCriblLakeHTTPType]
|
|
195
195
|
disabled: NotRequired[bool]
|
|
196
196
|
pipeline: NotRequired[str]
|
|
197
197
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -239,14 +239,16 @@ class InputCriblLakeHTTPTypedDict(TypedDict):
|
|
|
239
239
|
|
|
240
240
|
|
|
241
241
|
class InputCriblLakeHTTP(BaseModel):
|
|
242
|
-
type: Annotated[InputCriblLakeHTTPType, PlainValidator(validate_open_enum(False))]
|
|
243
|
-
|
|
244
242
|
port: float
|
|
245
243
|
r"""Port to listen on"""
|
|
246
244
|
|
|
247
245
|
id: Optional[str] = None
|
|
248
246
|
r"""Unique ID for this input"""
|
|
249
247
|
|
|
248
|
+
type: Annotated[
|
|
249
|
+
Optional[InputCriblLakeHTTPType], PlainValidator(validate_open_enum(False))
|
|
250
|
+
] = None
|
|
251
|
+
|
|
250
252
|
disabled: Optional[bool] = False
|
|
251
253
|
|
|
252
254
|
pipeline: Optional[str] = None
|
|
@@ -186,11 +186,11 @@ class InputCriblTCPMetadatum(BaseModel):
|
|
|
186
186
|
|
|
187
187
|
|
|
188
188
|
class InputCriblTCPTypedDict(TypedDict):
|
|
189
|
-
type: InputCriblTCPType
|
|
190
189
|
port: float
|
|
191
190
|
r"""Port to listen on"""
|
|
192
191
|
id: NotRequired[str]
|
|
193
192
|
r"""Unique ID for this input"""
|
|
193
|
+
type: NotRequired[InputCriblTCPType]
|
|
194
194
|
disabled: NotRequired[bool]
|
|
195
195
|
pipeline: NotRequired[str]
|
|
196
196
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -226,14 +226,16 @@ class InputCriblTCPTypedDict(TypedDict):
|
|
|
226
226
|
|
|
227
227
|
|
|
228
228
|
class InputCriblTCP(BaseModel):
|
|
229
|
-
type: Annotated[InputCriblTCPType, PlainValidator(validate_open_enum(False))]
|
|
230
|
-
|
|
231
229
|
port: float
|
|
232
230
|
r"""Port to listen on"""
|
|
233
231
|
|
|
234
232
|
id: Optional[str] = None
|
|
235
233
|
r"""Unique ID for this input"""
|
|
236
234
|
|
|
235
|
+
type: Annotated[
|
|
236
|
+
Optional[InputCriblTCPType], PlainValidator(validate_open_enum(False))
|
|
237
|
+
] = None
|
|
238
|
+
|
|
237
239
|
disabled: Optional[bool] = False
|
|
238
240
|
|
|
239
241
|
pipeline: Optional[str] = None
|
|
@@ -204,11 +204,11 @@ class InputDatadogAgentProxyMode(BaseModel):
|
|
|
204
204
|
|
|
205
205
|
|
|
206
206
|
class InputDatadogAgentTypedDict(TypedDict):
|
|
207
|
-
type: InputDatadogAgentType
|
|
208
207
|
port: float
|
|
209
208
|
r"""Port to listen on"""
|
|
210
209
|
id: NotRequired[str]
|
|
211
210
|
r"""Unique ID for this input"""
|
|
211
|
+
type: NotRequired[InputDatadogAgentType]
|
|
212
212
|
disabled: NotRequired[bool]
|
|
213
213
|
pipeline: NotRequired[str]
|
|
214
214
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -257,14 +257,16 @@ class InputDatadogAgentTypedDict(TypedDict):
|
|
|
257
257
|
|
|
258
258
|
|
|
259
259
|
class InputDatadogAgent(BaseModel):
|
|
260
|
-
type: Annotated[InputDatadogAgentType, PlainValidator(validate_open_enum(False))]
|
|
261
|
-
|
|
262
260
|
port: float
|
|
263
261
|
r"""Port to listen on"""
|
|
264
262
|
|
|
265
263
|
id: Optional[str] = None
|
|
266
264
|
r"""Unique ID for this input"""
|
|
267
265
|
|
|
266
|
+
type: Annotated[
|
|
267
|
+
Optional[InputDatadogAgentType], PlainValidator(validate_open_enum(False))
|
|
268
|
+
] = None
|
|
269
|
+
|
|
268
270
|
disabled: Optional[bool] = False
|
|
269
271
|
|
|
270
272
|
pipeline: Optional[str] = None
|
|
@@ -263,9 +263,9 @@ class PodFilter(BaseModel):
|
|
|
263
263
|
|
|
264
264
|
|
|
265
265
|
class InputEdgePrometheusTypedDict(TypedDict):
|
|
266
|
-
type: InputEdgePrometheusType
|
|
267
266
|
id: NotRequired[str]
|
|
268
267
|
r"""Unique ID for this input"""
|
|
268
|
+
type: NotRequired[InputEdgePrometheusType]
|
|
269
269
|
disabled: NotRequired[bool]
|
|
270
270
|
pipeline: NotRequired[str]
|
|
271
271
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -353,11 +353,13 @@ class InputEdgePrometheusTypedDict(TypedDict):
|
|
|
353
353
|
|
|
354
354
|
|
|
355
355
|
class InputEdgePrometheus(BaseModel):
|
|
356
|
-
type: Annotated[InputEdgePrometheusType, PlainValidator(validate_open_enum(False))]
|
|
357
|
-
|
|
358
356
|
id: Optional[str] = None
|
|
359
357
|
r"""Unique ID for this input"""
|
|
360
358
|
|
|
359
|
+
type: Annotated[
|
|
360
|
+
Optional[InputEdgePrometheusType], PlainValidator(validate_open_enum(False))
|
|
361
|
+
] = None
|
|
362
|
+
|
|
361
363
|
disabled: Optional[bool] = False
|
|
362
364
|
|
|
363
365
|
pipeline: Optional[str] = None
|
|
@@ -265,11 +265,11 @@ class InputElasticProxyMode(BaseModel):
|
|
|
265
265
|
|
|
266
266
|
|
|
267
267
|
class InputElasticTypedDict(TypedDict):
|
|
268
|
-
type: InputElasticType
|
|
269
268
|
port: float
|
|
270
269
|
r"""Port to listen on"""
|
|
271
270
|
id: NotRequired[str]
|
|
272
271
|
r"""Unique ID for this input"""
|
|
272
|
+
type: NotRequired[InputElasticType]
|
|
273
273
|
disabled: NotRequired[bool]
|
|
274
274
|
pipeline: NotRequired[str]
|
|
275
275
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -331,14 +331,16 @@ class InputElasticTypedDict(TypedDict):
|
|
|
331
331
|
|
|
332
332
|
|
|
333
333
|
class InputElastic(BaseModel):
|
|
334
|
-
type: Annotated[InputElasticType, PlainValidator(validate_open_enum(False))]
|
|
335
|
-
|
|
336
334
|
port: float
|
|
337
335
|
r"""Port to listen on"""
|
|
338
336
|
|
|
339
337
|
id: Optional[str] = None
|
|
340
338
|
r"""Unique ID for this input"""
|
|
341
339
|
|
|
340
|
+
type: Annotated[
|
|
341
|
+
Optional[InputElasticType], PlainValidator(validate_open_enum(False))
|
|
342
|
+
] = None
|
|
343
|
+
|
|
342
344
|
disabled: Optional[bool] = False
|
|
343
345
|
|
|
344
346
|
pipeline: Optional[str] = None
|
|
@@ -141,13 +141,13 @@ class InputEventhubMetadatum(BaseModel):
|
|
|
141
141
|
|
|
142
142
|
|
|
143
143
|
class InputEventhubTypedDict(TypedDict):
|
|
144
|
-
type: InputEventhubType
|
|
145
144
|
brokers: List[str]
|
|
146
145
|
r"""List of Event Hubs Kafka brokers to connect to (example: yourdomain.servicebus.windows.net:9093). The hostname can be found in the host portion of the primary or secondary connection string in Shared Access Policies."""
|
|
147
146
|
topics: List[str]
|
|
148
147
|
r"""The name of the Event Hub (Kafka topic) to subscribe to. Warning: To optimize performance, Cribl suggests subscribing each Event Hubs Source to only a single topic."""
|
|
149
148
|
id: NotRequired[str]
|
|
150
149
|
r"""Unique ID for this input"""
|
|
150
|
+
type: NotRequired[InputEventhubType]
|
|
151
151
|
disabled: NotRequired[bool]
|
|
152
152
|
pipeline: NotRequired[str]
|
|
153
153
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -219,8 +219,6 @@ class InputEventhubTypedDict(TypedDict):
|
|
|
219
219
|
|
|
220
220
|
|
|
221
221
|
class InputEventhub(BaseModel):
|
|
222
|
-
type: Annotated[InputEventhubType, PlainValidator(validate_open_enum(False))]
|
|
223
|
-
|
|
224
222
|
brokers: List[str]
|
|
225
223
|
r"""List of Event Hubs Kafka brokers to connect to (example: yourdomain.servicebus.windows.net:9093). The hostname can be found in the host portion of the primary or secondary connection string in Shared Access Policies."""
|
|
226
224
|
|
|
@@ -230,6 +228,10 @@ class InputEventhub(BaseModel):
|
|
|
230
228
|
id: Optional[str] = None
|
|
231
229
|
r"""Unique ID for this input"""
|
|
232
230
|
|
|
231
|
+
type: Annotated[
|
|
232
|
+
Optional[InputEventhubType], PlainValidator(validate_open_enum(False))
|
|
233
|
+
] = None
|
|
234
|
+
|
|
233
235
|
disabled: Optional[bool] = False
|
|
234
236
|
|
|
235
237
|
pipeline: Optional[str] = None
|
|
@@ -186,11 +186,11 @@ class InputFirehoseMetadatum(BaseModel):
|
|
|
186
186
|
|
|
187
187
|
|
|
188
188
|
class InputFirehoseTypedDict(TypedDict):
|
|
189
|
-
type: InputFirehoseType
|
|
190
189
|
port: float
|
|
191
190
|
r"""Port to listen on"""
|
|
192
191
|
id: NotRequired[str]
|
|
193
192
|
r"""Unique ID for this input"""
|
|
193
|
+
type: NotRequired[InputFirehoseType]
|
|
194
194
|
disabled: NotRequired[bool]
|
|
195
195
|
pipeline: NotRequired[str]
|
|
196
196
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -238,14 +238,16 @@ class InputFirehoseTypedDict(TypedDict):
|
|
|
238
238
|
|
|
239
239
|
|
|
240
240
|
class InputFirehose(BaseModel):
|
|
241
|
-
type: Annotated[InputFirehoseType, PlainValidator(validate_open_enum(False))]
|
|
242
|
-
|
|
243
241
|
port: float
|
|
244
242
|
r"""Port to listen on"""
|
|
245
243
|
|
|
246
244
|
id: Optional[str] = None
|
|
247
245
|
r"""Unique ID for this input"""
|
|
248
246
|
|
|
247
|
+
type: Annotated[
|
|
248
|
+
Optional[InputFirehoseType], PlainValidator(validate_open_enum(False))
|
|
249
|
+
] = None
|
|
250
|
+
|
|
249
251
|
disabled: Optional[bool] = False
|
|
250
252
|
|
|
251
253
|
pipeline: Optional[str] = None
|
|
@@ -115,13 +115,13 @@ class InputGooglePubsubMetadatum(BaseModel):
|
|
|
115
115
|
|
|
116
116
|
|
|
117
117
|
class InputGooglePubsubTypedDict(TypedDict):
|
|
118
|
-
type: InputGooglePubsubType
|
|
119
118
|
topic_name: str
|
|
120
119
|
r"""ID of the topic to receive events from"""
|
|
121
120
|
subscription_name: str
|
|
122
121
|
r"""ID of the subscription to use when receiving events"""
|
|
123
122
|
id: NotRequired[str]
|
|
124
123
|
r"""Unique ID for this input"""
|
|
124
|
+
type: NotRequired[InputGooglePubsubType]
|
|
125
125
|
disabled: NotRequired[bool]
|
|
126
126
|
pipeline: NotRequired[str]
|
|
127
127
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -162,8 +162,6 @@ class InputGooglePubsubTypedDict(TypedDict):
|
|
|
162
162
|
|
|
163
163
|
|
|
164
164
|
class InputGooglePubsub(BaseModel):
|
|
165
|
-
type: Annotated[InputGooglePubsubType, PlainValidator(validate_open_enum(False))]
|
|
166
|
-
|
|
167
165
|
topic_name: Annotated[str, pydantic.Field(alias="topicName")]
|
|
168
166
|
r"""ID of the topic to receive events from"""
|
|
169
167
|
|
|
@@ -173,6 +171,10 @@ class InputGooglePubsub(BaseModel):
|
|
|
173
171
|
id: Optional[str] = None
|
|
174
172
|
r"""Unique ID for this input"""
|
|
175
173
|
|
|
174
|
+
type: Annotated[
|
|
175
|
+
Optional[InputGooglePubsubType], PlainValidator(validate_open_enum(False))
|
|
176
|
+
] = None
|
|
177
|
+
|
|
176
178
|
disabled: Optional[bool] = False
|
|
177
179
|
|
|
178
180
|
pipeline: Optional[str] = None
|
|
@@ -452,11 +452,11 @@ class InputGrafanaMetadatum2(BaseModel):
|
|
|
452
452
|
|
|
453
453
|
|
|
454
454
|
class InputGrafana2TypedDict(TypedDict):
|
|
455
|
-
type: InputGrafanaType2
|
|
456
455
|
port: float
|
|
457
456
|
r"""Port to listen on"""
|
|
458
457
|
id: NotRequired[str]
|
|
459
458
|
r"""Unique ID for this input"""
|
|
459
|
+
type: NotRequired[InputGrafanaType2]
|
|
460
460
|
disabled: NotRequired[bool]
|
|
461
461
|
pipeline: NotRequired[str]
|
|
462
462
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -508,14 +508,16 @@ class InputGrafana2TypedDict(TypedDict):
|
|
|
508
508
|
|
|
509
509
|
|
|
510
510
|
class InputGrafana2(BaseModel):
|
|
511
|
-
type: Annotated[InputGrafanaType2, PlainValidator(validate_open_enum(False))]
|
|
512
|
-
|
|
513
511
|
port: float
|
|
514
512
|
r"""Port to listen on"""
|
|
515
513
|
|
|
516
514
|
id: Optional[str] = None
|
|
517
515
|
r"""Unique ID for this input"""
|
|
518
516
|
|
|
517
|
+
type: Annotated[
|
|
518
|
+
Optional[InputGrafanaType2], PlainValidator(validate_open_enum(False))
|
|
519
|
+
] = None
|
|
520
|
+
|
|
519
521
|
disabled: Optional[bool] = False
|
|
520
522
|
|
|
521
523
|
pipeline: Optional[str] = None
|
|
@@ -1065,11 +1067,11 @@ class InputGrafanaMetadatum1(BaseModel):
|
|
|
1065
1067
|
|
|
1066
1068
|
|
|
1067
1069
|
class InputGrafana1TypedDict(TypedDict):
|
|
1068
|
-
type: InputGrafanaType1
|
|
1069
1070
|
port: float
|
|
1070
1071
|
r"""Port to listen on"""
|
|
1071
1072
|
id: NotRequired[str]
|
|
1072
1073
|
r"""Unique ID for this input"""
|
|
1074
|
+
type: NotRequired[InputGrafanaType1]
|
|
1073
1075
|
disabled: NotRequired[bool]
|
|
1074
1076
|
pipeline: NotRequired[str]
|
|
1075
1077
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -1121,14 +1123,16 @@ class InputGrafana1TypedDict(TypedDict):
|
|
|
1121
1123
|
|
|
1122
1124
|
|
|
1123
1125
|
class InputGrafana1(BaseModel):
|
|
1124
|
-
type: Annotated[InputGrafanaType1, PlainValidator(validate_open_enum(False))]
|
|
1125
|
-
|
|
1126
1126
|
port: float
|
|
1127
1127
|
r"""Port to listen on"""
|
|
1128
1128
|
|
|
1129
1129
|
id: Optional[str] = None
|
|
1130
1130
|
r"""Unique ID for this input"""
|
|
1131
1131
|
|
|
1132
|
+
type: Annotated[
|
|
1133
|
+
Optional[InputGrafanaType1], PlainValidator(validate_open_enum(False))
|
|
1134
|
+
] = None
|
|
1135
|
+
|
|
1132
1136
|
disabled: Optional[bool] = False
|
|
1133
1137
|
|
|
1134
1138
|
pipeline: Optional[str] = None
|
|
@@ -217,11 +217,11 @@ class InputHTTPAuthTokensExt(BaseModel):
|
|
|
217
217
|
|
|
218
218
|
|
|
219
219
|
class InputHTTPTypedDict(TypedDict):
|
|
220
|
-
type: InputHTTPType
|
|
221
220
|
port: float
|
|
222
221
|
r"""Port to listen on"""
|
|
223
222
|
id: NotRequired[str]
|
|
224
223
|
r"""Unique ID for this input"""
|
|
224
|
+
type: NotRequired[InputHTTPType]
|
|
225
225
|
disabled: NotRequired[bool]
|
|
226
226
|
pipeline: NotRequired[str]
|
|
227
227
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -278,14 +278,16 @@ class InputHTTPTypedDict(TypedDict):
|
|
|
278
278
|
|
|
279
279
|
|
|
280
280
|
class InputHTTP(BaseModel):
|
|
281
|
-
type: Annotated[InputHTTPType, PlainValidator(validate_open_enum(False))]
|
|
282
|
-
|
|
283
281
|
port: float
|
|
284
282
|
r"""Port to listen on"""
|
|
285
283
|
|
|
286
284
|
id: Optional[str] = None
|
|
287
285
|
r"""Unique ID for this input"""
|
|
288
286
|
|
|
287
|
+
type: Annotated[
|
|
288
|
+
Optional[InputHTTPType], PlainValidator(validate_open_enum(False))
|
|
289
|
+
] = None
|
|
290
|
+
|
|
289
291
|
disabled: Optional[bool] = False
|
|
290
292
|
|
|
291
293
|
pipeline: Optional[str] = None
|
|
@@ -217,11 +217,11 @@ class InputHTTPRawAuthTokensExt(BaseModel):
|
|
|
217
217
|
|
|
218
218
|
|
|
219
219
|
class InputHTTPRawTypedDict(TypedDict):
|
|
220
|
-
type: InputHTTPRawType
|
|
221
220
|
port: float
|
|
222
221
|
r"""Port to listen on"""
|
|
223
222
|
id: NotRequired[str]
|
|
224
223
|
r"""Unique ID for this input"""
|
|
224
|
+
type: NotRequired[InputHTTPRawType]
|
|
225
225
|
disabled: NotRequired[bool]
|
|
226
226
|
pipeline: NotRequired[str]
|
|
227
227
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -279,14 +279,16 @@ class InputHTTPRawTypedDict(TypedDict):
|
|
|
279
279
|
|
|
280
280
|
|
|
281
281
|
class InputHTTPRaw(BaseModel):
|
|
282
|
-
type: Annotated[InputHTTPRawType, PlainValidator(validate_open_enum(False))]
|
|
283
|
-
|
|
284
282
|
port: float
|
|
285
283
|
r"""Port to listen on"""
|
|
286
284
|
|
|
287
285
|
id: Optional[str] = None
|
|
288
286
|
r"""Unique ID for this input"""
|
|
289
287
|
|
|
288
|
+
type: Annotated[
|
|
289
|
+
Optional[InputHTTPRawType], PlainValidator(validate_open_enum(False))
|
|
290
|
+
] = None
|
|
291
|
+
|
|
290
292
|
disabled: Optional[bool] = False
|
|
291
293
|
|
|
292
294
|
pipeline: Optional[str] = None
|