cribl-control-plane 0.0.24__py3-none-any.whl → 0.0.26__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +3 -3
- cribl_control_plane/destinations.py +10 -8
- cribl_control_plane/errors/healthstatus_error.py +1 -1
- cribl_control_plane/groups_sdk.py +10 -10
- cribl_control_plane/{healthinfo.py → health.py} +3 -1
- cribl_control_plane/models/__init__.py +21 -27
- cribl_control_plane/models/healthstatus.py +3 -3
- cribl_control_plane/models/input.py +4 -4
- cribl_control_plane/models/inputappscope.py +5 -5
- cribl_control_plane/models/inputcollection.py +2 -2
- cribl_control_plane/models/inputconfluentcloud.py +17 -3
- cribl_control_plane/models/inputcribl.py +5 -5
- cribl_control_plane/models/inputcriblhttp.py +3 -3
- cribl_control_plane/models/inputcribllakehttp.py +3 -3
- cribl_control_plane/models/inputcriblmetrics.py +5 -5
- cribl_control_plane/models/inputcribltcp.py +3 -3
- cribl_control_plane/models/inputdatadogagent.py +3 -3
- cribl_control_plane/models/inputedgeprometheus.py +3 -3
- cribl_control_plane/models/inputelastic.py +3 -3
- cribl_control_plane/models/inputeventhub.py +3 -3
- cribl_control_plane/models/inputfile.py +5 -5
- cribl_control_plane/models/inputfirehose.py +3 -3
- cribl_control_plane/models/inputgooglepubsub.py +16 -9
- cribl_control_plane/models/inputgrafana.py +20 -6
- cribl_control_plane/models/inputhttp.py +3 -3
- cribl_control_plane/models/inputhttpraw.py +3 -3
- cribl_control_plane/models/inputjournalfiles.py +3 -3
- cribl_control_plane/models/inputkafka.py +17 -3
- cribl_control_plane/models/inputkinesis.py +3 -3
- cribl_control_plane/models/inputkubeevents.py +5 -5
- cribl_control_plane/models/inputkubelogs.py +5 -5
- cribl_control_plane/models/inputkubemetrics.py +5 -5
- cribl_control_plane/models/inputloki.py +10 -3
- cribl_control_plane/models/inputmodeldriventelemetry.py +3 -3
- cribl_control_plane/models/inputmsk.py +17 -3
- cribl_control_plane/models/inputnetflow.py +3 -3
- cribl_control_plane/models/inputoffice365mgmt.py +3 -3
- cribl_control_plane/models/inputoffice365msgtrace.py +3 -3
- cribl_control_plane/models/inputoffice365service.py +3 -3
- cribl_control_plane/models/inputopentelemetry.py +3 -3
- cribl_control_plane/models/inputprometheus.py +3 -3
- cribl_control_plane/models/inputprometheusrw.py +3 -3
- cribl_control_plane/models/inputrawudp.py +3 -3
- cribl_control_plane/models/inputsnmp.py +3 -3
- cribl_control_plane/models/inputsplunk.py +3 -3
- cribl_control_plane/models/inputsplunkhec.py +3 -3
- cribl_control_plane/models/inputsplunksearch.py +3 -3
- cribl_control_plane/models/inputsqs.py +3 -3
- cribl_control_plane/models/inputsystemmetrics.py +5 -5
- cribl_control_plane/models/inputsystemstate.py +5 -5
- cribl_control_plane/models/inputtcp.py +3 -3
- cribl_control_plane/models/inputtcpjson.py +3 -3
- cribl_control_plane/models/inputwef.py +3 -3
- cribl_control_plane/models/inputwindowsmetrics.py +5 -5
- cribl_control_plane/models/inputwiz.py +3 -3
- cribl_control_plane/models/inputzscalerhec.py +3 -3
- cribl_control_plane/models/output.py +14 -14
- cribl_control_plane/models/outputazureblob.py +3 -3
- cribl_control_plane/models/outputazuredataexplorer.py +3 -3
- cribl_control_plane/models/outputazureeventhub.py +3 -3
- cribl_control_plane/models/outputclickhouse.py +3 -3
- cribl_control_plane/models/outputcloudwatch.py +3 -3
- cribl_control_plane/models/outputconfluentcloud.py +17 -3
- cribl_control_plane/models/outputcriblhttp.py +5 -5
- cribl_control_plane/models/outputcribllake.py +5 -5
- cribl_control_plane/models/outputcribltcp.py +5 -5
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +3 -3
- cribl_control_plane/models/outputdatadog.py +5 -5
- cribl_control_plane/models/outputdataset.py +5 -5
- cribl_control_plane/models/outputdevnull.py +5 -5
- cribl_control_plane/models/outputdiskspool.py +5 -5
- cribl_control_plane/models/outputdls3.py +5 -5
- cribl_control_plane/models/outputdynatracehttp.py +3 -3
- cribl_control_plane/models/outputdynatraceotlp.py +3 -3
- cribl_control_plane/models/outputelasticcloud.py +3 -3
- cribl_control_plane/models/outputexabeam.py +3 -3
- cribl_control_plane/models/outputgooglecloudlogging.py +3 -3
- cribl_control_plane/models/outputgooglecloudstorage.py +5 -5
- cribl_control_plane/models/outputgrafanacloud.py +24 -10
- cribl_control_plane/models/outputgraphite.py +3 -3
- cribl_control_plane/models/outputhumiohec.py +3 -3
- cribl_control_plane/models/outputkafka.py +17 -3
- cribl_control_plane/models/outputkinesis.py +3 -3
- cribl_control_plane/models/outputloki.py +14 -0
- cribl_control_plane/models/outputminio.py +3 -3
- cribl_control_plane/models/outputmsk.py +17 -3
- cribl_control_plane/models/outputnewrelic.py +5 -5
- cribl_control_plane/models/outputnewrelicevents.py +3 -3
- cribl_control_plane/models/outputring.py +5 -5
- cribl_control_plane/models/outputs3.py +5 -5
- cribl_control_plane/models/outputsecuritylake.py +3 -3
- cribl_control_plane/models/outputsentinel.py +3 -3
- cribl_control_plane/models/outputsentineloneaisiem.py +3 -3
- cribl_control_plane/models/outputservicenow.py +3 -3
- cribl_control_plane/models/outputsns.py +3 -3
- cribl_control_plane/models/outputsplunk.py +3 -3
- cribl_control_plane/models/outputsplunkhec.py +5 -5
- cribl_control_plane/models/outputsqs.py +3 -3
- cribl_control_plane/models/outputstatsd.py +3 -3
- cribl_control_plane/models/outputstatsdext.py +3 -3
- cribl_control_plane/models/outputsyslog.py +5 -5
- cribl_control_plane/models/outputtcpjson.py +5 -5
- cribl_control_plane/models/outputwebhook.py +5 -5
- cribl_control_plane/models/outputxsiam.py +5 -5
- cribl_control_plane/nodes.py +100 -90
- cribl_control_plane/pipelines.py +20 -20
- cribl_control_plane/sdk.py +6 -6
- cribl_control_plane/sources.py +2 -0
- cribl_control_plane/versioning.py +14 -14
- {cribl_control_plane-0.0.24.dist-info → cribl_control_plane-0.0.26.dist-info}/METADATA +24 -28
- {cribl_control_plane-0.0.24.dist-info → cribl_control_plane-0.0.26.dist-info}/RECORD +112 -115
- cribl_control_plane/deployments.py +0 -185
- cribl_control_plane/models/restartresponse.py +0 -26
- cribl_control_plane/models/updateworkersrestartop.py +0 -24
- {cribl_control_plane-0.0.24.dist-info → cribl_control_plane-0.0.26.dist-info}/WHEEL +0 -0
|
@@ -151,9 +151,9 @@ class InputKubeLogsDiskSpooling(BaseModel):
|
|
|
151
151
|
|
|
152
152
|
|
|
153
153
|
class InputKubeLogsTypedDict(TypedDict):
|
|
154
|
-
id: str
|
|
155
|
-
r"""Unique ID for this input"""
|
|
156
154
|
type: InputKubeLogsType
|
|
155
|
+
id: NotRequired[str]
|
|
156
|
+
r"""Unique ID for this input"""
|
|
157
157
|
disabled: NotRequired[bool]
|
|
158
158
|
pipeline: NotRequired[str]
|
|
159
159
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -187,11 +187,11 @@ class InputKubeLogsTypedDict(TypedDict):
|
|
|
187
187
|
|
|
188
188
|
|
|
189
189
|
class InputKubeLogs(BaseModel):
|
|
190
|
-
id: str
|
|
191
|
-
r"""Unique ID for this input"""
|
|
192
|
-
|
|
193
190
|
type: InputKubeLogsType
|
|
194
191
|
|
|
192
|
+
id: Optional[str] = None
|
|
193
|
+
r"""Unique ID for this input"""
|
|
194
|
+
|
|
195
195
|
disabled: Optional[bool] = False
|
|
196
196
|
|
|
197
197
|
pipeline: Optional[str] = None
|
|
@@ -154,9 +154,9 @@ class InputKubeMetricsPersistence(BaseModel):
|
|
|
154
154
|
|
|
155
155
|
|
|
156
156
|
class InputKubeMetricsTypedDict(TypedDict):
|
|
157
|
-
id: str
|
|
158
|
-
r"""Unique ID for this input"""
|
|
159
157
|
type: InputKubeMetricsType
|
|
158
|
+
id: NotRequired[str]
|
|
159
|
+
r"""Unique ID for this input"""
|
|
160
160
|
disabled: NotRequired[bool]
|
|
161
161
|
pipeline: NotRequired[str]
|
|
162
162
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -182,11 +182,11 @@ class InputKubeMetricsTypedDict(TypedDict):
|
|
|
182
182
|
|
|
183
183
|
|
|
184
184
|
class InputKubeMetrics(BaseModel):
|
|
185
|
-
id: str
|
|
186
|
-
r"""Unique ID for this input"""
|
|
187
|
-
|
|
188
185
|
type: InputKubeMetricsType
|
|
189
186
|
|
|
187
|
+
id: Optional[str] = None
|
|
188
|
+
r"""Unique ID for this input"""
|
|
189
|
+
|
|
190
190
|
disabled: Optional[bool] = False
|
|
191
191
|
|
|
192
192
|
pipeline: Optional[str] = None
|
|
@@ -212,11 +212,11 @@ class InputLokiOauthHeader(BaseModel):
|
|
|
212
212
|
|
|
213
213
|
|
|
214
214
|
class InputLokiTypedDict(TypedDict):
|
|
215
|
+
type: InputLokiType
|
|
215
216
|
port: float
|
|
216
217
|
r"""Port to listen on"""
|
|
217
218
|
id: NotRequired[str]
|
|
218
219
|
r"""Unique ID for this input"""
|
|
219
|
-
type: NotRequired[InputLokiType]
|
|
220
220
|
disabled: NotRequired[bool]
|
|
221
221
|
pipeline: NotRequired[str]
|
|
222
222
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -258,6 +258,8 @@ class InputLokiTypedDict(TypedDict):
|
|
|
258
258
|
r"""Messages from matched IP addresses will be ignored. This takes precedence over the allowlist."""
|
|
259
259
|
loki_api: NotRequired[str]
|
|
260
260
|
r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'."""
|
|
261
|
+
extract_structured_metadata: NotRequired[bool]
|
|
262
|
+
r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
|
|
261
263
|
auth_type: NotRequired[InputLokiAuthenticationType]
|
|
262
264
|
r"""Loki logs authentication type"""
|
|
263
265
|
metadata: NotRequired[List[InputLokiMetadatumTypedDict]]
|
|
@@ -290,14 +292,14 @@ class InputLokiTypedDict(TypedDict):
|
|
|
290
292
|
|
|
291
293
|
|
|
292
294
|
class InputLoki(BaseModel):
|
|
295
|
+
type: InputLokiType
|
|
296
|
+
|
|
293
297
|
port: float
|
|
294
298
|
r"""Port to listen on"""
|
|
295
299
|
|
|
296
300
|
id: Optional[str] = None
|
|
297
301
|
r"""Unique ID for this input"""
|
|
298
302
|
|
|
299
|
-
type: Optional[InputLokiType] = None
|
|
300
|
-
|
|
301
303
|
disabled: Optional[bool] = False
|
|
302
304
|
|
|
303
305
|
pipeline: Optional[str] = None
|
|
@@ -387,6 +389,11 @@ class InputLoki(BaseModel):
|
|
|
387
389
|
)
|
|
388
390
|
r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'."""
|
|
389
391
|
|
|
392
|
+
extract_structured_metadata: Annotated[
|
|
393
|
+
Optional[bool], pydantic.Field(alias="extractStructuredMetadata")
|
|
394
|
+
] = False
|
|
395
|
+
r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
|
|
396
|
+
|
|
390
397
|
auth_type: Annotated[
|
|
391
398
|
Optional[InputLokiAuthenticationType], pydantic.Field(alias="authType")
|
|
392
399
|
] = InputLokiAuthenticationType.NONE
|
|
@@ -170,9 +170,9 @@ class InputModelDrivenTelemetryMetadatum(BaseModel):
|
|
|
170
170
|
|
|
171
171
|
|
|
172
172
|
class InputModelDrivenTelemetryTypedDict(TypedDict):
|
|
173
|
+
type: InputModelDrivenTelemetryType
|
|
173
174
|
id: NotRequired[str]
|
|
174
175
|
r"""Unique ID for this input"""
|
|
175
|
-
type: NotRequired[InputModelDrivenTelemetryType]
|
|
176
176
|
disabled: NotRequired[bool]
|
|
177
177
|
pipeline: NotRequired[str]
|
|
178
178
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -202,11 +202,11 @@ class InputModelDrivenTelemetryTypedDict(TypedDict):
|
|
|
202
202
|
|
|
203
203
|
|
|
204
204
|
class InputModelDrivenTelemetry(BaseModel):
|
|
205
|
+
type: InputModelDrivenTelemetryType
|
|
206
|
+
|
|
205
207
|
id: Optional[str] = None
|
|
206
208
|
r"""Unique ID for this input"""
|
|
207
209
|
|
|
208
|
-
type: Optional[InputModelDrivenTelemetryType] = None
|
|
209
|
-
|
|
210
210
|
disabled: Optional[bool] = False
|
|
211
211
|
|
|
212
212
|
pipeline: Optional[str] = None
|
|
@@ -96,6 +96,13 @@ class InputMskMetadatum(BaseModel):
|
|
|
96
96
|
r"""JavaScript expression to compute field's value, enclosed in quotes or backticks. (Can evaluate to a constant.)"""
|
|
97
97
|
|
|
98
98
|
|
|
99
|
+
class InputMskSchemaType(str, Enum):
|
|
100
|
+
r"""The schema format used to encode and decode event data"""
|
|
101
|
+
|
|
102
|
+
AVRO = "avro"
|
|
103
|
+
JSON = "json"
|
|
104
|
+
|
|
105
|
+
|
|
99
106
|
class InputMskAuthTypedDict(TypedDict):
|
|
100
107
|
r"""Credentials to use when authenticating with the schema registry using basic HTTP authentication"""
|
|
101
108
|
|
|
@@ -196,6 +203,8 @@ class InputMskKafkaSchemaRegistryAuthenticationTypedDict(TypedDict):
|
|
|
196
203
|
disabled: NotRequired[bool]
|
|
197
204
|
schema_registry_url: NotRequired[str]
|
|
198
205
|
r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
|
|
206
|
+
schema_type: NotRequired[InputMskSchemaType]
|
|
207
|
+
r"""The schema format used to encode and decode event data"""
|
|
199
208
|
connection_timeout: NotRequired[float]
|
|
200
209
|
r"""Maximum time to wait for a Schema Registry connection to complete successfully"""
|
|
201
210
|
request_timeout: NotRequired[float]
|
|
@@ -215,6 +224,11 @@ class InputMskKafkaSchemaRegistryAuthentication(BaseModel):
|
|
|
215
224
|
] = "http://localhost:8081"
|
|
216
225
|
r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
|
|
217
226
|
|
|
227
|
+
schema_type: Annotated[
|
|
228
|
+
Optional[InputMskSchemaType], pydantic.Field(alias="schemaType")
|
|
229
|
+
] = InputMskSchemaType.AVRO
|
|
230
|
+
r"""The schema format used to encode and decode event data"""
|
|
231
|
+
|
|
218
232
|
connection_timeout: Annotated[
|
|
219
233
|
Optional[float], pydantic.Field(alias="connectionTimeout")
|
|
220
234
|
] = 30000
|
|
@@ -325,6 +339,7 @@ class InputMskTLSSettingsClientSide(BaseModel):
|
|
|
325
339
|
|
|
326
340
|
|
|
327
341
|
class InputMskTypedDict(TypedDict):
|
|
342
|
+
type: InputMskType
|
|
328
343
|
brokers: List[str]
|
|
329
344
|
r"""Enter each Kafka bootstrap server you want to use. Specify the hostname and port (such as mykafkabroker:9092) or just the hostname (in which case @{product} will assign port 9092)."""
|
|
330
345
|
topics: List[str]
|
|
@@ -333,7 +348,6 @@ class InputMskTypedDict(TypedDict):
|
|
|
333
348
|
r"""Region where the MSK cluster is located"""
|
|
334
349
|
id: NotRequired[str]
|
|
335
350
|
r"""Unique ID for this input"""
|
|
336
|
-
type: NotRequired[InputMskType]
|
|
337
351
|
disabled: NotRequired[bool]
|
|
338
352
|
pipeline: NotRequired[str]
|
|
339
353
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -427,6 +441,8 @@ class InputMskTypedDict(TypedDict):
|
|
|
427
441
|
|
|
428
442
|
|
|
429
443
|
class InputMsk(BaseModel):
|
|
444
|
+
type: InputMskType
|
|
445
|
+
|
|
430
446
|
brokers: List[str]
|
|
431
447
|
r"""Enter each Kafka bootstrap server you want to use. Specify the hostname and port (such as mykafkabroker:9092) or just the hostname (in which case @{product} will assign port 9092)."""
|
|
432
448
|
|
|
@@ -439,8 +455,6 @@ class InputMsk(BaseModel):
|
|
|
439
455
|
id: Optional[str] = None
|
|
440
456
|
r"""Unique ID for this input"""
|
|
441
457
|
|
|
442
|
-
type: Optional[InputMskType] = None
|
|
443
|
-
|
|
444
458
|
disabled: Optional[bool] = False
|
|
445
459
|
|
|
446
460
|
pipeline: Optional[str] = None
|
|
@@ -97,9 +97,9 @@ class InputNetflowMetadatum(BaseModel):
|
|
|
97
97
|
|
|
98
98
|
|
|
99
99
|
class InputNetflowTypedDict(TypedDict):
|
|
100
|
+
type: InputNetflowType
|
|
100
101
|
id: NotRequired[str]
|
|
101
102
|
r"""Unique ID for this input"""
|
|
102
|
-
type: NotRequired[InputNetflowType]
|
|
103
103
|
disabled: NotRequired[bool]
|
|
104
104
|
pipeline: NotRequired[str]
|
|
105
105
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -140,11 +140,11 @@ class InputNetflowTypedDict(TypedDict):
|
|
|
140
140
|
|
|
141
141
|
|
|
142
142
|
class InputNetflow(BaseModel):
|
|
143
|
+
type: InputNetflowType
|
|
144
|
+
|
|
143
145
|
id: Optional[str] = None
|
|
144
146
|
r"""Unique ID for this input"""
|
|
145
147
|
|
|
146
|
-
type: Optional[InputNetflowType] = None
|
|
147
|
-
|
|
148
148
|
disabled: Optional[bool] = False
|
|
149
149
|
|
|
150
150
|
pipeline: Optional[str] = None
|
|
@@ -211,13 +211,13 @@ class InputOffice365MgmtAuthenticationMethod(str, Enum):
|
|
|
211
211
|
|
|
212
212
|
|
|
213
213
|
class InputOffice365MgmtTypedDict(TypedDict):
|
|
214
|
+
type: InputOffice365MgmtType
|
|
214
215
|
tenant_id: str
|
|
215
216
|
r"""Office 365 Azure Tenant ID"""
|
|
216
217
|
app_id: str
|
|
217
218
|
r"""Office 365 Azure Application ID"""
|
|
218
219
|
id: NotRequired[str]
|
|
219
220
|
r"""Unique ID for this input"""
|
|
220
|
-
type: NotRequired[InputOffice365MgmtType]
|
|
221
221
|
disabled: NotRequired[bool]
|
|
222
222
|
pipeline: NotRequired[str]
|
|
223
223
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -265,6 +265,8 @@ class InputOffice365MgmtTypedDict(TypedDict):
|
|
|
265
265
|
|
|
266
266
|
|
|
267
267
|
class InputOffice365Mgmt(BaseModel):
|
|
268
|
+
type: InputOffice365MgmtType
|
|
269
|
+
|
|
268
270
|
tenant_id: Annotated[str, pydantic.Field(alias="tenantId")]
|
|
269
271
|
r"""Office 365 Azure Tenant ID"""
|
|
270
272
|
|
|
@@ -274,8 +276,6 @@ class InputOffice365Mgmt(BaseModel):
|
|
|
274
276
|
id: Optional[str] = None
|
|
275
277
|
r"""Unique ID for this input"""
|
|
276
278
|
|
|
277
|
-
type: Optional[InputOffice365MgmtType] = None
|
|
278
|
-
|
|
279
279
|
disabled: Optional[bool] = False
|
|
280
280
|
|
|
281
281
|
pipeline: Optional[str] = None
|
|
@@ -216,9 +216,9 @@ class CertOptions(BaseModel):
|
|
|
216
216
|
|
|
217
217
|
|
|
218
218
|
class InputOffice365MsgTraceTypedDict(TypedDict):
|
|
219
|
+
type: InputOffice365MsgTraceType
|
|
219
220
|
id: NotRequired[str]
|
|
220
221
|
r"""Unique ID for this input"""
|
|
221
|
-
type: NotRequired[InputOffice365MsgTraceType]
|
|
222
222
|
disabled: NotRequired[bool]
|
|
223
223
|
pipeline: NotRequired[str]
|
|
224
224
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -289,11 +289,11 @@ class InputOffice365MsgTraceTypedDict(TypedDict):
|
|
|
289
289
|
|
|
290
290
|
|
|
291
291
|
class InputOffice365MsgTrace(BaseModel):
|
|
292
|
+
type: InputOffice365MsgTraceType
|
|
293
|
+
|
|
292
294
|
id: Optional[str] = None
|
|
293
295
|
r"""Unique ID for this input"""
|
|
294
296
|
|
|
295
|
-
type: Optional[InputOffice365MsgTraceType] = None
|
|
296
|
-
|
|
297
297
|
disabled: Optional[bool] = False
|
|
298
298
|
|
|
299
299
|
pipeline: Optional[str] = None
|
|
@@ -213,13 +213,13 @@ class InputOffice365ServiceAuthenticationMethod(str, Enum):
|
|
|
213
213
|
|
|
214
214
|
|
|
215
215
|
class InputOffice365ServiceTypedDict(TypedDict):
|
|
216
|
+
type: InputOffice365ServiceType
|
|
216
217
|
tenant_id: str
|
|
217
218
|
r"""Office 365 Azure Tenant ID"""
|
|
218
219
|
app_id: str
|
|
219
220
|
r"""Office 365 Azure Application ID"""
|
|
220
221
|
id: NotRequired[str]
|
|
221
222
|
r"""Unique ID for this input"""
|
|
222
|
-
type: NotRequired[InputOffice365ServiceType]
|
|
223
223
|
disabled: NotRequired[bool]
|
|
224
224
|
pipeline: NotRequired[str]
|
|
225
225
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -263,6 +263,8 @@ class InputOffice365ServiceTypedDict(TypedDict):
|
|
|
263
263
|
|
|
264
264
|
|
|
265
265
|
class InputOffice365Service(BaseModel):
|
|
266
|
+
type: InputOffice365ServiceType
|
|
267
|
+
|
|
266
268
|
tenant_id: Annotated[str, pydantic.Field(alias="tenantId")]
|
|
267
269
|
r"""Office 365 Azure Tenant ID"""
|
|
268
270
|
|
|
@@ -272,8 +274,6 @@ class InputOffice365Service(BaseModel):
|
|
|
272
274
|
id: Optional[str] = None
|
|
273
275
|
r"""Unique ID for this input"""
|
|
274
276
|
|
|
275
|
-
type: Optional[InputOffice365ServiceType] = None
|
|
276
|
-
|
|
277
277
|
disabled: Optional[bool] = False
|
|
278
278
|
|
|
279
279
|
pipeline: Optional[str] = None
|
|
@@ -230,9 +230,9 @@ class InputOpenTelemetryOauthHeader(BaseModel):
|
|
|
230
230
|
|
|
231
231
|
|
|
232
232
|
class InputOpenTelemetryTypedDict(TypedDict):
|
|
233
|
+
type: InputOpenTelemetryType
|
|
233
234
|
id: NotRequired[str]
|
|
234
235
|
r"""Unique ID for this input"""
|
|
235
|
-
type: NotRequired[InputOpenTelemetryType]
|
|
236
236
|
disabled: NotRequired[bool]
|
|
237
237
|
pipeline: NotRequired[str]
|
|
238
238
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -315,11 +315,11 @@ class InputOpenTelemetryTypedDict(TypedDict):
|
|
|
315
315
|
|
|
316
316
|
|
|
317
317
|
class InputOpenTelemetry(BaseModel):
|
|
318
|
+
type: InputOpenTelemetryType
|
|
319
|
+
|
|
318
320
|
id: Optional[str] = None
|
|
319
321
|
r"""Unique ID for this input"""
|
|
320
322
|
|
|
321
|
-
type: Optional[InputOpenTelemetryType] = None
|
|
322
|
-
|
|
323
323
|
disabled: Optional[bool] = False
|
|
324
324
|
|
|
325
325
|
pipeline: Optional[str] = None
|
|
@@ -166,9 +166,9 @@ class InputPrometheusSignatureVersion(str, Enum):
|
|
|
166
166
|
|
|
167
167
|
|
|
168
168
|
class InputPrometheusTypedDict(TypedDict):
|
|
169
|
+
type: InputPrometheusType
|
|
169
170
|
id: NotRequired[str]
|
|
170
171
|
r"""Unique ID for this input"""
|
|
171
|
-
type: NotRequired[InputPrometheusType]
|
|
172
172
|
disabled: NotRequired[bool]
|
|
173
173
|
pipeline: NotRequired[str]
|
|
174
174
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -254,11 +254,11 @@ class InputPrometheusTypedDict(TypedDict):
|
|
|
254
254
|
|
|
255
255
|
|
|
256
256
|
class InputPrometheus(BaseModel):
|
|
257
|
+
type: InputPrometheusType
|
|
258
|
+
|
|
257
259
|
id: Optional[str] = None
|
|
258
260
|
r"""Unique ID for this input"""
|
|
259
261
|
|
|
260
|
-
type: Optional[InputPrometheusType] = None
|
|
261
|
-
|
|
262
262
|
disabled: Optional[bool] = False
|
|
263
263
|
|
|
264
264
|
pipeline: Optional[str] = None
|
|
@@ -212,11 +212,11 @@ class InputPrometheusRwOauthHeader(BaseModel):
|
|
|
212
212
|
|
|
213
213
|
|
|
214
214
|
class InputPrometheusRwTypedDict(TypedDict):
|
|
215
|
+
type: InputPrometheusRwType
|
|
215
216
|
port: float
|
|
216
217
|
r"""Port to listen on"""
|
|
217
218
|
id: NotRequired[str]
|
|
218
219
|
r"""Unique ID for this input"""
|
|
219
|
-
type: NotRequired[InputPrometheusRwType]
|
|
220
220
|
disabled: NotRequired[bool]
|
|
221
221
|
pipeline: NotRequired[str]
|
|
222
222
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -290,14 +290,14 @@ class InputPrometheusRwTypedDict(TypedDict):
|
|
|
290
290
|
|
|
291
291
|
|
|
292
292
|
class InputPrometheusRw(BaseModel):
|
|
293
|
+
type: InputPrometheusRwType
|
|
294
|
+
|
|
293
295
|
port: float
|
|
294
296
|
r"""Port to listen on"""
|
|
295
297
|
|
|
296
298
|
id: Optional[str] = None
|
|
297
299
|
r"""Unique ID for this input"""
|
|
298
300
|
|
|
299
|
-
type: Optional[InputPrometheusRwType] = None
|
|
300
|
-
|
|
301
301
|
disabled: Optional[bool] = False
|
|
302
302
|
|
|
303
303
|
pipeline: Optional[str] = None
|
|
@@ -97,11 +97,11 @@ class InputRawUDPMetadatum(BaseModel):
|
|
|
97
97
|
|
|
98
98
|
|
|
99
99
|
class InputRawUDPTypedDict(TypedDict):
|
|
100
|
+
type: InputRawUDPType
|
|
100
101
|
port: float
|
|
101
102
|
r"""Port to listen on"""
|
|
102
103
|
id: NotRequired[str]
|
|
103
104
|
r"""Unique ID for this input"""
|
|
104
|
-
type: NotRequired[InputRawUDPType]
|
|
105
105
|
disabled: NotRequired[bool]
|
|
106
106
|
pipeline: NotRequired[str]
|
|
107
107
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -134,14 +134,14 @@ class InputRawUDPTypedDict(TypedDict):
|
|
|
134
134
|
|
|
135
135
|
|
|
136
136
|
class InputRawUDP(BaseModel):
|
|
137
|
+
type: InputRawUDPType
|
|
138
|
+
|
|
137
139
|
port: float
|
|
138
140
|
r"""Port to listen on"""
|
|
139
141
|
|
|
140
142
|
id: Optional[str] = None
|
|
141
143
|
r"""Unique ID for this input"""
|
|
142
144
|
|
|
143
|
-
type: Optional[InputRawUDPType] = None
|
|
144
|
-
|
|
145
145
|
disabled: Optional[bool] = False
|
|
146
146
|
|
|
147
147
|
pipeline: Optional[str] = None
|
|
@@ -154,9 +154,9 @@ class InputSnmpMetadatum(BaseModel):
|
|
|
154
154
|
|
|
155
155
|
|
|
156
156
|
class InputSnmpTypedDict(TypedDict):
|
|
157
|
+
type: InputSnmpType
|
|
157
158
|
id: NotRequired[str]
|
|
158
159
|
r"""Unique ID for this input"""
|
|
159
|
-
type: NotRequired[InputSnmpType]
|
|
160
160
|
disabled: NotRequired[bool]
|
|
161
161
|
pipeline: NotRequired[str]
|
|
162
162
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -193,11 +193,11 @@ class InputSnmpTypedDict(TypedDict):
|
|
|
193
193
|
|
|
194
194
|
|
|
195
195
|
class InputSnmp(BaseModel):
|
|
196
|
+
type: InputSnmpType
|
|
197
|
+
|
|
196
198
|
id: Optional[str] = None
|
|
197
199
|
r"""Unique ID for this input"""
|
|
198
200
|
|
|
199
|
-
type: Optional[InputSnmpType] = None
|
|
200
|
-
|
|
201
201
|
disabled: Optional[bool] = False
|
|
202
202
|
|
|
203
203
|
pipeline: Optional[str] = None
|
|
@@ -199,11 +199,11 @@ class InputSplunkCompression(str, Enum):
|
|
|
199
199
|
|
|
200
200
|
|
|
201
201
|
class InputSplunkTypedDict(TypedDict):
|
|
202
|
+
type: InputSplunkType
|
|
202
203
|
port: float
|
|
203
204
|
r"""Port to listen on"""
|
|
204
205
|
id: NotRequired[str]
|
|
205
206
|
r"""Unique ID for this input"""
|
|
206
|
-
type: NotRequired[InputSplunkType]
|
|
207
207
|
disabled: NotRequired[bool]
|
|
208
208
|
pipeline: NotRequired[str]
|
|
209
209
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -255,14 +255,14 @@ class InputSplunkTypedDict(TypedDict):
|
|
|
255
255
|
|
|
256
256
|
|
|
257
257
|
class InputSplunk(BaseModel):
|
|
258
|
+
type: InputSplunkType
|
|
259
|
+
|
|
258
260
|
port: float
|
|
259
261
|
r"""Port to listen on"""
|
|
260
262
|
|
|
261
263
|
id: Optional[str] = None
|
|
262
264
|
r"""Unique ID for this input"""
|
|
263
265
|
|
|
264
|
-
type: Optional[InputSplunkType] = None
|
|
265
|
-
|
|
266
266
|
disabled: Optional[bool] = False
|
|
267
267
|
|
|
268
268
|
pipeline: Optional[str] = None
|
|
@@ -229,11 +229,11 @@ class InputSplunkHecMetadatum(BaseModel):
|
|
|
229
229
|
|
|
230
230
|
|
|
231
231
|
class InputSplunkHecTypedDict(TypedDict):
|
|
232
|
+
type: InputSplunkHecType
|
|
232
233
|
port: float
|
|
233
234
|
r"""Port to listen on"""
|
|
234
235
|
id: NotRequired[str]
|
|
235
236
|
r"""Unique ID for this input"""
|
|
236
|
-
type: NotRequired[InputSplunkHecType]
|
|
237
237
|
disabled: NotRequired[bool]
|
|
238
238
|
pipeline: NotRequired[str]
|
|
239
239
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -302,14 +302,14 @@ class InputSplunkHecTypedDict(TypedDict):
|
|
|
302
302
|
|
|
303
303
|
|
|
304
304
|
class InputSplunkHec(BaseModel):
|
|
305
|
+
type: InputSplunkHecType
|
|
306
|
+
|
|
305
307
|
port: float
|
|
306
308
|
r"""Port to listen on"""
|
|
307
309
|
|
|
308
310
|
id: Optional[str] = None
|
|
309
311
|
r"""Unique ID for this input"""
|
|
310
312
|
|
|
311
|
-
type: Optional[InputSplunkHecType] = None
|
|
312
|
-
|
|
313
313
|
disabled: Optional[bool] = False
|
|
314
314
|
|
|
315
315
|
pipeline: Optional[str] = None
|
|
@@ -239,11 +239,11 @@ class InputSplunkSearchOauthHeader(BaseModel):
|
|
|
239
239
|
|
|
240
240
|
|
|
241
241
|
class InputSplunkSearchTypedDict(TypedDict):
|
|
242
|
+
type: InputSplunkSearchType
|
|
242
243
|
search: str
|
|
243
244
|
r"""Enter Splunk search here. Examples: 'index=myAppLogs level=error channel=myApp' OR '| mstats avg(myStat) as myStat WHERE index=myStatsIndex.'"""
|
|
244
245
|
id: NotRequired[str]
|
|
245
246
|
r"""Unique ID for this input"""
|
|
246
|
-
type: NotRequired[InputSplunkSearchType]
|
|
247
247
|
disabled: NotRequired[bool]
|
|
248
248
|
pipeline: NotRequired[str]
|
|
249
249
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -331,14 +331,14 @@ class InputSplunkSearchTypedDict(TypedDict):
|
|
|
331
331
|
|
|
332
332
|
|
|
333
333
|
class InputSplunkSearch(BaseModel):
|
|
334
|
+
type: InputSplunkSearchType
|
|
335
|
+
|
|
334
336
|
search: str
|
|
335
337
|
r"""Enter Splunk search here. Examples: 'index=myAppLogs level=error channel=myApp' OR '| mstats avg(myStat) as myStat WHERE index=myStatsIndex.'"""
|
|
336
338
|
|
|
337
339
|
id: Optional[str] = None
|
|
338
340
|
r"""Unique ID for this input"""
|
|
339
341
|
|
|
340
|
-
type: Optional[InputSplunkSearchType] = None
|
|
341
|
-
|
|
342
342
|
disabled: Optional[bool] = False
|
|
343
343
|
|
|
344
344
|
pipeline: Optional[str] = None
|
|
@@ -119,11 +119,11 @@ class InputSqsMetadatum(BaseModel):
|
|
|
119
119
|
|
|
120
120
|
|
|
121
121
|
class InputSqsTypedDict(TypedDict):
|
|
122
|
+
type: InputSqsType
|
|
122
123
|
queue_name: str
|
|
123
124
|
r"""The name, URL, or ARN of the SQS queue to read events from. When a non-AWS URL is specified, format must be: '{url}/myQueueName'. Example: 'https://host:port/myQueueName'. Value must be a JavaScript expression (which can evaluate to a constant value), enclosed in quotes or backticks. Can only be evaluated at init time. Example referencing a Global Variable: `https://host:port/myQueue-${C.vars.myVar}`."""
|
|
124
125
|
id: NotRequired[str]
|
|
125
126
|
r"""Unique ID for this input"""
|
|
126
|
-
type: NotRequired[InputSqsType]
|
|
127
127
|
disabled: NotRequired[bool]
|
|
128
128
|
pipeline: NotRequired[str]
|
|
129
129
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -182,14 +182,14 @@ class InputSqsTypedDict(TypedDict):
|
|
|
182
182
|
|
|
183
183
|
|
|
184
184
|
class InputSqs(BaseModel):
|
|
185
|
+
type: InputSqsType
|
|
186
|
+
|
|
185
187
|
queue_name: Annotated[str, pydantic.Field(alias="queueName")]
|
|
186
188
|
r"""The name, URL, or ARN of the SQS queue to read events from. When a non-AWS URL is specified, format must be: '{url}/myQueueName'. Example: 'https://host:port/myQueueName'. Value must be a JavaScript expression (which can evaluate to a constant value), enclosed in quotes or backticks. Can only be evaluated at init time. Example referencing a Global Variable: `https://host:port/myQueue-${C.vars.myVar}`."""
|
|
187
189
|
|
|
188
190
|
id: Optional[str] = None
|
|
189
191
|
r"""Unique ID for this input"""
|
|
190
192
|
|
|
191
|
-
type: Optional[InputSqsType] = None
|
|
192
|
-
|
|
193
193
|
disabled: Optional[bool] = False
|
|
194
194
|
|
|
195
195
|
pipeline: Optional[str] = None
|
|
@@ -434,9 +434,9 @@ class InputSystemMetricsPersistence(BaseModel):
|
|
|
434
434
|
|
|
435
435
|
|
|
436
436
|
class InputSystemMetricsTypedDict(TypedDict):
|
|
437
|
-
id: str
|
|
438
|
-
r"""Unique ID for this input"""
|
|
439
437
|
type: InputSystemMetricsType
|
|
438
|
+
id: NotRequired[str]
|
|
439
|
+
r"""Unique ID for this input"""
|
|
440
440
|
disabled: NotRequired[bool]
|
|
441
441
|
pipeline: NotRequired[str]
|
|
442
442
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -463,11 +463,11 @@ class InputSystemMetricsTypedDict(TypedDict):
|
|
|
463
463
|
|
|
464
464
|
|
|
465
465
|
class InputSystemMetrics(BaseModel):
|
|
466
|
-
id: str
|
|
467
|
-
r"""Unique ID for this input"""
|
|
468
|
-
|
|
469
466
|
type: InputSystemMetricsType
|
|
470
467
|
|
|
468
|
+
id: Optional[str] = None
|
|
469
|
+
r"""Unique ID for this input"""
|
|
470
|
+
|
|
471
471
|
disabled: Optional[bool] = False
|
|
472
472
|
|
|
473
473
|
pipeline: Optional[str] = None
|
|
@@ -333,9 +333,9 @@ class InputSystemStatePersistence(BaseModel):
|
|
|
333
333
|
|
|
334
334
|
|
|
335
335
|
class InputSystemStateTypedDict(TypedDict):
|
|
336
|
-
id: str
|
|
337
|
-
r"""Unique ID for this input"""
|
|
338
336
|
type: InputSystemStateType
|
|
337
|
+
id: NotRequired[str]
|
|
338
|
+
r"""Unique ID for this input"""
|
|
339
339
|
disabled: NotRequired[bool]
|
|
340
340
|
pipeline: NotRequired[str]
|
|
341
341
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -362,11 +362,11 @@ class InputSystemStateTypedDict(TypedDict):
|
|
|
362
362
|
|
|
363
363
|
|
|
364
364
|
class InputSystemState(BaseModel):
|
|
365
|
-
id: str
|
|
366
|
-
r"""Unique ID for this input"""
|
|
367
|
-
|
|
368
365
|
type: InputSystemStateType
|
|
369
366
|
|
|
367
|
+
id: Optional[str] = None
|
|
368
|
+
r"""Unique ID for this input"""
|
|
369
|
+
|
|
370
370
|
disabled: Optional[bool] = False
|
|
371
371
|
|
|
372
372
|
pipeline: Optional[str] = None
|
|
@@ -196,11 +196,11 @@ class InputTCPAuthenticationMethod(str, Enum):
|
|
|
196
196
|
|
|
197
197
|
|
|
198
198
|
class InputTCPTypedDict(TypedDict):
|
|
199
|
+
type: InputTCPType
|
|
199
200
|
port: float
|
|
200
201
|
r"""Port to listen on"""
|
|
201
202
|
id: NotRequired[str]
|
|
202
203
|
r"""Unique ID for this input"""
|
|
203
|
-
type: NotRequired[InputTCPType]
|
|
204
204
|
disabled: NotRequired[bool]
|
|
205
205
|
pipeline: NotRequired[str]
|
|
206
206
|
r"""Pipeline to process data from this Source before sending it through the Routes"""
|
|
@@ -245,14 +245,14 @@ class InputTCPTypedDict(TypedDict):
|
|
|
245
245
|
|
|
246
246
|
|
|
247
247
|
class InputTCP(BaseModel):
|
|
248
|
+
type: InputTCPType
|
|
249
|
+
|
|
248
250
|
port: float
|
|
249
251
|
r"""Port to listen on"""
|
|
250
252
|
|
|
251
253
|
id: Optional[str] = None
|
|
252
254
|
r"""Unique ID for this input"""
|
|
253
255
|
|
|
254
|
-
type: Optional[InputTCPType] = None
|
|
255
|
-
|
|
256
256
|
disabled: Optional[bool] = False
|
|
257
257
|
|
|
258
258
|
pipeline: Optional[str] = None
|