cribl-control-plane 0.0.24__py3-none-any.whl → 0.0.26__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (115) hide show
  1. cribl_control_plane/_version.py +3 -3
  2. cribl_control_plane/destinations.py +10 -8
  3. cribl_control_plane/errors/healthstatus_error.py +1 -1
  4. cribl_control_plane/groups_sdk.py +10 -10
  5. cribl_control_plane/{healthinfo.py → health.py} +3 -1
  6. cribl_control_plane/models/__init__.py +21 -27
  7. cribl_control_plane/models/healthstatus.py +3 -3
  8. cribl_control_plane/models/input.py +4 -4
  9. cribl_control_plane/models/inputappscope.py +5 -5
  10. cribl_control_plane/models/inputcollection.py +2 -2
  11. cribl_control_plane/models/inputconfluentcloud.py +17 -3
  12. cribl_control_plane/models/inputcribl.py +5 -5
  13. cribl_control_plane/models/inputcriblhttp.py +3 -3
  14. cribl_control_plane/models/inputcribllakehttp.py +3 -3
  15. cribl_control_plane/models/inputcriblmetrics.py +5 -5
  16. cribl_control_plane/models/inputcribltcp.py +3 -3
  17. cribl_control_plane/models/inputdatadogagent.py +3 -3
  18. cribl_control_plane/models/inputedgeprometheus.py +3 -3
  19. cribl_control_plane/models/inputelastic.py +3 -3
  20. cribl_control_plane/models/inputeventhub.py +3 -3
  21. cribl_control_plane/models/inputfile.py +5 -5
  22. cribl_control_plane/models/inputfirehose.py +3 -3
  23. cribl_control_plane/models/inputgooglepubsub.py +16 -9
  24. cribl_control_plane/models/inputgrafana.py +20 -6
  25. cribl_control_plane/models/inputhttp.py +3 -3
  26. cribl_control_plane/models/inputhttpraw.py +3 -3
  27. cribl_control_plane/models/inputjournalfiles.py +3 -3
  28. cribl_control_plane/models/inputkafka.py +17 -3
  29. cribl_control_plane/models/inputkinesis.py +3 -3
  30. cribl_control_plane/models/inputkubeevents.py +5 -5
  31. cribl_control_plane/models/inputkubelogs.py +5 -5
  32. cribl_control_plane/models/inputkubemetrics.py +5 -5
  33. cribl_control_plane/models/inputloki.py +10 -3
  34. cribl_control_plane/models/inputmodeldriventelemetry.py +3 -3
  35. cribl_control_plane/models/inputmsk.py +17 -3
  36. cribl_control_plane/models/inputnetflow.py +3 -3
  37. cribl_control_plane/models/inputoffice365mgmt.py +3 -3
  38. cribl_control_plane/models/inputoffice365msgtrace.py +3 -3
  39. cribl_control_plane/models/inputoffice365service.py +3 -3
  40. cribl_control_plane/models/inputopentelemetry.py +3 -3
  41. cribl_control_plane/models/inputprometheus.py +3 -3
  42. cribl_control_plane/models/inputprometheusrw.py +3 -3
  43. cribl_control_plane/models/inputrawudp.py +3 -3
  44. cribl_control_plane/models/inputsnmp.py +3 -3
  45. cribl_control_plane/models/inputsplunk.py +3 -3
  46. cribl_control_plane/models/inputsplunkhec.py +3 -3
  47. cribl_control_plane/models/inputsplunksearch.py +3 -3
  48. cribl_control_plane/models/inputsqs.py +3 -3
  49. cribl_control_plane/models/inputsystemmetrics.py +5 -5
  50. cribl_control_plane/models/inputsystemstate.py +5 -5
  51. cribl_control_plane/models/inputtcp.py +3 -3
  52. cribl_control_plane/models/inputtcpjson.py +3 -3
  53. cribl_control_plane/models/inputwef.py +3 -3
  54. cribl_control_plane/models/inputwindowsmetrics.py +5 -5
  55. cribl_control_plane/models/inputwiz.py +3 -3
  56. cribl_control_plane/models/inputzscalerhec.py +3 -3
  57. cribl_control_plane/models/output.py +14 -14
  58. cribl_control_plane/models/outputazureblob.py +3 -3
  59. cribl_control_plane/models/outputazuredataexplorer.py +3 -3
  60. cribl_control_plane/models/outputazureeventhub.py +3 -3
  61. cribl_control_plane/models/outputclickhouse.py +3 -3
  62. cribl_control_plane/models/outputcloudwatch.py +3 -3
  63. cribl_control_plane/models/outputconfluentcloud.py +17 -3
  64. cribl_control_plane/models/outputcriblhttp.py +5 -5
  65. cribl_control_plane/models/outputcribllake.py +5 -5
  66. cribl_control_plane/models/outputcribltcp.py +5 -5
  67. cribl_control_plane/models/outputcrowdstrikenextgensiem.py +3 -3
  68. cribl_control_plane/models/outputdatadog.py +5 -5
  69. cribl_control_plane/models/outputdataset.py +5 -5
  70. cribl_control_plane/models/outputdevnull.py +5 -5
  71. cribl_control_plane/models/outputdiskspool.py +5 -5
  72. cribl_control_plane/models/outputdls3.py +5 -5
  73. cribl_control_plane/models/outputdynatracehttp.py +3 -3
  74. cribl_control_plane/models/outputdynatraceotlp.py +3 -3
  75. cribl_control_plane/models/outputelasticcloud.py +3 -3
  76. cribl_control_plane/models/outputexabeam.py +3 -3
  77. cribl_control_plane/models/outputgooglecloudlogging.py +3 -3
  78. cribl_control_plane/models/outputgooglecloudstorage.py +5 -5
  79. cribl_control_plane/models/outputgrafanacloud.py +24 -10
  80. cribl_control_plane/models/outputgraphite.py +3 -3
  81. cribl_control_plane/models/outputhumiohec.py +3 -3
  82. cribl_control_plane/models/outputkafka.py +17 -3
  83. cribl_control_plane/models/outputkinesis.py +3 -3
  84. cribl_control_plane/models/outputloki.py +14 -0
  85. cribl_control_plane/models/outputminio.py +3 -3
  86. cribl_control_plane/models/outputmsk.py +17 -3
  87. cribl_control_plane/models/outputnewrelic.py +5 -5
  88. cribl_control_plane/models/outputnewrelicevents.py +3 -3
  89. cribl_control_plane/models/outputring.py +5 -5
  90. cribl_control_plane/models/outputs3.py +5 -5
  91. cribl_control_plane/models/outputsecuritylake.py +3 -3
  92. cribl_control_plane/models/outputsentinel.py +3 -3
  93. cribl_control_plane/models/outputsentineloneaisiem.py +3 -3
  94. cribl_control_plane/models/outputservicenow.py +3 -3
  95. cribl_control_plane/models/outputsns.py +3 -3
  96. cribl_control_plane/models/outputsplunk.py +3 -3
  97. cribl_control_plane/models/outputsplunkhec.py +5 -5
  98. cribl_control_plane/models/outputsqs.py +3 -3
  99. cribl_control_plane/models/outputstatsd.py +3 -3
  100. cribl_control_plane/models/outputstatsdext.py +3 -3
  101. cribl_control_plane/models/outputsyslog.py +5 -5
  102. cribl_control_plane/models/outputtcpjson.py +5 -5
  103. cribl_control_plane/models/outputwebhook.py +5 -5
  104. cribl_control_plane/models/outputxsiam.py +5 -5
  105. cribl_control_plane/nodes.py +100 -90
  106. cribl_control_plane/pipelines.py +20 -20
  107. cribl_control_plane/sdk.py +6 -6
  108. cribl_control_plane/sources.py +2 -0
  109. cribl_control_plane/versioning.py +14 -14
  110. {cribl_control_plane-0.0.24.dist-info → cribl_control_plane-0.0.26.dist-info}/METADATA +24 -28
  111. {cribl_control_plane-0.0.24.dist-info → cribl_control_plane-0.0.26.dist-info}/RECORD +112 -115
  112. cribl_control_plane/deployments.py +0 -185
  113. cribl_control_plane/models/restartresponse.py +0 -26
  114. cribl_control_plane/models/updateworkersrestartop.py +0 -24
  115. {cribl_control_plane-0.0.24.dist-info → cribl_control_plane-0.0.26.dist-info}/WHEEL +0 -0
@@ -151,9 +151,9 @@ class InputKubeLogsDiskSpooling(BaseModel):
151
151
 
152
152
 
153
153
  class InputKubeLogsTypedDict(TypedDict):
154
- id: str
155
- r"""Unique ID for this input"""
156
154
  type: InputKubeLogsType
155
+ id: NotRequired[str]
156
+ r"""Unique ID for this input"""
157
157
  disabled: NotRequired[bool]
158
158
  pipeline: NotRequired[str]
159
159
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -187,11 +187,11 @@ class InputKubeLogsTypedDict(TypedDict):
187
187
 
188
188
 
189
189
  class InputKubeLogs(BaseModel):
190
- id: str
191
- r"""Unique ID for this input"""
192
-
193
190
  type: InputKubeLogsType
194
191
 
192
+ id: Optional[str] = None
193
+ r"""Unique ID for this input"""
194
+
195
195
  disabled: Optional[bool] = False
196
196
 
197
197
  pipeline: Optional[str] = None
@@ -154,9 +154,9 @@ class InputKubeMetricsPersistence(BaseModel):
154
154
 
155
155
 
156
156
  class InputKubeMetricsTypedDict(TypedDict):
157
- id: str
158
- r"""Unique ID for this input"""
159
157
  type: InputKubeMetricsType
158
+ id: NotRequired[str]
159
+ r"""Unique ID for this input"""
160
160
  disabled: NotRequired[bool]
161
161
  pipeline: NotRequired[str]
162
162
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -182,11 +182,11 @@ class InputKubeMetricsTypedDict(TypedDict):
182
182
 
183
183
 
184
184
  class InputKubeMetrics(BaseModel):
185
- id: str
186
- r"""Unique ID for this input"""
187
-
188
185
  type: InputKubeMetricsType
189
186
 
187
+ id: Optional[str] = None
188
+ r"""Unique ID for this input"""
189
+
190
190
  disabled: Optional[bool] = False
191
191
 
192
192
  pipeline: Optional[str] = None
@@ -212,11 +212,11 @@ class InputLokiOauthHeader(BaseModel):
212
212
 
213
213
 
214
214
  class InputLokiTypedDict(TypedDict):
215
+ type: InputLokiType
215
216
  port: float
216
217
  r"""Port to listen on"""
217
218
  id: NotRequired[str]
218
219
  r"""Unique ID for this input"""
219
- type: NotRequired[InputLokiType]
220
220
  disabled: NotRequired[bool]
221
221
  pipeline: NotRequired[str]
222
222
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -258,6 +258,8 @@ class InputLokiTypedDict(TypedDict):
258
258
  r"""Messages from matched IP addresses will be ignored. This takes precedence over the allowlist."""
259
259
  loki_api: NotRequired[str]
260
260
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'."""
261
+ extract_structured_metadata: NotRequired[bool]
262
+ r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
261
263
  auth_type: NotRequired[InputLokiAuthenticationType]
262
264
  r"""Loki logs authentication type"""
263
265
  metadata: NotRequired[List[InputLokiMetadatumTypedDict]]
@@ -290,14 +292,14 @@ class InputLokiTypedDict(TypedDict):
290
292
 
291
293
 
292
294
  class InputLoki(BaseModel):
295
+ type: InputLokiType
296
+
293
297
  port: float
294
298
  r"""Port to listen on"""
295
299
 
296
300
  id: Optional[str] = None
297
301
  r"""Unique ID for this input"""
298
302
 
299
- type: Optional[InputLokiType] = None
300
-
301
303
  disabled: Optional[bool] = False
302
304
 
303
305
  pipeline: Optional[str] = None
@@ -387,6 +389,11 @@ class InputLoki(BaseModel):
387
389
  )
388
390
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'."""
389
391
 
392
+ extract_structured_metadata: Annotated[
393
+ Optional[bool], pydantic.Field(alias="extractStructuredMetadata")
394
+ ] = False
395
+ r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
396
+
390
397
  auth_type: Annotated[
391
398
  Optional[InputLokiAuthenticationType], pydantic.Field(alias="authType")
392
399
  ] = InputLokiAuthenticationType.NONE
@@ -170,9 +170,9 @@ class InputModelDrivenTelemetryMetadatum(BaseModel):
170
170
 
171
171
 
172
172
  class InputModelDrivenTelemetryTypedDict(TypedDict):
173
+ type: InputModelDrivenTelemetryType
173
174
  id: NotRequired[str]
174
175
  r"""Unique ID for this input"""
175
- type: NotRequired[InputModelDrivenTelemetryType]
176
176
  disabled: NotRequired[bool]
177
177
  pipeline: NotRequired[str]
178
178
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -202,11 +202,11 @@ class InputModelDrivenTelemetryTypedDict(TypedDict):
202
202
 
203
203
 
204
204
  class InputModelDrivenTelemetry(BaseModel):
205
+ type: InputModelDrivenTelemetryType
206
+
205
207
  id: Optional[str] = None
206
208
  r"""Unique ID for this input"""
207
209
 
208
- type: Optional[InputModelDrivenTelemetryType] = None
209
-
210
210
  disabled: Optional[bool] = False
211
211
 
212
212
  pipeline: Optional[str] = None
@@ -96,6 +96,13 @@ class InputMskMetadatum(BaseModel):
96
96
  r"""JavaScript expression to compute field's value, enclosed in quotes or backticks. (Can evaluate to a constant.)"""
97
97
 
98
98
 
99
+ class InputMskSchemaType(str, Enum):
100
+ r"""The schema format used to encode and decode event data"""
101
+
102
+ AVRO = "avro"
103
+ JSON = "json"
104
+
105
+
99
106
  class InputMskAuthTypedDict(TypedDict):
100
107
  r"""Credentials to use when authenticating with the schema registry using basic HTTP authentication"""
101
108
 
@@ -196,6 +203,8 @@ class InputMskKafkaSchemaRegistryAuthenticationTypedDict(TypedDict):
196
203
  disabled: NotRequired[bool]
197
204
  schema_registry_url: NotRequired[str]
198
205
  r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
206
+ schema_type: NotRequired[InputMskSchemaType]
207
+ r"""The schema format used to encode and decode event data"""
199
208
  connection_timeout: NotRequired[float]
200
209
  r"""Maximum time to wait for a Schema Registry connection to complete successfully"""
201
210
  request_timeout: NotRequired[float]
@@ -215,6 +224,11 @@ class InputMskKafkaSchemaRegistryAuthentication(BaseModel):
215
224
  ] = "http://localhost:8081"
216
225
  r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
217
226
 
227
+ schema_type: Annotated[
228
+ Optional[InputMskSchemaType], pydantic.Field(alias="schemaType")
229
+ ] = InputMskSchemaType.AVRO
230
+ r"""The schema format used to encode and decode event data"""
231
+
218
232
  connection_timeout: Annotated[
219
233
  Optional[float], pydantic.Field(alias="connectionTimeout")
220
234
  ] = 30000
@@ -325,6 +339,7 @@ class InputMskTLSSettingsClientSide(BaseModel):
325
339
 
326
340
 
327
341
  class InputMskTypedDict(TypedDict):
342
+ type: InputMskType
328
343
  brokers: List[str]
329
344
  r"""Enter each Kafka bootstrap server you want to use. Specify the hostname and port (such as mykafkabroker:9092) or just the hostname (in which case @{product} will assign port 9092)."""
330
345
  topics: List[str]
@@ -333,7 +348,6 @@ class InputMskTypedDict(TypedDict):
333
348
  r"""Region where the MSK cluster is located"""
334
349
  id: NotRequired[str]
335
350
  r"""Unique ID for this input"""
336
- type: NotRequired[InputMskType]
337
351
  disabled: NotRequired[bool]
338
352
  pipeline: NotRequired[str]
339
353
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -427,6 +441,8 @@ class InputMskTypedDict(TypedDict):
427
441
 
428
442
 
429
443
  class InputMsk(BaseModel):
444
+ type: InputMskType
445
+
430
446
  brokers: List[str]
431
447
  r"""Enter each Kafka bootstrap server you want to use. Specify the hostname and port (such as mykafkabroker:9092) or just the hostname (in which case @{product} will assign port 9092)."""
432
448
 
@@ -439,8 +455,6 @@ class InputMsk(BaseModel):
439
455
  id: Optional[str] = None
440
456
  r"""Unique ID for this input"""
441
457
 
442
- type: Optional[InputMskType] = None
443
-
444
458
  disabled: Optional[bool] = False
445
459
 
446
460
  pipeline: Optional[str] = None
@@ -97,9 +97,9 @@ class InputNetflowMetadatum(BaseModel):
97
97
 
98
98
 
99
99
  class InputNetflowTypedDict(TypedDict):
100
+ type: InputNetflowType
100
101
  id: NotRequired[str]
101
102
  r"""Unique ID for this input"""
102
- type: NotRequired[InputNetflowType]
103
103
  disabled: NotRequired[bool]
104
104
  pipeline: NotRequired[str]
105
105
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -140,11 +140,11 @@ class InputNetflowTypedDict(TypedDict):
140
140
 
141
141
 
142
142
  class InputNetflow(BaseModel):
143
+ type: InputNetflowType
144
+
143
145
  id: Optional[str] = None
144
146
  r"""Unique ID for this input"""
145
147
 
146
- type: Optional[InputNetflowType] = None
147
-
148
148
  disabled: Optional[bool] = False
149
149
 
150
150
  pipeline: Optional[str] = None
@@ -211,13 +211,13 @@ class InputOffice365MgmtAuthenticationMethod(str, Enum):
211
211
 
212
212
 
213
213
  class InputOffice365MgmtTypedDict(TypedDict):
214
+ type: InputOffice365MgmtType
214
215
  tenant_id: str
215
216
  r"""Office 365 Azure Tenant ID"""
216
217
  app_id: str
217
218
  r"""Office 365 Azure Application ID"""
218
219
  id: NotRequired[str]
219
220
  r"""Unique ID for this input"""
220
- type: NotRequired[InputOffice365MgmtType]
221
221
  disabled: NotRequired[bool]
222
222
  pipeline: NotRequired[str]
223
223
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -265,6 +265,8 @@ class InputOffice365MgmtTypedDict(TypedDict):
265
265
 
266
266
 
267
267
  class InputOffice365Mgmt(BaseModel):
268
+ type: InputOffice365MgmtType
269
+
268
270
  tenant_id: Annotated[str, pydantic.Field(alias="tenantId")]
269
271
  r"""Office 365 Azure Tenant ID"""
270
272
 
@@ -274,8 +276,6 @@ class InputOffice365Mgmt(BaseModel):
274
276
  id: Optional[str] = None
275
277
  r"""Unique ID for this input"""
276
278
 
277
- type: Optional[InputOffice365MgmtType] = None
278
-
279
279
  disabled: Optional[bool] = False
280
280
 
281
281
  pipeline: Optional[str] = None
@@ -216,9 +216,9 @@ class CertOptions(BaseModel):
216
216
 
217
217
 
218
218
  class InputOffice365MsgTraceTypedDict(TypedDict):
219
+ type: InputOffice365MsgTraceType
219
220
  id: NotRequired[str]
220
221
  r"""Unique ID for this input"""
221
- type: NotRequired[InputOffice365MsgTraceType]
222
222
  disabled: NotRequired[bool]
223
223
  pipeline: NotRequired[str]
224
224
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -289,11 +289,11 @@ class InputOffice365MsgTraceTypedDict(TypedDict):
289
289
 
290
290
 
291
291
  class InputOffice365MsgTrace(BaseModel):
292
+ type: InputOffice365MsgTraceType
293
+
292
294
  id: Optional[str] = None
293
295
  r"""Unique ID for this input"""
294
296
 
295
- type: Optional[InputOffice365MsgTraceType] = None
296
-
297
297
  disabled: Optional[bool] = False
298
298
 
299
299
  pipeline: Optional[str] = None
@@ -213,13 +213,13 @@ class InputOffice365ServiceAuthenticationMethod(str, Enum):
213
213
 
214
214
 
215
215
  class InputOffice365ServiceTypedDict(TypedDict):
216
+ type: InputOffice365ServiceType
216
217
  tenant_id: str
217
218
  r"""Office 365 Azure Tenant ID"""
218
219
  app_id: str
219
220
  r"""Office 365 Azure Application ID"""
220
221
  id: NotRequired[str]
221
222
  r"""Unique ID for this input"""
222
- type: NotRequired[InputOffice365ServiceType]
223
223
  disabled: NotRequired[bool]
224
224
  pipeline: NotRequired[str]
225
225
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -263,6 +263,8 @@ class InputOffice365ServiceTypedDict(TypedDict):
263
263
 
264
264
 
265
265
  class InputOffice365Service(BaseModel):
266
+ type: InputOffice365ServiceType
267
+
266
268
  tenant_id: Annotated[str, pydantic.Field(alias="tenantId")]
267
269
  r"""Office 365 Azure Tenant ID"""
268
270
 
@@ -272,8 +274,6 @@ class InputOffice365Service(BaseModel):
272
274
  id: Optional[str] = None
273
275
  r"""Unique ID for this input"""
274
276
 
275
- type: Optional[InputOffice365ServiceType] = None
276
-
277
277
  disabled: Optional[bool] = False
278
278
 
279
279
  pipeline: Optional[str] = None
@@ -230,9 +230,9 @@ class InputOpenTelemetryOauthHeader(BaseModel):
230
230
 
231
231
 
232
232
  class InputOpenTelemetryTypedDict(TypedDict):
233
+ type: InputOpenTelemetryType
233
234
  id: NotRequired[str]
234
235
  r"""Unique ID for this input"""
235
- type: NotRequired[InputOpenTelemetryType]
236
236
  disabled: NotRequired[bool]
237
237
  pipeline: NotRequired[str]
238
238
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -315,11 +315,11 @@ class InputOpenTelemetryTypedDict(TypedDict):
315
315
 
316
316
 
317
317
  class InputOpenTelemetry(BaseModel):
318
+ type: InputOpenTelemetryType
319
+
318
320
  id: Optional[str] = None
319
321
  r"""Unique ID for this input"""
320
322
 
321
- type: Optional[InputOpenTelemetryType] = None
322
-
323
323
  disabled: Optional[bool] = False
324
324
 
325
325
  pipeline: Optional[str] = None
@@ -166,9 +166,9 @@ class InputPrometheusSignatureVersion(str, Enum):
166
166
 
167
167
 
168
168
  class InputPrometheusTypedDict(TypedDict):
169
+ type: InputPrometheusType
169
170
  id: NotRequired[str]
170
171
  r"""Unique ID for this input"""
171
- type: NotRequired[InputPrometheusType]
172
172
  disabled: NotRequired[bool]
173
173
  pipeline: NotRequired[str]
174
174
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -254,11 +254,11 @@ class InputPrometheusTypedDict(TypedDict):
254
254
 
255
255
 
256
256
  class InputPrometheus(BaseModel):
257
+ type: InputPrometheusType
258
+
257
259
  id: Optional[str] = None
258
260
  r"""Unique ID for this input"""
259
261
 
260
- type: Optional[InputPrometheusType] = None
261
-
262
262
  disabled: Optional[bool] = False
263
263
 
264
264
  pipeline: Optional[str] = None
@@ -212,11 +212,11 @@ class InputPrometheusRwOauthHeader(BaseModel):
212
212
 
213
213
 
214
214
  class InputPrometheusRwTypedDict(TypedDict):
215
+ type: InputPrometheusRwType
215
216
  port: float
216
217
  r"""Port to listen on"""
217
218
  id: NotRequired[str]
218
219
  r"""Unique ID for this input"""
219
- type: NotRequired[InputPrometheusRwType]
220
220
  disabled: NotRequired[bool]
221
221
  pipeline: NotRequired[str]
222
222
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -290,14 +290,14 @@ class InputPrometheusRwTypedDict(TypedDict):
290
290
 
291
291
 
292
292
  class InputPrometheusRw(BaseModel):
293
+ type: InputPrometheusRwType
294
+
293
295
  port: float
294
296
  r"""Port to listen on"""
295
297
 
296
298
  id: Optional[str] = None
297
299
  r"""Unique ID for this input"""
298
300
 
299
- type: Optional[InputPrometheusRwType] = None
300
-
301
301
  disabled: Optional[bool] = False
302
302
 
303
303
  pipeline: Optional[str] = None
@@ -97,11 +97,11 @@ class InputRawUDPMetadatum(BaseModel):
97
97
 
98
98
 
99
99
  class InputRawUDPTypedDict(TypedDict):
100
+ type: InputRawUDPType
100
101
  port: float
101
102
  r"""Port to listen on"""
102
103
  id: NotRequired[str]
103
104
  r"""Unique ID for this input"""
104
- type: NotRequired[InputRawUDPType]
105
105
  disabled: NotRequired[bool]
106
106
  pipeline: NotRequired[str]
107
107
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -134,14 +134,14 @@ class InputRawUDPTypedDict(TypedDict):
134
134
 
135
135
 
136
136
  class InputRawUDP(BaseModel):
137
+ type: InputRawUDPType
138
+
137
139
  port: float
138
140
  r"""Port to listen on"""
139
141
 
140
142
  id: Optional[str] = None
141
143
  r"""Unique ID for this input"""
142
144
 
143
- type: Optional[InputRawUDPType] = None
144
-
145
145
  disabled: Optional[bool] = False
146
146
 
147
147
  pipeline: Optional[str] = None
@@ -154,9 +154,9 @@ class InputSnmpMetadatum(BaseModel):
154
154
 
155
155
 
156
156
  class InputSnmpTypedDict(TypedDict):
157
+ type: InputSnmpType
157
158
  id: NotRequired[str]
158
159
  r"""Unique ID for this input"""
159
- type: NotRequired[InputSnmpType]
160
160
  disabled: NotRequired[bool]
161
161
  pipeline: NotRequired[str]
162
162
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -193,11 +193,11 @@ class InputSnmpTypedDict(TypedDict):
193
193
 
194
194
 
195
195
  class InputSnmp(BaseModel):
196
+ type: InputSnmpType
197
+
196
198
  id: Optional[str] = None
197
199
  r"""Unique ID for this input"""
198
200
 
199
- type: Optional[InputSnmpType] = None
200
-
201
201
  disabled: Optional[bool] = False
202
202
 
203
203
  pipeline: Optional[str] = None
@@ -199,11 +199,11 @@ class InputSplunkCompression(str, Enum):
199
199
 
200
200
 
201
201
  class InputSplunkTypedDict(TypedDict):
202
+ type: InputSplunkType
202
203
  port: float
203
204
  r"""Port to listen on"""
204
205
  id: NotRequired[str]
205
206
  r"""Unique ID for this input"""
206
- type: NotRequired[InputSplunkType]
207
207
  disabled: NotRequired[bool]
208
208
  pipeline: NotRequired[str]
209
209
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -255,14 +255,14 @@ class InputSplunkTypedDict(TypedDict):
255
255
 
256
256
 
257
257
  class InputSplunk(BaseModel):
258
+ type: InputSplunkType
259
+
258
260
  port: float
259
261
  r"""Port to listen on"""
260
262
 
261
263
  id: Optional[str] = None
262
264
  r"""Unique ID for this input"""
263
265
 
264
- type: Optional[InputSplunkType] = None
265
-
266
266
  disabled: Optional[bool] = False
267
267
 
268
268
  pipeline: Optional[str] = None
@@ -229,11 +229,11 @@ class InputSplunkHecMetadatum(BaseModel):
229
229
 
230
230
 
231
231
  class InputSplunkHecTypedDict(TypedDict):
232
+ type: InputSplunkHecType
232
233
  port: float
233
234
  r"""Port to listen on"""
234
235
  id: NotRequired[str]
235
236
  r"""Unique ID for this input"""
236
- type: NotRequired[InputSplunkHecType]
237
237
  disabled: NotRequired[bool]
238
238
  pipeline: NotRequired[str]
239
239
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -302,14 +302,14 @@ class InputSplunkHecTypedDict(TypedDict):
302
302
 
303
303
 
304
304
  class InputSplunkHec(BaseModel):
305
+ type: InputSplunkHecType
306
+
305
307
  port: float
306
308
  r"""Port to listen on"""
307
309
 
308
310
  id: Optional[str] = None
309
311
  r"""Unique ID for this input"""
310
312
 
311
- type: Optional[InputSplunkHecType] = None
312
-
313
313
  disabled: Optional[bool] = False
314
314
 
315
315
  pipeline: Optional[str] = None
@@ -239,11 +239,11 @@ class InputSplunkSearchOauthHeader(BaseModel):
239
239
 
240
240
 
241
241
  class InputSplunkSearchTypedDict(TypedDict):
242
+ type: InputSplunkSearchType
242
243
  search: str
243
244
  r"""Enter Splunk search here. Examples: 'index=myAppLogs level=error channel=myApp' OR '| mstats avg(myStat) as myStat WHERE index=myStatsIndex.'"""
244
245
  id: NotRequired[str]
245
246
  r"""Unique ID for this input"""
246
- type: NotRequired[InputSplunkSearchType]
247
247
  disabled: NotRequired[bool]
248
248
  pipeline: NotRequired[str]
249
249
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -331,14 +331,14 @@ class InputSplunkSearchTypedDict(TypedDict):
331
331
 
332
332
 
333
333
  class InputSplunkSearch(BaseModel):
334
+ type: InputSplunkSearchType
335
+
334
336
  search: str
335
337
  r"""Enter Splunk search here. Examples: 'index=myAppLogs level=error channel=myApp' OR '| mstats avg(myStat) as myStat WHERE index=myStatsIndex.'"""
336
338
 
337
339
  id: Optional[str] = None
338
340
  r"""Unique ID for this input"""
339
341
 
340
- type: Optional[InputSplunkSearchType] = None
341
-
342
342
  disabled: Optional[bool] = False
343
343
 
344
344
  pipeline: Optional[str] = None
@@ -119,11 +119,11 @@ class InputSqsMetadatum(BaseModel):
119
119
 
120
120
 
121
121
  class InputSqsTypedDict(TypedDict):
122
+ type: InputSqsType
122
123
  queue_name: str
123
124
  r"""The name, URL, or ARN of the SQS queue to read events from. When a non-AWS URL is specified, format must be: '{url}/myQueueName'. Example: 'https://host:port/myQueueName'. Value must be a JavaScript expression (which can evaluate to a constant value), enclosed in quotes or backticks. Can only be evaluated at init time. Example referencing a Global Variable: `https://host:port/myQueue-${C.vars.myVar}`."""
124
125
  id: NotRequired[str]
125
126
  r"""Unique ID for this input"""
126
- type: NotRequired[InputSqsType]
127
127
  disabled: NotRequired[bool]
128
128
  pipeline: NotRequired[str]
129
129
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -182,14 +182,14 @@ class InputSqsTypedDict(TypedDict):
182
182
 
183
183
 
184
184
  class InputSqs(BaseModel):
185
+ type: InputSqsType
186
+
185
187
  queue_name: Annotated[str, pydantic.Field(alias="queueName")]
186
188
  r"""The name, URL, or ARN of the SQS queue to read events from. When a non-AWS URL is specified, format must be: '{url}/myQueueName'. Example: 'https://host:port/myQueueName'. Value must be a JavaScript expression (which can evaluate to a constant value), enclosed in quotes or backticks. Can only be evaluated at init time. Example referencing a Global Variable: `https://host:port/myQueue-${C.vars.myVar}`."""
187
189
 
188
190
  id: Optional[str] = None
189
191
  r"""Unique ID for this input"""
190
192
 
191
- type: Optional[InputSqsType] = None
192
-
193
193
  disabled: Optional[bool] = False
194
194
 
195
195
  pipeline: Optional[str] = None
@@ -434,9 +434,9 @@ class InputSystemMetricsPersistence(BaseModel):
434
434
 
435
435
 
436
436
  class InputSystemMetricsTypedDict(TypedDict):
437
- id: str
438
- r"""Unique ID for this input"""
439
437
  type: InputSystemMetricsType
438
+ id: NotRequired[str]
439
+ r"""Unique ID for this input"""
440
440
  disabled: NotRequired[bool]
441
441
  pipeline: NotRequired[str]
442
442
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -463,11 +463,11 @@ class InputSystemMetricsTypedDict(TypedDict):
463
463
 
464
464
 
465
465
  class InputSystemMetrics(BaseModel):
466
- id: str
467
- r"""Unique ID for this input"""
468
-
469
466
  type: InputSystemMetricsType
470
467
 
468
+ id: Optional[str] = None
469
+ r"""Unique ID for this input"""
470
+
471
471
  disabled: Optional[bool] = False
472
472
 
473
473
  pipeline: Optional[str] = None
@@ -333,9 +333,9 @@ class InputSystemStatePersistence(BaseModel):
333
333
 
334
334
 
335
335
  class InputSystemStateTypedDict(TypedDict):
336
- id: str
337
- r"""Unique ID for this input"""
338
336
  type: InputSystemStateType
337
+ id: NotRequired[str]
338
+ r"""Unique ID for this input"""
339
339
  disabled: NotRequired[bool]
340
340
  pipeline: NotRequired[str]
341
341
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -362,11 +362,11 @@ class InputSystemStateTypedDict(TypedDict):
362
362
 
363
363
 
364
364
  class InputSystemState(BaseModel):
365
- id: str
366
- r"""Unique ID for this input"""
367
-
368
365
  type: InputSystemStateType
369
366
 
367
+ id: Optional[str] = None
368
+ r"""Unique ID for this input"""
369
+
370
370
  disabled: Optional[bool] = False
371
371
 
372
372
  pipeline: Optional[str] = None
@@ -196,11 +196,11 @@ class InputTCPAuthenticationMethod(str, Enum):
196
196
 
197
197
 
198
198
  class InputTCPTypedDict(TypedDict):
199
+ type: InputTCPType
199
200
  port: float
200
201
  r"""Port to listen on"""
201
202
  id: NotRequired[str]
202
203
  r"""Unique ID for this input"""
203
- type: NotRequired[InputTCPType]
204
204
  disabled: NotRequired[bool]
205
205
  pipeline: NotRequired[str]
206
206
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -245,14 +245,14 @@ class InputTCPTypedDict(TypedDict):
245
245
 
246
246
 
247
247
  class InputTCP(BaseModel):
248
+ type: InputTCPType
249
+
248
250
  port: float
249
251
  r"""Port to listen on"""
250
252
 
251
253
  id: Optional[str] = None
252
254
  r"""Unique ID for this input"""
253
255
 
254
- type: Optional[InputTCPType] = None
255
-
256
256
  disabled: Optional[bool] = False
257
257
 
258
258
  pipeline: Optional[str] = None