cribl-control-plane 0.0.26a1__py3-none-any.whl → 0.0.27__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (127) hide show
  1. cribl_control_plane/_version.py +3 -3
  2. cribl_control_plane/acl.py +203 -0
  3. cribl_control_plane/auth_sdk.py +9 -176
  4. cribl_control_plane/branches.py +335 -0
  5. cribl_control_plane/commits.py +1141 -0
  6. cribl_control_plane/commits_files.py +371 -0
  7. cribl_control_plane/configs_versions.py +189 -0
  8. cribl_control_plane/destinations.py +18 -729
  9. cribl_control_plane/destinations_pq.py +359 -0
  10. cribl_control_plane/errors/healthstatus_error.py +1 -1
  11. cribl_control_plane/groups_configs.py +17 -0
  12. cribl_control_plane/groups_sdk.py +18 -551
  13. cribl_control_plane/{healthinfo.py → health.py} +3 -1
  14. cribl_control_plane/hectokens.py +479 -0
  15. cribl_control_plane/models/__init__.py +3 -35
  16. cribl_control_plane/models/healthstatus.py +3 -3
  17. cribl_control_plane/models/inputappscope.py +5 -5
  18. cribl_control_plane/models/inputcollection.py +2 -2
  19. cribl_control_plane/models/inputconfluentcloud.py +3 -3
  20. cribl_control_plane/models/inputcribl.py +5 -5
  21. cribl_control_plane/models/inputcriblhttp.py +3 -3
  22. cribl_control_plane/models/inputcribllakehttp.py +3 -3
  23. cribl_control_plane/models/inputcriblmetrics.py +5 -5
  24. cribl_control_plane/models/inputcribltcp.py +3 -3
  25. cribl_control_plane/models/inputdatadogagent.py +3 -3
  26. cribl_control_plane/models/inputedgeprometheus.py +3 -3
  27. cribl_control_plane/models/inputelastic.py +3 -3
  28. cribl_control_plane/models/inputeventhub.py +3 -3
  29. cribl_control_plane/models/inputfile.py +5 -5
  30. cribl_control_plane/models/inputfirehose.py +3 -3
  31. cribl_control_plane/models/inputgooglepubsub.py +3 -3
  32. cribl_control_plane/models/inputgrafana.py +6 -6
  33. cribl_control_plane/models/inputhttp.py +3 -3
  34. cribl_control_plane/models/inputhttpraw.py +3 -3
  35. cribl_control_plane/models/inputjournalfiles.py +3 -3
  36. cribl_control_plane/models/inputkafka.py +3 -3
  37. cribl_control_plane/models/inputkinesis.py +3 -3
  38. cribl_control_plane/models/inputkubeevents.py +5 -5
  39. cribl_control_plane/models/inputkubelogs.py +5 -5
  40. cribl_control_plane/models/inputkubemetrics.py +5 -5
  41. cribl_control_plane/models/inputloki.py +3 -3
  42. cribl_control_plane/models/inputmodeldriventelemetry.py +3 -3
  43. cribl_control_plane/models/inputmsk.py +3 -3
  44. cribl_control_plane/models/inputnetflow.py +3 -3
  45. cribl_control_plane/models/inputoffice365mgmt.py +3 -3
  46. cribl_control_plane/models/inputoffice365msgtrace.py +3 -3
  47. cribl_control_plane/models/inputoffice365service.py +3 -3
  48. cribl_control_plane/models/inputopentelemetry.py +3 -3
  49. cribl_control_plane/models/inputprometheus.py +3 -3
  50. cribl_control_plane/models/inputprometheusrw.py +3 -3
  51. cribl_control_plane/models/inputrawudp.py +3 -3
  52. cribl_control_plane/models/inputsnmp.py +3 -3
  53. cribl_control_plane/models/inputsplunk.py +3 -3
  54. cribl_control_plane/models/inputsplunkhec.py +3 -3
  55. cribl_control_plane/models/inputsplunksearch.py +3 -3
  56. cribl_control_plane/models/inputsqs.py +3 -3
  57. cribl_control_plane/models/inputsystemmetrics.py +5 -5
  58. cribl_control_plane/models/inputsystemstate.py +5 -5
  59. cribl_control_plane/models/inputtcp.py +3 -3
  60. cribl_control_plane/models/inputtcpjson.py +3 -3
  61. cribl_control_plane/models/inputwef.py +3 -3
  62. cribl_control_plane/models/inputwindowsmetrics.py +5 -5
  63. cribl_control_plane/models/inputwiz.py +3 -3
  64. cribl_control_plane/models/inputzscalerhec.py +3 -3
  65. cribl_control_plane/models/outputazureblob.py +3 -3
  66. cribl_control_plane/models/outputazuredataexplorer.py +3 -3
  67. cribl_control_plane/models/outputazureeventhub.py +3 -3
  68. cribl_control_plane/models/outputclickhouse.py +3 -3
  69. cribl_control_plane/models/outputcloudwatch.py +3 -3
  70. cribl_control_plane/models/outputconfluentcloud.py +3 -3
  71. cribl_control_plane/models/outputcriblhttp.py +5 -5
  72. cribl_control_plane/models/outputcribllake.py +5 -5
  73. cribl_control_plane/models/outputcribltcp.py +5 -5
  74. cribl_control_plane/models/outputcrowdstrikenextgensiem.py +3 -3
  75. cribl_control_plane/models/outputdatadog.py +5 -5
  76. cribl_control_plane/models/outputdataset.py +5 -5
  77. cribl_control_plane/models/outputdevnull.py +5 -5
  78. cribl_control_plane/models/outputdiskspool.py +5 -5
  79. cribl_control_plane/models/outputdls3.py +3 -3
  80. cribl_control_plane/models/outputdynatracehttp.py +3 -3
  81. cribl_control_plane/models/outputdynatraceotlp.py +3 -3
  82. cribl_control_plane/models/outputelasticcloud.py +3 -3
  83. cribl_control_plane/models/outputexabeam.py +3 -3
  84. cribl_control_plane/models/outputgooglecloudlogging.py +3 -3
  85. cribl_control_plane/models/outputgooglecloudstorage.py +3 -3
  86. cribl_control_plane/models/outputgrafanacloud.py +10 -10
  87. cribl_control_plane/models/outputgraphite.py +3 -3
  88. cribl_control_plane/models/outputhumiohec.py +3 -3
  89. cribl_control_plane/models/outputkafka.py +3 -3
  90. cribl_control_plane/models/outputkinesis.py +3 -3
  91. cribl_control_plane/models/outputminio.py +3 -3
  92. cribl_control_plane/models/outputmsk.py +3 -3
  93. cribl_control_plane/models/outputnewrelic.py +5 -5
  94. cribl_control_plane/models/outputnewrelicevents.py +3 -3
  95. cribl_control_plane/models/outputring.py +5 -5
  96. cribl_control_plane/models/outputs3.py +3 -3
  97. cribl_control_plane/models/outputsecuritylake.py +3 -3
  98. cribl_control_plane/models/outputsentinel.py +3 -3
  99. cribl_control_plane/models/outputsentineloneaisiem.py +3 -3
  100. cribl_control_plane/models/outputservicenow.py +3 -3
  101. cribl_control_plane/models/outputsns.py +3 -3
  102. cribl_control_plane/models/outputsplunk.py +3 -3
  103. cribl_control_plane/models/outputsplunkhec.py +5 -5
  104. cribl_control_plane/models/outputsqs.py +3 -3
  105. cribl_control_plane/models/outputstatsd.py +3 -3
  106. cribl_control_plane/models/outputstatsdext.py +3 -3
  107. cribl_control_plane/models/outputsyslog.py +5 -5
  108. cribl_control_plane/models/outputtcpjson.py +5 -5
  109. cribl_control_plane/models/outputwebhook.py +5 -5
  110. cribl_control_plane/models/outputxsiam.py +5 -5
  111. cribl_control_plane/nodes.py +90 -68
  112. cribl_control_plane/samples.py +391 -0
  113. cribl_control_plane/sdk.py +11 -17
  114. cribl_control_plane/sources.py +15 -469
  115. cribl_control_plane/{workers_sdk.py → statuses.py} +23 -25
  116. cribl_control_plane/{deployments.py → summaries.py} +3 -3
  117. cribl_control_plane/teams.py +201 -0
  118. cribl_control_plane/tokens.py +182 -0
  119. cribl_control_plane/versions.py +26 -0
  120. cribl_control_plane/versions_configs.py +171 -0
  121. {cribl_control_plane-0.0.26a1.dist-info → cribl_control_plane-0.0.27.dist-info}/METADATA +67 -36
  122. {cribl_control_plane-0.0.26a1.dist-info → cribl_control_plane-0.0.27.dist-info}/RECORD +123 -114
  123. cribl_control_plane/models/createversionsyncop.py +0 -23
  124. cribl_control_plane/models/restartresponse.py +0 -26
  125. cribl_control_plane/models/updateworkersrestartop.py +0 -24
  126. cribl_control_plane/versioning.py +0 -2309
  127. {cribl_control_plane-0.0.26a1.dist-info → cribl_control_plane-0.0.27.dist-info}/WHEEL +0 -0
@@ -354,13 +354,13 @@ class InputConfluentCloudMetadatum(BaseModel):
354
354
 
355
355
 
356
356
  class InputConfluentCloudTypedDict(TypedDict):
357
+ type: InputConfluentCloudType
357
358
  brokers: List[str]
358
359
  r"""List of Confluent Cloud bootstrap servers to use, such as yourAccount.confluent.cloud:9092"""
359
360
  topics: List[str]
360
361
  r"""Topic to subscribe to. Warning: To optimize performance, Cribl suggests subscribing each Kafka Source to a single topic only."""
361
362
  id: NotRequired[str]
362
363
  r"""Unique ID for this input"""
363
- type: NotRequired[InputConfluentCloudType]
364
364
  disabled: NotRequired[bool]
365
365
  pipeline: NotRequired[str]
366
366
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -434,6 +434,8 @@ class InputConfluentCloudTypedDict(TypedDict):
434
434
 
435
435
 
436
436
  class InputConfluentCloud(BaseModel):
437
+ type: InputConfluentCloudType
438
+
437
439
  brokers: List[str]
438
440
  r"""List of Confluent Cloud bootstrap servers to use, such as yourAccount.confluent.cloud:9092"""
439
441
 
@@ -443,8 +445,6 @@ class InputConfluentCloud(BaseModel):
443
445
  id: Optional[str] = None
444
446
  r"""Unique ID for this input"""
445
447
 
446
- type: Optional[InputConfluentCloudType] = None
447
-
448
448
  disabled: Optional[bool] = False
449
449
 
450
450
  pipeline: Optional[str] = None
@@ -97,9 +97,9 @@ class InputCriblMetadatum(BaseModel):
97
97
 
98
98
 
99
99
  class InputCriblTypedDict(TypedDict):
100
- id: str
101
- r"""Unique ID for this input"""
102
100
  type: InputCriblType
101
+ id: NotRequired[str]
102
+ r"""Unique ID for this input"""
103
103
  disabled: NotRequired[bool]
104
104
  pipeline: NotRequired[str]
105
105
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -121,11 +121,11 @@ class InputCriblTypedDict(TypedDict):
121
121
 
122
122
 
123
123
  class InputCribl(BaseModel):
124
- id: str
125
- r"""Unique ID for this input"""
126
-
127
124
  type: InputCriblType
128
125
 
126
+ id: Optional[str] = None
127
+ r"""Unique ID for this input"""
128
+
129
129
  disabled: Optional[bool] = False
130
130
 
131
131
  pipeline: Optional[str] = None
@@ -171,11 +171,11 @@ class InputCriblHTTPMetadatum(BaseModel):
171
171
 
172
172
 
173
173
  class InputCriblHTTPTypedDict(TypedDict):
174
+ type: InputCriblHTTPType
174
175
  port: float
175
176
  r"""Port to listen on"""
176
177
  id: NotRequired[str]
177
178
  r"""Unique ID for this input"""
178
- type: NotRequired[InputCriblHTTPType]
179
179
  disabled: NotRequired[bool]
180
180
  pipeline: NotRequired[str]
181
181
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -223,14 +223,14 @@ class InputCriblHTTPTypedDict(TypedDict):
223
223
 
224
224
 
225
225
  class InputCriblHTTP(BaseModel):
226
+ type: InputCriblHTTPType
227
+
226
228
  port: float
227
229
  r"""Port to listen on"""
228
230
 
229
231
  id: Optional[str] = None
230
232
  r"""Unique ID for this input"""
231
233
 
232
- type: Optional[InputCriblHTTPType] = None
233
-
234
234
  disabled: Optional[bool] = False
235
235
 
236
236
  pipeline: Optional[str] = None
@@ -175,11 +175,11 @@ class InputCriblLakeHTTPMetadatum(BaseModel):
175
175
 
176
176
 
177
177
  class InputCriblLakeHTTPTypedDict(TypedDict):
178
+ type: InputCriblLakeHTTPType
178
179
  port: float
179
180
  r"""Port to listen on"""
180
181
  id: NotRequired[str]
181
182
  r"""Unique ID for this input"""
182
- type: NotRequired[InputCriblLakeHTTPType]
183
183
  disabled: NotRequired[bool]
184
184
  pipeline: NotRequired[str]
185
185
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -227,14 +227,14 @@ class InputCriblLakeHTTPTypedDict(TypedDict):
227
227
 
228
228
 
229
229
  class InputCriblLakeHTTP(BaseModel):
230
+ type: InputCriblLakeHTTPType
231
+
230
232
  port: float
231
233
  r"""Port to listen on"""
232
234
 
233
235
  id: Optional[str] = None
234
236
  r"""Unique ID for this input"""
235
237
 
236
- type: Optional[InputCriblLakeHTTPType] = None
237
-
238
238
  disabled: Optional[bool] = False
239
239
 
240
240
  pipeline: Optional[str] = None
@@ -97,9 +97,9 @@ class InputCriblmetricsMetadatum(BaseModel):
97
97
 
98
98
 
99
99
  class InputCriblmetricsTypedDict(TypedDict):
100
- id: str
101
- r"""Unique ID for this input"""
102
100
  type: InputCriblmetricsType
101
+ id: NotRequired[str]
102
+ r"""Unique ID for this input"""
103
103
  disabled: NotRequired[bool]
104
104
  pipeline: NotRequired[str]
105
105
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -124,11 +124,11 @@ class InputCriblmetricsTypedDict(TypedDict):
124
124
 
125
125
 
126
126
  class InputCriblmetrics(BaseModel):
127
- id: str
128
- r"""Unique ID for this input"""
129
-
130
127
  type: InputCriblmetricsType
131
128
 
129
+ id: Optional[str] = None
130
+ r"""Unique ID for this input"""
131
+
132
132
  disabled: Optional[bool] = False
133
133
 
134
134
  pipeline: Optional[str] = None
@@ -171,11 +171,11 @@ class InputCriblTCPMetadatum(BaseModel):
171
171
 
172
172
 
173
173
  class InputCriblTCPTypedDict(TypedDict):
174
+ type: InputCriblTCPType
174
175
  port: float
175
176
  r"""Port to listen on"""
176
177
  id: NotRequired[str]
177
178
  r"""Unique ID for this input"""
178
- type: NotRequired[InputCriblTCPType]
179
179
  disabled: NotRequired[bool]
180
180
  pipeline: NotRequired[str]
181
181
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -211,14 +211,14 @@ class InputCriblTCPTypedDict(TypedDict):
211
211
 
212
212
 
213
213
  class InputCriblTCP(BaseModel):
214
+ type: InputCriblTCPType
215
+
214
216
  port: float
215
217
  r"""Port to listen on"""
216
218
 
217
219
  id: Optional[str] = None
218
220
  r"""Unique ID for this input"""
219
221
 
220
- type: Optional[InputCriblTCPType] = None
221
-
222
222
  disabled: Optional[bool] = False
223
223
 
224
224
  pipeline: Optional[str] = None
@@ -188,11 +188,11 @@ class InputDatadogAgentProxyMode(BaseModel):
188
188
 
189
189
 
190
190
  class InputDatadogAgentTypedDict(TypedDict):
191
+ type: InputDatadogAgentType
191
192
  port: float
192
193
  r"""Port to listen on"""
193
194
  id: NotRequired[str]
194
195
  r"""Unique ID for this input"""
195
- type: NotRequired[InputDatadogAgentType]
196
196
  disabled: NotRequired[bool]
197
197
  pipeline: NotRequired[str]
198
198
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -241,14 +241,14 @@ class InputDatadogAgentTypedDict(TypedDict):
241
241
 
242
242
 
243
243
  class InputDatadogAgent(BaseModel):
244
+ type: InputDatadogAgentType
245
+
244
246
  port: float
245
247
  r"""Port to listen on"""
246
248
 
247
249
  id: Optional[str] = None
248
250
  r"""Unique ID for this input"""
249
251
 
250
- type: Optional[InputDatadogAgentType] = None
251
-
252
252
  disabled: Optional[bool] = False
253
253
 
254
254
  pipeline: Optional[str] = None
@@ -248,9 +248,9 @@ class PodFilter(BaseModel):
248
248
 
249
249
 
250
250
  class InputEdgePrometheusTypedDict(TypedDict):
251
+ type: InputEdgePrometheusType
251
252
  id: NotRequired[str]
252
253
  r"""Unique ID for this input"""
253
- type: NotRequired[InputEdgePrometheusType]
254
254
  disabled: NotRequired[bool]
255
255
  pipeline: NotRequired[str]
256
256
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -338,11 +338,11 @@ class InputEdgePrometheusTypedDict(TypedDict):
338
338
 
339
339
 
340
340
  class InputEdgePrometheus(BaseModel):
341
+ type: InputEdgePrometheusType
342
+
341
343
  id: Optional[str] = None
342
344
  r"""Unique ID for this input"""
343
345
 
344
- type: Optional[InputEdgePrometheusType] = None
345
-
346
346
  disabled: Optional[bool] = False
347
347
 
348
348
  pipeline: Optional[str] = None
@@ -246,11 +246,11 @@ class InputElasticProxyMode(BaseModel):
246
246
 
247
247
 
248
248
  class InputElasticTypedDict(TypedDict):
249
+ type: InputElasticType
249
250
  port: float
250
251
  r"""Port to listen on"""
251
252
  id: NotRequired[str]
252
253
  r"""Unique ID for this input"""
253
- type: NotRequired[InputElasticType]
254
254
  disabled: NotRequired[bool]
255
255
  pipeline: NotRequired[str]
256
256
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -312,14 +312,14 @@ class InputElasticTypedDict(TypedDict):
312
312
 
313
313
 
314
314
  class InputElastic(BaseModel):
315
+ type: InputElasticType
316
+
315
317
  port: float
316
318
  r"""Port to listen on"""
317
319
 
318
320
  id: Optional[str] = None
319
321
  r"""Unique ID for this input"""
320
322
 
321
- type: Optional[InputElasticType] = None
322
-
323
323
  disabled: Optional[bool] = False
324
324
 
325
325
  pipeline: Optional[str] = None
@@ -132,13 +132,13 @@ class InputEventhubMetadatum(BaseModel):
132
132
 
133
133
 
134
134
  class InputEventhubTypedDict(TypedDict):
135
+ type: InputEventhubType
135
136
  brokers: List[str]
136
137
  r"""List of Event Hubs Kafka brokers to connect to (example: yourdomain.servicebus.windows.net:9093). The hostname can be found in the host portion of the primary or secondary connection string in Shared Access Policies."""
137
138
  topics: List[str]
138
139
  r"""The name of the Event Hub (Kafka topic) to subscribe to. Warning: To optimize performance, Cribl suggests subscribing each Event Hubs Source to only a single topic."""
139
140
  id: NotRequired[str]
140
141
  r"""Unique ID for this input"""
141
- type: NotRequired[InputEventhubType]
142
142
  disabled: NotRequired[bool]
143
143
  pipeline: NotRequired[str]
144
144
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -210,6 +210,8 @@ class InputEventhubTypedDict(TypedDict):
210
210
 
211
211
 
212
212
  class InputEventhub(BaseModel):
213
+ type: InputEventhubType
214
+
213
215
  brokers: List[str]
214
216
  r"""List of Event Hubs Kafka brokers to connect to (example: yourdomain.servicebus.windows.net:9093). The hostname can be found in the host portion of the primary or secondary connection string in Shared Access Policies."""
215
217
 
@@ -219,8 +221,6 @@ class InputEventhub(BaseModel):
219
221
  id: Optional[str] = None
220
222
  r"""Unique ID for this input"""
221
223
 
222
- type: Optional[InputEventhubType] = None
223
-
224
224
  disabled: Optional[bool] = False
225
225
 
226
226
  pipeline: Optional[str] = None
@@ -104,9 +104,9 @@ class InputFileMetadatum(BaseModel):
104
104
 
105
105
 
106
106
  class InputFileTypedDict(TypedDict):
107
- id: str
108
- r"""Unique ID for this input"""
109
107
  type: InputFileType
108
+ id: NotRequired[str]
109
+ r"""Unique ID for this input"""
110
110
  disabled: NotRequired[bool]
111
111
  pipeline: NotRequired[str]
112
112
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -158,11 +158,11 @@ class InputFileTypedDict(TypedDict):
158
158
 
159
159
 
160
160
  class InputFile(BaseModel):
161
- id: str
162
- r"""Unique ID for this input"""
163
-
164
161
  type: InputFileType
165
162
 
163
+ id: Optional[str] = None
164
+ r"""Unique ID for this input"""
165
+
166
166
  disabled: Optional[bool] = False
167
167
 
168
168
  pipeline: Optional[str] = None
@@ -171,11 +171,11 @@ class InputFirehoseMetadatum(BaseModel):
171
171
 
172
172
 
173
173
  class InputFirehoseTypedDict(TypedDict):
174
+ type: InputFirehoseType
174
175
  port: float
175
176
  r"""Port to listen on"""
176
177
  id: NotRequired[str]
177
178
  r"""Unique ID for this input"""
178
- type: NotRequired[InputFirehoseType]
179
179
  disabled: NotRequired[bool]
180
180
  pipeline: NotRequired[str]
181
181
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -223,14 +223,14 @@ class InputFirehoseTypedDict(TypedDict):
223
223
 
224
224
 
225
225
  class InputFirehose(BaseModel):
226
+ type: InputFirehoseType
227
+
226
228
  port: float
227
229
  r"""Port to listen on"""
228
230
 
229
231
  id: Optional[str] = None
230
232
  r"""Unique ID for this input"""
231
233
 
232
- type: Optional[InputFirehoseType] = None
233
-
234
234
  disabled: Optional[bool] = False
235
235
 
236
236
  pipeline: Optional[str] = None
@@ -105,11 +105,11 @@ class InputGooglePubsubMetadatum(BaseModel):
105
105
 
106
106
 
107
107
  class InputGooglePubsubTypedDict(TypedDict):
108
+ type: InputGooglePubsubType
108
109
  subscription_name: str
109
110
  r"""ID of the subscription to use when receiving events. When Monitor subscription is enabled, the fully qualified subscription name must be entered. Example: projects/myProject/subscriptions/mySubscription"""
110
111
  id: NotRequired[str]
111
112
  r"""Unique ID for this input"""
112
- type: NotRequired[InputGooglePubsubType]
113
113
  disabled: NotRequired[bool]
114
114
  pipeline: NotRequired[str]
115
115
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -154,14 +154,14 @@ class InputGooglePubsubTypedDict(TypedDict):
154
154
 
155
155
 
156
156
  class InputGooglePubsub(BaseModel):
157
+ type: InputGooglePubsubType
158
+
157
159
  subscription_name: Annotated[str, pydantic.Field(alias="subscriptionName")]
158
160
  r"""ID of the subscription to use when receiving events. When Monitor subscription is enabled, the fully qualified subscription name must be entered. Example: projects/myProject/subscriptions/mySubscription"""
159
161
 
160
162
  id: Optional[str] = None
161
163
  r"""Unique ID for this input"""
162
164
 
163
- type: Optional[InputGooglePubsubType] = None
164
-
165
165
  disabled: Optional[bool] = False
166
166
 
167
167
  pipeline: Optional[str] = None
@@ -429,11 +429,11 @@ class InputGrafanaMetadatum2(BaseModel):
429
429
 
430
430
 
431
431
  class InputGrafanaGrafana2TypedDict(TypedDict):
432
+ type: InputGrafanaType2
432
433
  port: float
433
434
  r"""Port to listen on"""
434
435
  id: NotRequired[str]
435
436
  r"""Unique ID for this input"""
436
- type: NotRequired[InputGrafanaType2]
437
437
  disabled: NotRequired[bool]
438
438
  pipeline: NotRequired[str]
439
439
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -487,14 +487,14 @@ class InputGrafanaGrafana2TypedDict(TypedDict):
487
487
 
488
488
 
489
489
  class InputGrafanaGrafana2(BaseModel):
490
+ type: InputGrafanaType2
491
+
490
492
  port: float
491
493
  r"""Port to listen on"""
492
494
 
493
495
  id: Optional[str] = None
494
496
  r"""Unique ID for this input"""
495
497
 
496
- type: Optional[InputGrafanaType2] = None
497
-
498
498
  disabled: Optional[bool] = False
499
499
 
500
500
  pipeline: Optional[str] = None
@@ -1029,11 +1029,11 @@ class InputGrafanaMetadatum1(BaseModel):
1029
1029
 
1030
1030
 
1031
1031
  class InputGrafanaGrafana1TypedDict(TypedDict):
1032
+ type: InputGrafanaType1
1032
1033
  port: float
1033
1034
  r"""Port to listen on"""
1034
1035
  id: NotRequired[str]
1035
1036
  r"""Unique ID for this input"""
1036
- type: NotRequired[InputGrafanaType1]
1037
1037
  disabled: NotRequired[bool]
1038
1038
  pipeline: NotRequired[str]
1039
1039
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -1087,14 +1087,14 @@ class InputGrafanaGrafana1TypedDict(TypedDict):
1087
1087
 
1088
1088
 
1089
1089
  class InputGrafanaGrafana1(BaseModel):
1090
+ type: InputGrafanaType1
1091
+
1090
1092
  port: float
1091
1093
  r"""Port to listen on"""
1092
1094
 
1093
1095
  id: Optional[str] = None
1094
1096
  r"""Unique ID for this input"""
1095
1097
 
1096
- type: Optional[InputGrafanaType1] = None
1097
-
1098
1098
  disabled: Optional[bool] = False
1099
1099
 
1100
1100
  pipeline: Optional[str] = None
@@ -202,11 +202,11 @@ class InputHTTPAuthTokensExt(BaseModel):
202
202
 
203
203
 
204
204
  class InputHTTPTypedDict(TypedDict):
205
+ type: InputHTTPType
205
206
  port: float
206
207
  r"""Port to listen on"""
207
208
  id: NotRequired[str]
208
209
  r"""Unique ID for this input"""
209
- type: NotRequired[InputHTTPType]
210
210
  disabled: NotRequired[bool]
211
211
  pipeline: NotRequired[str]
212
212
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -263,14 +263,14 @@ class InputHTTPTypedDict(TypedDict):
263
263
 
264
264
 
265
265
  class InputHTTP(BaseModel):
266
+ type: InputHTTPType
267
+
266
268
  port: float
267
269
  r"""Port to listen on"""
268
270
 
269
271
  id: Optional[str] = None
270
272
  r"""Unique ID for this input"""
271
273
 
272
- type: Optional[InputHTTPType] = None
273
-
274
274
  disabled: Optional[bool] = False
275
275
 
276
276
  pipeline: Optional[str] = None
@@ -202,11 +202,11 @@ class InputHTTPRawAuthTokensExt(BaseModel):
202
202
 
203
203
 
204
204
  class InputHTTPRawTypedDict(TypedDict):
205
+ type: InputHTTPRawType
205
206
  port: float
206
207
  r"""Port to listen on"""
207
208
  id: NotRequired[str]
208
209
  r"""Unique ID for this input"""
209
- type: NotRequired[InputHTTPRawType]
210
210
  disabled: NotRequired[bool]
211
211
  pipeline: NotRequired[str]
212
212
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -264,14 +264,14 @@ class InputHTTPRawTypedDict(TypedDict):
264
264
 
265
265
 
266
266
  class InputHTTPRaw(BaseModel):
267
+ type: InputHTTPRawType
268
+
267
269
  port: float
268
270
  r"""Port to listen on"""
269
271
 
270
272
  id: Optional[str] = None
271
273
  r"""Unique ID for this input"""
272
274
 
273
- type: Optional[InputHTTPRawType] = None
274
-
275
275
  disabled: Optional[bool] = False
276
276
 
277
277
  pipeline: Optional[str] = None
@@ -112,13 +112,13 @@ class InputJournalFilesMetadatum(BaseModel):
112
112
 
113
113
 
114
114
  class InputJournalFilesTypedDict(TypedDict):
115
+ type: InputJournalFilesType
115
116
  path: str
116
117
  r"""Directory path to search for journals. Environment variables will be resolved, e.g. $CRIBL_EDGE_FS_ROOT/var/log/journal/$MACHINE_ID."""
117
118
  journals: List[str]
118
119
  r"""The full path of discovered journals are matched against this wildcard list."""
119
120
  id: NotRequired[str]
120
121
  r"""Unique ID for this input"""
121
- type: NotRequired[InputJournalFilesType]
122
122
  disabled: NotRequired[bool]
123
123
  pipeline: NotRequired[str]
124
124
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -147,6 +147,8 @@ class InputJournalFilesTypedDict(TypedDict):
147
147
 
148
148
 
149
149
  class InputJournalFiles(BaseModel):
150
+ type: InputJournalFilesType
151
+
150
152
  path: str
151
153
  r"""Directory path to search for journals. Environment variables will be resolved, e.g. $CRIBL_EDGE_FS_ROOT/var/log/journal/$MACHINE_ID."""
152
154
 
@@ -156,8 +158,6 @@ class InputJournalFiles(BaseModel):
156
158
  id: Optional[str] = None
157
159
  r"""Unique ID for this input"""
158
160
 
159
- type: Optional[InputJournalFilesType] = None
160
-
161
161
  disabled: Optional[bool] = False
162
162
 
163
163
  pipeline: Optional[str] = None
@@ -346,13 +346,13 @@ class InputKafkaMetadatum(BaseModel):
346
346
 
347
347
 
348
348
  class InputKafkaTypedDict(TypedDict):
349
+ type: InputKafkaType
349
350
  brokers: List[str]
350
351
  r"""Enter each Kafka bootstrap server you want to use. Specify the hostname and port (such as mykafkabroker:9092) or just the hostname (in which case @{product} will assign port 9092)."""
351
352
  topics: List[str]
352
353
  r"""Topic to subscribe to. Warning: To optimize performance, Cribl suggests subscribing each Kafka Source to a single topic only."""
353
354
  id: NotRequired[str]
354
355
  r"""Unique ID for this input"""
355
- type: NotRequired[InputKafkaType]
356
356
  disabled: NotRequired[bool]
357
357
  pipeline: NotRequired[str]
358
358
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -426,6 +426,8 @@ class InputKafkaTypedDict(TypedDict):
426
426
 
427
427
 
428
428
  class InputKafka(BaseModel):
429
+ type: InputKafkaType
430
+
429
431
  brokers: List[str]
430
432
  r"""Enter each Kafka bootstrap server you want to use. Specify the hostname and port (such as mykafkabroker:9092) or just the hostname (in which case @{product} will assign port 9092)."""
431
433
 
@@ -435,8 +437,6 @@ class InputKafka(BaseModel):
435
437
  id: Optional[str] = None
436
438
  r"""Unique ID for this input"""
437
439
 
438
- type: Optional[InputKafkaType] = None
439
-
440
440
  disabled: Optional[bool] = False
441
441
 
442
442
  pipeline: Optional[str] = None
@@ -135,13 +135,13 @@ class InputKinesisMetadatum(BaseModel):
135
135
 
136
136
 
137
137
  class InputKinesisTypedDict(TypedDict):
138
+ type: InputKinesisType
138
139
  stream_name: str
139
140
  r"""Kinesis Data Stream to read data from"""
140
141
  region: str
141
142
  r"""Region where the Kinesis stream is located"""
142
143
  id: NotRequired[str]
143
144
  r"""Unique ID for this input"""
144
- type: NotRequired[InputKinesisType]
145
145
  disabled: NotRequired[bool]
146
146
  pipeline: NotRequired[str]
147
147
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -202,6 +202,8 @@ class InputKinesisTypedDict(TypedDict):
202
202
 
203
203
 
204
204
  class InputKinesis(BaseModel):
205
+ type: InputKinesisType
206
+
205
207
  stream_name: Annotated[str, pydantic.Field(alias="streamName")]
206
208
  r"""Kinesis Data Stream to read data from"""
207
209
 
@@ -211,8 +213,6 @@ class InputKinesis(BaseModel):
211
213
  id: Optional[str] = None
212
214
  r"""Unique ID for this input"""
213
215
 
214
- type: Optional[InputKinesisType] = None
215
-
216
216
  disabled: Optional[bool] = False
217
217
 
218
218
  pipeline: Optional[str] = None
@@ -112,9 +112,9 @@ class InputKubeEventsMetadatum(BaseModel):
112
112
 
113
113
 
114
114
  class InputKubeEventsTypedDict(TypedDict):
115
- id: str
116
- r"""Unique ID for this input"""
117
115
  type: InputKubeEventsType
116
+ id: NotRequired[str]
117
+ r"""Unique ID for this input"""
118
118
  disabled: NotRequired[bool]
119
119
  pipeline: NotRequired[str]
120
120
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -137,11 +137,11 @@ class InputKubeEventsTypedDict(TypedDict):
137
137
 
138
138
 
139
139
  class InputKubeEvents(BaseModel):
140
- id: str
141
- r"""Unique ID for this input"""
142
-
143
140
  type: InputKubeEventsType
144
141
 
142
+ id: Optional[str] = None
143
+ r"""Unique ID for this input"""
144
+
145
145
  disabled: Optional[bool] = False
146
146
 
147
147
  pipeline: Optional[str] = None
@@ -151,9 +151,9 @@ class InputKubeLogsDiskSpooling(BaseModel):
151
151
 
152
152
 
153
153
  class InputKubeLogsTypedDict(TypedDict):
154
- id: str
155
- r"""Unique ID for this input"""
156
154
  type: InputKubeLogsType
155
+ id: NotRequired[str]
156
+ r"""Unique ID for this input"""
157
157
  disabled: NotRequired[bool]
158
158
  pipeline: NotRequired[str]
159
159
  r"""Pipeline to process data from this Source before sending it through the Routes"""
@@ -187,11 +187,11 @@ class InputKubeLogsTypedDict(TypedDict):
187
187
 
188
188
 
189
189
  class InputKubeLogs(BaseModel):
190
- id: str
191
- r"""Unique ID for this input"""
192
-
193
190
  type: InputKubeLogsType
194
191
 
192
+ id: Optional[str] = None
193
+ r"""Unique ID for this input"""
194
+
195
195
  disabled: Optional[bool] = False
196
196
 
197
197
  pipeline: Optional[str] = None