cribl-control-plane 0.0.42__py3-none-any.whl → 0.0.44__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (87) hide show
  1. cribl_control_plane/_hooks/clientcredentials.py +1 -1
  2. cribl_control_plane/_version.py +4 -4
  3. cribl_control_plane/acl.py +5 -3
  4. cribl_control_plane/auth_sdk.py +6 -3
  5. cribl_control_plane/basesdk.py +11 -1
  6. cribl_control_plane/commits.py +5 -3
  7. cribl_control_plane/destinations.py +6 -4
  8. cribl_control_plane/errors/__init__.py +15 -3
  9. cribl_control_plane/groups_configs.py +8 -3
  10. cribl_control_plane/groups_sdk.py +6 -4
  11. cribl_control_plane/models/__init__.py +401 -1
  12. cribl_control_plane/models/input.py +2 -2
  13. cribl_control_plane/models/inputappscope.py +13 -0
  14. cribl_control_plane/models/inputazureblob.py +13 -0
  15. cribl_control_plane/models/inputcollection.py +13 -0
  16. cribl_control_plane/models/inputconfluentcloud.py +20 -0
  17. cribl_control_plane/models/inputcribl.py +13 -0
  18. cribl_control_plane/models/inputcriblhttp.py +13 -0
  19. cribl_control_plane/models/inputcribllakehttp.py +38 -4
  20. cribl_control_plane/models/inputcriblmetrics.py +13 -0
  21. cribl_control_plane/models/inputcribltcp.py +13 -0
  22. cribl_control_plane/models/inputcrowdstrike.py +13 -0
  23. cribl_control_plane/models/inputdatadogagent.py +13 -0
  24. cribl_control_plane/models/inputdatagen.py +13 -0
  25. cribl_control_plane/models/inputedgeprometheus.py +13 -0
  26. cribl_control_plane/models/inputelastic.py +13 -0
  27. cribl_control_plane/models/inputeventhub.py +13 -0
  28. cribl_control_plane/models/inputexec.py +13 -0
  29. cribl_control_plane/models/inputfile.py +13 -0
  30. cribl_control_plane/models/inputfirehose.py +13 -0
  31. cribl_control_plane/models/inputgooglepubsub.py +13 -0
  32. cribl_control_plane/models/inputgrafana.py +26 -14
  33. cribl_control_plane/models/inputhttp.py +13 -0
  34. cribl_control_plane/models/inputhttpraw.py +13 -0
  35. cribl_control_plane/models/inputjournalfiles.py +13 -0
  36. cribl_control_plane/models/inputkafka.py +20 -0
  37. cribl_control_plane/models/inputkinesis.py +13 -0
  38. cribl_control_plane/models/inputkubeevents.py +13 -0
  39. cribl_control_plane/models/inputkubelogs.py +13 -0
  40. cribl_control_plane/models/inputkubemetrics.py +13 -0
  41. cribl_control_plane/models/inputloki.py +13 -7
  42. cribl_control_plane/models/inputmetrics.py +13 -0
  43. cribl_control_plane/models/inputmodeldriventelemetry.py +14 -0
  44. cribl_control_plane/models/inputmsk.py +13 -0
  45. cribl_control_plane/models/inputnetflow.py +13 -0
  46. cribl_control_plane/models/inputoffice365mgmt.py +13 -0
  47. cribl_control_plane/models/inputoffice365msgtrace.py +13 -0
  48. cribl_control_plane/models/inputoffice365service.py +13 -0
  49. cribl_control_plane/models/inputopentelemetry.py +13 -0
  50. cribl_control_plane/models/inputprometheus.py +13 -0
  51. cribl_control_plane/models/inputprometheusrw.py +13 -0
  52. cribl_control_plane/models/inputrawudp.py +13 -0
  53. cribl_control_plane/models/inputs3.py +13 -0
  54. cribl_control_plane/models/inputs3inventory.py +13 -0
  55. cribl_control_plane/models/inputsecuritylake.py +13 -0
  56. cribl_control_plane/models/inputsnmp.py +13 -0
  57. cribl_control_plane/models/inputsplunk.py +13 -0
  58. cribl_control_plane/models/inputsplunkhec.py +13 -0
  59. cribl_control_plane/models/inputsplunksearch.py +13 -0
  60. cribl_control_plane/models/inputsqs.py +13 -0
  61. cribl_control_plane/models/inputsyslog.py +26 -0
  62. cribl_control_plane/models/inputsystemmetrics.py +13 -0
  63. cribl_control_plane/models/inputsystemstate.py +13 -0
  64. cribl_control_plane/models/inputtcp.py +13 -0
  65. cribl_control_plane/models/inputtcpjson.py +13 -0
  66. cribl_control_plane/models/inputwef.py +13 -0
  67. cribl_control_plane/models/inputwindowsmetrics.py +13 -0
  68. cribl_control_plane/models/inputwineventlogs.py +13 -0
  69. cribl_control_plane/models/inputwiz.py +13 -0
  70. cribl_control_plane/models/inputwizwebhook.py +13 -0
  71. cribl_control_plane/models/inputzscalerhec.py +13 -0
  72. cribl_control_plane/models/output.py +18 -18
  73. cribl_control_plane/models/outputazuredataexplorer.py +7 -0
  74. cribl_control_plane/models/outputconfluentcloud.py +7 -0
  75. cribl_control_plane/models/outputgrafanacloud.py +0 -14
  76. cribl_control_plane/models/outputkafka.py +7 -0
  77. cribl_control_plane/models/outputloki.py +0 -7
  78. cribl_control_plane/models/outputsyslog.py +68 -0
  79. cribl_control_plane/nodes.py +5 -3
  80. cribl_control_plane/sdk.py +15 -2
  81. cribl_control_plane/sources.py +5 -3
  82. cribl_control_plane/utils/__init__.py +15 -3
  83. cribl_control_plane/utils/eventstreaming.py +10 -0
  84. cribl_control_plane/versions.py +11 -6
  85. {cribl_control_plane-0.0.42.dist-info → cribl_control_plane-0.0.44.dist-info}/METADATA +1 -1
  86. {cribl_control_plane-0.0.42.dist-info → cribl_control_plane-0.0.44.dist-info}/RECORD +87 -87
  87. {cribl_control_plane-0.0.42.dist-info → cribl_control_plane-0.0.44.dist-info}/WHEEL +0 -0
@@ -37,6 +37,14 @@ class InputFirehoseCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputFirehosePqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputFirehosePqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputFirehosePqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputFirehoseMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputFirehosePqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputFirehoseCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputFirehosePqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputFirehosePq(BaseModel):
@@ -82,6 +91,10 @@ class InputFirehosePq(BaseModel):
82
91
  compress: Optional[InputFirehoseCompression] = InputFirehoseCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputFirehosePqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputFirehoseMinimumTLSVersion(str, Enum):
87
100
  TL_SV1 = "TLSv1"
@@ -37,6 +37,14 @@ class InputGooglePubsubCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputGooglePubsubPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputGooglePubsubPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputGooglePubsubPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputGooglePubsubMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputGooglePubsubPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputGooglePubsubCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputGooglePubsubPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputGooglePubsubPq(BaseModel):
@@ -82,6 +91,10 @@ class InputGooglePubsubPq(BaseModel):
82
91
  compress: Optional[InputGooglePubsubCompression] = InputGooglePubsubCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputGooglePubsubPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputGooglePubsubGoogleAuthenticationMethod(str, Enum):
87
100
  r"""Choose Auto to use Google Application Default Credentials (ADC), Manual to enter Google service account credentials directly, or Secret to select or create a stored secret that references Google service account credentials."""
@@ -37,6 +37,14 @@ class InputGrafanaCompression2(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputGrafanaPqControls2TypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputGrafanaPqControls2(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputGrafanaPq2TypedDict(TypedDict):
41
49
  mode: NotRequired[InputGrafanaMode2]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputGrafanaPq2TypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputGrafanaCompression2]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputGrafanaPqControls2TypedDict]
55
64
 
56
65
 
57
66
  class InputGrafanaPq2(BaseModel):
@@ -82,6 +91,10 @@ class InputGrafanaPq2(BaseModel):
82
91
  compress: Optional[InputGrafanaCompression2] = InputGrafanaCompression2.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputGrafanaPqControls2], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputGrafanaMinimumTLSVersion2(str, Enum):
87
100
  TL_SV1 = "TLSv1"
@@ -477,8 +490,6 @@ class InputGrafanaGrafana2TypedDict(TypedDict):
477
490
  r"""Absolute path on which to listen for Grafana Agent's Remote Write requests. Defaults to /api/prom/push, which will expand as: 'http://<your‑upstream‑URL>:<your‑port>/api/prom/push'. Either this field or 'Logs API endpoint' must be configured."""
478
491
  loki_api: NotRequired[str]
479
492
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'. Either this field or 'Remote Write API endpoint' must be configured."""
480
- extract_structured_metadata: NotRequired[bool]
481
- r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
482
493
  prometheus_auth: NotRequired[InputGrafanaPrometheusAuth2TypedDict]
483
494
  loki_auth: NotRequired[InputGrafanaLokiAuth2TypedDict]
484
495
  metadata: NotRequired[List[InputGrafanaMetadatum2TypedDict]]
@@ -589,11 +600,6 @@ class InputGrafanaGrafana2(BaseModel):
589
600
  )
590
601
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'. Either this field or 'Remote Write API endpoint' must be configured."""
591
602
 
592
- extract_structured_metadata: Annotated[
593
- Optional[bool], pydantic.Field(alias="extractStructuredMetadata")
594
- ] = False
595
- r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
596
-
597
603
  prometheus_auth: Annotated[
598
604
  Optional[InputGrafanaPrometheusAuth2], pydantic.Field(alias="prometheusAuth")
599
605
  ] = None
@@ -637,6 +643,14 @@ class InputGrafanaCompression1(str, Enum):
637
643
  GZIP = "gzip"
638
644
 
639
645
 
646
+ class InputGrafanaPqControls1TypedDict(TypedDict):
647
+ pass
648
+
649
+
650
+ class InputGrafanaPqControls1(BaseModel):
651
+ pass
652
+
653
+
640
654
  class InputGrafanaPq1TypedDict(TypedDict):
641
655
  mode: NotRequired[InputGrafanaMode1]
642
656
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -652,6 +666,7 @@ class InputGrafanaPq1TypedDict(TypedDict):
652
666
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
653
667
  compress: NotRequired[InputGrafanaCompression1]
654
668
  r"""Codec to use to compress the persisted data"""
669
+ pq_controls: NotRequired[InputGrafanaPqControls1TypedDict]
655
670
 
656
671
 
657
672
  class InputGrafanaPq1(BaseModel):
@@ -682,6 +697,10 @@ class InputGrafanaPq1(BaseModel):
682
697
  compress: Optional[InputGrafanaCompression1] = InputGrafanaCompression1.NONE
683
698
  r"""Codec to use to compress the persisted data"""
684
699
 
700
+ pq_controls: Annotated[
701
+ Optional[InputGrafanaPqControls1], pydantic.Field(alias="pqControls")
702
+ ] = None
703
+
685
704
 
686
705
  class InputGrafanaMinimumTLSVersion1(str, Enum):
687
706
  TL_SV1 = "TLSv1"
@@ -1077,8 +1096,6 @@ class InputGrafanaGrafana1TypedDict(TypedDict):
1077
1096
  r"""Absolute path on which to listen for Grafana Agent's Remote Write requests. Defaults to /api/prom/push, which will expand as: 'http://<your‑upstream‑URL>:<your‑port>/api/prom/push'. Either this field or 'Logs API endpoint' must be configured."""
1078
1097
  loki_api: NotRequired[str]
1079
1098
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'. Either this field or 'Remote Write API endpoint' must be configured."""
1080
- extract_structured_metadata: NotRequired[bool]
1081
- r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
1082
1099
  prometheus_auth: NotRequired[InputGrafanaPrometheusAuth1TypedDict]
1083
1100
  loki_auth: NotRequired[InputGrafanaLokiAuth1TypedDict]
1084
1101
  metadata: NotRequired[List[InputGrafanaMetadatum1TypedDict]]
@@ -1189,11 +1206,6 @@ class InputGrafanaGrafana1(BaseModel):
1189
1206
  )
1190
1207
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'. Either this field or 'Remote Write API endpoint' must be configured."""
1191
1208
 
1192
- extract_structured_metadata: Annotated[
1193
- Optional[bool], pydantic.Field(alias="extractStructuredMetadata")
1194
- ] = False
1195
- r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
1196
-
1197
1209
  prometheus_auth: Annotated[
1198
1210
  Optional[InputGrafanaPrometheusAuth1], pydantic.Field(alias="prometheusAuth")
1199
1211
  ] = None
@@ -37,6 +37,14 @@ class InputHTTPCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputHTTPPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputHTTPPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputHTTPPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputHTTPMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputHTTPPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputHTTPCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputHTTPPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputHTTPPq(BaseModel):
@@ -82,6 +91,10 @@ class InputHTTPPq(BaseModel):
82
91
  compress: Optional[InputHTTPCompression] = InputHTTPCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputHTTPPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputHTTPMinimumTLSVersion(str, Enum):
87
100
  TL_SV1 = "TLSv1"
@@ -37,6 +37,14 @@ class InputHTTPRawCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputHTTPRawPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputHTTPRawPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputHTTPRawPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputHTTPRawMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputHTTPRawPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputHTTPRawCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputHTTPRawPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputHTTPRawPq(BaseModel):
@@ -82,6 +91,10 @@ class InputHTTPRawPq(BaseModel):
82
91
  compress: Optional[InputHTTPRawCompression] = InputHTTPRawCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputHTTPRawPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputHTTPRawMinimumTLSVersion(str, Enum):
87
100
  TL_SV1 = "TLSv1"
@@ -37,6 +37,14 @@ class InputJournalFilesCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputJournalFilesPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputJournalFilesPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputJournalFilesPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputJournalFilesMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputJournalFilesPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputJournalFilesCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputJournalFilesPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputJournalFilesPq(BaseModel):
@@ -82,6 +91,10 @@ class InputJournalFilesPq(BaseModel):
82
91
  compress: Optional[InputJournalFilesCompression] = InputJournalFilesCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputJournalFilesPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputJournalFilesRuleTypedDict(TypedDict):
87
100
  filter_: str
@@ -37,6 +37,14 @@ class InputKafkaCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputKafkaPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputKafkaPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputKafkaPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputKafkaMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputKafkaPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputKafkaCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputKafkaPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputKafkaPq(BaseModel):
@@ -82,6 +91,10 @@ class InputKafkaPq(BaseModel):
82
91
  compress: Optional[InputKafkaCompression] = InputKafkaCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputKafkaPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputKafkaSchemaType(str, Enum):
87
100
  r"""The schema format used to encode and decode event data"""
@@ -247,6 +260,8 @@ class InputKafkaAuthenticationTypedDict(TypedDict):
247
260
 
248
261
  disabled: NotRequired[bool]
249
262
  mechanism: NotRequired[InputKafkaSASLMechanism]
263
+ oauth_enabled: NotRequired[bool]
264
+ r"""Enable OAuth authentication"""
250
265
 
251
266
 
252
267
  class InputKafkaAuthentication(BaseModel):
@@ -256,6 +271,11 @@ class InputKafkaAuthentication(BaseModel):
256
271
 
257
272
  mechanism: Optional[InputKafkaSASLMechanism] = InputKafkaSASLMechanism.PLAIN
258
273
 
274
+ oauth_enabled: Annotated[Optional[bool], pydantic.Field(alias="oauthEnabled")] = (
275
+ False
276
+ )
277
+ r"""Enable OAuth authentication"""
278
+
259
279
 
260
280
  class InputKafkaMinimumTLSVersion(str, Enum):
261
281
  TL_SV1 = "TLSv1"
@@ -37,6 +37,14 @@ class InputKinesisCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputKinesisPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputKinesisPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputKinesisPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputKinesisMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputKinesisPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputKinesisCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputKinesisPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputKinesisPq(BaseModel):
@@ -82,6 +91,10 @@ class InputKinesisPq(BaseModel):
82
91
  compress: Optional[InputKinesisCompression] = InputKinesisCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputKinesisPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class ShardIteratorStart(str, Enum):
87
100
  r"""Location at which to start reading a shard for the first time"""
@@ -37,6 +37,14 @@ class InputKubeEventsCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputKubeEventsPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputKubeEventsPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputKubeEventsPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputKubeEventsMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputKubeEventsPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputKubeEventsCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputKubeEventsPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputKubeEventsPq(BaseModel):
@@ -82,6 +91,10 @@ class InputKubeEventsPq(BaseModel):
82
91
  compress: Optional[InputKubeEventsCompression] = InputKubeEventsCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputKubeEventsPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputKubeEventsRuleTypedDict(TypedDict):
87
100
  filter_: str
@@ -37,6 +37,14 @@ class InputKubeLogsPqCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputKubeLogsPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputKubeLogsPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputKubeLogsPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputKubeLogsMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputKubeLogsPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputKubeLogsPqCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputKubeLogsPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputKubeLogsPq(BaseModel):
@@ -82,6 +91,10 @@ class InputKubeLogsPq(BaseModel):
82
91
  compress: Optional[InputKubeLogsPqCompression] = InputKubeLogsPqCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputKubeLogsPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputKubeLogsRuleTypedDict(TypedDict):
87
100
  filter_: str
@@ -37,6 +37,14 @@ class InputKubeMetricsCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputKubeMetricsPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputKubeMetricsPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputKubeMetricsPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputKubeMetricsMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputKubeMetricsPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputKubeMetricsCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputKubeMetricsPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputKubeMetricsPq(BaseModel):
@@ -82,6 +91,10 @@ class InputKubeMetricsPq(BaseModel):
82
91
  compress: Optional[InputKubeMetricsCompression] = InputKubeMetricsCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputKubeMetricsPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputKubeMetricsRuleTypedDict(TypedDict):
87
100
  filter_: str
@@ -37,6 +37,14 @@ class InputLokiCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputLokiPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputLokiPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputLokiPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputLokiMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputLokiPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputLokiCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputLokiPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputLokiPq(BaseModel):
@@ -82,6 +91,10 @@ class InputLokiPq(BaseModel):
82
91
  compress: Optional[InputLokiCompression] = InputLokiCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputLokiPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputLokiMinimumTLSVersion(str, Enum):
87
100
  TL_SV1 = "TLSv1"
@@ -258,8 +271,6 @@ class InputLokiTypedDict(TypedDict):
258
271
  r"""Messages from matched IP addresses will be ignored. This takes precedence over the allowlist."""
259
272
  loki_api: NotRequired[str]
260
273
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'."""
261
- extract_structured_metadata: NotRequired[bool]
262
- r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
263
274
  auth_type: NotRequired[InputLokiAuthenticationType]
264
275
  r"""Loki logs authentication type"""
265
276
  metadata: NotRequired[List[InputLokiMetadatumTypedDict]]
@@ -389,11 +400,6 @@ class InputLoki(BaseModel):
389
400
  )
390
401
  r"""Absolute path on which to listen for Loki logs requests. Defaults to /loki/api/v1/push, which will (in this example) expand as: 'http://<your‑upstream‑URL>:<your‑port>/loki/api/v1/push'."""
391
402
 
392
- extract_structured_metadata: Annotated[
393
- Optional[bool], pydantic.Field(alias="extractStructuredMetadata")
394
- ] = False
395
- r"""Extract structured metadata from the Loki 3.5.3+ format and place it in the __structuredMetadata field. When disabled, uses legacy Loki parsing for backward compatibility."""
396
-
397
403
  auth_type: Annotated[
398
404
  Optional[InputLokiAuthenticationType], pydantic.Field(alias="authType")
399
405
  ] = InputLokiAuthenticationType.NONE
@@ -37,6 +37,14 @@ class InputMetricsCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputMetricsPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputMetricsPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputMetricsPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputMetricsMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputMetricsPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputMetricsCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputMetricsPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputMetricsPq(BaseModel):
@@ -82,6 +91,10 @@ class InputMetricsPq(BaseModel):
82
91
  compress: Optional[InputMetricsCompression] = InputMetricsCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputMetricsPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputMetricsMinimumTLSVersion(str, Enum):
87
100
  TL_SV1 = "TLSv1"
@@ -37,6 +37,14 @@ class InputModelDrivenTelemetryCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputModelDrivenTelemetryPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputModelDrivenTelemetryPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputModelDrivenTelemetryPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputModelDrivenTelemetryMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputModelDrivenTelemetryPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputModelDrivenTelemetryCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputModelDrivenTelemetryPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputModelDrivenTelemetryPq(BaseModel):
@@ -84,6 +93,11 @@ class InputModelDrivenTelemetryPq(BaseModel):
84
93
  )
85
94
  r"""Codec to use to compress the persisted data"""
86
95
 
96
+ pq_controls: Annotated[
97
+ Optional[InputModelDrivenTelemetryPqControls],
98
+ pydantic.Field(alias="pqControls"),
99
+ ] = None
100
+
87
101
 
88
102
  class InputModelDrivenTelemetryMinimumTLSVersion(str, Enum):
89
103
  TL_SV1 = "TLSv1"
@@ -37,6 +37,14 @@ class InputMskCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputMskPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputMskPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputMskPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputMskMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputMskPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputMskCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputMskPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputMskPq(BaseModel):
@@ -82,6 +91,10 @@ class InputMskPq(BaseModel):
82
91
  compress: Optional[InputMskCompression] = InputMskCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputMskPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputMskMetadatumTypedDict(TypedDict):
87
100
  name: str
@@ -37,6 +37,14 @@ class InputNetflowCompression(str, Enum):
37
37
  GZIP = "gzip"
38
38
 
39
39
 
40
+ class InputNetflowPqControlsTypedDict(TypedDict):
41
+ pass
42
+
43
+
44
+ class InputNetflowPqControls(BaseModel):
45
+ pass
46
+
47
+
40
48
  class InputNetflowPqTypedDict(TypedDict):
41
49
  mode: NotRequired[InputNetflowMode]
42
50
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
@@ -52,6 +60,7 @@ class InputNetflowPqTypedDict(TypedDict):
52
60
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
53
61
  compress: NotRequired[InputNetflowCompression]
54
62
  r"""Codec to use to compress the persisted data"""
63
+ pq_controls: NotRequired[InputNetflowPqControlsTypedDict]
55
64
 
56
65
 
57
66
  class InputNetflowPq(BaseModel):
@@ -82,6 +91,10 @@ class InputNetflowPq(BaseModel):
82
91
  compress: Optional[InputNetflowCompression] = InputNetflowCompression.NONE
83
92
  r"""Codec to use to compress the persisted data"""
84
93
 
94
+ pq_controls: Annotated[
95
+ Optional[InputNetflowPqControls], pydantic.Field(alias="pqControls")
96
+ ] = None
97
+
85
98
 
86
99
  class InputNetflowMetadatumTypedDict(TypedDict):
87
100
  name: str