cribl-control-plane 0.0.44a2__py3-none-any.whl → 0.0.45__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (158) hide show
  1. cribl_control_plane/_version.py +3 -3
  2. cribl_control_plane/errors/healthstatus_error.py +2 -8
  3. cribl_control_plane/models/__init__.py +3 -3
  4. cribl_control_plane/models/appmode.py +1 -2
  5. cribl_control_plane/models/cacheconnection.py +2 -10
  6. cribl_control_plane/models/cacheconnectionbackfillstatus.py +1 -2
  7. cribl_control_plane/models/cloudprovider.py +1 -2
  8. cribl_control_plane/models/configgroup.py +2 -7
  9. cribl_control_plane/models/configgroupcloud.py +2 -6
  10. cribl_control_plane/models/createconfiggroupbyproductop.py +2 -8
  11. cribl_control_plane/models/cribllakedataset.py +2 -8
  12. cribl_control_plane/models/datasetmetadata.py +2 -8
  13. cribl_control_plane/models/deleteconfiggroupbyproductandidop.py +2 -7
  14. cribl_control_plane/models/getconfiggroupaclbyproductandidop.py +2 -4
  15. cribl_control_plane/models/getconfiggroupaclteamsbyproductandidop.py +2 -4
  16. cribl_control_plane/models/getconfiggroupbyproductandidop.py +1 -3
  17. cribl_control_plane/models/getconfiggroupconfigversionbyproductandidop.py +2 -7
  18. cribl_control_plane/models/getsummaryop.py +2 -7
  19. cribl_control_plane/models/hbcriblinfo.py +6 -6
  20. cribl_control_plane/models/healthstatus.py +4 -7
  21. cribl_control_plane/models/inputappscope.py +14 -34
  22. cribl_control_plane/models/inputazureblob.py +6 -17
  23. cribl_control_plane/models/inputcollection.py +4 -11
  24. cribl_control_plane/models/inputconfluentcloud.py +20 -47
  25. cribl_control_plane/models/inputcribl.py +4 -11
  26. cribl_control_plane/models/inputcriblhttp.py +8 -23
  27. cribl_control_plane/models/inputcribllakehttp.py +10 -22
  28. cribl_control_plane/models/inputcriblmetrics.py +4 -12
  29. cribl_control_plane/models/inputcribltcp.py +8 -23
  30. cribl_control_plane/models/inputcrowdstrike.py +10 -26
  31. cribl_control_plane/models/inputdatadogagent.py +8 -24
  32. cribl_control_plane/models/inputdatagen.py +4 -11
  33. cribl_control_plane/models/inputedgeprometheus.py +24 -58
  34. cribl_control_plane/models/inputelastic.py +14 -40
  35. cribl_control_plane/models/inputeventhub.py +6 -15
  36. cribl_control_plane/models/inputexec.py +6 -14
  37. cribl_control_plane/models/inputfile.py +6 -15
  38. cribl_control_plane/models/inputfirehose.py +8 -23
  39. cribl_control_plane/models/inputgooglepubsub.py +6 -19
  40. cribl_control_plane/models/inputgrafana.py +24 -67
  41. cribl_control_plane/models/inputhttp.py +8 -23
  42. cribl_control_plane/models/inputhttpraw.py +8 -23
  43. cribl_control_plane/models/inputjournalfiles.py +4 -12
  44. cribl_control_plane/models/inputkafka.py +16 -46
  45. cribl_control_plane/models/inputkinesis.py +14 -38
  46. cribl_control_plane/models/inputkubeevents.py +4 -11
  47. cribl_control_plane/models/inputkubelogs.py +8 -16
  48. cribl_control_plane/models/inputkubemetrics.py +8 -16
  49. cribl_control_plane/models/inputloki.py +10 -29
  50. cribl_control_plane/models/inputmetrics.py +8 -23
  51. cribl_control_plane/models/inputmodeldriventelemetry.py +10 -27
  52. cribl_control_plane/models/inputmsk.py +18 -53
  53. cribl_control_plane/models/inputnetflow.py +4 -11
  54. cribl_control_plane/models/inputoffice365mgmt.py +14 -33
  55. cribl_control_plane/models/inputoffice365msgtrace.py +16 -35
  56. cribl_control_plane/models/inputoffice365service.py +16 -35
  57. cribl_control_plane/models/inputopentelemetry.py +16 -38
  58. cribl_control_plane/models/inputprometheus.py +18 -50
  59. cribl_control_plane/models/inputprometheusrw.py +10 -30
  60. cribl_control_plane/models/inputrawudp.py +4 -11
  61. cribl_control_plane/models/inputs3.py +8 -21
  62. cribl_control_plane/models/inputs3inventory.py +10 -26
  63. cribl_control_plane/models/inputsecuritylake.py +10 -27
  64. cribl_control_plane/models/inputsnmp.py +6 -16
  65. cribl_control_plane/models/inputsplunk.py +12 -33
  66. cribl_control_plane/models/inputsplunkhec.py +10 -29
  67. cribl_control_plane/models/inputsplunksearch.py +14 -33
  68. cribl_control_plane/models/inputsqs.py +10 -27
  69. cribl_control_plane/models/inputsyslog.py +16 -43
  70. cribl_control_plane/models/inputsystemmetrics.py +24 -48
  71. cribl_control_plane/models/inputsystemstate.py +8 -16
  72. cribl_control_plane/models/inputtcp.py +10 -29
  73. cribl_control_plane/models/inputtcpjson.py +10 -29
  74. cribl_control_plane/models/inputwef.py +14 -37
  75. cribl_control_plane/models/inputwindowsmetrics.py +24 -44
  76. cribl_control_plane/models/inputwineventlogs.py +10 -20
  77. cribl_control_plane/models/inputwiz.py +8 -21
  78. cribl_control_plane/models/inputwizwebhook.py +8 -23
  79. cribl_control_plane/models/inputzscalerhec.py +10 -29
  80. cribl_control_plane/models/lakehouseconnectiontype.py +1 -2
  81. cribl_control_plane/models/listconfiggroupbyproductop.py +1 -3
  82. cribl_control_plane/models/masterworkerentry.py +2 -7
  83. cribl_control_plane/models/nodeactiveupgradestatus.py +1 -2
  84. cribl_control_plane/models/nodefailedupgradestatus.py +1 -2
  85. cribl_control_plane/models/nodeskippedupgradestatus.py +1 -2
  86. cribl_control_plane/models/nodeupgradestate.py +1 -2
  87. cribl_control_plane/models/nodeupgradestatus.py +5 -13
  88. cribl_control_plane/models/outputazureblob.py +18 -48
  89. cribl_control_plane/models/outputazuredataexplorer.py +28 -73
  90. cribl_control_plane/models/outputazureeventhub.py +18 -40
  91. cribl_control_plane/models/outputazurelogs.py +12 -35
  92. cribl_control_plane/models/outputclickhouse.py +20 -55
  93. cribl_control_plane/models/outputcloudwatch.py +10 -29
  94. cribl_control_plane/models/outputconfluentcloud.py +32 -77
  95. cribl_control_plane/models/outputcriblhttp.py +16 -44
  96. cribl_control_plane/models/outputcribllake.py +16 -46
  97. cribl_control_plane/models/outputcribltcp.py +18 -45
  98. cribl_control_plane/models/outputcrowdstrikenextgensiem.py +14 -49
  99. cribl_control_plane/models/outputdatadog.py +20 -48
  100. cribl_control_plane/models/outputdataset.py +18 -46
  101. cribl_control_plane/models/outputdiskspool.py +2 -7
  102. cribl_control_plane/models/outputdls3.py +24 -68
  103. cribl_control_plane/models/outputdynatracehttp.py +20 -53
  104. cribl_control_plane/models/outputdynatraceotlp.py +22 -55
  105. cribl_control_plane/models/outputelastic.py +18 -43
  106. cribl_control_plane/models/outputelasticcloud.py +12 -36
  107. cribl_control_plane/models/outputexabeam.py +10 -29
  108. cribl_control_plane/models/outputfilesystem.py +14 -39
  109. cribl_control_plane/models/outputgooglechronicle.py +16 -50
  110. cribl_control_plane/models/outputgooglecloudlogging.py +14 -41
  111. cribl_control_plane/models/outputgooglecloudstorage.py +24 -66
  112. cribl_control_plane/models/outputgooglepubsub.py +10 -31
  113. cribl_control_plane/models/outputgrafanacloud.py +32 -97
  114. cribl_control_plane/models/outputgraphite.py +14 -31
  115. cribl_control_plane/models/outputhoneycomb.py +12 -35
  116. cribl_control_plane/models/outputhumiohec.py +16 -43
  117. cribl_control_plane/models/outputinfluxdb.py +16 -42
  118. cribl_control_plane/models/outputkafka.py +28 -74
  119. cribl_control_plane/models/outputkinesis.py +16 -40
  120. cribl_control_plane/models/outputloki.py +16 -41
  121. cribl_control_plane/models/outputminio.py +24 -65
  122. cribl_control_plane/models/outputmsk.py +30 -82
  123. cribl_control_plane/models/outputnewrelic.py +18 -43
  124. cribl_control_plane/models/outputnewrelicevents.py +14 -41
  125. cribl_control_plane/models/outputopentelemetry.py +26 -67
  126. cribl_control_plane/models/outputprometheus.py +12 -35
  127. cribl_control_plane/models/outputring.py +8 -19
  128. cribl_control_plane/models/outputs3.py +26 -68
  129. cribl_control_plane/models/outputsecuritylake.py +18 -52
  130. cribl_control_plane/models/outputsentinel.py +18 -45
  131. cribl_control_plane/models/outputsentineloneaisiem.py +18 -50
  132. cribl_control_plane/models/outputservicenow.py +24 -60
  133. cribl_control_plane/models/outputsignalfx.py +14 -37
  134. cribl_control_plane/models/outputsns.py +14 -36
  135. cribl_control_plane/models/outputsplunk.py +24 -60
  136. cribl_control_plane/models/outputsplunkhec.py +12 -35
  137. cribl_control_plane/models/outputsplunklb.py +30 -77
  138. cribl_control_plane/models/outputsqs.py +16 -41
  139. cribl_control_plane/models/outputstatsd.py +14 -30
  140. cribl_control_plane/models/outputstatsdext.py +12 -29
  141. cribl_control_plane/models/outputsumologic.py +12 -35
  142. cribl_control_plane/models/outputsyslog.py +24 -58
  143. cribl_control_plane/models/outputtcpjson.py +20 -52
  144. cribl_control_plane/models/outputwavefront.py +12 -35
  145. cribl_control_plane/models/outputwebhook.py +22 -58
  146. cribl_control_plane/models/outputxsiam.py +14 -35
  147. cribl_control_plane/models/productscore.py +1 -2
  148. cribl_control_plane/models/rbacresource.py +1 -2
  149. cribl_control_plane/models/resourcepolicy.py +2 -4
  150. cribl_control_plane/models/runnablejobcollection.py +13 -30
  151. cribl_control_plane/models/runnablejobexecutor.py +4 -13
  152. cribl_control_plane/models/runnablejobscheduledsearch.py +2 -7
  153. cribl_control_plane/models/updateconfiggroupbyproductandidop.py +2 -8
  154. cribl_control_plane/models/updateconfiggroupdeploybyproductandidop.py +2 -8
  155. cribl_control_plane/models/workertypes.py +1 -2
  156. {cribl_control_plane-0.0.44a2.dist-info → cribl_control_plane-0.0.45.dist-info}/METADATA +1 -1
  157. {cribl_control_plane-0.0.44a2.dist-info → cribl_control_plane-0.0.45.dist-info}/RECORD +158 -158
  158. {cribl_control_plane-0.0.44a2.dist-info → cribl_control_plane-0.0.45.dist-info}/WHEEL +0 -0
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import Any, List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputHTTPConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputHTTPMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputHTTPMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputHTTPCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputHTTPCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputHTTPPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputHTTPPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputHTTPMode], PlainValidator(validate_open_enum(False))
72
- ] = InputHTTPMode.ALWAYS
67
+ mode: Optional[InputHTTPMode] = InputHTTPMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,9 +88,7 @@ class InputHTTPPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputHTTPCompression], PlainValidator(validate_open_enum(False))
98
- ] = InputHTTPCompression.NONE
91
+ compress: Optional[InputHTTPCompression] = InputHTTPCompression.NONE
99
92
  r"""Codec to use to compress the persisted data"""
100
93
 
101
94
  pq_controls: Annotated[
@@ -103,14 +96,14 @@ class InputHTTPPq(BaseModel):
103
96
  ] = None
104
97
 
105
98
 
106
- class InputHTTPMinimumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
99
+ class InputHTTPMinimumTLSVersion(str, Enum):
107
100
  TL_SV1 = "TLSv1"
108
101
  TL_SV1_1 = "TLSv1.1"
109
102
  TL_SV1_2 = "TLSv1.2"
110
103
  TL_SV1_3 = "TLSv1.3"
111
104
 
112
105
 
113
- class InputHTTPMaximumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
106
+ class InputHTTPMaximumTLSVersion(str, Enum):
114
107
  TL_SV1 = "TLSv1"
115
108
  TL_SV1_1 = "TLSv1.1"
116
109
  TL_SV1_2 = "TLSv1.2"
@@ -169,19 +162,11 @@ class InputHTTPTLSSettingsServerSide(BaseModel):
169
162
  ] = None
170
163
 
171
164
  min_version: Annotated[
172
- Annotated[
173
- Optional[InputHTTPMinimumTLSVersion],
174
- PlainValidator(validate_open_enum(False)),
175
- ],
176
- pydantic.Field(alias="minVersion"),
165
+ Optional[InputHTTPMinimumTLSVersion], pydantic.Field(alias="minVersion")
177
166
  ] = None
178
167
 
179
168
  max_version: Annotated[
180
- Annotated[
181
- Optional[InputHTTPMaximumTLSVersion],
182
- PlainValidator(validate_open_enum(False)),
183
- ],
184
- pydantic.Field(alias="maxVersion"),
169
+ Optional[InputHTTPMaximumTLSVersion], pydantic.Field(alias="maxVersion")
185
170
  ] = None
186
171
 
187
172
 
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import Any, List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputHTTPRawConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputHTTPRawMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputHTTPRawMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputHTTPRawCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputHTTPRawCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputHTTPRawPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputHTTPRawPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputHTTPRawMode], PlainValidator(validate_open_enum(False))
72
- ] = InputHTTPRawMode.ALWAYS
67
+ mode: Optional[InputHTTPRawMode] = InputHTTPRawMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,9 +88,7 @@ class InputHTTPRawPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputHTTPRawCompression], PlainValidator(validate_open_enum(False))
98
- ] = InputHTTPRawCompression.NONE
91
+ compress: Optional[InputHTTPRawCompression] = InputHTTPRawCompression.NONE
99
92
  r"""Codec to use to compress the persisted data"""
100
93
 
101
94
  pq_controls: Annotated[
@@ -103,14 +96,14 @@ class InputHTTPRawPq(BaseModel):
103
96
  ] = None
104
97
 
105
98
 
106
- class InputHTTPRawMinimumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
99
+ class InputHTTPRawMinimumTLSVersion(str, Enum):
107
100
  TL_SV1 = "TLSv1"
108
101
  TL_SV1_1 = "TLSv1.1"
109
102
  TL_SV1_2 = "TLSv1.2"
110
103
  TL_SV1_3 = "TLSv1.3"
111
104
 
112
105
 
113
- class InputHTTPRawMaximumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
106
+ class InputHTTPRawMaximumTLSVersion(str, Enum):
114
107
  TL_SV1 = "TLSv1"
115
108
  TL_SV1_1 = "TLSv1.1"
116
109
  TL_SV1_2 = "TLSv1.2"
@@ -169,19 +162,11 @@ class InputHTTPRawTLSSettingsServerSide(BaseModel):
169
162
  ] = None
170
163
 
171
164
  min_version: Annotated[
172
- Annotated[
173
- Optional[InputHTTPRawMinimumTLSVersion],
174
- PlainValidator(validate_open_enum(False)),
175
- ],
176
- pydantic.Field(alias="minVersion"),
165
+ Optional[InputHTTPRawMinimumTLSVersion], pydantic.Field(alias="minVersion")
177
166
  ] = None
178
167
 
179
168
  max_version: Annotated[
180
- Annotated[
181
- Optional[InputHTTPRawMaximumTLSVersion],
182
- PlainValidator(validate_open_enum(False)),
183
- ],
184
- pydantic.Field(alias="maxVersion"),
169
+ Optional[InputHTTPRawMaximumTLSVersion], pydantic.Field(alias="maxVersion")
185
170
  ] = None
186
171
 
187
172
 
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputJournalFilesConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputJournalFilesMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputJournalFilesMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputJournalFilesCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputJournalFilesCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputJournalFilesPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputJournalFilesPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputJournalFilesMode], PlainValidator(validate_open_enum(False))
72
- ] = InputJournalFilesMode.ALWAYS
67
+ mode: Optional[InputJournalFilesMode] = InputJournalFilesMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,10 +88,7 @@ class InputJournalFilesPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputJournalFilesCompression],
98
- PlainValidator(validate_open_enum(False)),
99
- ] = InputJournalFilesCompression.NONE
91
+ compress: Optional[InputJournalFilesCompression] = InputJournalFilesCompression.NONE
100
92
  r"""Codec to use to compress the persisted data"""
101
93
 
102
94
  pq_controls: Annotated[
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputKafkaConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputKafkaMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputKafkaMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputKafkaCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputKafkaCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputKafkaPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputKafkaPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputKafkaMode], PlainValidator(validate_open_enum(False))
72
- ] = InputKafkaMode.ALWAYS
67
+ mode: Optional[InputKafkaMode] = InputKafkaMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,9 +88,7 @@ class InputKafkaPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputKafkaCompression], PlainValidator(validate_open_enum(False))
98
- ] = InputKafkaCompression.NONE
91
+ compress: Optional[InputKafkaCompression] = InputKafkaCompression.NONE
99
92
  r"""Codec to use to compress the persisted data"""
100
93
 
101
94
  pq_controls: Annotated[
@@ -103,7 +96,7 @@ class InputKafkaPq(BaseModel):
103
96
  ] = None
104
97
 
105
98
 
106
- class InputKafkaSchemaType(str, Enum, metaclass=utils.OpenEnumMeta):
99
+ class InputKafkaSchemaType(str, Enum):
107
100
  r"""The schema format used to encode and decode event data"""
108
101
 
109
102
  AVRO = "avro"
@@ -129,18 +122,14 @@ class InputKafkaAuth(BaseModel):
129
122
  r"""Select or create a secret that references your credentials"""
130
123
 
131
124
 
132
- class InputKafkaKafkaSchemaRegistryMinimumTLSVersion(
133
- str, Enum, metaclass=utils.OpenEnumMeta
134
- ):
125
+ class InputKafkaKafkaSchemaRegistryMinimumTLSVersion(str, Enum):
135
126
  TL_SV1 = "TLSv1"
136
127
  TL_SV1_1 = "TLSv1.1"
137
128
  TL_SV1_2 = "TLSv1.2"
138
129
  TL_SV1_3 = "TLSv1.3"
139
130
 
140
131
 
141
- class InputKafkaKafkaSchemaRegistryMaximumTLSVersion(
142
- str, Enum, metaclass=utils.OpenEnumMeta
143
- ):
132
+ class InputKafkaKafkaSchemaRegistryMaximumTLSVersion(str, Enum):
144
133
  TL_SV1 = "TLSv1"
145
134
  TL_SV1_1 = "TLSv1.1"
146
135
  TL_SV1_2 = "TLSv1.2"
@@ -200,18 +189,12 @@ class InputKafkaKafkaSchemaRegistryTLSSettingsClientSide(BaseModel):
200
189
  r"""Passphrase to use to decrypt private key"""
201
190
 
202
191
  min_version: Annotated[
203
- Annotated[
204
- Optional[InputKafkaKafkaSchemaRegistryMinimumTLSVersion],
205
- PlainValidator(validate_open_enum(False)),
206
- ],
192
+ Optional[InputKafkaKafkaSchemaRegistryMinimumTLSVersion],
207
193
  pydantic.Field(alias="minVersion"),
208
194
  ] = None
209
195
 
210
196
  max_version: Annotated[
211
- Annotated[
212
- Optional[InputKafkaKafkaSchemaRegistryMaximumTLSVersion],
213
- PlainValidator(validate_open_enum(False)),
214
- ],
197
+ Optional[InputKafkaKafkaSchemaRegistryMaximumTLSVersion],
215
198
  pydantic.Field(alias="maxVersion"),
216
199
  ] = None
217
200
 
@@ -242,10 +225,7 @@ class InputKafkaKafkaSchemaRegistryAuthentication(BaseModel):
242
225
  r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
243
226
 
244
227
  schema_type: Annotated[
245
- Annotated[
246
- Optional[InputKafkaSchemaType], PlainValidator(validate_open_enum(False))
247
- ],
248
- pydantic.Field(alias="schemaType"),
228
+ Optional[InputKafkaSchemaType], pydantic.Field(alias="schemaType")
249
229
  ] = InputKafkaSchemaType.AVRO
250
230
  r"""The schema format used to encode and decode event data"""
251
231
 
@@ -268,7 +248,7 @@ class InputKafkaKafkaSchemaRegistryAuthentication(BaseModel):
268
248
  tls: Optional[InputKafkaKafkaSchemaRegistryTLSSettingsClientSide] = None
269
249
 
270
250
 
271
- class InputKafkaSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta):
251
+ class InputKafkaSASLMechanism(str, Enum):
272
252
  PLAIN = "plain"
273
253
  SCRAM_SHA_256 = "scram-sha-256"
274
254
  SCRAM_SHA_512 = "scram-sha-512"
@@ -289,9 +269,7 @@ class InputKafkaAuthentication(BaseModel):
289
269
 
290
270
  disabled: Optional[bool] = True
291
271
 
292
- mechanism: Annotated[
293
- Optional[InputKafkaSASLMechanism], PlainValidator(validate_open_enum(False))
294
- ] = InputKafkaSASLMechanism.PLAIN
272
+ mechanism: Optional[InputKafkaSASLMechanism] = InputKafkaSASLMechanism.PLAIN
295
273
 
296
274
  oauth_enabled: Annotated[Optional[bool], pydantic.Field(alias="oauthEnabled")] = (
297
275
  False
@@ -299,14 +277,14 @@ class InputKafkaAuthentication(BaseModel):
299
277
  r"""Enable OAuth authentication"""
300
278
 
301
279
 
302
- class InputKafkaMinimumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
280
+ class InputKafkaMinimumTLSVersion(str, Enum):
303
281
  TL_SV1 = "TLSv1"
304
282
  TL_SV1_1 = "TLSv1.1"
305
283
  TL_SV1_2 = "TLSv1.2"
306
284
  TL_SV1_3 = "TLSv1.3"
307
285
 
308
286
 
309
- class InputKafkaMaximumTLSVersion(str, Enum, metaclass=utils.OpenEnumMeta):
287
+ class InputKafkaMaximumTLSVersion(str, Enum):
310
288
  TL_SV1 = "TLSv1"
311
289
  TL_SV1_1 = "TLSv1.1"
312
290
  TL_SV1_2 = "TLSv1.2"
@@ -366,19 +344,11 @@ class InputKafkaTLSSettingsClientSide(BaseModel):
366
344
  r"""Passphrase to use to decrypt private key"""
367
345
 
368
346
  min_version: Annotated[
369
- Annotated[
370
- Optional[InputKafkaMinimumTLSVersion],
371
- PlainValidator(validate_open_enum(False)),
372
- ],
373
- pydantic.Field(alias="minVersion"),
347
+ Optional[InputKafkaMinimumTLSVersion], pydantic.Field(alias="minVersion")
374
348
  ] = None
375
349
 
376
350
  max_version: Annotated[
377
- Annotated[
378
- Optional[InputKafkaMaximumTLSVersion],
379
- PlainValidator(validate_open_enum(False)),
380
- ],
381
- pydantic.Field(alias="maxVersion"),
351
+ Optional[InputKafkaMaximumTLSVersion], pydantic.Field(alias="maxVersion")
382
352
  ] = None
383
353
 
384
354
 
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputKinesisConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputKinesisMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputKinesisMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputKinesisCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputKinesisCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputKinesisPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputKinesisPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputKinesisMode], PlainValidator(validate_open_enum(False))
72
- ] = InputKinesisMode.ALWAYS
67
+ mode: Optional[InputKinesisMode] = InputKinesisMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,9 +88,7 @@ class InputKinesisPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputKinesisCompression], PlainValidator(validate_open_enum(False))
98
- ] = InputKinesisCompression.NONE
91
+ compress: Optional[InputKinesisCompression] = InputKinesisCompression.NONE
99
92
  r"""Codec to use to compress the persisted data"""
100
93
 
101
94
  pq_controls: Annotated[
@@ -103,14 +96,14 @@ class InputKinesisPq(BaseModel):
103
96
  ] = None
104
97
 
105
98
 
106
- class ShardIteratorStart(str, Enum, metaclass=utils.OpenEnumMeta):
99
+ class ShardIteratorStart(str, Enum):
107
100
  r"""Location at which to start reading a shard for the first time"""
108
101
 
109
102
  TRIM_HORIZON = "TRIM_HORIZON"
110
103
  LATEST = "LATEST"
111
104
 
112
105
 
113
- class InputKinesisRecordDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
106
+ class InputKinesisRecordDataFormat(str, Enum):
114
107
  r"""Format of data inside the Kinesis Stream records. Gzip compression is automatically detected."""
115
108
 
116
109
  CRIBL = "cribl"
@@ -119,14 +112,14 @@ class InputKinesisRecordDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
119
112
  LINE = "line"
120
113
 
121
114
 
122
- class ShardLoadBalancing(str, Enum, metaclass=utils.OpenEnumMeta):
115
+ class ShardLoadBalancing(str, Enum):
123
116
  r"""The load-balancing algorithm to use for spreading out shards across Workers and Worker Processes"""
124
117
 
125
118
  CONSISTENT_HASHING = "ConsistentHashing"
126
119
  ROUND_ROBIN = "RoundRobin"
127
120
 
128
121
 
129
- class InputKinesisAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
122
+ class InputKinesisAuthenticationMethod(str, Enum):
130
123
  r"""AWS authentication method. Choose Auto to use IAM roles."""
131
124
 
132
125
  AUTO = "auto"
@@ -134,7 +127,7 @@ class InputKinesisAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
134
127
  SECRET = "secret"
135
128
 
136
129
 
137
- class InputKinesisSignatureVersion(str, Enum, metaclass=utils.OpenEnumMeta):
130
+ class InputKinesisSignatureVersion(str, Enum):
138
131
  r"""Signature version to use for signing Kinesis stream requests"""
139
132
 
140
133
  V2 = "v2"
@@ -266,19 +259,12 @@ class InputKinesis(BaseModel):
266
259
  r"""A JavaScript expression to be called with each shardId for the stream. If the expression evaluates to a truthy value, the shard will be processed."""
267
260
 
268
261
  shard_iterator_type: Annotated[
269
- Annotated[
270
- Optional[ShardIteratorStart], PlainValidator(validate_open_enum(False))
271
- ],
272
- pydantic.Field(alias="shardIteratorType"),
262
+ Optional[ShardIteratorStart], pydantic.Field(alias="shardIteratorType")
273
263
  ] = ShardIteratorStart.TRIM_HORIZON
274
264
  r"""Location at which to start reading a shard for the first time"""
275
265
 
276
266
  payload_format: Annotated[
277
- Annotated[
278
- Optional[InputKinesisRecordDataFormat],
279
- PlainValidator(validate_open_enum(False)),
280
- ],
281
- pydantic.Field(alias="payloadFormat"),
267
+ Optional[InputKinesisRecordDataFormat], pydantic.Field(alias="payloadFormat")
282
268
  ] = InputKinesisRecordDataFormat.CRIBL
283
269
  r"""Format of data inside the Kinesis Stream records. Gzip compression is automatically detected."""
284
270
 
@@ -293,18 +279,12 @@ class InputKinesis(BaseModel):
293
279
  r"""Maximum number of records, across all shards, to pull down at once per Worker Process"""
294
280
 
295
281
  load_balancing_algorithm: Annotated[
296
- Annotated[
297
- Optional[ShardLoadBalancing], PlainValidator(validate_open_enum(False))
298
- ],
299
- pydantic.Field(alias="loadBalancingAlgorithm"),
282
+ Optional[ShardLoadBalancing], pydantic.Field(alias="loadBalancingAlgorithm")
300
283
  ] = ShardLoadBalancing.CONSISTENT_HASHING
301
284
  r"""The load-balancing algorithm to use for spreading out shards across Workers and Worker Processes"""
302
285
 
303
286
  aws_authentication_method: Annotated[
304
- Annotated[
305
- Optional[InputKinesisAuthenticationMethod],
306
- PlainValidator(validate_open_enum(False)),
307
- ],
287
+ Optional[InputKinesisAuthenticationMethod],
308
288
  pydantic.Field(alias="awsAuthenticationMethod"),
309
289
  ] = InputKinesisAuthenticationMethod.AUTO
310
290
  r"""AWS authentication method. Choose Auto to use IAM roles."""
@@ -317,11 +297,7 @@ class InputKinesis(BaseModel):
317
297
  r"""Kinesis stream service endpoint. If empty, defaults to the AWS Region-specific endpoint. Otherwise, it must point to Kinesis stream-compatible endpoint."""
318
298
 
319
299
  signature_version: Annotated[
320
- Annotated[
321
- Optional[InputKinesisSignatureVersion],
322
- PlainValidator(validate_open_enum(False)),
323
- ],
324
- pydantic.Field(alias="signatureVersion"),
300
+ Optional[InputKinesisSignatureVersion], pydantic.Field(alias="signatureVersion")
325
301
  ] = InputKinesisSignatureVersion.V4
326
302
  r"""Signature version to use for signing Kinesis stream requests"""
327
303
 
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputKubeEventsConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputKubeEventsMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputKubeEventsMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputKubeEventsCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputKubeEventsCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputKubeEventsPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputKubeEventsPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputKubeEventsMode], PlainValidator(validate_open_enum(False))
72
- ] = InputKubeEventsMode.ALWAYS
67
+ mode: Optional[InputKubeEventsMode] = InputKubeEventsMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,9 +88,7 @@ class InputKubeEventsPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputKubeEventsCompression], PlainValidator(validate_open_enum(False))
98
- ] = InputKubeEventsCompression.NONE
91
+ compress: Optional[InputKubeEventsCompression] = InputKubeEventsCompression.NONE
99
92
  r"""Codec to use to compress the persisted data"""
100
93
 
101
94
  pq_controls: Annotated[
@@ -1,12 +1,9 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
- from cribl_control_plane import utils
5
4
  from cribl_control_plane.types import BaseModel
6
- from cribl_control_plane.utils import validate_open_enum
7
5
  from enum import Enum
8
6
  import pydantic
9
- from pydantic.functional_validators import PlainValidator
10
7
  from typing import List, Optional
11
8
  from typing_extensions import Annotated, NotRequired, TypedDict
12
9
 
@@ -26,14 +23,14 @@ class InputKubeLogsConnection(BaseModel):
26
23
  pipeline: Optional[str] = None
27
24
 
28
25
 
29
- class InputKubeLogsMode(str, Enum, metaclass=utils.OpenEnumMeta):
26
+ class InputKubeLogsMode(str, Enum):
30
27
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
31
28
 
32
29
  SMART = "smart"
33
30
  ALWAYS = "always"
34
31
 
35
32
 
36
- class InputKubeLogsPqCompression(str, Enum, metaclass=utils.OpenEnumMeta):
33
+ class InputKubeLogsPqCompression(str, Enum):
37
34
  r"""Codec to use to compress the persisted data"""
38
35
 
39
36
  NONE = "none"
@@ -67,9 +64,7 @@ class InputKubeLogsPqTypedDict(TypedDict):
67
64
 
68
65
 
69
66
  class InputKubeLogsPq(BaseModel):
70
- mode: Annotated[
71
- Optional[InputKubeLogsMode], PlainValidator(validate_open_enum(False))
72
- ] = InputKubeLogsMode.ALWAYS
67
+ mode: Optional[InputKubeLogsMode] = InputKubeLogsMode.ALWAYS
73
68
  r"""With Smart mode, PQ will write events to the filesystem only when it detects backpressure from the processing engine. With Always On mode, PQ will always write events directly to the queue before forwarding them to the processing engine."""
74
69
 
75
70
  max_buffer_size: Annotated[
@@ -93,9 +88,7 @@ class InputKubeLogsPq(BaseModel):
93
88
  path: Optional[str] = "$CRIBL_HOME/state/queues"
94
89
  r"""The location for the persistent queue files. To this field's value, the system will append: /<worker-id>/inputs/<input-id>"""
95
90
 
96
- compress: Annotated[
97
- Optional[InputKubeLogsPqCompression], PlainValidator(validate_open_enum(False))
98
- ] = InputKubeLogsPqCompression.NONE
91
+ compress: Optional[InputKubeLogsPqCompression] = InputKubeLogsPqCompression.NONE
99
92
  r"""Codec to use to compress the persisted data"""
100
93
 
101
94
  pq_controls: Annotated[
@@ -131,7 +124,7 @@ class InputKubeLogsMetadatum(BaseModel):
131
124
  r"""JavaScript expression to compute field's value, enclosed in quotes or backticks. (Can evaluate to a constant.)"""
132
125
 
133
126
 
134
- class InputKubeLogsPersistenceCompression(str, Enum, metaclass=utils.OpenEnumMeta):
127
+ class InputKubeLogsPersistenceCompression(str, Enum):
135
128
  r"""Data compression format. Default is gzip."""
136
129
 
137
130
  NONE = "none"
@@ -164,10 +157,9 @@ class InputKubeLogsDiskSpooling(BaseModel):
164
157
  max_data_time: Annotated[Optional[str], pydantic.Field(alias="maxDataTime")] = "24h"
165
158
  r"""Maximum amount of time to retain data before older buckets are deleted. Examples: 2h, 4d. Default is 24h."""
166
159
 
167
- compress: Annotated[
168
- Optional[InputKubeLogsPersistenceCompression],
169
- PlainValidator(validate_open_enum(False)),
170
- ] = InputKubeLogsPersistenceCompression.GZIP
160
+ compress: Optional[InputKubeLogsPersistenceCompression] = (
161
+ InputKubeLogsPersistenceCompression.GZIP
162
+ )
171
163
  r"""Data compression format. Default is gzip."""
172
164
 
173
165