cribl-control-plane 0.2.1rc6__py3-none-any.whl → 0.3.0a1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (157) hide show
  1. cribl_control_plane/_version.py +4 -4
  2. cribl_control_plane/errors/__init__.py +5 -8
  3. cribl_control_plane/errors/{healthserverstatus_error.py → healthstatus_error.py} +9 -10
  4. cribl_control_plane/groups_sdk.py +30 -54
  5. cribl_control_plane/health.py +16 -22
  6. cribl_control_plane/models/__init__.py +50 -103
  7. cribl_control_plane/models/appmode.py +14 -0
  8. cribl_control_plane/models/authtoken.py +1 -5
  9. cribl_control_plane/models/configgroup.py +6 -35
  10. cribl_control_plane/models/createconfiggroupbyproductop.py +7 -8
  11. cribl_control_plane/models/createroutesappendbyidop.py +2 -2
  12. cribl_control_plane/models/createversionundoop.py +3 -3
  13. cribl_control_plane/models/deleteoutputpqbyidop.py +2 -2
  14. cribl_control_plane/models/distributedsummary.py +0 -6
  15. cribl_control_plane/models/error.py +16 -0
  16. cribl_control_plane/models/gethealthinfoop.py +17 -0
  17. cribl_control_plane/models/hbcriblinfo.py +3 -14
  18. cribl_control_plane/models/{healthserverstatus.py → healthstatus.py} +7 -7
  19. cribl_control_plane/models/heartbeatmetadata.py +0 -3
  20. cribl_control_plane/models/input.py +63 -65
  21. cribl_control_plane/models/inputappscope.py +0 -4
  22. cribl_control_plane/models/inputazureblob.py +0 -4
  23. cribl_control_plane/models/inputcollection.py +0 -4
  24. cribl_control_plane/models/inputconfluentcloud.py +18 -8
  25. cribl_control_plane/models/inputcribl.py +0 -4
  26. cribl_control_plane/models/inputcriblhttp.py +0 -4
  27. cribl_control_plane/models/inputcribllakehttp.py +0 -4
  28. cribl_control_plane/models/inputcriblmetrics.py +0 -4
  29. cribl_control_plane/models/inputcribltcp.py +0 -4
  30. cribl_control_plane/models/inputcrowdstrike.py +0 -7
  31. cribl_control_plane/models/inputdatadogagent.py +0 -4
  32. cribl_control_plane/models/inputdatagen.py +0 -4
  33. cribl_control_plane/models/inputedgeprometheus.py +0 -12
  34. cribl_control_plane/models/inputelastic.py +0 -11
  35. cribl_control_plane/models/inputeventhub.py +0 -6
  36. cribl_control_plane/models/inputexec.py +0 -4
  37. cribl_control_plane/models/inputfile.py +0 -6
  38. cribl_control_plane/models/inputfirehose.py +0 -4
  39. cribl_control_plane/models/inputgooglepubsub.py +0 -7
  40. cribl_control_plane/models/inputgrafana.py +0 -8
  41. cribl_control_plane/models/inputhttp.py +0 -4
  42. cribl_control_plane/models/inputhttpraw.py +0 -4
  43. cribl_control_plane/models/inputjournalfiles.py +0 -4
  44. cribl_control_plane/models/inputkafka.py +17 -8
  45. cribl_control_plane/models/inputkinesis.py +0 -15
  46. cribl_control_plane/models/inputkubeevents.py +0 -4
  47. cribl_control_plane/models/inputkubelogs.py +0 -4
  48. cribl_control_plane/models/inputkubemetrics.py +0 -4
  49. cribl_control_plane/models/inputloki.py +0 -4
  50. cribl_control_plane/models/inputmetrics.py +0 -4
  51. cribl_control_plane/models/inputmodeldriventelemetry.py +0 -4
  52. cribl_control_plane/models/inputmsk.py +17 -7
  53. cribl_control_plane/models/inputnetflow.py +0 -4
  54. cribl_control_plane/models/inputoffice365mgmt.py +0 -11
  55. cribl_control_plane/models/inputoffice365msgtrace.py +0 -11
  56. cribl_control_plane/models/inputoffice365service.py +0 -11
  57. cribl_control_plane/models/inputopentelemetry.py +0 -8
  58. cribl_control_plane/models/inputprometheus.py +0 -10
  59. cribl_control_plane/models/inputprometheusrw.py +0 -4
  60. cribl_control_plane/models/inputrawudp.py +0 -4
  61. cribl_control_plane/models/inputs3.py +0 -7
  62. cribl_control_plane/models/inputs3inventory.py +0 -7
  63. cribl_control_plane/models/inputsecuritylake.py +0 -7
  64. cribl_control_plane/models/inputsnmp.py +0 -11
  65. cribl_control_plane/models/inputsplunk.py +0 -9
  66. cribl_control_plane/models/inputsplunkhec.py +0 -4
  67. cribl_control_plane/models/inputsplunksearch.py +0 -7
  68. cribl_control_plane/models/inputsqs.py +0 -9
  69. cribl_control_plane/models/inputsyslog.py +0 -8
  70. cribl_control_plane/models/inputsystemmetrics.py +0 -32
  71. cribl_control_plane/models/inputsystemstate.py +0 -4
  72. cribl_control_plane/models/inputtcp.py +0 -4
  73. cribl_control_plane/models/inputtcpjson.py +0 -4
  74. cribl_control_plane/models/inputwef.py +0 -6
  75. cribl_control_plane/models/inputwindowsmetrics.py +0 -28
  76. cribl_control_plane/models/inputwineventlogs.py +0 -8
  77. cribl_control_plane/models/inputwiz.py +0 -7
  78. cribl_control_plane/models/inputwizwebhook.py +0 -4
  79. cribl_control_plane/models/inputzscalerhec.py +0 -4
  80. cribl_control_plane/models/jobinfo.py +1 -4
  81. cribl_control_plane/models/logininfo.py +3 -3
  82. cribl_control_plane/models/nodeprovidedinfo.py +1 -11
  83. cribl_control_plane/models/output.py +72 -77
  84. cribl_control_plane/models/outputazureblob.py +0 -20
  85. cribl_control_plane/models/outputazuredataexplorer.py +0 -28
  86. cribl_control_plane/models/outputazureeventhub.py +0 -17
  87. cribl_control_plane/models/outputazurelogs.py +0 -13
  88. cribl_control_plane/models/outputchronicle.py +0 -13
  89. cribl_control_plane/models/outputclickhouse.py +0 -17
  90. cribl_control_plane/models/outputcloudwatch.py +0 -13
  91. cribl_control_plane/models/outputconfluentcloud.py +18 -24
  92. cribl_control_plane/models/outputcriblhttp.py +0 -15
  93. cribl_control_plane/models/outputcribllake.py +0 -21
  94. cribl_control_plane/models/outputcribltcp.py +0 -12
  95. cribl_control_plane/models/outputcrowdstrikenextgensiem.py +0 -15
  96. cribl_control_plane/models/outputdatadog.py +0 -30
  97. cribl_control_plane/models/outputdataset.py +0 -23
  98. cribl_control_plane/models/outputdls3.py +0 -35
  99. cribl_control_plane/models/outputdynatracehttp.py +0 -22
  100. cribl_control_plane/models/outputdynatraceotlp.py +0 -22
  101. cribl_control_plane/models/outputelastic.py +0 -18
  102. cribl_control_plane/models/outputelasticcloud.py +0 -13
  103. cribl_control_plane/models/outputexabeam.py +0 -14
  104. cribl_control_plane/models/outputfilesystem.py +0 -15
  105. cribl_control_plane/models/outputgooglechronicle.py +0 -21
  106. cribl_control_plane/models/outputgooglecloudlogging.py +0 -19
  107. cribl_control_plane/models/outputgooglecloudstorage.py +0 -28
  108. cribl_control_plane/models/outputgooglepubsub.py +0 -13
  109. cribl_control_plane/models/outputgrafanacloud.py +0 -50
  110. cribl_control_plane/models/outputgraphite.py +0 -12
  111. cribl_control_plane/models/outputhoneycomb.py +0 -13
  112. cribl_control_plane/models/outputhumiohec.py +0 -15
  113. cribl_control_plane/models/outputinfluxdb.py +0 -19
  114. cribl_control_plane/models/outputkafka.py +17 -24
  115. cribl_control_plane/models/outputkinesis.py +0 -15
  116. cribl_control_plane/models/outputloki.py +0 -20
  117. cribl_control_plane/models/outputminio.py +0 -28
  118. cribl_control_plane/models/outputmsk.py +17 -23
  119. cribl_control_plane/models/outputnewrelic.py +0 -16
  120. cribl_control_plane/models/outputnewrelicevents.py +0 -16
  121. cribl_control_plane/models/outputopentelemetry.py +0 -22
  122. cribl_control_plane/models/outputprometheus.py +0 -13
  123. cribl_control_plane/models/outputring.py +0 -2
  124. cribl_control_plane/models/outputs3.py +0 -35
  125. cribl_control_plane/models/outputsecuritylake.py +0 -29
  126. cribl_control_plane/models/outputsentinel.py +0 -15
  127. cribl_control_plane/models/outputsentineloneaisiem.py +0 -13
  128. cribl_control_plane/models/outputservicenow.py +0 -21
  129. cribl_control_plane/models/outputsignalfx.py +0 -13
  130. cribl_control_plane/models/outputsns.py +0 -13
  131. cribl_control_plane/models/outputsplunk.py +0 -15
  132. cribl_control_plane/models/outputsplunkhec.py +0 -13
  133. cribl_control_plane/models/outputsplunklb.py +0 -15
  134. cribl_control_plane/models/outputsqs.py +0 -15
  135. cribl_control_plane/models/outputstatsd.py +0 -12
  136. cribl_control_plane/models/outputstatsdext.py +0 -12
  137. cribl_control_plane/models/outputsumologic.py +0 -15
  138. cribl_control_plane/models/outputsyslog.py +0 -24
  139. cribl_control_plane/models/outputtcpjson.py +0 -12
  140. cribl_control_plane/models/outputwavefront.py +0 -13
  141. cribl_control_plane/models/outputwebhook.py +0 -23
  142. cribl_control_plane/models/outputxsiam.py +0 -13
  143. cribl_control_plane/models/packinfo.py +5 -8
  144. cribl_control_plane/models/packinstallinfo.py +5 -8
  145. cribl_control_plane/models/{uploadpackresponse.py → routecloneconf.py} +4 -4
  146. cribl_control_plane/models/routeconf.py +4 -3
  147. cribl_control_plane/models/runnablejobcollection.py +0 -4
  148. cribl_control_plane/packs.py +7 -202
  149. cribl_control_plane/routes_sdk.py +6 -6
  150. cribl_control_plane/tokens.py +15 -23
  151. {cribl_control_plane-0.2.1rc6.dist-info → cribl_control_plane-0.3.0a1.dist-info}/METADATA +9 -50
  152. {cribl_control_plane-0.2.1rc6.dist-info → cribl_control_plane-0.3.0a1.dist-info}/RECORD +153 -154
  153. cribl_control_plane/models/groupcreaterequest.py +0 -152
  154. cribl_control_plane/models/outpostnodeinfo.py +0 -16
  155. cribl_control_plane/models/outputdatabricks.py +0 -411
  156. cribl_control_plane/models/updatepacksop.py +0 -25
  157. {cribl_control_plane-0.2.1rc6.dist-info → cribl_control_plane-0.3.0a1.dist-info}/WHEEL +0 -0
@@ -16,9 +16,7 @@ class OutputAzureDataExplorerType(str, Enum):
16
16
 
17
17
 
18
18
  class IngestionMode(str, Enum, metaclass=utils.OpenEnumMeta):
19
- # Batching
20
19
  BATCHING = "batching"
21
- # Streaming
22
20
  STREAMING = "streaming"
23
21
 
24
22
 
@@ -35,11 +33,8 @@ class OutputAzureDataExplorerAuthenticationMethod(
35
33
  ):
36
34
  r"""The type of OAuth 2.0 client credentials grant flow to use"""
37
35
 
38
- # Client secret
39
36
  CLIENT_SECRET = "clientSecret"
40
- # Client secret (text secret)
41
37
  CLIENT_TEXT_SECRET = "clientTextSecret"
42
- # Certificate
43
38
  CERTIFICATE = "certificate"
44
39
 
45
40
 
@@ -60,22 +55,16 @@ class OutputAzureDataExplorerBackpressureBehavior(
60
55
  ):
61
56
  r"""How to handle events when all receivers are exerting backpressure"""
62
57
 
63
- # Block
64
58
  BLOCK = "block"
65
- # Drop
66
59
  DROP = "drop"
67
- # Persistent Queue
68
60
  QUEUE = "queue"
69
61
 
70
62
 
71
63
  class OutputAzureDataExplorerDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
72
64
  r"""Format of the output data"""
73
65
 
74
- # JSON
75
66
  JSON = "json"
76
- # Raw
77
67
  RAW = "raw"
78
- # Parquet
79
68
  PARQUET = "parquet"
80
69
 
81
70
 
@@ -84,16 +73,12 @@ class OutputAzureDataExplorerDiskSpaceProtection(
84
73
  ):
85
74
  r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
86
75
 
87
- # Block
88
76
  BLOCK = "block"
89
- # Drop
90
77
  DROP = "drop"
91
78
 
92
79
 
93
80
  class PrefixOptional(str, Enum, metaclass=utils.OpenEnumMeta):
94
- # drop-by
95
81
  DROP_BY = "dropBy"
96
- # ingest-by
97
82
  INGEST_BY = "ingestBy"
98
83
 
99
84
 
@@ -121,22 +106,16 @@ class IngestIfNotExist(BaseModel):
121
106
  class ReportLevel(str, Enum, metaclass=utils.OpenEnumMeta):
122
107
  r"""Level of ingestion status reporting. Defaults to FailuresOnly."""
123
108
 
124
- # FailuresOnly
125
109
  FAILURES_ONLY = "failuresOnly"
126
- # DoNotReport
127
110
  DO_NOT_REPORT = "doNotReport"
128
- # FailuresAndSuccesses
129
111
  FAILURES_AND_SUCCESSES = "failuresAndSuccesses"
130
112
 
131
113
 
132
114
  class ReportMethod(str, Enum, metaclass=utils.OpenEnumMeta):
133
115
  r"""Target of the ingestion status reporting. Defaults to Queue."""
134
116
 
135
- # Queue
136
117
  QUEUE = "queue"
137
- # Table
138
118
  TABLE = "table"
139
- # QueueAndTable
140
119
  QUEUE_AND_TABLE = "queueAndTable"
141
120
 
142
121
 
@@ -219,29 +198,22 @@ class OutputAzureDataExplorerPqCompressCompression(
219
198
  ):
220
199
  r"""Codec to use to compress the persisted data"""
221
200
 
222
- # None
223
201
  NONE = "none"
224
- # Gzip
225
202
  GZIP = "gzip"
226
203
 
227
204
 
228
205
  class OutputAzureDataExplorerQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
229
206
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
230
207
 
231
- # Block
232
208
  BLOCK = "block"
233
- # Drop new data
234
209
  DROP = "drop"
235
210
 
236
211
 
237
212
  class OutputAzureDataExplorerMode(str, Enum, metaclass=utils.OpenEnumMeta):
238
213
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
239
214
 
240
- # Error
241
215
  ERROR = "error"
242
- # Backpressure
243
216
  BACKPRESSURE = "backpressure"
244
- # Always On
245
217
  ALWAYS = "always"
246
218
 
247
219
 
@@ -18,27 +18,20 @@ class OutputAzureEventhubType(str, Enum):
18
18
  class OutputAzureEventhubAcknowledgments(int, Enum, metaclass=utils.OpenEnumMeta):
19
19
  r"""Control the number of required acknowledgments"""
20
20
 
21
- # Leader
22
21
  ONE = 1
23
- # None
24
22
  ZERO = 0
25
- # All
26
23
  MINUS_1 = -1
27
24
 
28
25
 
29
26
  class OutputAzureEventhubRecordDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
30
27
  r"""Format to use to serialize events before writing to the Event Hubs Kafka brokers"""
31
28
 
32
- # JSON
33
29
  JSON = "json"
34
- # Field _raw
35
30
  RAW = "raw"
36
31
 
37
32
 
38
33
  class OutputAzureEventhubSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta):
39
- # PLAIN
40
34
  PLAIN = "plain"
41
- # OAUTHBEARER
42
35
  OAUTHBEARER = "oauthbearer"
43
36
 
44
37
 
@@ -78,40 +71,30 @@ class OutputAzureEventhubTLSSettingsClientSide(BaseModel):
78
71
  class OutputAzureEventhubBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
79
72
  r"""How to handle events when all receivers are exerting backpressure"""
80
73
 
81
- # Block
82
74
  BLOCK = "block"
83
- # Drop
84
75
  DROP = "drop"
85
- # Persistent Queue
86
76
  QUEUE = "queue"
87
77
 
88
78
 
89
79
  class OutputAzureEventhubCompression(str, Enum, metaclass=utils.OpenEnumMeta):
90
80
  r"""Codec to use to compress the persisted data"""
91
81
 
92
- # None
93
82
  NONE = "none"
94
- # Gzip
95
83
  GZIP = "gzip"
96
84
 
97
85
 
98
86
  class OutputAzureEventhubQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
99
87
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
100
88
 
101
- # Block
102
89
  BLOCK = "block"
103
- # Drop new data
104
90
  DROP = "drop"
105
91
 
106
92
 
107
93
  class OutputAzureEventhubMode(str, Enum, metaclass=utils.OpenEnumMeta):
108
94
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
109
95
 
110
- # Error
111
96
  ERROR = "error"
112
- # Backpressure
113
97
  BACKPRESSURE = "backpressure"
114
- # Always On
115
98
  ALWAYS = "always"
116
99
 
117
100
 
@@ -29,11 +29,8 @@ class OutputAzureLogsExtraHTTPHeader(BaseModel):
29
29
  class OutputAzureLogsFailedRequestLoggingMode(str, Enum, metaclass=utils.OpenEnumMeta):
30
30
  r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
31
31
 
32
- # Payload
33
32
  PAYLOAD = "payload"
34
- # Payload + Headers
35
33
  PAYLOAD_AND_HEADERS = "payloadAndHeaders"
36
- # None
37
34
  NONE = "none"
38
35
 
39
36
 
@@ -94,11 +91,8 @@ class OutputAzureLogsTimeoutRetrySettings(BaseModel):
94
91
  class OutputAzureLogsBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
95
92
  r"""How to handle events when all receivers are exerting backpressure"""
96
93
 
97
- # Block
98
94
  BLOCK = "block"
99
- # Drop
100
95
  DROP = "drop"
101
- # Persistent Queue
102
96
  QUEUE = "queue"
103
97
 
104
98
 
@@ -112,29 +106,22 @@ class OutputAzureLogsAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMet
112
106
  class OutputAzureLogsCompression(str, Enum, metaclass=utils.OpenEnumMeta):
113
107
  r"""Codec to use to compress the persisted data"""
114
108
 
115
- # None
116
109
  NONE = "none"
117
- # Gzip
118
110
  GZIP = "gzip"
119
111
 
120
112
 
121
113
  class OutputAzureLogsQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
122
114
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
123
115
 
124
- # Block
125
116
  BLOCK = "block"
126
- # Drop new data
127
117
  DROP = "drop"
128
118
 
129
119
 
130
120
  class OutputAzureLogsMode(str, Enum, metaclass=utils.OpenEnumMeta):
131
121
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
132
122
 
133
- # Error
134
123
  ERROR = "error"
135
- # Backpressure
136
124
  BACKPRESSURE = "backpressure"
137
- # Always On
138
125
  ALWAYS = "always"
139
126
 
140
127
 
@@ -88,22 +88,16 @@ class OutputChronicleExtraHTTPHeader(BaseModel):
88
88
  class OutputChronicleFailedRequestLoggingMode(str, Enum, metaclass=utils.OpenEnumMeta):
89
89
  r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
90
90
 
91
- # Payload
92
91
  PAYLOAD = "payload"
93
- # Payload + Headers
94
92
  PAYLOAD_AND_HEADERS = "payloadAndHeaders"
95
- # None
96
93
  NONE = "none"
97
94
 
98
95
 
99
96
  class OutputChronicleBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
100
97
  r"""How to handle events when all receivers are exerting backpressure"""
101
98
 
102
- # Block
103
99
  BLOCK = "block"
104
- # Drop
105
100
  DROP = "drop"
106
- # Persistent Queue
107
101
  QUEUE = "queue"
108
102
 
109
103
 
@@ -121,29 +115,22 @@ class OutputChronicleCustomLabel(BaseModel):
121
115
  class OutputChronicleCompression(str, Enum, metaclass=utils.OpenEnumMeta):
122
116
  r"""Codec to use to compress the persisted data"""
123
117
 
124
- # None
125
118
  NONE = "none"
126
- # Gzip
127
119
  GZIP = "gzip"
128
120
 
129
121
 
130
122
  class OutputChronicleQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
131
123
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
132
124
 
133
- # Block
134
125
  BLOCK = "block"
135
- # Drop new data
136
126
  DROP = "drop"
137
127
 
138
128
 
139
129
  class OutputChronicleMode(str, Enum, metaclass=utils.OpenEnumMeta):
140
130
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
141
131
 
142
- # Error
143
132
  ERROR = "error"
144
- # Backpressure
145
133
  BACKPRESSURE = "backpressure"
146
- # Always On
147
134
  ALWAYS = "always"
148
135
 
149
136
 
@@ -28,18 +28,14 @@ class OutputClickHouseAuthenticationType(str, Enum, metaclass=utils.OpenEnumMeta
28
28
  class OutputClickHouseFormat(str, Enum, metaclass=utils.OpenEnumMeta):
29
29
  r"""Data format to use when sending data to ClickHouse. Defaults to JSON Compact."""
30
30
 
31
- # JSONCompactEachRowWithNames
32
31
  JSON_COMPACT_EACH_ROW_WITH_NAMES = "json-compact-each-row-with-names"
33
- # JSONEachRow
34
32
  JSON_EACH_ROW = "json-each-row"
35
33
 
36
34
 
37
35
  class MappingType(str, Enum, metaclass=utils.OpenEnumMeta):
38
36
  r"""How event fields are mapped to ClickHouse columns."""
39
37
 
40
- # Automatic
41
38
  AUTOMATIC = "automatic"
42
- # Custom
43
39
  CUSTOM = "custom"
44
40
 
45
41
 
@@ -129,11 +125,8 @@ class OutputClickHouseExtraHTTPHeader(BaseModel):
129
125
  class OutputClickHouseFailedRequestLoggingMode(str, Enum, metaclass=utils.OpenEnumMeta):
130
126
  r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
131
127
 
132
- # Payload
133
128
  PAYLOAD = "payload"
134
- # Payload + Headers
135
129
  PAYLOAD_AND_HEADERS = "payloadAndHeaders"
136
- # None
137
130
  NONE = "none"
138
131
 
139
132
 
@@ -194,11 +187,8 @@ class OutputClickHouseTimeoutRetrySettings(BaseModel):
194
187
  class OutputClickHouseBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
195
188
  r"""How to handle events when all receivers are exerting backpressure"""
196
189
 
197
- # Block
198
190
  BLOCK = "block"
199
- # Drop
200
191
  DROP = "drop"
201
- # Persistent Queue
202
192
  QUEUE = "queue"
203
193
 
204
194
 
@@ -257,29 +247,22 @@ class ColumnMapping(BaseModel):
257
247
  class OutputClickHouseCompression(str, Enum, metaclass=utils.OpenEnumMeta):
258
248
  r"""Codec to use to compress the persisted data"""
259
249
 
260
- # None
261
250
  NONE = "none"
262
- # Gzip
263
251
  GZIP = "gzip"
264
252
 
265
253
 
266
254
  class OutputClickHouseQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
267
255
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
268
256
 
269
- # Block
270
257
  BLOCK = "block"
271
- # Drop new data
272
258
  DROP = "drop"
273
259
 
274
260
 
275
261
  class OutputClickHouseMode(str, Enum, metaclass=utils.OpenEnumMeta):
276
262
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
277
263
 
278
- # Error
279
264
  ERROR = "error"
280
- # Backpressure
281
265
  BACKPRESSURE = "backpressure"
282
- # Always On
283
266
  ALWAYS = "always"
284
267
 
285
268
 
@@ -18,51 +18,38 @@ class OutputCloudwatchType(str, Enum):
18
18
  class OutputCloudwatchAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMeta):
19
19
  r"""AWS authentication method. Choose Auto to use IAM roles."""
20
20
 
21
- # Auto
22
21
  AUTO = "auto"
23
- # Manual
24
22
  MANUAL = "manual"
25
- # Secret Key pair
26
23
  SECRET = "secret"
27
24
 
28
25
 
29
26
  class OutputCloudwatchBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
30
27
  r"""How to handle events when all receivers are exerting backpressure"""
31
28
 
32
- # Block
33
29
  BLOCK = "block"
34
- # Drop
35
30
  DROP = "drop"
36
- # Persistent Queue
37
31
  QUEUE = "queue"
38
32
 
39
33
 
40
34
  class OutputCloudwatchCompression(str, Enum, metaclass=utils.OpenEnumMeta):
41
35
  r"""Codec to use to compress the persisted data"""
42
36
 
43
- # None
44
37
  NONE = "none"
45
- # Gzip
46
38
  GZIP = "gzip"
47
39
 
48
40
 
49
41
  class OutputCloudwatchQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
50
42
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
51
43
 
52
- # Block
53
44
  BLOCK = "block"
54
- # Drop new data
55
45
  DROP = "drop"
56
46
 
57
47
 
58
48
  class OutputCloudwatchMode(str, Enum, metaclass=utils.OpenEnumMeta):
59
49
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
60
50
 
61
- # Error
62
51
  ERROR = "error"
63
- # Backpressure
64
52
  BACKPRESSURE = "backpressure"
65
- # Always On
66
53
  ALWAYS = "always"
67
54
 
68
55
 
@@ -101,38 +101,35 @@ class OutputConfluentCloudTLSSettingsClientSide(BaseModel):
101
101
  class OutputConfluentCloudAcknowledgments(int, Enum, metaclass=utils.OpenEnumMeta):
102
102
  r"""Control the number of required acknowledgments."""
103
103
 
104
- # Leader
105
104
  ONE = 1
106
- # None
107
105
  ZERO = 0
108
- # All
109
106
  MINUS_1 = -1
110
107
 
111
108
 
112
109
  class OutputConfluentCloudRecordDataFormat(str, Enum, metaclass=utils.OpenEnumMeta):
113
110
  r"""Format to use to serialize events before writing to Kafka."""
114
111
 
115
- # JSON
116
112
  JSON = "json"
117
- # Field _raw
118
113
  RAW = "raw"
119
- # Protobuf
120
114
  PROTOBUF = "protobuf"
121
115
 
122
116
 
123
117
  class OutputConfluentCloudCompression(str, Enum, metaclass=utils.OpenEnumMeta):
124
118
  r"""Codec to use to compress the data before sending to Kafka"""
125
119
 
126
- # None
127
120
  NONE = "none"
128
- # Gzip
129
121
  GZIP = "gzip"
130
- # Snappy
131
122
  SNAPPY = "snappy"
132
- # LZ4
133
123
  LZ4 = "lz4"
134
124
 
135
125
 
126
+ class OutputConfluentCloudSchemaType(str, Enum, metaclass=utils.OpenEnumMeta):
127
+ r"""The schema format used to encode and decode event data"""
128
+
129
+ AVRO = "avro"
130
+ JSON = "json"
131
+
132
+
136
133
  class OutputConfluentCloudAuthTypedDict(TypedDict):
137
134
  r"""Credentials to use when authenticating with the schema registry using basic HTTP authentication"""
138
135
 
@@ -243,6 +240,8 @@ class OutputConfluentCloudKafkaSchemaRegistryAuthenticationTypedDict(TypedDict):
243
240
  disabled: NotRequired[bool]
244
241
  schema_registry_url: NotRequired[str]
245
242
  r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
243
+ schema_type: NotRequired[OutputConfluentCloudSchemaType]
244
+ r"""The schema format used to encode and decode event data"""
246
245
  connection_timeout: NotRequired[float]
247
246
  r"""Maximum time to wait for a Schema Registry connection to complete successfully"""
248
247
  request_timeout: NotRequired[float]
@@ -268,6 +267,15 @@ class OutputConfluentCloudKafkaSchemaRegistryAuthentication(BaseModel):
268
267
  ] = "http://localhost:8081"
269
268
  r"""URL for accessing the Confluent Schema Registry. Example: http://localhost:8081. To connect over TLS, use https instead of http."""
270
269
 
270
+ schema_type: Annotated[
271
+ Annotated[
272
+ Optional[OutputConfluentCloudSchemaType],
273
+ PlainValidator(validate_open_enum(False)),
274
+ ],
275
+ pydantic.Field(alias="schemaType"),
276
+ ] = OutputConfluentCloudSchemaType.AVRO
277
+ r"""The schema format used to encode and decode event data"""
278
+
271
279
  connection_timeout: Annotated[
272
280
  Optional[float], pydantic.Field(alias="connectionTimeout")
273
281
  ] = 30000
@@ -298,13 +306,9 @@ class OutputConfluentCloudKafkaSchemaRegistryAuthentication(BaseModel):
298
306
 
299
307
 
300
308
  class OutputConfluentCloudSASLMechanism(str, Enum, metaclass=utils.OpenEnumMeta):
301
- # PLAIN
302
309
  PLAIN = "plain"
303
- # SCRAM-SHA-256
304
310
  SCRAM_SHA_256 = "scram-sha-256"
305
- # SCRAM-SHA-512
306
311
  SCRAM_SHA_512 = "scram-sha-512"
307
- # GSSAPI/Kerberos
308
312
  KERBEROS = "kerberos"
309
313
 
310
314
 
@@ -336,11 +340,8 @@ class OutputConfluentCloudAuthentication(BaseModel):
336
340
  class OutputConfluentCloudBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
337
341
  r"""How to handle events when all receivers are exerting backpressure"""
338
342
 
339
- # Block
340
343
  BLOCK = "block"
341
- # Drop
342
344
  DROP = "drop"
343
- # Persistent Queue
344
345
  QUEUE = "queue"
345
346
 
346
347
 
@@ -349,29 +350,22 @@ class OutputConfluentCloudPqCompressCompression(
349
350
  ):
350
351
  r"""Codec to use to compress the persisted data"""
351
352
 
352
- # None
353
353
  NONE = "none"
354
- # Gzip
355
354
  GZIP = "gzip"
356
355
 
357
356
 
358
357
  class OutputConfluentCloudQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
359
358
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
360
359
 
361
- # Block
362
360
  BLOCK = "block"
363
- # Drop new data
364
361
  DROP = "drop"
365
362
 
366
363
 
367
364
  class OutputConfluentCloudMode(str, Enum, metaclass=utils.OpenEnumMeta):
368
365
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
369
366
 
370
- # Error
371
367
  ERROR = "error"
372
- # Backpressure
373
368
  BACKPRESSURE = "backpressure"
374
- # Always On
375
369
  ALWAYS = "always"
376
370
 
377
371
 
@@ -101,9 +101,7 @@ class OutputCriblHTTPTLSSettingsClientSide(BaseModel):
101
101
  class OutputCriblHTTPCompression(str, Enum, metaclass=utils.OpenEnumMeta):
102
102
  r"""Codec to use to compress the data before sending"""
103
103
 
104
- # None
105
104
  NONE = "none"
106
- # Gzip
107
105
  GZIP = "gzip"
108
106
 
109
107
 
@@ -121,11 +119,8 @@ class OutputCriblHTTPExtraHTTPHeader(BaseModel):
121
119
  class OutputCriblHTTPFailedRequestLoggingMode(str, Enum, metaclass=utils.OpenEnumMeta):
122
120
  r"""Data to log when a request fails. All headers are redacted by default, unless listed as safe headers below."""
123
121
 
124
- # Payload
125
122
  PAYLOAD = "payload"
126
- # Payload + Headers
127
123
  PAYLOAD_AND_HEADERS = "payloadAndHeaders"
128
- # None
129
124
  NONE = "none"
130
125
 
131
126
 
@@ -186,11 +181,8 @@ class OutputCriblHTTPTimeoutRetrySettings(BaseModel):
186
181
  class OutputCriblHTTPBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
187
182
  r"""How to handle events when all receivers are exerting backpressure"""
188
183
 
189
- # Block
190
184
  BLOCK = "block"
191
- # Drop
192
185
  DROP = "drop"
193
- # Persistent Queue
194
186
  QUEUE = "queue"
195
187
 
196
188
 
@@ -212,29 +204,22 @@ class OutputCriblHTTPURL(BaseModel):
212
204
  class OutputCriblHTTPPqCompressCompression(str, Enum, metaclass=utils.OpenEnumMeta):
213
205
  r"""Codec to use to compress the persisted data"""
214
206
 
215
- # None
216
207
  NONE = "none"
217
- # Gzip
218
208
  GZIP = "gzip"
219
209
 
220
210
 
221
211
  class OutputCriblHTTPQueueFullBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
222
212
  r"""How to handle events when the queue is exerting backpressure (full capacity or low disk). 'Block' is the same behavior as non-PQ blocking. 'Drop new data' throws away incoming data, while leaving the contents of the PQ unchanged."""
223
213
 
224
- # Block
225
214
  BLOCK = "block"
226
- # Drop new data
227
215
  DROP = "drop"
228
216
 
229
217
 
230
218
  class OutputCriblHTTPMode(str, Enum, metaclass=utils.OpenEnumMeta):
231
219
  r"""In Error mode, PQ writes events to the filesystem if the Destination is unavailable. In Backpressure mode, PQ writes events to the filesystem when it detects backpressure from the Destination. In Always On mode, PQ always writes events to the filesystem."""
232
220
 
233
- # Error
234
221
  ERROR = "error"
235
- # Backpressure
236
222
  BACKPRESSURE = "backpressure"
237
- # Always On
238
223
  ALWAYS = "always"
239
224
 
240
225
 
@@ -25,67 +25,46 @@ class OutputCriblLakeSignatureVersion(str, Enum, metaclass=utils.OpenEnumMeta):
25
25
  class OutputCriblLakeObjectACL(str, Enum, metaclass=utils.OpenEnumMeta):
26
26
  r"""Object ACL to assign to uploaded objects"""
27
27
 
28
- # Private
29
28
  PRIVATE = "private"
30
- # Public Read Only
31
29
  PUBLIC_READ = "public-read"
32
- # Public Read/Write
33
30
  PUBLIC_READ_WRITE = "public-read-write"
34
- # Authenticated Read Only
35
31
  AUTHENTICATED_READ = "authenticated-read"
36
- # AWS EC2 AMI Read Only
37
32
  AWS_EXEC_READ = "aws-exec-read"
38
- # Bucket Owner Read Only
39
33
  BUCKET_OWNER_READ = "bucket-owner-read"
40
- # Bucket Owner Full Control
41
34
  BUCKET_OWNER_FULL_CONTROL = "bucket-owner-full-control"
42
35
 
43
36
 
44
37
  class OutputCriblLakeStorageClass(str, Enum, metaclass=utils.OpenEnumMeta):
45
38
  r"""Storage class to select for uploaded objects"""
46
39
 
47
- # Standard
48
40
  STANDARD = "STANDARD"
49
- # Reduced Redundancy Storage
50
41
  REDUCED_REDUNDANCY = "REDUCED_REDUNDANCY"
51
- # Standard, Infrequent Access
52
42
  STANDARD_IA = "STANDARD_IA"
53
- # One Zone, Infrequent Access
54
43
  ONEZONE_IA = "ONEZONE_IA"
55
- # Intelligent Tiering
56
44
  INTELLIGENT_TIERING = "INTELLIGENT_TIERING"
57
- # Glacier Flexible Retrieval
58
45
  GLACIER = "GLACIER"
59
- # Glacier Instant Retrieval
60
46
  GLACIER_IR = "GLACIER_IR"
61
- # Glacier Deep Archive
62
47
  DEEP_ARCHIVE = "DEEP_ARCHIVE"
63
48
 
64
49
 
65
50
  class OutputCriblLakeServerSideEncryptionForUploadedObjects(
66
51
  str, Enum, metaclass=utils.OpenEnumMeta
67
52
  ):
68
- # Amazon S3 Managed Key
69
53
  AES256 = "AES256"
70
- # AWS KMS Managed Key
71
54
  AWS_KMS = "aws:kms"
72
55
 
73
56
 
74
57
  class OutputCriblLakeBackpressureBehavior(str, Enum, metaclass=utils.OpenEnumMeta):
75
58
  r"""How to handle events when all receivers are exerting backpressure"""
76
59
 
77
- # Block
78
60
  BLOCK = "block"
79
- # Drop
80
61
  DROP = "drop"
81
62
 
82
63
 
83
64
  class OutputCriblLakeDiskSpaceProtection(str, Enum, metaclass=utils.OpenEnumMeta):
84
65
  r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
85
66
 
86
- # Block
87
67
  BLOCK = "block"
88
- # Drop
89
68
  DROP = "drop"
90
69
 
91
70