cribl-control-plane 0.1.0b1__py3-none-any.whl → 0.1.0b2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

@@ -3,10 +3,10 @@
3
3
  import importlib.metadata
4
4
 
5
5
  __title__: str = "cribl-control-plane"
6
- __version__: str = "0.1.0b1"
7
- __openapi_doc_version__: str = "4.15.0-alpha.1758872649120-b192666f"
6
+ __version__: str = "0.1.0b2"
7
+ __openapi_doc_version__: str = "4.15.0-alpha.1759327501989-3e9845e9"
8
8
  __gen_version__: str = "2.716.16"
9
- __user_agent__: str = "speakeasy-sdk/python 0.1.0b1 2.716.16 4.15.0-alpha.1758872649120-b192666f cribl-control-plane"
9
+ __user_agent__: str = "speakeasy-sdk/python 0.1.0b2 2.716.16 4.15.0-alpha.1759327501989-3e9845e9 cribl-control-plane"
10
10
 
11
11
  try:
12
12
  if __package__ is not None:
@@ -224,7 +224,7 @@ class GroupsSDK(BaseSDK):
224
224
  config_version: Optional[str] = None,
225
225
  deploying_worker_count: Optional[float] = None,
226
226
  description: Optional[str] = None,
227
- estimated_ingest_rate: Optional[float] = None,
227
+ estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
228
228
  git: Optional[Union[models.Git, models.GitTypedDict]] = None,
229
229
  incompatible_worker_count: Optional[float] = None,
230
230
  inherits: Optional[str] = None,
@@ -395,7 +395,7 @@ class GroupsSDK(BaseSDK):
395
395
  config_version: Optional[str] = None,
396
396
  deploying_worker_count: Optional[float] = None,
397
397
  description: Optional[str] = None,
398
- estimated_ingest_rate: Optional[float] = None,
398
+ estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
399
399
  git: Optional[Union[models.Git, models.GitTypedDict]] = None,
400
400
  incompatible_worker_count: Optional[float] = None,
401
401
  inherits: Optional[str] = None,
@@ -757,7 +757,7 @@ class GroupsSDK(BaseSDK):
757
757
  config_version: Optional[str] = None,
758
758
  deploying_worker_count: Optional[float] = None,
759
759
  description: Optional[str] = None,
760
- estimated_ingest_rate: Optional[float] = None,
760
+ estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
761
761
  git: Optional[Union[models.Git, models.GitTypedDict]] = None,
762
762
  incompatible_worker_count: Optional[float] = None,
763
763
  inherits: Optional[str] = None,
@@ -931,7 +931,7 @@ class GroupsSDK(BaseSDK):
931
931
  config_version: Optional[str] = None,
932
932
  deploying_worker_count: Optional[float] = None,
933
933
  description: Optional[str] = None,
934
- estimated_ingest_rate: Optional[float] = None,
934
+ estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
935
935
  git: Optional[Union[models.Git, models.GitTypedDict]] = None,
936
936
  incompatible_worker_count: Optional[float] = None,
937
937
  inherits: Optional[str] = None,
@@ -22,6 +22,7 @@ if TYPE_CHECKING:
22
22
  ConfigGroup,
23
23
  ConfigGroupType,
24
24
  ConfigGroupTypedDict,
25
+ EstimatedIngestRate,
25
26
  Git,
26
27
  GitTypedDict,
27
28
  )
@@ -2073,14 +2074,8 @@ if TYPE_CHECKING:
2073
2074
  OutputDatabricks,
2074
2075
  OutputDatabricksAuthenticationMethod,
2075
2076
  OutputDatabricksBackpressureBehavior,
2076
- OutputDatabricksCompression,
2077
- OutputDatabricksCompressionLevel,
2078
2077
  OutputDatabricksDataFormat,
2079
- OutputDatabricksDataPageVersion,
2080
2078
  OutputDatabricksDiskSpaceProtection,
2081
- OutputDatabricksKeyValueMetadatum,
2082
- OutputDatabricksKeyValueMetadatumTypedDict,
2083
- OutputDatabricksParquetVersion,
2084
2079
  OutputDatabricksType,
2085
2080
  OutputDatabricksTypedDict,
2086
2081
  )
@@ -3470,6 +3465,7 @@ __all__ = [
3470
3465
  "EndpointType",
3471
3466
  "Error",
3472
3467
  "ErrorTypedDict",
3468
+ "EstimatedIngestRate",
3473
3469
  "EventFormat",
3474
3470
  "Executor",
3475
3471
  "ExecutorSpecificSettings",
@@ -5131,14 +5127,8 @@ __all__ = [
5131
5127
  "OutputDatabricks",
5132
5128
  "OutputDatabricksAuthenticationMethod",
5133
5129
  "OutputDatabricksBackpressureBehavior",
5134
- "OutputDatabricksCompression",
5135
- "OutputDatabricksCompressionLevel",
5136
5130
  "OutputDatabricksDataFormat",
5137
- "OutputDatabricksDataPageVersion",
5138
5131
  "OutputDatabricksDiskSpaceProtection",
5139
- "OutputDatabricksKeyValueMetadatum",
5140
- "OutputDatabricksKeyValueMetadatumTypedDict",
5141
- "OutputDatabricksParquetVersion",
5142
5132
  "OutputDatabricksType",
5143
5133
  "OutputDatabricksTypedDict",
5144
5134
  "OutputDatadog",
@@ -6246,6 +6236,7 @@ _dynamic_imports: dict[str, str] = {
6246
6236
  "ConfigGroup": ".configgroup",
6247
6237
  "ConfigGroupType": ".configgroup",
6248
6238
  "ConfigGroupTypedDict": ".configgroup",
6239
+ "EstimatedIngestRate": ".configgroup",
6249
6240
  "Git": ".configgroup",
6250
6241
  "GitTypedDict": ".configgroup",
6251
6242
  "ConfigGroupCloud": ".configgroupcloud",
@@ -8078,14 +8069,8 @@ _dynamic_imports: dict[str, str] = {
8078
8069
  "OutputDatabricks": ".outputdatabricks",
8079
8070
  "OutputDatabricksAuthenticationMethod": ".outputdatabricks",
8080
8071
  "OutputDatabricksBackpressureBehavior": ".outputdatabricks",
8081
- "OutputDatabricksCompression": ".outputdatabricks",
8082
- "OutputDatabricksCompressionLevel": ".outputdatabricks",
8083
8072
  "OutputDatabricksDataFormat": ".outputdatabricks",
8084
- "OutputDatabricksDataPageVersion": ".outputdatabricks",
8085
8073
  "OutputDatabricksDiskSpaceProtection": ".outputdatabricks",
8086
- "OutputDatabricksKeyValueMetadatum": ".outputdatabricks",
8087
- "OutputDatabricksKeyValueMetadatumTypedDict": ".outputdatabricks",
8088
- "OutputDatabricksParquetVersion": ".outputdatabricks",
8089
8074
  "OutputDatabricksType": ".outputdatabricks",
8090
8075
  "OutputDatabricksTypedDict": ".outputdatabricks",
8091
8076
  "DatadogSite": ".outputdatadog",
@@ -14,6 +14,18 @@ from typing import List, Optional
14
14
  from typing_extensions import Annotated, NotRequired, TypedDict
15
15
 
16
16
 
17
+ class EstimatedIngestRate(int, Enum, metaclass=utils.OpenEnumMeta):
18
+ ONE_THOUSAND_AND_TWENTY_FOUR = 1024
19
+ FOUR_THOUSAND_AND_NINETY_SIX = 4096
20
+ TEN_THOUSAND_TWO_HUNDRED_AND_FORTY = 10240
21
+ TWO_THOUSAND_AND_FORTY_EIGHT = 2048
22
+ THREE_THOUSAND_AND_SEVENTY_TWO = 3072
23
+ FIVE_THOUSAND_ONE_HUNDRED_AND_TWENTY = 5120
24
+ SEVEN_THOUSAND_ONE_HUNDRED_AND_SIXTY_EIGHT = 7168
25
+ THIRTEEN_THOUSAND_THREE_HUNDRED_AND_TWELVE = 13312
26
+ FIFTEEN_THOUSAND_THREE_HUNDRED_AND_SIXTY = 15360
27
+
28
+
17
29
  class GitTypedDict(TypedDict):
18
30
  commit: NotRequired[str]
19
31
  local_changes: NotRequired[float]
@@ -40,7 +52,7 @@ class ConfigGroupTypedDict(TypedDict):
40
52
  config_version: NotRequired[str]
41
53
  deploying_worker_count: NotRequired[float]
42
54
  description: NotRequired[str]
43
- estimated_ingest_rate: NotRequired[float]
55
+ estimated_ingest_rate: NotRequired[EstimatedIngestRate]
44
56
  git: NotRequired[GitTypedDict]
45
57
  incompatible_worker_count: NotRequired[float]
46
58
  inherits: NotRequired[str]
@@ -75,7 +87,10 @@ class ConfigGroup(BaseModel):
75
87
  description: Optional[str] = None
76
88
 
77
89
  estimated_ingest_rate: Annotated[
78
- Optional[float], pydantic.Field(alias="estimatedIngestRate")
90
+ Annotated[
91
+ Optional[EstimatedIngestRate], PlainValidator(validate_open_enum(True))
92
+ ],
93
+ pydantic.Field(alias="estimatedIngestRate"),
79
94
  ] = None
80
95
 
81
96
  git: Optional[Git] = None
@@ -91,31 +91,32 @@ OutputTypedDict = TypeAliasType(
91
91
  OutputDevnullTypedDict,
92
92
  OutputDefaultTypedDict,
93
93
  OutputRouterTypedDict,
94
- OutputSnmpTypedDict,
95
94
  OutputNetflowTypedDict,
95
+ OutputSnmpTypedDict,
96
96
  OutputDiskSpoolTypedDict,
97
97
  OutputRingTypedDict,
98
- OutputStatsdExtTypedDict,
99
98
  OutputGraphiteTypedDict,
99
+ OutputStatsdExtTypedDict,
100
100
  OutputStatsdTypedDict,
101
101
  OutputGooglePubsubTypedDict,
102
- OutputCriblTCPTypedDict,
103
102
  OutputSplunkTypedDict,
103
+ OutputCriblTCPTypedDict,
104
104
  OutputSnsTypedDict,
105
105
  OutputCloudwatchTypedDict,
106
106
  OutputAzureEventhubTypedDict,
107
- OutputWavefrontTypedDict,
108
107
  OutputSignalfxTypedDict,
108
+ OutputWavefrontTypedDict,
109
109
  OutputHoneycombTypedDict,
110
- OutputSumoLogicTypedDict,
111
- OutputCrowdstrikeNextGenSiemTypedDict,
112
110
  OutputHumioHecTypedDict,
113
111
  OutputTcpjsonTypedDict,
112
+ OutputSumoLogicTypedDict,
113
+ OutputCrowdstrikeNextGenSiemTypedDict,
114
114
  OutputElasticCloudTypedDict,
115
- OutputKinesisTypedDict,
116
- OutputConfluentCloudTypedDict,
117
- OutputKafkaTypedDict,
118
115
  OutputExabeamTypedDict,
116
+ OutputKafkaTypedDict,
117
+ OutputConfluentCloudTypedDict,
118
+ OutputKinesisTypedDict,
119
+ OutputDatabricksTypedDict,
119
120
  OutputNewrelicEventsTypedDict,
120
121
  OutputAzureLogsTypedDict,
121
122
  OutputSplunkLbTypedDict,
@@ -131,14 +132,13 @@ OutputTypedDict = TypeAliasType(
131
132
  OutputDynatraceHTTPTypedDict,
132
133
  OutputServiceNowTypedDict,
133
134
  OutputDynatraceOtlpTypedDict,
134
- OutputElasticTypedDict,
135
135
  OutputGoogleChronicleTypedDict,
136
- OutputCriblLakeTypedDict,
136
+ OutputElasticTypedDict,
137
137
  OutputDatadogTypedDict,
138
+ OutputCriblLakeTypedDict,
138
139
  OutputPrometheusTypedDict,
139
140
  OutputMskTypedDict,
140
141
  OutputSentinelOneAiSiemTypedDict,
141
- OutputDatabricksTypedDict,
142
142
  OutputSentinelTypedDict,
143
143
  OutputInfluxdbTypedDict,
144
144
  OutputGoogleCloudStorageTypedDict,
@@ -163,31 +163,32 @@ Output = TypeAliasType(
163
163
  OutputDevnull,
164
164
  OutputDefault,
165
165
  OutputRouter,
166
- OutputSnmp,
167
166
  OutputNetflow,
167
+ OutputSnmp,
168
168
  OutputDiskSpool,
169
169
  OutputRing,
170
- OutputStatsdExt,
171
170
  OutputGraphite,
171
+ OutputStatsdExt,
172
172
  OutputStatsd,
173
173
  OutputGooglePubsub,
174
- OutputCriblTCP,
175
174
  OutputSplunk,
175
+ OutputCriblTCP,
176
176
  OutputSns,
177
177
  OutputCloudwatch,
178
178
  OutputAzureEventhub,
179
- OutputWavefront,
180
179
  OutputSignalfx,
180
+ OutputWavefront,
181
181
  OutputHoneycomb,
182
- OutputSumoLogic,
183
- OutputCrowdstrikeNextGenSiem,
184
182
  OutputHumioHec,
185
183
  OutputTcpjson,
184
+ OutputSumoLogic,
185
+ OutputCrowdstrikeNextGenSiem,
186
186
  OutputElasticCloud,
187
- OutputKinesis,
188
- OutputConfluentCloud,
189
- OutputKafka,
190
187
  OutputExabeam,
188
+ OutputKafka,
189
+ OutputConfluentCloud,
190
+ OutputKinesis,
191
+ OutputDatabricks,
191
192
  OutputNewrelicEvents,
192
193
  OutputAzureLogs,
193
194
  OutputSplunkLb,
@@ -203,14 +204,13 @@ Output = TypeAliasType(
203
204
  OutputDynatraceHTTP,
204
205
  OutputServiceNow,
205
206
  OutputDynatraceOtlp,
206
- OutputElastic,
207
207
  OutputGoogleChronicle,
208
- OutputCriblLake,
208
+ OutputElastic,
209
209
  OutputDatadog,
210
+ OutputCriblLake,
210
211
  OutputPrometheus,
211
212
  OutputMsk,
212
213
  OutputSentinelOneAiSiem,
213
- OutputDatabricks,
214
214
  OutputSentinel,
215
215
  OutputInfluxdb,
216
216
  OutputGoogleCloudStorage,
@@ -44,49 +44,12 @@ class OutputDatabricksAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMe
44
44
  SECRET = "secret"
45
45
 
46
46
 
47
- class OutputDatabricksCompression(str, Enum, metaclass=utils.OpenEnumMeta):
48
- r"""Data compression format to apply to HTTP content before it is delivered"""
49
-
50
- NONE = "none"
51
- GZIP = "gzip"
52
-
53
-
54
- class OutputDatabricksCompressionLevel(str, Enum, metaclass=utils.OpenEnumMeta):
55
- r"""Compression level to apply before moving files to final destination"""
56
-
57
- BEST_SPEED = "best_speed"
58
- NORMAL = "normal"
59
- BEST_COMPRESSION = "best_compression"
60
-
61
-
62
- class OutputDatabricksParquetVersion(str, Enum, metaclass=utils.OpenEnumMeta):
63
- r"""Determines which data types are supported and how they are represented"""
64
-
65
- PARQUET_1_0 = "PARQUET_1_0"
66
- PARQUET_2_4 = "PARQUET_2_4"
67
- PARQUET_2_6 = "PARQUET_2_6"
68
-
69
-
70
- class OutputDatabricksDataPageVersion(str, Enum, metaclass=utils.OpenEnumMeta):
71
- r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
72
-
73
- DATA_PAGE_V1 = "DATA_PAGE_V1"
74
- DATA_PAGE_V2 = "DATA_PAGE_V2"
75
-
76
-
77
- class OutputDatabricksKeyValueMetadatumTypedDict(TypedDict):
78
- value: str
79
- key: NotRequired[str]
80
-
81
-
82
- class OutputDatabricksKeyValueMetadatum(BaseModel):
83
- value: str
84
-
85
- key: Optional[str] = ""
86
-
87
-
88
47
  class OutputDatabricksTypedDict(TypedDict):
89
48
  type: OutputDatabricksType
49
+ login_url: str
50
+ r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
51
+ client_id: str
52
+ r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
90
53
  id: NotRequired[str]
91
54
  r"""Unique ID for this output"""
92
55
  pipeline: NotRequired[str]
@@ -133,14 +96,6 @@ class OutputDatabricksTypedDict(TypedDict):
133
96
  r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
134
97
  unity_auth_method: NotRequired[OutputDatabricksAuthenticationMethod]
135
98
  r"""Unity Catalog authentication method. Choose Manual to enter credentials directly, or Secret to use a stored secret."""
136
- login_url: NotRequired[str]
137
- r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
138
- client_id: NotRequired[str]
139
- r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
140
- client_secret: NotRequired[str]
141
- r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
142
- client_text_secret: NotRequired[str]
143
- r"""Select or create a stored secret that references your Client ID and Client Secret"""
144
99
  scope: NotRequired[str]
145
100
  r"""OAuth scope for Unity Catalog authentication"""
146
101
  token_timeout_secs: NotRequired[float]
@@ -154,41 +109,21 @@ class OutputDatabricksTypedDict(TypedDict):
154
109
  over_write_files: NotRequired[bool]
155
110
  r"""Uploaded files should be overwritten if they already exist. If disabled, upload will fail if a file already exists."""
156
111
  description: NotRequired[str]
157
- compress: NotRequired[OutputDatabricksCompression]
158
- r"""Data compression format to apply to HTTP content before it is delivered"""
159
- compression_level: NotRequired[OutputDatabricksCompressionLevel]
160
- r"""Compression level to apply before moving files to final destination"""
161
- automatic_schema: NotRequired[bool]
162
- r"""Automatically calculate the schema based on the events of each Parquet file generated"""
163
- parquet_version: NotRequired[OutputDatabricksParquetVersion]
164
- r"""Determines which data types are supported and how they are represented"""
165
- parquet_data_page_version: NotRequired[OutputDatabricksDataPageVersion]
166
- r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
167
- parquet_row_group_length: NotRequired[float]
168
- r"""The number of rows that every group will contain. The final group can contain a smaller number of rows."""
169
- parquet_page_size: NotRequired[str]
170
- r"""Target memory size for page segments, such as 1MB or 128MB. Generally, lower values improve reading speed, while higher values improve compression."""
171
- should_log_invalid_rows: NotRequired[bool]
172
- r"""Log up to 3 rows that @{product} skips due to data mismatch"""
173
- key_value_metadata: NotRequired[List[OutputDatabricksKeyValueMetadatumTypedDict]]
174
- r"""The metadata of files the Destination writes will include the properties you add here as key-value pairs. Useful for tagging. Examples: \"key\":\"OCSF Event Class\", \"value\":\"9001\" """
175
- enable_statistics: NotRequired[bool]
176
- r"""Statistics profile an entire file in terms of minimum/maximum values within data, numbers of nulls, etc. You can use Parquet tools to view statistics."""
177
- enable_write_page_index: NotRequired[bool]
178
- r"""One page index contains statistics for one data page. Parquet readers use statistics to enable page skipping."""
179
- enable_page_checksum: NotRequired[bool]
180
- r"""Parquet tools can use the checksum of a Parquet page to verify data integrity"""
181
- empty_dir_cleanup_sec: NotRequired[float]
182
- r"""How frequently, in seconds, to clean up empty directories"""
183
- deadletter_path: NotRequired[str]
184
- r"""Storage location for files that fail to reach their final destination after maximum retries are exceeded"""
185
- max_retry_num: NotRequired[float]
186
- r"""The maximum number of times a file will attempt to move to its final destination before being dead-lettered"""
112
+ client_secret: NotRequired[str]
113
+ r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
114
+ client_text_secret: NotRequired[str]
115
+ r"""Select or create a stored text secret"""
187
116
 
188
117
 
189
118
  class OutputDatabricks(BaseModel):
190
119
  type: OutputDatabricksType
191
120
 
121
+ login_url: Annotated[str, pydantic.Field(alias="loginUrl")]
122
+ r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
123
+
124
+ client_id: Annotated[str, pydantic.Field(alias="clientId")]
125
+ r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
126
+
192
127
  id: Optional[str] = None
193
128
  r"""Unique ID for this output"""
194
129
 
@@ -308,20 +243,6 @@ class OutputDatabricks(BaseModel):
308
243
  ] = OutputDatabricksAuthenticationMethod.MANUAL
309
244
  r"""Unity Catalog authentication method. Choose Manual to enter credentials directly, or Secret to use a stored secret."""
310
245
 
311
- login_url: Annotated[Optional[str], pydantic.Field(alias="loginUrl")] = None
312
- r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
313
-
314
- client_id: Annotated[Optional[str], pydantic.Field(alias="clientId")] = None
315
- r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
316
-
317
- client_secret: Annotated[Optional[str], pydantic.Field(alias="clientSecret")] = None
318
- r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
319
-
320
- client_text_secret: Annotated[
321
- Optional[str], pydantic.Field(alias="clientTextSecret")
322
- ] = None
323
- r"""Select or create a stored secret that references your Client ID and Client Secret"""
324
-
325
246
  scope: Optional[str] = "all-apis"
326
247
  r"""OAuth scope for Unity Catalog authentication"""
327
248
 
@@ -352,88 +273,10 @@ class OutputDatabricks(BaseModel):
352
273
 
353
274
  description: Optional[str] = None
354
275
 
355
- compress: Annotated[
356
- Optional[OutputDatabricksCompression], PlainValidator(validate_open_enum(False))
357
- ] = OutputDatabricksCompression.GZIP
358
- r"""Data compression format to apply to HTTP content before it is delivered"""
359
-
360
- compression_level: Annotated[
361
- Annotated[
362
- Optional[OutputDatabricksCompressionLevel],
363
- PlainValidator(validate_open_enum(False)),
364
- ],
365
- pydantic.Field(alias="compressionLevel"),
366
- ] = OutputDatabricksCompressionLevel.BEST_SPEED
367
- r"""Compression level to apply before moving files to final destination"""
368
-
369
- automatic_schema: Annotated[
370
- Optional[bool], pydantic.Field(alias="automaticSchema")
371
- ] = False
372
- r"""Automatically calculate the schema based on the events of each Parquet file generated"""
373
-
374
- parquet_version: Annotated[
375
- Annotated[
376
- Optional[OutputDatabricksParquetVersion],
377
- PlainValidator(validate_open_enum(False)),
378
- ],
379
- pydantic.Field(alias="parquetVersion"),
380
- ] = OutputDatabricksParquetVersion.PARQUET_2_6
381
- r"""Determines which data types are supported and how they are represented"""
382
-
383
- parquet_data_page_version: Annotated[
384
- Annotated[
385
- Optional[OutputDatabricksDataPageVersion],
386
- PlainValidator(validate_open_enum(False)),
387
- ],
388
- pydantic.Field(alias="parquetDataPageVersion"),
389
- ] = OutputDatabricksDataPageVersion.DATA_PAGE_V2
390
- r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
391
-
392
- parquet_row_group_length: Annotated[
393
- Optional[float], pydantic.Field(alias="parquetRowGroupLength")
394
- ] = 10000
395
- r"""The number of rows that every group will contain. The final group can contain a smaller number of rows."""
396
-
397
- parquet_page_size: Annotated[
398
- Optional[str], pydantic.Field(alias="parquetPageSize")
399
- ] = "1MB"
400
- r"""Target memory size for page segments, such as 1MB or 128MB. Generally, lower values improve reading speed, while higher values improve compression."""
401
-
402
- should_log_invalid_rows: Annotated[
403
- Optional[bool], pydantic.Field(alias="shouldLogInvalidRows")
404
- ] = None
405
- r"""Log up to 3 rows that @{product} skips due to data mismatch"""
276
+ client_secret: Annotated[Optional[str], pydantic.Field(alias="clientSecret")] = None
277
+ r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
406
278
 
407
- key_value_metadata: Annotated[
408
- Optional[List[OutputDatabricksKeyValueMetadatum]],
409
- pydantic.Field(alias="keyValueMetadata"),
279
+ client_text_secret: Annotated[
280
+ Optional[str], pydantic.Field(alias="clientTextSecret")
410
281
  ] = None
411
- r"""The metadata of files the Destination writes will include the properties you add here as key-value pairs. Useful for tagging. Examples: \"key\":\"OCSF Event Class\", \"value\":\"9001\" """
412
-
413
- enable_statistics: Annotated[
414
- Optional[bool], pydantic.Field(alias="enableStatistics")
415
- ] = True
416
- r"""Statistics profile an entire file in terms of minimum/maximum values within data, numbers of nulls, etc. You can use Parquet tools to view statistics."""
417
-
418
- enable_write_page_index: Annotated[
419
- Optional[bool], pydantic.Field(alias="enableWritePageIndex")
420
- ] = True
421
- r"""One page index contains statistics for one data page. Parquet readers use statistics to enable page skipping."""
422
-
423
- enable_page_checksum: Annotated[
424
- Optional[bool], pydantic.Field(alias="enablePageChecksum")
425
- ] = False
426
- r"""Parquet tools can use the checksum of a Parquet page to verify data integrity"""
427
-
428
- empty_dir_cleanup_sec: Annotated[
429
- Optional[float], pydantic.Field(alias="emptyDirCleanupSec")
430
- ] = 300
431
- r"""How frequently, in seconds, to clean up empty directories"""
432
-
433
- deadletter_path: Annotated[
434
- Optional[str], pydantic.Field(alias="deadletterPath")
435
- ] = "$CRIBL_HOME/state/outputs/dead-letter"
436
- r"""Storage location for files that fail to reach their final destination after maximum retries are exceeded"""
437
-
438
- max_retry_num: Annotated[Optional[float], pydantic.Field(alias="maxRetryNum")] = 20
439
- r"""The maximum number of times a file will attempt to move to its final destination before being dead-lettered"""
282
+ r"""Select or create a stored text secret"""
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: cribl-control-plane
3
- Version: 0.1.0b1
3
+ Version: 0.1.0b2
4
4
  Summary: Python Client SDK Generated by Speakeasy.
5
5
  Author: Speakeasy
6
6
  Requires-Python: >=3.9.2
@@ -4,7 +4,7 @@ cribl_control_plane/_hooks/clientcredentials.py,sha256=p1WN7LL3PHrAf4AxXrsOZF_NB
4
4
  cribl_control_plane/_hooks/registration.py,sha256=1QZB41w6If7I9dXiOSQx6dhSc6BPWrnI5Q5bMOr4iVA,624
5
5
  cribl_control_plane/_hooks/sdkhooks.py,sha256=ggXjME1_Rdv8CVCg1XHqB83eYtbxzKyhXyfQ36Yc1gA,2816
6
6
  cribl_control_plane/_hooks/types.py,sha256=Tw_C4zTZm01rW_89VDEUpvQ8KQr1WxN0Gu_-s_fYSPc,2998
7
- cribl_control_plane/_version.py,sha256=fEJ_XGOE2S0NaLK3-Wh2Bp8ixPeJOi1-yjcxCI4YbvE,546
7
+ cribl_control_plane/_version.py,sha256=BXnoleFws7YHijH4Jb3Mnb7Ffa-2mD7o8h-uXu7BnZs,546
8
8
  cribl_control_plane/acl.py,sha256=8lvYOKAli4PzsQhOVaCU6YCwblPMh9jQo04L0r4HJuQ,9025
9
9
  cribl_control_plane/auth_sdk.py,sha256=3sjf1VoyWwfhSyuMDQLixgWISSf03BOZwmkiT8g5Ruw,626
10
10
  cribl_control_plane/basesdk.py,sha256=y4yIXSNVXLMd0sLS2htBFdTCI3gkPQbIWd-C671kg1I,12249
@@ -22,12 +22,12 @@ cribl_control_plane/errors/healthstatus_error.py,sha256=hakt3RO4e4mWnmiDy1rzBWCZ
22
22
  cribl_control_plane/errors/no_response_error.py,sha256=Kb7hmMtDo72KrLSjUEDNeQxvzZiVxUjOZym8TPdZp5Y,462
23
23
  cribl_control_plane/errors/responsevalidationerror.py,sha256=7vHiNVp0rm6nrult0NCC73VDabehUhM861LrWfaObvA,778
24
24
  cribl_control_plane/groups_configs.py,sha256=dgi-W0ElnyygaVKXqk5df2ldAAgj9YmXRPCez2hP7yc,695
25
- cribl_control_plane/groups_sdk.py,sha256=EkviXQbbtP9HIv8tSkRtyOTPqxVTySgzMlgx_zhudig,61835
25
+ cribl_control_plane/groups_sdk.py,sha256=pZy0owBWP-RFNp9CwA2fopDZSceUX1CFlAtaTsFctDc,61919
26
26
  cribl_control_plane/health.py,sha256=N8pX8RHkJVtLFd4nZ8ypJPrzT_JezciEVry9s9qvCRc,7019
27
27
  cribl_control_plane/hectokens.py,sha256=0EGgGGrM83m1YmTZwkN5S4xFkHQGnw1IZe3y6uMwmLw,19151
28
28
  cribl_control_plane/httpclient.py,sha256=Eu73urOAiZQtdUIyOUnPccxCiBbWEKrXG-JrRG3SLM4,3946
29
29
  cribl_control_plane/lakedatasets.py,sha256=7WYWcjXMzliDW1j3TQlgikc_h54IUq4lsysVy_39l38,46578
30
- cribl_control_plane/models/__init__.py,sha256=IkETDpbByd70u_j2-Pz0sWPhTJneGxWwQw9HyYcOA_8,388599
30
+ cribl_control_plane/models/__init__.py,sha256=8x_C90VeiA7FF5Z14TmOwCjYLHWQ-syjaWXzOJTS19M,387831
31
31
  cribl_control_plane/models/addhectokenrequest.py,sha256=mzQLKrMWlwxNheqEs5SM_yrT-gyenfCWgHKhmb5oXFQ,800
32
32
  cribl_control_plane/models/authtoken.py,sha256=uW0aIs8j14CQzFM2ueY5GIWFulna91cigBWQ3oPlDgY,295
33
33
  cribl_control_plane/models/branchinfo.py,sha256=jCX31O5TMG9jTjqigPvvUiBwpgPpVxHtSuhYrNykXiI,291
@@ -35,7 +35,7 @@ cribl_control_plane/models/cacheconnection.py,sha256=IaqcKQhOxuY_SYdMpD2FqBGMbra
35
35
  cribl_control_plane/models/cacheconnectionbackfillstatus.py,sha256=EFJtxJ8EULR0JI4SjDoIsYSxx7uGHk-ULOeSxSzTCSc,380
36
36
  cribl_control_plane/models/cloudprovider.py,sha256=OwlC4oXKrv5AjIgotkcSSTGgods-2QWeR0N5UgJ20Yw,268
37
37
  cribl_control_plane/models/commit.py,sha256=wXQkjOYsffxWURHTrfU9kJ4HF2H65QfD1R9-vH0clvQ,641
38
- cribl_control_plane/models/configgroup.py,sha256=UxscJdSzek2Bq6jHi2I_Jum-iz4OxV40LfjUBh87mpE,3701
38
+ cribl_control_plane/models/configgroup.py,sha256=5t0IF-5BVrtJLM8XPBBrfriY9RXP9TxCzK2m46mDWiA,4302
39
39
  cribl_control_plane/models/configgroupcloud.py,sha256=xme7fTZesBGwpma0huL-EXyPNaDYhP2CeYnvxBRxwU8,1589
40
40
  cribl_control_plane/models/configgrouplookups.py,sha256=1z1DlvlVehqfD6hZMXG0XedZTfoCIpYd0cHav45tiRw,830
41
41
  cribl_control_plane/models/createconfiggroupbyproductop.py,sha256=OZtUSrA4Lf81IhqH-PnO5nqbcWJ_plFLJJbe-vIBsSI,1728
@@ -188,7 +188,7 @@ cribl_control_plane/models/nodeprovidedinfo.py,sha256=XElCf2bdE4ahiO7MaVmbtGziKg
188
188
  cribl_control_plane/models/nodeskippedupgradestatus.py,sha256=EY-U3cUPwMa3H-X-hn5gdaEBmSAP3hB9gRPdiQZs5yU,294
189
189
  cribl_control_plane/models/nodeupgradestate.py,sha256=EerzMMQeFl-iHKHsJwEIxRroH6w97S7-em9YoY2-ASk,286
190
190
  cribl_control_plane/models/nodeupgradestatus.py,sha256=Ygdb7jTFOvD6M3Fjl3brliLCKbkdX3aCwkPYjTE4Dw0,1346
191
- cribl_control_plane/models/output.py,sha256=iok_sCpufe4AGE1O2mFQfz7fOX56UqSpEXbj5ps3Ixs,8830
191
+ cribl_control_plane/models/output.py,sha256=BN8ePTnt10hGAIyZ_WmiUBHdsQ6EMHE0_Tu9pOhiuRA,8830
192
192
  cribl_control_plane/models/outputazureblob.py,sha256=l3N9XWrzsjgpoV8AZ59lmHXL7ZNrGek1VyO2bA13DKQ,23132
193
193
  cribl_control_plane/models/outputazuredataexplorer.py,sha256=dHMj1qBpoNwusSety4OG3KgaKKNixTaqcdKSsxDAo2Q,31977
194
194
  cribl_control_plane/models/outputazureeventhub.py,sha256=mccmtlZjwoeyr3AagtPRleOpZKB6roJvwekRjScqa8w,15710
@@ -200,7 +200,7 @@ cribl_control_plane/models/outputcriblhttp.py,sha256=qH_DLlLfUDMTgyjLJ-EyMytmHAJ
200
200
  cribl_control_plane/models/outputcribllake.py,sha256=h-J1uPLPV3teRWJIyrJLV1oWjxiWlQAiwDoV3rNX-Ks,17914
201
201
  cribl_control_plane/models/outputcribltcp.py,sha256=wpiSsXcSBFXe2ZO0muijlA58TFVgL1Fw7FTt_zodf_8,17482
202
202
  cribl_control_plane/models/outputcrowdstrikenextgensiem.py,sha256=mFJowiqrVy6peiPXOcLZTyGtxGTLLuIpoG56gw81dMY,19864
203
- cribl_control_plane/models/outputdatabricks.py,sha256=SsZhmfnZ-VMxa_dcuLj0_y98NASgKHl_CDQi5UzQuGs,21879
203
+ cribl_control_plane/models/outputdatabricks.py,sha256=dmxj_K4nbO8nHAcWhOGZJIqs6KWlo1Q1RpO8Twez3WM,13934
204
204
  cribl_control_plane/models/outputdatadog.py,sha256=KUFjLBY9FtfzSuS6XiWsvru_HmAgnSfeQFw_J-T3OVk,23291
205
205
  cribl_control_plane/models/outputdataset.py,sha256=-iXA-e7awghBU6Plu-774F1P2TQN2w7oZX-JexG7hS4,21606
206
206
  cribl_control_plane/models/outputdefault.py,sha256=2tjMKYSksR-0qWLd_u3PPLXL0gZiSlUdj9JTPYeYMps,1952
@@ -322,6 +322,6 @@ cribl_control_plane/utils/url.py,sha256=BgGPgcTA6MRK4bF8fjP2dUopN3NzEzxWMXPBVg8N
322
322
  cribl_control_plane/utils/values.py,sha256=CcaCXEa3xHhkUDROyXZocN8f0bdITftv9Y0P9lTf0YM,3517
323
323
  cribl_control_plane/versions.py,sha256=4xdTYbM84Xyjr5qkixqNpgn2q6V8aXVYXkEPDU2Ele0,1156
324
324
  cribl_control_plane/versions_configs.py,sha256=5CKcfN4SzuyFgggrx6O8H_h3GhNyKSbfdVhSkVGZKi4,7284
325
- cribl_control_plane-0.1.0b1.dist-info/METADATA,sha256=xKgYEPVnci4W3V8TLkmV4KtMnf8iFAUQUc1fKI7TsyA,38885
326
- cribl_control_plane-0.1.0b1.dist-info/WHEEL,sha256=zp0Cn7JsFoX2ATtOhtaFYIiE2rmFAD4OcMhtUki8W3U,88
327
- cribl_control_plane-0.1.0b1.dist-info/RECORD,,
325
+ cribl_control_plane-0.1.0b2.dist-info/METADATA,sha256=w5QuXBbKaZra4HltGYPIeJCH720jumBaSVy-kqYo1Dk,38885
326
+ cribl_control_plane-0.1.0b2.dist-info/WHEEL,sha256=zp0Cn7JsFoX2ATtOhtaFYIiE2rmFAD4OcMhtUki8W3U,88
327
+ cribl_control_plane-0.1.0b2.dist-info/RECORD,,