cribl-control-plane 0.1.0b1__py3-none-any.whl → 0.1.0b2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +3 -3
- cribl_control_plane/groups_sdk.py +4 -4
- cribl_control_plane/models/__init__.py +3 -18
- cribl_control_plane/models/configgroup.py +17 -2
- cribl_control_plane/models/output.py +24 -24
- cribl_control_plane/models/outputdatabricks.py +19 -176
- {cribl_control_plane-0.1.0b1.dist-info → cribl_control_plane-0.1.0b2.dist-info}/METADATA +1 -1
- {cribl_control_plane-0.1.0b1.dist-info → cribl_control_plane-0.1.0b2.dist-info}/RECORD +9 -9
- {cribl_control_plane-0.1.0b1.dist-info → cribl_control_plane-0.1.0b2.dist-info}/WHEEL +0 -0
cribl_control_plane/_version.py
CHANGED
|
@@ -3,10 +3,10 @@
|
|
|
3
3
|
import importlib.metadata
|
|
4
4
|
|
|
5
5
|
__title__: str = "cribl-control-plane"
|
|
6
|
-
__version__: str = "0.1.
|
|
7
|
-
__openapi_doc_version__: str = "4.15.0-alpha.
|
|
6
|
+
__version__: str = "0.1.0b2"
|
|
7
|
+
__openapi_doc_version__: str = "4.15.0-alpha.1759327501989-3e9845e9"
|
|
8
8
|
__gen_version__: str = "2.716.16"
|
|
9
|
-
__user_agent__: str = "speakeasy-sdk/python 0.1.
|
|
9
|
+
__user_agent__: str = "speakeasy-sdk/python 0.1.0b2 2.716.16 4.15.0-alpha.1759327501989-3e9845e9 cribl-control-plane"
|
|
10
10
|
|
|
11
11
|
try:
|
|
12
12
|
if __package__ is not None:
|
|
@@ -224,7 +224,7 @@ class GroupsSDK(BaseSDK):
|
|
|
224
224
|
config_version: Optional[str] = None,
|
|
225
225
|
deploying_worker_count: Optional[float] = None,
|
|
226
226
|
description: Optional[str] = None,
|
|
227
|
-
estimated_ingest_rate: Optional[
|
|
227
|
+
estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
|
|
228
228
|
git: Optional[Union[models.Git, models.GitTypedDict]] = None,
|
|
229
229
|
incompatible_worker_count: Optional[float] = None,
|
|
230
230
|
inherits: Optional[str] = None,
|
|
@@ -395,7 +395,7 @@ class GroupsSDK(BaseSDK):
|
|
|
395
395
|
config_version: Optional[str] = None,
|
|
396
396
|
deploying_worker_count: Optional[float] = None,
|
|
397
397
|
description: Optional[str] = None,
|
|
398
|
-
estimated_ingest_rate: Optional[
|
|
398
|
+
estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
|
|
399
399
|
git: Optional[Union[models.Git, models.GitTypedDict]] = None,
|
|
400
400
|
incompatible_worker_count: Optional[float] = None,
|
|
401
401
|
inherits: Optional[str] = None,
|
|
@@ -757,7 +757,7 @@ class GroupsSDK(BaseSDK):
|
|
|
757
757
|
config_version: Optional[str] = None,
|
|
758
758
|
deploying_worker_count: Optional[float] = None,
|
|
759
759
|
description: Optional[str] = None,
|
|
760
|
-
estimated_ingest_rate: Optional[
|
|
760
|
+
estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
|
|
761
761
|
git: Optional[Union[models.Git, models.GitTypedDict]] = None,
|
|
762
762
|
incompatible_worker_count: Optional[float] = None,
|
|
763
763
|
inherits: Optional[str] = None,
|
|
@@ -931,7 +931,7 @@ class GroupsSDK(BaseSDK):
|
|
|
931
931
|
config_version: Optional[str] = None,
|
|
932
932
|
deploying_worker_count: Optional[float] = None,
|
|
933
933
|
description: Optional[str] = None,
|
|
934
|
-
estimated_ingest_rate: Optional[
|
|
934
|
+
estimated_ingest_rate: Optional[models.EstimatedIngestRate] = None,
|
|
935
935
|
git: Optional[Union[models.Git, models.GitTypedDict]] = None,
|
|
936
936
|
incompatible_worker_count: Optional[float] = None,
|
|
937
937
|
inherits: Optional[str] = None,
|
|
@@ -22,6 +22,7 @@ if TYPE_CHECKING:
|
|
|
22
22
|
ConfigGroup,
|
|
23
23
|
ConfigGroupType,
|
|
24
24
|
ConfigGroupTypedDict,
|
|
25
|
+
EstimatedIngestRate,
|
|
25
26
|
Git,
|
|
26
27
|
GitTypedDict,
|
|
27
28
|
)
|
|
@@ -2073,14 +2074,8 @@ if TYPE_CHECKING:
|
|
|
2073
2074
|
OutputDatabricks,
|
|
2074
2075
|
OutputDatabricksAuthenticationMethod,
|
|
2075
2076
|
OutputDatabricksBackpressureBehavior,
|
|
2076
|
-
OutputDatabricksCompression,
|
|
2077
|
-
OutputDatabricksCompressionLevel,
|
|
2078
2077
|
OutputDatabricksDataFormat,
|
|
2079
|
-
OutputDatabricksDataPageVersion,
|
|
2080
2078
|
OutputDatabricksDiskSpaceProtection,
|
|
2081
|
-
OutputDatabricksKeyValueMetadatum,
|
|
2082
|
-
OutputDatabricksKeyValueMetadatumTypedDict,
|
|
2083
|
-
OutputDatabricksParquetVersion,
|
|
2084
2079
|
OutputDatabricksType,
|
|
2085
2080
|
OutputDatabricksTypedDict,
|
|
2086
2081
|
)
|
|
@@ -3470,6 +3465,7 @@ __all__ = [
|
|
|
3470
3465
|
"EndpointType",
|
|
3471
3466
|
"Error",
|
|
3472
3467
|
"ErrorTypedDict",
|
|
3468
|
+
"EstimatedIngestRate",
|
|
3473
3469
|
"EventFormat",
|
|
3474
3470
|
"Executor",
|
|
3475
3471
|
"ExecutorSpecificSettings",
|
|
@@ -5131,14 +5127,8 @@ __all__ = [
|
|
|
5131
5127
|
"OutputDatabricks",
|
|
5132
5128
|
"OutputDatabricksAuthenticationMethod",
|
|
5133
5129
|
"OutputDatabricksBackpressureBehavior",
|
|
5134
|
-
"OutputDatabricksCompression",
|
|
5135
|
-
"OutputDatabricksCompressionLevel",
|
|
5136
5130
|
"OutputDatabricksDataFormat",
|
|
5137
|
-
"OutputDatabricksDataPageVersion",
|
|
5138
5131
|
"OutputDatabricksDiskSpaceProtection",
|
|
5139
|
-
"OutputDatabricksKeyValueMetadatum",
|
|
5140
|
-
"OutputDatabricksKeyValueMetadatumTypedDict",
|
|
5141
|
-
"OutputDatabricksParquetVersion",
|
|
5142
5132
|
"OutputDatabricksType",
|
|
5143
5133
|
"OutputDatabricksTypedDict",
|
|
5144
5134
|
"OutputDatadog",
|
|
@@ -6246,6 +6236,7 @@ _dynamic_imports: dict[str, str] = {
|
|
|
6246
6236
|
"ConfigGroup": ".configgroup",
|
|
6247
6237
|
"ConfigGroupType": ".configgroup",
|
|
6248
6238
|
"ConfigGroupTypedDict": ".configgroup",
|
|
6239
|
+
"EstimatedIngestRate": ".configgroup",
|
|
6249
6240
|
"Git": ".configgroup",
|
|
6250
6241
|
"GitTypedDict": ".configgroup",
|
|
6251
6242
|
"ConfigGroupCloud": ".configgroupcloud",
|
|
@@ -8078,14 +8069,8 @@ _dynamic_imports: dict[str, str] = {
|
|
|
8078
8069
|
"OutputDatabricks": ".outputdatabricks",
|
|
8079
8070
|
"OutputDatabricksAuthenticationMethod": ".outputdatabricks",
|
|
8080
8071
|
"OutputDatabricksBackpressureBehavior": ".outputdatabricks",
|
|
8081
|
-
"OutputDatabricksCompression": ".outputdatabricks",
|
|
8082
|
-
"OutputDatabricksCompressionLevel": ".outputdatabricks",
|
|
8083
8072
|
"OutputDatabricksDataFormat": ".outputdatabricks",
|
|
8084
|
-
"OutputDatabricksDataPageVersion": ".outputdatabricks",
|
|
8085
8073
|
"OutputDatabricksDiskSpaceProtection": ".outputdatabricks",
|
|
8086
|
-
"OutputDatabricksKeyValueMetadatum": ".outputdatabricks",
|
|
8087
|
-
"OutputDatabricksKeyValueMetadatumTypedDict": ".outputdatabricks",
|
|
8088
|
-
"OutputDatabricksParquetVersion": ".outputdatabricks",
|
|
8089
8074
|
"OutputDatabricksType": ".outputdatabricks",
|
|
8090
8075
|
"OutputDatabricksTypedDict": ".outputdatabricks",
|
|
8091
8076
|
"DatadogSite": ".outputdatadog",
|
|
@@ -14,6 +14,18 @@ from typing import List, Optional
|
|
|
14
14
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
15
15
|
|
|
16
16
|
|
|
17
|
+
class EstimatedIngestRate(int, Enum, metaclass=utils.OpenEnumMeta):
|
|
18
|
+
ONE_THOUSAND_AND_TWENTY_FOUR = 1024
|
|
19
|
+
FOUR_THOUSAND_AND_NINETY_SIX = 4096
|
|
20
|
+
TEN_THOUSAND_TWO_HUNDRED_AND_FORTY = 10240
|
|
21
|
+
TWO_THOUSAND_AND_FORTY_EIGHT = 2048
|
|
22
|
+
THREE_THOUSAND_AND_SEVENTY_TWO = 3072
|
|
23
|
+
FIVE_THOUSAND_ONE_HUNDRED_AND_TWENTY = 5120
|
|
24
|
+
SEVEN_THOUSAND_ONE_HUNDRED_AND_SIXTY_EIGHT = 7168
|
|
25
|
+
THIRTEEN_THOUSAND_THREE_HUNDRED_AND_TWELVE = 13312
|
|
26
|
+
FIFTEEN_THOUSAND_THREE_HUNDRED_AND_SIXTY = 15360
|
|
27
|
+
|
|
28
|
+
|
|
17
29
|
class GitTypedDict(TypedDict):
|
|
18
30
|
commit: NotRequired[str]
|
|
19
31
|
local_changes: NotRequired[float]
|
|
@@ -40,7 +52,7 @@ class ConfigGroupTypedDict(TypedDict):
|
|
|
40
52
|
config_version: NotRequired[str]
|
|
41
53
|
deploying_worker_count: NotRequired[float]
|
|
42
54
|
description: NotRequired[str]
|
|
43
|
-
estimated_ingest_rate: NotRequired[
|
|
55
|
+
estimated_ingest_rate: NotRequired[EstimatedIngestRate]
|
|
44
56
|
git: NotRequired[GitTypedDict]
|
|
45
57
|
incompatible_worker_count: NotRequired[float]
|
|
46
58
|
inherits: NotRequired[str]
|
|
@@ -75,7 +87,10 @@ class ConfigGroup(BaseModel):
|
|
|
75
87
|
description: Optional[str] = None
|
|
76
88
|
|
|
77
89
|
estimated_ingest_rate: Annotated[
|
|
78
|
-
|
|
90
|
+
Annotated[
|
|
91
|
+
Optional[EstimatedIngestRate], PlainValidator(validate_open_enum(True))
|
|
92
|
+
],
|
|
93
|
+
pydantic.Field(alias="estimatedIngestRate"),
|
|
79
94
|
] = None
|
|
80
95
|
|
|
81
96
|
git: Optional[Git] = None
|
|
@@ -91,31 +91,32 @@ OutputTypedDict = TypeAliasType(
|
|
|
91
91
|
OutputDevnullTypedDict,
|
|
92
92
|
OutputDefaultTypedDict,
|
|
93
93
|
OutputRouterTypedDict,
|
|
94
|
-
OutputSnmpTypedDict,
|
|
95
94
|
OutputNetflowTypedDict,
|
|
95
|
+
OutputSnmpTypedDict,
|
|
96
96
|
OutputDiskSpoolTypedDict,
|
|
97
97
|
OutputRingTypedDict,
|
|
98
|
-
OutputStatsdExtTypedDict,
|
|
99
98
|
OutputGraphiteTypedDict,
|
|
99
|
+
OutputStatsdExtTypedDict,
|
|
100
100
|
OutputStatsdTypedDict,
|
|
101
101
|
OutputGooglePubsubTypedDict,
|
|
102
|
-
OutputCriblTCPTypedDict,
|
|
103
102
|
OutputSplunkTypedDict,
|
|
103
|
+
OutputCriblTCPTypedDict,
|
|
104
104
|
OutputSnsTypedDict,
|
|
105
105
|
OutputCloudwatchTypedDict,
|
|
106
106
|
OutputAzureEventhubTypedDict,
|
|
107
|
-
OutputWavefrontTypedDict,
|
|
108
107
|
OutputSignalfxTypedDict,
|
|
108
|
+
OutputWavefrontTypedDict,
|
|
109
109
|
OutputHoneycombTypedDict,
|
|
110
|
-
OutputSumoLogicTypedDict,
|
|
111
|
-
OutputCrowdstrikeNextGenSiemTypedDict,
|
|
112
110
|
OutputHumioHecTypedDict,
|
|
113
111
|
OutputTcpjsonTypedDict,
|
|
112
|
+
OutputSumoLogicTypedDict,
|
|
113
|
+
OutputCrowdstrikeNextGenSiemTypedDict,
|
|
114
114
|
OutputElasticCloudTypedDict,
|
|
115
|
-
OutputKinesisTypedDict,
|
|
116
|
-
OutputConfluentCloudTypedDict,
|
|
117
|
-
OutputKafkaTypedDict,
|
|
118
115
|
OutputExabeamTypedDict,
|
|
116
|
+
OutputKafkaTypedDict,
|
|
117
|
+
OutputConfluentCloudTypedDict,
|
|
118
|
+
OutputKinesisTypedDict,
|
|
119
|
+
OutputDatabricksTypedDict,
|
|
119
120
|
OutputNewrelicEventsTypedDict,
|
|
120
121
|
OutputAzureLogsTypedDict,
|
|
121
122
|
OutputSplunkLbTypedDict,
|
|
@@ -131,14 +132,13 @@ OutputTypedDict = TypeAliasType(
|
|
|
131
132
|
OutputDynatraceHTTPTypedDict,
|
|
132
133
|
OutputServiceNowTypedDict,
|
|
133
134
|
OutputDynatraceOtlpTypedDict,
|
|
134
|
-
OutputElasticTypedDict,
|
|
135
135
|
OutputGoogleChronicleTypedDict,
|
|
136
|
-
|
|
136
|
+
OutputElasticTypedDict,
|
|
137
137
|
OutputDatadogTypedDict,
|
|
138
|
+
OutputCriblLakeTypedDict,
|
|
138
139
|
OutputPrometheusTypedDict,
|
|
139
140
|
OutputMskTypedDict,
|
|
140
141
|
OutputSentinelOneAiSiemTypedDict,
|
|
141
|
-
OutputDatabricksTypedDict,
|
|
142
142
|
OutputSentinelTypedDict,
|
|
143
143
|
OutputInfluxdbTypedDict,
|
|
144
144
|
OutputGoogleCloudStorageTypedDict,
|
|
@@ -163,31 +163,32 @@ Output = TypeAliasType(
|
|
|
163
163
|
OutputDevnull,
|
|
164
164
|
OutputDefault,
|
|
165
165
|
OutputRouter,
|
|
166
|
-
OutputSnmp,
|
|
167
166
|
OutputNetflow,
|
|
167
|
+
OutputSnmp,
|
|
168
168
|
OutputDiskSpool,
|
|
169
169
|
OutputRing,
|
|
170
|
-
OutputStatsdExt,
|
|
171
170
|
OutputGraphite,
|
|
171
|
+
OutputStatsdExt,
|
|
172
172
|
OutputStatsd,
|
|
173
173
|
OutputGooglePubsub,
|
|
174
|
-
OutputCriblTCP,
|
|
175
174
|
OutputSplunk,
|
|
175
|
+
OutputCriblTCP,
|
|
176
176
|
OutputSns,
|
|
177
177
|
OutputCloudwatch,
|
|
178
178
|
OutputAzureEventhub,
|
|
179
|
-
OutputWavefront,
|
|
180
179
|
OutputSignalfx,
|
|
180
|
+
OutputWavefront,
|
|
181
181
|
OutputHoneycomb,
|
|
182
|
-
OutputSumoLogic,
|
|
183
|
-
OutputCrowdstrikeNextGenSiem,
|
|
184
182
|
OutputHumioHec,
|
|
185
183
|
OutputTcpjson,
|
|
184
|
+
OutputSumoLogic,
|
|
185
|
+
OutputCrowdstrikeNextGenSiem,
|
|
186
186
|
OutputElasticCloud,
|
|
187
|
-
OutputKinesis,
|
|
188
|
-
OutputConfluentCloud,
|
|
189
|
-
OutputKafka,
|
|
190
187
|
OutputExabeam,
|
|
188
|
+
OutputKafka,
|
|
189
|
+
OutputConfluentCloud,
|
|
190
|
+
OutputKinesis,
|
|
191
|
+
OutputDatabricks,
|
|
191
192
|
OutputNewrelicEvents,
|
|
192
193
|
OutputAzureLogs,
|
|
193
194
|
OutputSplunkLb,
|
|
@@ -203,14 +204,13 @@ Output = TypeAliasType(
|
|
|
203
204
|
OutputDynatraceHTTP,
|
|
204
205
|
OutputServiceNow,
|
|
205
206
|
OutputDynatraceOtlp,
|
|
206
|
-
OutputElastic,
|
|
207
207
|
OutputGoogleChronicle,
|
|
208
|
-
|
|
208
|
+
OutputElastic,
|
|
209
209
|
OutputDatadog,
|
|
210
|
+
OutputCriblLake,
|
|
210
211
|
OutputPrometheus,
|
|
211
212
|
OutputMsk,
|
|
212
213
|
OutputSentinelOneAiSiem,
|
|
213
|
-
OutputDatabricks,
|
|
214
214
|
OutputSentinel,
|
|
215
215
|
OutputInfluxdb,
|
|
216
216
|
OutputGoogleCloudStorage,
|
|
@@ -44,49 +44,12 @@ class OutputDatabricksAuthenticationMethod(str, Enum, metaclass=utils.OpenEnumMe
|
|
|
44
44
|
SECRET = "secret"
|
|
45
45
|
|
|
46
46
|
|
|
47
|
-
class OutputDatabricksCompression(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
48
|
-
r"""Data compression format to apply to HTTP content before it is delivered"""
|
|
49
|
-
|
|
50
|
-
NONE = "none"
|
|
51
|
-
GZIP = "gzip"
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
class OutputDatabricksCompressionLevel(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
55
|
-
r"""Compression level to apply before moving files to final destination"""
|
|
56
|
-
|
|
57
|
-
BEST_SPEED = "best_speed"
|
|
58
|
-
NORMAL = "normal"
|
|
59
|
-
BEST_COMPRESSION = "best_compression"
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
class OutputDatabricksParquetVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
63
|
-
r"""Determines which data types are supported and how they are represented"""
|
|
64
|
-
|
|
65
|
-
PARQUET_1_0 = "PARQUET_1_0"
|
|
66
|
-
PARQUET_2_4 = "PARQUET_2_4"
|
|
67
|
-
PARQUET_2_6 = "PARQUET_2_6"
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
class OutputDatabricksDataPageVersion(str, Enum, metaclass=utils.OpenEnumMeta):
|
|
71
|
-
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
72
|
-
|
|
73
|
-
DATA_PAGE_V1 = "DATA_PAGE_V1"
|
|
74
|
-
DATA_PAGE_V2 = "DATA_PAGE_V2"
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
class OutputDatabricksKeyValueMetadatumTypedDict(TypedDict):
|
|
78
|
-
value: str
|
|
79
|
-
key: NotRequired[str]
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
class OutputDatabricksKeyValueMetadatum(BaseModel):
|
|
83
|
-
value: str
|
|
84
|
-
|
|
85
|
-
key: Optional[str] = ""
|
|
86
|
-
|
|
87
|
-
|
|
88
47
|
class OutputDatabricksTypedDict(TypedDict):
|
|
89
48
|
type: OutputDatabricksType
|
|
49
|
+
login_url: str
|
|
50
|
+
r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
|
|
51
|
+
client_id: str
|
|
52
|
+
r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
|
|
90
53
|
id: NotRequired[str]
|
|
91
54
|
r"""Unique ID for this output"""
|
|
92
55
|
pipeline: NotRequired[str]
|
|
@@ -133,14 +96,6 @@ class OutputDatabricksTypedDict(TypedDict):
|
|
|
133
96
|
r"""How to handle events when disk space is below the global 'Min free disk space' limit"""
|
|
134
97
|
unity_auth_method: NotRequired[OutputDatabricksAuthenticationMethod]
|
|
135
98
|
r"""Unity Catalog authentication method. Choose Manual to enter credentials directly, or Secret to use a stored secret."""
|
|
136
|
-
login_url: NotRequired[str]
|
|
137
|
-
r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
|
|
138
|
-
client_id: NotRequired[str]
|
|
139
|
-
r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
|
|
140
|
-
client_secret: NotRequired[str]
|
|
141
|
-
r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
|
|
142
|
-
client_text_secret: NotRequired[str]
|
|
143
|
-
r"""Select or create a stored secret that references your Client ID and Client Secret"""
|
|
144
99
|
scope: NotRequired[str]
|
|
145
100
|
r"""OAuth scope for Unity Catalog authentication"""
|
|
146
101
|
token_timeout_secs: NotRequired[float]
|
|
@@ -154,41 +109,21 @@ class OutputDatabricksTypedDict(TypedDict):
|
|
|
154
109
|
over_write_files: NotRequired[bool]
|
|
155
110
|
r"""Uploaded files should be overwritten if they already exist. If disabled, upload will fail if a file already exists."""
|
|
156
111
|
description: NotRequired[str]
|
|
157
|
-
|
|
158
|
-
r"""
|
|
159
|
-
|
|
160
|
-
r"""
|
|
161
|
-
automatic_schema: NotRequired[bool]
|
|
162
|
-
r"""Automatically calculate the schema based on the events of each Parquet file generated"""
|
|
163
|
-
parquet_version: NotRequired[OutputDatabricksParquetVersion]
|
|
164
|
-
r"""Determines which data types are supported and how they are represented"""
|
|
165
|
-
parquet_data_page_version: NotRequired[OutputDatabricksDataPageVersion]
|
|
166
|
-
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
167
|
-
parquet_row_group_length: NotRequired[float]
|
|
168
|
-
r"""The number of rows that every group will contain. The final group can contain a smaller number of rows."""
|
|
169
|
-
parquet_page_size: NotRequired[str]
|
|
170
|
-
r"""Target memory size for page segments, such as 1MB or 128MB. Generally, lower values improve reading speed, while higher values improve compression."""
|
|
171
|
-
should_log_invalid_rows: NotRequired[bool]
|
|
172
|
-
r"""Log up to 3 rows that @{product} skips due to data mismatch"""
|
|
173
|
-
key_value_metadata: NotRequired[List[OutputDatabricksKeyValueMetadatumTypedDict]]
|
|
174
|
-
r"""The metadata of files the Destination writes will include the properties you add here as key-value pairs. Useful for tagging. Examples: \"key\":\"OCSF Event Class\", \"value\":\"9001\" """
|
|
175
|
-
enable_statistics: NotRequired[bool]
|
|
176
|
-
r"""Statistics profile an entire file in terms of minimum/maximum values within data, numbers of nulls, etc. You can use Parquet tools to view statistics."""
|
|
177
|
-
enable_write_page_index: NotRequired[bool]
|
|
178
|
-
r"""One page index contains statistics for one data page. Parquet readers use statistics to enable page skipping."""
|
|
179
|
-
enable_page_checksum: NotRequired[bool]
|
|
180
|
-
r"""Parquet tools can use the checksum of a Parquet page to verify data integrity"""
|
|
181
|
-
empty_dir_cleanup_sec: NotRequired[float]
|
|
182
|
-
r"""How frequently, in seconds, to clean up empty directories"""
|
|
183
|
-
deadletter_path: NotRequired[str]
|
|
184
|
-
r"""Storage location for files that fail to reach their final destination after maximum retries are exceeded"""
|
|
185
|
-
max_retry_num: NotRequired[float]
|
|
186
|
-
r"""The maximum number of times a file will attempt to move to its final destination before being dead-lettered"""
|
|
112
|
+
client_secret: NotRequired[str]
|
|
113
|
+
r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
|
|
114
|
+
client_text_secret: NotRequired[str]
|
|
115
|
+
r"""Select or create a stored text secret"""
|
|
187
116
|
|
|
188
117
|
|
|
189
118
|
class OutputDatabricks(BaseModel):
|
|
190
119
|
type: OutputDatabricksType
|
|
191
120
|
|
|
121
|
+
login_url: Annotated[str, pydantic.Field(alias="loginUrl")]
|
|
122
|
+
r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
|
|
123
|
+
|
|
124
|
+
client_id: Annotated[str, pydantic.Field(alias="clientId")]
|
|
125
|
+
r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
|
|
126
|
+
|
|
192
127
|
id: Optional[str] = None
|
|
193
128
|
r"""Unique ID for this output"""
|
|
194
129
|
|
|
@@ -308,20 +243,6 @@ class OutputDatabricks(BaseModel):
|
|
|
308
243
|
] = OutputDatabricksAuthenticationMethod.MANUAL
|
|
309
244
|
r"""Unity Catalog authentication method. Choose Manual to enter credentials directly, or Secret to use a stored secret."""
|
|
310
245
|
|
|
311
|
-
login_url: Annotated[Optional[str], pydantic.Field(alias="loginUrl")] = None
|
|
312
|
-
r"""URL for Unity Catalog OAuth token endpoint (example: 'https://your-workspace.cloud.databricks.com/oauth/token')"""
|
|
313
|
-
|
|
314
|
-
client_id: Annotated[Optional[str], pydantic.Field(alias="clientId")] = None
|
|
315
|
-
r"""JavaScript expression to compute the OAuth client ID for Unity Catalog authentication. Can be a constant."""
|
|
316
|
-
|
|
317
|
-
client_secret: Annotated[Optional[str], pydantic.Field(alias="clientSecret")] = None
|
|
318
|
-
r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
|
|
319
|
-
|
|
320
|
-
client_text_secret: Annotated[
|
|
321
|
-
Optional[str], pydantic.Field(alias="clientTextSecret")
|
|
322
|
-
] = None
|
|
323
|
-
r"""Select or create a stored secret that references your Client ID and Client Secret"""
|
|
324
|
-
|
|
325
246
|
scope: Optional[str] = "all-apis"
|
|
326
247
|
r"""OAuth scope for Unity Catalog authentication"""
|
|
327
248
|
|
|
@@ -352,88 +273,10 @@ class OutputDatabricks(BaseModel):
|
|
|
352
273
|
|
|
353
274
|
description: Optional[str] = None
|
|
354
275
|
|
|
355
|
-
|
|
356
|
-
|
|
357
|
-
] = OutputDatabricksCompression.GZIP
|
|
358
|
-
r"""Data compression format to apply to HTTP content before it is delivered"""
|
|
359
|
-
|
|
360
|
-
compression_level: Annotated[
|
|
361
|
-
Annotated[
|
|
362
|
-
Optional[OutputDatabricksCompressionLevel],
|
|
363
|
-
PlainValidator(validate_open_enum(False)),
|
|
364
|
-
],
|
|
365
|
-
pydantic.Field(alias="compressionLevel"),
|
|
366
|
-
] = OutputDatabricksCompressionLevel.BEST_SPEED
|
|
367
|
-
r"""Compression level to apply before moving files to final destination"""
|
|
368
|
-
|
|
369
|
-
automatic_schema: Annotated[
|
|
370
|
-
Optional[bool], pydantic.Field(alias="automaticSchema")
|
|
371
|
-
] = False
|
|
372
|
-
r"""Automatically calculate the schema based on the events of each Parquet file generated"""
|
|
373
|
-
|
|
374
|
-
parquet_version: Annotated[
|
|
375
|
-
Annotated[
|
|
376
|
-
Optional[OutputDatabricksParquetVersion],
|
|
377
|
-
PlainValidator(validate_open_enum(False)),
|
|
378
|
-
],
|
|
379
|
-
pydantic.Field(alias="parquetVersion"),
|
|
380
|
-
] = OutputDatabricksParquetVersion.PARQUET_2_6
|
|
381
|
-
r"""Determines which data types are supported and how they are represented"""
|
|
382
|
-
|
|
383
|
-
parquet_data_page_version: Annotated[
|
|
384
|
-
Annotated[
|
|
385
|
-
Optional[OutputDatabricksDataPageVersion],
|
|
386
|
-
PlainValidator(validate_open_enum(False)),
|
|
387
|
-
],
|
|
388
|
-
pydantic.Field(alias="parquetDataPageVersion"),
|
|
389
|
-
] = OutputDatabricksDataPageVersion.DATA_PAGE_V2
|
|
390
|
-
r"""Serialization format of data pages. Note that some reader implementations use Data page V2's attributes to work more efficiently, while others ignore it."""
|
|
391
|
-
|
|
392
|
-
parquet_row_group_length: Annotated[
|
|
393
|
-
Optional[float], pydantic.Field(alias="parquetRowGroupLength")
|
|
394
|
-
] = 10000
|
|
395
|
-
r"""The number of rows that every group will contain. The final group can contain a smaller number of rows."""
|
|
396
|
-
|
|
397
|
-
parquet_page_size: Annotated[
|
|
398
|
-
Optional[str], pydantic.Field(alias="parquetPageSize")
|
|
399
|
-
] = "1MB"
|
|
400
|
-
r"""Target memory size for page segments, such as 1MB or 128MB. Generally, lower values improve reading speed, while higher values improve compression."""
|
|
401
|
-
|
|
402
|
-
should_log_invalid_rows: Annotated[
|
|
403
|
-
Optional[bool], pydantic.Field(alias="shouldLogInvalidRows")
|
|
404
|
-
] = None
|
|
405
|
-
r"""Log up to 3 rows that @{product} skips due to data mismatch"""
|
|
276
|
+
client_secret: Annotated[Optional[str], pydantic.Field(alias="clientSecret")] = None
|
|
277
|
+
r"""JavaScript expression to compute the OAuth client secret for Unity Catalog authentication. Can be a constant."""
|
|
406
278
|
|
|
407
|
-
|
|
408
|
-
Optional[
|
|
409
|
-
pydantic.Field(alias="keyValueMetadata"),
|
|
279
|
+
client_text_secret: Annotated[
|
|
280
|
+
Optional[str], pydantic.Field(alias="clientTextSecret")
|
|
410
281
|
] = None
|
|
411
|
-
r"""
|
|
412
|
-
|
|
413
|
-
enable_statistics: Annotated[
|
|
414
|
-
Optional[bool], pydantic.Field(alias="enableStatistics")
|
|
415
|
-
] = True
|
|
416
|
-
r"""Statistics profile an entire file in terms of minimum/maximum values within data, numbers of nulls, etc. You can use Parquet tools to view statistics."""
|
|
417
|
-
|
|
418
|
-
enable_write_page_index: Annotated[
|
|
419
|
-
Optional[bool], pydantic.Field(alias="enableWritePageIndex")
|
|
420
|
-
] = True
|
|
421
|
-
r"""One page index contains statistics for one data page. Parquet readers use statistics to enable page skipping."""
|
|
422
|
-
|
|
423
|
-
enable_page_checksum: Annotated[
|
|
424
|
-
Optional[bool], pydantic.Field(alias="enablePageChecksum")
|
|
425
|
-
] = False
|
|
426
|
-
r"""Parquet tools can use the checksum of a Parquet page to verify data integrity"""
|
|
427
|
-
|
|
428
|
-
empty_dir_cleanup_sec: Annotated[
|
|
429
|
-
Optional[float], pydantic.Field(alias="emptyDirCleanupSec")
|
|
430
|
-
] = 300
|
|
431
|
-
r"""How frequently, in seconds, to clean up empty directories"""
|
|
432
|
-
|
|
433
|
-
deadletter_path: Annotated[
|
|
434
|
-
Optional[str], pydantic.Field(alias="deadletterPath")
|
|
435
|
-
] = "$CRIBL_HOME/state/outputs/dead-letter"
|
|
436
|
-
r"""Storage location for files that fail to reach their final destination after maximum retries are exceeded"""
|
|
437
|
-
|
|
438
|
-
max_retry_num: Annotated[Optional[float], pydantic.Field(alias="maxRetryNum")] = 20
|
|
439
|
-
r"""The maximum number of times a file will attempt to move to its final destination before being dead-lettered"""
|
|
282
|
+
r"""Select or create a stored text secret"""
|
|
@@ -4,7 +4,7 @@ cribl_control_plane/_hooks/clientcredentials.py,sha256=p1WN7LL3PHrAf4AxXrsOZF_NB
|
|
|
4
4
|
cribl_control_plane/_hooks/registration.py,sha256=1QZB41w6If7I9dXiOSQx6dhSc6BPWrnI5Q5bMOr4iVA,624
|
|
5
5
|
cribl_control_plane/_hooks/sdkhooks.py,sha256=ggXjME1_Rdv8CVCg1XHqB83eYtbxzKyhXyfQ36Yc1gA,2816
|
|
6
6
|
cribl_control_plane/_hooks/types.py,sha256=Tw_C4zTZm01rW_89VDEUpvQ8KQr1WxN0Gu_-s_fYSPc,2998
|
|
7
|
-
cribl_control_plane/_version.py,sha256=
|
|
7
|
+
cribl_control_plane/_version.py,sha256=BXnoleFws7YHijH4Jb3Mnb7Ffa-2mD7o8h-uXu7BnZs,546
|
|
8
8
|
cribl_control_plane/acl.py,sha256=8lvYOKAli4PzsQhOVaCU6YCwblPMh9jQo04L0r4HJuQ,9025
|
|
9
9
|
cribl_control_plane/auth_sdk.py,sha256=3sjf1VoyWwfhSyuMDQLixgWISSf03BOZwmkiT8g5Ruw,626
|
|
10
10
|
cribl_control_plane/basesdk.py,sha256=y4yIXSNVXLMd0sLS2htBFdTCI3gkPQbIWd-C671kg1I,12249
|
|
@@ -22,12 +22,12 @@ cribl_control_plane/errors/healthstatus_error.py,sha256=hakt3RO4e4mWnmiDy1rzBWCZ
|
|
|
22
22
|
cribl_control_plane/errors/no_response_error.py,sha256=Kb7hmMtDo72KrLSjUEDNeQxvzZiVxUjOZym8TPdZp5Y,462
|
|
23
23
|
cribl_control_plane/errors/responsevalidationerror.py,sha256=7vHiNVp0rm6nrult0NCC73VDabehUhM861LrWfaObvA,778
|
|
24
24
|
cribl_control_plane/groups_configs.py,sha256=dgi-W0ElnyygaVKXqk5df2ldAAgj9YmXRPCez2hP7yc,695
|
|
25
|
-
cribl_control_plane/groups_sdk.py,sha256=
|
|
25
|
+
cribl_control_plane/groups_sdk.py,sha256=pZy0owBWP-RFNp9CwA2fopDZSceUX1CFlAtaTsFctDc,61919
|
|
26
26
|
cribl_control_plane/health.py,sha256=N8pX8RHkJVtLFd4nZ8ypJPrzT_JezciEVry9s9qvCRc,7019
|
|
27
27
|
cribl_control_plane/hectokens.py,sha256=0EGgGGrM83m1YmTZwkN5S4xFkHQGnw1IZe3y6uMwmLw,19151
|
|
28
28
|
cribl_control_plane/httpclient.py,sha256=Eu73urOAiZQtdUIyOUnPccxCiBbWEKrXG-JrRG3SLM4,3946
|
|
29
29
|
cribl_control_plane/lakedatasets.py,sha256=7WYWcjXMzliDW1j3TQlgikc_h54IUq4lsysVy_39l38,46578
|
|
30
|
-
cribl_control_plane/models/__init__.py,sha256=
|
|
30
|
+
cribl_control_plane/models/__init__.py,sha256=8x_C90VeiA7FF5Z14TmOwCjYLHWQ-syjaWXzOJTS19M,387831
|
|
31
31
|
cribl_control_plane/models/addhectokenrequest.py,sha256=mzQLKrMWlwxNheqEs5SM_yrT-gyenfCWgHKhmb5oXFQ,800
|
|
32
32
|
cribl_control_plane/models/authtoken.py,sha256=uW0aIs8j14CQzFM2ueY5GIWFulna91cigBWQ3oPlDgY,295
|
|
33
33
|
cribl_control_plane/models/branchinfo.py,sha256=jCX31O5TMG9jTjqigPvvUiBwpgPpVxHtSuhYrNykXiI,291
|
|
@@ -35,7 +35,7 @@ cribl_control_plane/models/cacheconnection.py,sha256=IaqcKQhOxuY_SYdMpD2FqBGMbra
|
|
|
35
35
|
cribl_control_plane/models/cacheconnectionbackfillstatus.py,sha256=EFJtxJ8EULR0JI4SjDoIsYSxx7uGHk-ULOeSxSzTCSc,380
|
|
36
36
|
cribl_control_plane/models/cloudprovider.py,sha256=OwlC4oXKrv5AjIgotkcSSTGgods-2QWeR0N5UgJ20Yw,268
|
|
37
37
|
cribl_control_plane/models/commit.py,sha256=wXQkjOYsffxWURHTrfU9kJ4HF2H65QfD1R9-vH0clvQ,641
|
|
38
|
-
cribl_control_plane/models/configgroup.py,sha256=
|
|
38
|
+
cribl_control_plane/models/configgroup.py,sha256=5t0IF-5BVrtJLM8XPBBrfriY9RXP9TxCzK2m46mDWiA,4302
|
|
39
39
|
cribl_control_plane/models/configgroupcloud.py,sha256=xme7fTZesBGwpma0huL-EXyPNaDYhP2CeYnvxBRxwU8,1589
|
|
40
40
|
cribl_control_plane/models/configgrouplookups.py,sha256=1z1DlvlVehqfD6hZMXG0XedZTfoCIpYd0cHav45tiRw,830
|
|
41
41
|
cribl_control_plane/models/createconfiggroupbyproductop.py,sha256=OZtUSrA4Lf81IhqH-PnO5nqbcWJ_plFLJJbe-vIBsSI,1728
|
|
@@ -188,7 +188,7 @@ cribl_control_plane/models/nodeprovidedinfo.py,sha256=XElCf2bdE4ahiO7MaVmbtGziKg
|
|
|
188
188
|
cribl_control_plane/models/nodeskippedupgradestatus.py,sha256=EY-U3cUPwMa3H-X-hn5gdaEBmSAP3hB9gRPdiQZs5yU,294
|
|
189
189
|
cribl_control_plane/models/nodeupgradestate.py,sha256=EerzMMQeFl-iHKHsJwEIxRroH6w97S7-em9YoY2-ASk,286
|
|
190
190
|
cribl_control_plane/models/nodeupgradestatus.py,sha256=Ygdb7jTFOvD6M3Fjl3brliLCKbkdX3aCwkPYjTE4Dw0,1346
|
|
191
|
-
cribl_control_plane/models/output.py,sha256=
|
|
191
|
+
cribl_control_plane/models/output.py,sha256=BN8ePTnt10hGAIyZ_WmiUBHdsQ6EMHE0_Tu9pOhiuRA,8830
|
|
192
192
|
cribl_control_plane/models/outputazureblob.py,sha256=l3N9XWrzsjgpoV8AZ59lmHXL7ZNrGek1VyO2bA13DKQ,23132
|
|
193
193
|
cribl_control_plane/models/outputazuredataexplorer.py,sha256=dHMj1qBpoNwusSety4OG3KgaKKNixTaqcdKSsxDAo2Q,31977
|
|
194
194
|
cribl_control_plane/models/outputazureeventhub.py,sha256=mccmtlZjwoeyr3AagtPRleOpZKB6roJvwekRjScqa8w,15710
|
|
@@ -200,7 +200,7 @@ cribl_control_plane/models/outputcriblhttp.py,sha256=qH_DLlLfUDMTgyjLJ-EyMytmHAJ
|
|
|
200
200
|
cribl_control_plane/models/outputcribllake.py,sha256=h-J1uPLPV3teRWJIyrJLV1oWjxiWlQAiwDoV3rNX-Ks,17914
|
|
201
201
|
cribl_control_plane/models/outputcribltcp.py,sha256=wpiSsXcSBFXe2ZO0muijlA58TFVgL1Fw7FTt_zodf_8,17482
|
|
202
202
|
cribl_control_plane/models/outputcrowdstrikenextgensiem.py,sha256=mFJowiqrVy6peiPXOcLZTyGtxGTLLuIpoG56gw81dMY,19864
|
|
203
|
-
cribl_control_plane/models/outputdatabricks.py,sha256=
|
|
203
|
+
cribl_control_plane/models/outputdatabricks.py,sha256=dmxj_K4nbO8nHAcWhOGZJIqs6KWlo1Q1RpO8Twez3WM,13934
|
|
204
204
|
cribl_control_plane/models/outputdatadog.py,sha256=KUFjLBY9FtfzSuS6XiWsvru_HmAgnSfeQFw_J-T3OVk,23291
|
|
205
205
|
cribl_control_plane/models/outputdataset.py,sha256=-iXA-e7awghBU6Plu-774F1P2TQN2w7oZX-JexG7hS4,21606
|
|
206
206
|
cribl_control_plane/models/outputdefault.py,sha256=2tjMKYSksR-0qWLd_u3PPLXL0gZiSlUdj9JTPYeYMps,1952
|
|
@@ -322,6 +322,6 @@ cribl_control_plane/utils/url.py,sha256=BgGPgcTA6MRK4bF8fjP2dUopN3NzEzxWMXPBVg8N
|
|
|
322
322
|
cribl_control_plane/utils/values.py,sha256=CcaCXEa3xHhkUDROyXZocN8f0bdITftv9Y0P9lTf0YM,3517
|
|
323
323
|
cribl_control_plane/versions.py,sha256=4xdTYbM84Xyjr5qkixqNpgn2q6V8aXVYXkEPDU2Ele0,1156
|
|
324
324
|
cribl_control_plane/versions_configs.py,sha256=5CKcfN4SzuyFgggrx6O8H_h3GhNyKSbfdVhSkVGZKi4,7284
|
|
325
|
-
cribl_control_plane-0.1.
|
|
326
|
-
cribl_control_plane-0.1.
|
|
327
|
-
cribl_control_plane-0.1.
|
|
325
|
+
cribl_control_plane-0.1.0b2.dist-info/METADATA,sha256=w5QuXBbKaZra4HltGYPIeJCH720jumBaSVy-kqYo1Dk,38885
|
|
326
|
+
cribl_control_plane-0.1.0b2.dist-info/WHEEL,sha256=zp0Cn7JsFoX2ATtOhtaFYIiE2rmFAD4OcMhtUki8W3U,88
|
|
327
|
+
cribl_control_plane-0.1.0b2.dist-info/RECORD,,
|
|
File without changes
|