yandexcloud 0.294.0__py3-none-any.whl → 0.296.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of yandexcloud might be problematic. Click here for more details.
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_pb2.py +5 -3
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_pb2.pyi +28 -1
- yandex/cloud/backup/v1/resource_pb2.py +16 -14
- yandex/cloud/backup/v1/resource_pb2.pyi +38 -1
- yandex/cloud/billing/v1/customer_service_pb2.pyi +1 -1
- yandex/cloud/billing/v1/sku_service_pb2.pyi +1 -0
- yandex/cloud/compute/v1/instance_pb2.pyi +1 -3
- yandex/cloud/compute/v1/instance_service_pb2.py +26 -26
- yandex/cloud/compute/v1/instance_service_pb2.pyi +5 -1
- yandex/cloud/datasphere/v2/jobs/jobs_pb2.py +40 -32
- yandex/cloud/datasphere/v2/jobs/jobs_pb2.pyi +78 -1
- yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.py +56 -54
- yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.pyi +21 -6
- yandex/cloud/datasphere/v2/project_service_pb2.py +34 -30
- yandex/cloud/datasphere/v2/project_service_pb2.pyi +13 -3
- yandex/cloud/iam/v1/api_key_service_pb2.py +13 -3
- yandex/cloud/iam/v1/api_key_service_pb2.pyi +53 -0
- yandex/cloud/iam/v1/api_key_service_pb2_grpc.py +34 -0
- yandex/cloud/iam/v1/api_key_service_pb2_grpc.pyi +20 -0
- yandex/cloud/iot/devices/v1/registry_pb2.py +13 -13
- yandex/cloud/iot/devices/v1/registry_pb2.pyi +4 -0
- yandex/cloud/iot/devices/v1/registry_service_pb2.py +23 -3
- yandex/cloud/iot/devices/v1/registry_service_pb2.pyi +64 -0
- yandex/cloud/iot/devices/v1/registry_service_pb2_grpc.py +68 -0
- yandex/cloud/iot/devices/v1/registry_service_pb2_grpc.pyi +40 -0
- yandex/cloud/mdb/greenplum/v1/pxf_pb2.pyi +226 -52
- yandex/cloud/mdb/kafka/v1/cluster_pb2.py +42 -35
- yandex/cloud/mdb/kafka/v1/cluster_pb2.pyi +32 -2
- yandex/cloud/mdb/mongodb/v1/cluster_service_pb2.pyi +13 -7
- yandex/cloud/mdb/postgresql/v1/user_pb2.pyi +33 -16
- yandex/cloud/organizationmanager/v1/os_login_service_pb2.py +20 -20
- yandex/cloud/organizationmanager/v1/os_login_service_pb2.pyi +2 -2
- yandex/cloud/serverless/functions/v1/function_pb2.py +30 -28
- yandex/cloud/serverless/functions/v1/function_pb2.pyi +5 -1
- yandex/cloud/serverless/functions/v1/function_service_pb2.py +38 -36
- yandex/cloud/serverless/functions/v1/function_service_pb2.pyi +5 -1
- yandex/cloud/smartcaptcha/__init__.py +0 -0
- yandex/cloud/smartcaptcha/v1/__init__.py +0 -0
- yandex/cloud/smartcaptcha/v1/captcha_pb2.py +100 -0
- yandex/cloud/smartcaptcha/v1/captcha_pb2.pyi +500 -0
- yandex/cloud/smartcaptcha/v1/captcha_pb2_grpc.py +4 -0
- yandex/cloud/smartcaptcha/v1/captcha_pb2_grpc.pyi +17 -0
- yandex/cloud/smartcaptcha/v1/captcha_service_pb2.py +76 -0
- yandex/cloud/smartcaptcha/v1/captcha_service_pb2.pyi +268 -0
- yandex/cloud/smartcaptcha/v1/captcha_service_pb2_grpc.py +242 -0
- yandex/cloud/smartcaptcha/v1/captcha_service_pb2_grpc.pyi +152 -0
- yandex/cloud/smartwebsecurity/__init__.py +0 -0
- yandex/cloud/smartwebsecurity/v1/__init__.py +0 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2.py +110 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2.pyi +520 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2_grpc.py +4 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2_grpc.pyi +17 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.py +78 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.pyi +260 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2_grpc.py +208 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2_grpc.pyi +132 -0
- yandex/cloud/speechsense/v1/analysis/__init__.py +0 -0
- yandex/cloud/speechsense/v1/analysis/conversation_statistics_pb2.py +28 -0
- yandex/cloud/speechsense/v1/analysis/conversation_statistics_pb2.pyi +40 -0
- yandex/cloud/speechsense/v1/analysis/conversation_statistics_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/conversation_statistics_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/interrupts_statistics_pb2.py +29 -0
- yandex/cloud/speechsense/v1/analysis/interrupts_statistics_pb2.pyi +66 -0
- yandex/cloud/speechsense/v1/analysis/interrupts_statistics_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/interrupts_statistics_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/points_pb2.py +29 -0
- yandex/cloud/speechsense/v1/analysis/points_pb2.pyi +53 -0
- yandex/cloud/speechsense/v1/analysis/points_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/points_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/predefined_classifiers_pb2.py +30 -0
- yandex/cloud/speechsense/v1/analysis/predefined_classifiers_pb2.pyi +93 -0
- yandex/cloud/speechsense/v1/analysis/predefined_classifiers_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/predefined_classifiers_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/silence_statistics_pb2.py +27 -0
- yandex/cloud/speechsense/v1/analysis/silence_statistics_pb2.pyi +41 -0
- yandex/cloud/speechsense/v1/analysis/silence_statistics_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/silence_statistics_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/speaker_statistics_pb2.py +28 -0
- yandex/cloud/speechsense/v1/analysis/speaker_statistics_pb2.pyi +58 -0
- yandex/cloud/speechsense/v1/analysis/speaker_statistics_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/speaker_statistics_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/speech_statistics_pb2.py +27 -0
- yandex/cloud/speechsense/v1/analysis/speech_statistics_pb2.pyi +43 -0
- yandex/cloud/speechsense/v1/analysis/speech_statistics_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/speech_statistics_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/statistics_common_pb2.py +30 -0
- yandex/cloud/speechsense/v1/analysis/statistics_common_pb2.pyi +91 -0
- yandex/cloud/speechsense/v1/analysis/statistics_common_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/statistics_common_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/text_classifiers_pb2.py +33 -0
- yandex/cloud/speechsense/v1/analysis/text_classifiers_pb2.pyi +102 -0
- yandex/cloud/speechsense/v1/analysis/text_classifiers_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/text_classifiers_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/transcription_pb2.py +41 -0
- yandex/cloud/speechsense/v1/analysis/transcription_pb2.pyi +186 -0
- yandex/cloud/speechsense/v1/analysis/transcription_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/transcription_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/analysis/utterance_statistics_pb2.py +27 -0
- yandex/cloud/speechsense/v1/analysis/utterance_statistics_pb2.pyi +70 -0
- yandex/cloud/speechsense/v1/analysis/utterance_statistics_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/analysis/utterance_statistics_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/search_pb2.py +51 -0
- yandex/cloud/speechsense/v1/search_pb2.pyi +324 -0
- yandex/cloud/speechsense/v1/search_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/search_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/talk_pb2.py +38 -0
- yandex/cloud/speechsense/v1/talk_pb2.pyi +161 -0
- yandex/cloud/speechsense/v1/talk_pb2_grpc.py +4 -0
- yandex/cloud/speechsense/v1/talk_pb2_grpc.pyi +17 -0
- yandex/cloud/speechsense/v1/talk_service_pb2.py +34 -18
- yandex/cloud/speechsense/v1/talk_service_pb2.pyi +143 -0
- yandex/cloud/speechsense/v1/talk_service_pb2_grpc.py +68 -0
- yandex/cloud/speechsense/v1/talk_service_pb2_grpc.pyi +40 -0
- {yandexcloud-0.294.0.dist-info → yandexcloud-0.296.0.dist-info}/METADATA +8 -8
- {yandexcloud-0.294.0.dist-info → yandexcloud-0.296.0.dist-info}/RECORD +119 -46
- {yandexcloud-0.294.0.dist-info → yandexcloud-0.296.0.dist-info}/AUTHORS +0 -0
- {yandexcloud-0.294.0.dist-info → yandexcloud-0.296.0.dist-info}/LICENSE +0 -0
- {yandexcloud-0.294.0.dist-info → yandexcloud-0.296.0.dist-info}/WHEEL +0 -0
- {yandexcloud-0.294.0.dist-info → yandexcloud-0.296.0.dist-info}/top_level.txt +0 -0
|
@@ -28,28 +28,52 @@ class PXFConfig(google.protobuf.message.Message):
|
|
|
28
28
|
XMS_FIELD_NUMBER: builtins.int
|
|
29
29
|
@property
|
|
30
30
|
def connection_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
31
|
-
"""
|
|
31
|
+
"""Timeout for connection to the Apache Tomcat® server when making read requests.
|
|
32
|
+
|
|
33
|
+
Specify values in seconds.
|
|
34
|
+
"""
|
|
32
35
|
|
|
33
36
|
@property
|
|
34
|
-
def upload_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
37
|
+
def upload_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
38
|
+
"""Timeout for connection to the Apache Tomcat® server when making write requests.
|
|
39
|
+
|
|
40
|
+
Specify the values in seconds.
|
|
41
|
+
"""
|
|
42
|
+
|
|
35
43
|
@property
|
|
36
44
|
def max_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
37
|
-
"""
|
|
45
|
+
"""Maximum number of the Apache Tomcat® threads.
|
|
46
|
+
|
|
47
|
+
To prevent situations when requests get stuck or fail due to running out of memory or malfunctioning of the Java garbage collector, specify the number of the Apache Tomcat® threads. Learn more about adjusting the number of threads in the [VMware Greenplum® Platform Extension Framework](https://docs.vmware.com/en/VMware-Greenplum-Platform-Extension-Framework/6.9/greenplum-platform-extension-framework/cfg_mem.html) documentation.
|
|
48
|
+
"""
|
|
38
49
|
|
|
39
50
|
@property
|
|
40
|
-
def pool_allow_core_thread_timeout(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
51
|
+
def pool_allow_core_thread_timeout(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
52
|
+
"""Determines whether the timeout for core streaming threads is permitted."""
|
|
53
|
+
|
|
41
54
|
@property
|
|
42
|
-
def pool_core_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
55
|
+
def pool_core_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
56
|
+
"""Number of core streaming threads per pool."""
|
|
57
|
+
|
|
43
58
|
@property
|
|
44
|
-
def pool_queue_capacity(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
59
|
+
def pool_queue_capacity(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
60
|
+
"""Maximum number of requests you can add to a pool queue for core streaming threads.
|
|
61
|
+
|
|
62
|
+
If `0`, no pool queue is generated.
|
|
63
|
+
"""
|
|
64
|
+
|
|
45
65
|
@property
|
|
46
|
-
def pool_max_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
66
|
+
def pool_max_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
67
|
+
"""Maximum allowed number of core streaming threads."""
|
|
68
|
+
|
|
47
69
|
@property
|
|
48
70
|
def xmx(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
49
|
-
"""JVM"""
|
|
71
|
+
"""Initial size, in megabytes, of the JVM heap for the PXF daemon."""
|
|
50
72
|
|
|
51
73
|
@property
|
|
52
|
-
def xms(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
74
|
+
def xms(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
75
|
+
"""Maximum size, in megabytes, of the JVM heap for the PXF daemon."""
|
|
76
|
+
|
|
53
77
|
def __init__(
|
|
54
78
|
self,
|
|
55
79
|
*,
|
|
@@ -79,11 +103,11 @@ class PXFConfigSet(google.protobuf.message.Message):
|
|
|
79
103
|
def effective_config(self) -> global___PXFConfig: ...
|
|
80
104
|
@property
|
|
81
105
|
def user_config(self) -> global___PXFConfig:
|
|
82
|
-
"""User-defined settings"""
|
|
106
|
+
"""User-defined settings."""
|
|
83
107
|
|
|
84
108
|
@property
|
|
85
109
|
def default_config(self) -> global___PXFConfig:
|
|
86
|
-
"""Default configuration"""
|
|
110
|
+
"""Default configuration."""
|
|
87
111
|
|
|
88
112
|
def __init__(
|
|
89
113
|
self,
|
|
@@ -108,10 +132,18 @@ class PXFDatasourceS3(google.protobuf.message.Message):
|
|
|
108
132
|
FAST_UPLOAD_FIELD_NUMBER: builtins.int
|
|
109
133
|
ENDPOINT_FIELD_NUMBER: builtins.int
|
|
110
134
|
access_key: builtins.str
|
|
135
|
+
"""Public key to access S3 storage."""
|
|
111
136
|
secret_key: builtins.str
|
|
137
|
+
"""Secret key to access S3 storage."""
|
|
112
138
|
endpoint: builtins.str
|
|
139
|
+
"""S3 storage address. The default value is `storage.yandexcloud.net` used for Yandex Object Storage."""
|
|
113
140
|
@property
|
|
114
|
-
def fast_upload(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
141
|
+
def fast_upload(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
142
|
+
"""Manages a fast upload of big files to S3 storage. In case of the `false` value, the PXF generates files on disk before sending them to the S3 storage. In case of the `true` value, the PXF generates files in RAM (the PXF writes to disc only if there is not enough RAM).
|
|
143
|
+
|
|
144
|
+
The fast upload is enabled by default.
|
|
145
|
+
"""
|
|
146
|
+
|
|
115
147
|
def __init__(
|
|
116
148
|
self,
|
|
117
149
|
*,
|
|
@@ -142,44 +174,81 @@ class PXFDatasourceJDBC(google.protobuf.message.Message):
|
|
|
142
174
|
POOL_IDLE_TIMEOUT_FIELD_NUMBER: builtins.int
|
|
143
175
|
POOL_MINIMUM_IDLE_FIELD_NUMBER: builtins.int
|
|
144
176
|
driver: builtins.str
|
|
145
|
-
"""
|
|
177
|
+
"""JDBC driver class in Java. The possible values are the following:
|
|
178
|
+
|
|
179
|
+
* `com.simba.athena.jdbc.Driver`
|
|
180
|
+
* `com.clickhouse.jdbc.ClickHouseDriver`
|
|
181
|
+
* `com.ibm.as400.access.AS400JDBCDriver`
|
|
182
|
+
* `com.microsoft.sqlserver.jdbc.SQLServerDriver`
|
|
183
|
+
* `com.mysql.cj.jdbc.Driver`
|
|
184
|
+
* `org.postgresql.Driver`
|
|
185
|
+
* `oracle.jdbc.driver.OracleDriver`
|
|
186
|
+
* `net.snowflake.client.jdbc.SnowflakeDriver`
|
|
187
|
+
* `io.trino.jdbc.TrinoDriver`
|
|
188
|
+
"""
|
|
146
189
|
url: builtins.str
|
|
147
|
-
"""
|
|
190
|
+
"""URL that the JDBC driver uses to connect to the database. Examples:
|
|
191
|
+
|
|
192
|
+
* `jdbc:mysql://mysqlhost:3306/testdb`: Local MySQL DB.
|
|
193
|
+
* `jdbc:postgresql://c-<cluster_id>.rw.mdb.yandexcloud.net:6432/db1`: Managed Service for PostgreSQL cluster. The address contains the special FQDN of the cluster's master.
|
|
194
|
+
* `jdbc:oracle:thin:@host.example:1521:orcl`: Oracle DB.
|
|
195
|
+
"""
|
|
148
196
|
user: builtins.str
|
|
149
|
-
"""
|
|
197
|
+
"""Username of the DB owner."""
|
|
150
198
|
password: builtins.str
|
|
151
|
-
"""
|
|
199
|
+
"""Password of the DB owner."""
|
|
152
200
|
@property
|
|
153
201
|
def statement_batch_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
154
|
-
"""
|
|
202
|
+
"""Number of rows to read in an external table, in a batch.
|
|
203
|
+
|
|
204
|
+
The default value is `100`.
|
|
205
|
+
"""
|
|
155
206
|
|
|
156
207
|
@property
|
|
157
208
|
def statement_fetch_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
158
|
-
"""
|
|
209
|
+
"""Number of rows to fetch (buffer) when reading from an external table.
|
|
210
|
+
|
|
211
|
+
The default value is `1000`.
|
|
212
|
+
"""
|
|
159
213
|
|
|
160
214
|
@property
|
|
161
215
|
def statement_query_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
162
|
-
"""
|
|
216
|
+
"""Amount of time (in seconds) the JDBC driver waits for a statement to run. This timeout applies to statements created for both read and write operations.
|
|
217
|
+
|
|
218
|
+
The default value is `60`.
|
|
219
|
+
"""
|
|
163
220
|
|
|
164
221
|
@property
|
|
165
222
|
def pool_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
166
|
-
"""
|
|
223
|
+
"""Determines whether JDBC connection pooling is used in a server configuration. By default, it is used."""
|
|
167
224
|
|
|
168
225
|
@property
|
|
169
226
|
def pool_maximum_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
170
|
-
"""
|
|
227
|
+
"""Maximum number of connections to the DB backend.
|
|
228
|
+
|
|
229
|
+
The default value is `5`.
|
|
230
|
+
"""
|
|
171
231
|
|
|
172
232
|
@property
|
|
173
233
|
def pool_connection_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
174
|
-
"""
|
|
234
|
+
"""Maximum amount of time, in milliseconds, after which an inactive connection is considered idle.
|
|
235
|
+
|
|
236
|
+
The default value is `30000`.
|
|
237
|
+
"""
|
|
175
238
|
|
|
176
239
|
@property
|
|
177
240
|
def pool_idle_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
178
|
-
"""
|
|
241
|
+
"""Maximum amount of time, in milliseconds, after which an inactive connection is considered idle.
|
|
242
|
+
|
|
243
|
+
The default value is `30000`.
|
|
244
|
+
"""
|
|
179
245
|
|
|
180
246
|
@property
|
|
181
247
|
def pool_minimum_idle(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
182
|
-
"""
|
|
248
|
+
"""Minimum number of idle connections maintained in the connection pool.
|
|
249
|
+
|
|
250
|
+
The default value is `0`.
|
|
251
|
+
"""
|
|
183
252
|
|
|
184
253
|
def __init__(
|
|
185
254
|
self,
|
|
@@ -209,7 +278,9 @@ class PXFDatasourceCore(google.protobuf.message.Message):
|
|
|
209
278
|
DEFAULT_FS_FIELD_NUMBER: builtins.int
|
|
210
279
|
SECURITY_AUTH_TO_LOCAL_FIELD_NUMBER: builtins.int
|
|
211
280
|
default_fs: builtins.str
|
|
281
|
+
"""URI whose scheme and authority determine the file system implementation."""
|
|
212
282
|
security_auth_to_local: builtins.str
|
|
283
|
+
"""Rules for mapping Kerberos principals to operating system user accounts."""
|
|
213
284
|
def __init__(
|
|
214
285
|
self,
|
|
215
286
|
*,
|
|
@@ -232,14 +303,23 @@ class PXFDatasourceKerberos(google.protobuf.message.Message):
|
|
|
232
303
|
DEFAULT_DOMAIN_FIELD_NUMBER: builtins.int
|
|
233
304
|
KEYTAB_BASE64_FIELD_NUMBER: builtins.int
|
|
234
305
|
primary: builtins.str
|
|
306
|
+
"""Host of the primary KDC server (Key Distribution Center)."""
|
|
235
307
|
realm: builtins.str
|
|
308
|
+
"""Kerberos realm for a Greenplum® DB."""
|
|
236
309
|
admin_server: builtins.str
|
|
310
|
+
"""Administration server host. Usually, this is the primary Kerberos server."""
|
|
237
311
|
default_domain: builtins.str
|
|
312
|
+
"""Domain that is used for the host name extension. Applicable when Kerberos 4 service members become Kerberos 5 service members (for example, when rcmd.hostname is replaced with host/hostname.domain)."""
|
|
238
313
|
keytab_base64: builtins.str
|
|
314
|
+
"""Base64 encoded contents of the keytab file."""
|
|
239
315
|
@property
|
|
240
|
-
def enable(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
316
|
+
def enable(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
317
|
+
"""Determines whether the Kerberos authentication server is used. By default, it is not used."""
|
|
318
|
+
|
|
241
319
|
@property
|
|
242
|
-
def kdc_servers(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
320
|
+
def kdc_servers(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
321
|
+
"""KDC server hosts."""
|
|
322
|
+
|
|
243
323
|
def __init__(
|
|
244
324
|
self,
|
|
245
325
|
*,
|
|
@@ -310,13 +390,30 @@ class PXFDatasourceHDFSDfs(google.protobuf.message.Message):
|
|
|
310
390
|
nameservices: builtins.str
|
|
311
391
|
"""Corresponds well-known HDFS client setting "dfs.nameservices" for this datasource"""
|
|
312
392
|
@property
|
|
313
|
-
def ha_automatic_failover_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
393
|
+
def ha_automatic_failover_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
394
|
+
"""Determines whether automatic failover is enabled for the high availability of the file system.
|
|
395
|
+
|
|
396
|
+
The automatic failover is enabled by default.
|
|
397
|
+
"""
|
|
398
|
+
|
|
314
399
|
@property
|
|
315
|
-
def block_access_token_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
400
|
+
def block_access_token_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
401
|
+
"""If `true`, access tokens are used as capabilities for accessing datanodes. If `false`, no access tokens are checked on accessing datanodes.
|
|
402
|
+
|
|
403
|
+
The check of access tokens is enabled by default.
|
|
404
|
+
"""
|
|
405
|
+
|
|
316
406
|
@property
|
|
317
|
-
def use_datanode_hostname(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
407
|
+
def use_datanode_hostname(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
408
|
+
"""Determines whether the datanode hostname is used when connecting to datanodes."""
|
|
409
|
+
|
|
318
410
|
@property
|
|
319
|
-
def namenodes(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, global___PXFDatasourceHDFSDfsNamenode]:
|
|
411
|
+
def namenodes(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, global___PXFDatasourceHDFSDfsNamenode]:
|
|
412
|
+
"""List of HDFS service logical names.
|
|
413
|
+
|
|
414
|
+
Specify them separated by commas. The names can be arbitrary.
|
|
415
|
+
"""
|
|
416
|
+
|
|
320
417
|
def __init__(
|
|
321
418
|
self,
|
|
322
419
|
*,
|
|
@@ -389,14 +486,29 @@ class PXFDatasourceHDFSYarn(google.protobuf.message.Message):
|
|
|
389
486
|
RESOURCEMANAGER_CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
390
487
|
HA_RM_FIELD_NUMBER: builtins.int
|
|
391
488
|
resourcemanager_cluster_id: builtins.str
|
|
489
|
+
"""Cluster ID. Specify it, so the ResourceManager service does not become active for a different cluster."""
|
|
392
490
|
@property
|
|
393
|
-
def resourcemanager_ha_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
491
|
+
def resourcemanager_ha_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
492
|
+
"""Determines whether high availability is enabled for YARN's ResourceManager services.
|
|
493
|
+
|
|
494
|
+
The high availability is enabled by default.
|
|
495
|
+
"""
|
|
496
|
+
|
|
394
497
|
@property
|
|
395
|
-
def resourcemanager_ha_auto_failover_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
498
|
+
def resourcemanager_ha_auto_failover_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
499
|
+
"""Determines whether another ResourceManager should automatically become active when the active ResourceManager has failed and does not respond.
|
|
500
|
+
|
|
501
|
+
The switch of ResourceManagers is enabled by default if the high availability is enabled.
|
|
502
|
+
"""
|
|
503
|
+
|
|
396
504
|
@property
|
|
397
|
-
def resourcemanager_ha_auto_failover_embedded(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
505
|
+
def resourcemanager_ha_auto_failover_embedded(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
506
|
+
"""Determines whether the embedded ActiveStandbyElector method should be used for the election of the active ResourceManager. If the current active ResourceManager has failed and does not respond, the ActiveStandbyElector method makes another ResourceManager active which then takes over."""
|
|
507
|
+
|
|
398
508
|
@property
|
|
399
|
-
def ha_rm(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, global___PXFDatasourceHDFSYarnHaRm]:
|
|
509
|
+
def ha_rm(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, global___PXFDatasourceHDFSYarnHaRm]:
|
|
510
|
+
"""Highly available ResourceManager service."""
|
|
511
|
+
|
|
400
512
|
def __init__(
|
|
401
513
|
self,
|
|
402
514
|
*,
|
|
@@ -424,20 +536,44 @@ class PXFDatasourceHDFS(google.protobuf.message.Message):
|
|
|
424
536
|
DFS_FIELD_NUMBER: builtins.int
|
|
425
537
|
YARN_FIELD_NUMBER: builtins.int
|
|
426
538
|
username: builtins.str
|
|
539
|
+
"""Login username for the remote file storage or DBMS if authentication on behalf of the Greenplum® user is enabled."""
|
|
427
540
|
@property
|
|
428
|
-
def core(self) -> global___PXFDatasourceCore:
|
|
541
|
+
def core(self) -> global___PXFDatasourceCore:
|
|
542
|
+
"""Settings of the file system and security rules."""
|
|
543
|
+
|
|
429
544
|
@property
|
|
430
|
-
def kerberos(self) -> global___PXFDatasourceKerberos:
|
|
545
|
+
def kerberos(self) -> global___PXFDatasourceKerberos:
|
|
546
|
+
"""Settings of the Kerberos network authentication protocol."""
|
|
547
|
+
|
|
431
548
|
@property
|
|
432
|
-
def user_impersonation(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
549
|
+
def user_impersonation(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
550
|
+
"""Enables authentication on behalf of the Greenplum® user when connecting to the remote file storage or DBMS.
|
|
551
|
+
|
|
552
|
+
The authentication is disabled by default.
|
|
553
|
+
"""
|
|
554
|
+
|
|
433
555
|
@property
|
|
434
|
-
def sasl_connection_retries(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
556
|
+
def sasl_connection_retries(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
557
|
+
"""Maximum number of times that PXF retries a SASL connection request after a refused connection returns a `GSS initiate failed` error.
|
|
558
|
+
|
|
559
|
+
The default value is `5`.
|
|
560
|
+
"""
|
|
561
|
+
|
|
435
562
|
@property
|
|
436
|
-
def zk_hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
563
|
+
def zk_hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
564
|
+
"""ZooKeeper server hosts.
|
|
565
|
+
|
|
566
|
+
Specify values in the `<address>:<port>` format.
|
|
567
|
+
"""
|
|
568
|
+
|
|
437
569
|
@property
|
|
438
|
-
def dfs(self) -> global___PXFDatasourceHDFSDfs:
|
|
570
|
+
def dfs(self) -> global___PXFDatasourceHDFSDfs:
|
|
571
|
+
"""Settings of the distributed file system."""
|
|
572
|
+
|
|
439
573
|
@property
|
|
440
|
-
def yarn(self) -> global___PXFDatasourceHDFSYarn:
|
|
574
|
+
def yarn(self) -> global___PXFDatasourceHDFSYarn:
|
|
575
|
+
"""Settings of the ResourceManager service that is responsible for tracking resources in a cluster and scheduling applications (e.g., MapReduce jobs)."""
|
|
576
|
+
|
|
441
577
|
def __init__(
|
|
442
578
|
self,
|
|
443
579
|
*,
|
|
@@ -470,22 +606,51 @@ class PXFDatasourceHive(google.protobuf.message.Message):
|
|
|
470
606
|
METASTORE_KERBEROS_PRINCIPAL_FIELD_NUMBER: builtins.int
|
|
471
607
|
AUTH_KERBEROS_PRINCIPAL_FIELD_NUMBER: builtins.int
|
|
472
608
|
username: builtins.str
|
|
609
|
+
"""Login username for the remote file storage or DBMS if authentication on behalf of the Greenplum® user is enabled."""
|
|
473
610
|
metastore_kerberos_principal: builtins.str
|
|
611
|
+
"""Service principal for the Metastore Thrift server."""
|
|
474
612
|
auth_kerberos_principal: builtins.str
|
|
613
|
+
"""Kerberos server principal."""
|
|
475
614
|
@property
|
|
476
|
-
def core(self) -> global___PXFDatasourceCore:
|
|
615
|
+
def core(self) -> global___PXFDatasourceCore:
|
|
616
|
+
"""Settings of the file system and security rules."""
|
|
617
|
+
|
|
477
618
|
@property
|
|
478
|
-
def kerberos(self) -> global___PXFDatasourceKerberos:
|
|
619
|
+
def kerberos(self) -> global___PXFDatasourceKerberos:
|
|
620
|
+
"""Settings of the Kerberos network authentication protocol."""
|
|
621
|
+
|
|
479
622
|
@property
|
|
480
|
-
def user_impersonation(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
623
|
+
def user_impersonation(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
624
|
+
"""Enables authentication on behalf of the Greenplum® user when connecting to the remote file storage or DBMS.
|
|
625
|
+
|
|
626
|
+
The authentication is disabled by default.
|
|
627
|
+
"""
|
|
628
|
+
|
|
481
629
|
@property
|
|
482
|
-
def sasl_connection_retries(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
630
|
+
def sasl_connection_retries(self) -> google.protobuf.wrappers_pb2.Int64Value:
|
|
631
|
+
"""Maximum number of times that PXF retries a SASL connection request after a refused connection returns a `GSS initiate failed` error.
|
|
632
|
+
|
|
633
|
+
The default value is `5`.
|
|
634
|
+
"""
|
|
635
|
+
|
|
483
636
|
@property
|
|
484
|
-
def zk_hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
637
|
+
def zk_hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
638
|
+
"""ZooKeeper server hosts.
|
|
639
|
+
|
|
640
|
+
Specify values in the `<address>:<port>` format.
|
|
641
|
+
"""
|
|
642
|
+
|
|
485
643
|
@property
|
|
486
|
-
def ppd(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
644
|
+
def ppd(self) -> google.protobuf.wrappers_pb2.BoolValue:
|
|
645
|
+
"""Specifies if predicate pushdown is enabled for queries on external tables.
|
|
646
|
+
|
|
647
|
+
The predicate pushdown is enabled by default.
|
|
648
|
+
"""
|
|
649
|
+
|
|
487
650
|
@property
|
|
488
|
-
def metastore_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
651
|
+
def metastore_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
652
|
+
"""List of URIs separated by commas. To request metadata, the remote DBMS connects to Metastore by one of these URIs."""
|
|
653
|
+
|
|
489
654
|
def __init__(
|
|
490
655
|
self,
|
|
491
656
|
*,
|
|
@@ -515,14 +680,23 @@ class PXFDatasource(google.protobuf.message.Message):
|
|
|
515
680
|
HDFS_FIELD_NUMBER: builtins.int
|
|
516
681
|
HIVE_FIELD_NUMBER: builtins.int
|
|
517
682
|
name: builtins.str
|
|
683
|
+
"""Data source name."""
|
|
518
684
|
@property
|
|
519
|
-
def s3(self) -> global___PXFDatasourceS3:
|
|
685
|
+
def s3(self) -> global___PXFDatasourceS3:
|
|
686
|
+
"""Settings of an external S3 data source."""
|
|
687
|
+
|
|
520
688
|
@property
|
|
521
|
-
def jdbc(self) -> global___PXFDatasourceJDBC:
|
|
689
|
+
def jdbc(self) -> global___PXFDatasourceJDBC:
|
|
690
|
+
"""Settings of an external JDBC data source."""
|
|
691
|
+
|
|
522
692
|
@property
|
|
523
|
-
def hdfs(self) -> global___PXFDatasourceHDFS:
|
|
693
|
+
def hdfs(self) -> global___PXFDatasourceHDFS:
|
|
694
|
+
"""Settings of an external HDFS data source."""
|
|
695
|
+
|
|
524
696
|
@property
|
|
525
|
-
def hive(self) -> global___PXFDatasourceHive:
|
|
697
|
+
def hive(self) -> global___PXFDatasourceHive:
|
|
698
|
+
"""Settings of an external Hive data source."""
|
|
699
|
+
|
|
526
700
|
def __init__(
|
|
527
701
|
self,
|
|
528
702
|
*,
|
|
@@ -16,9 +16,10 @@ from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__
|
|
|
16
16
|
from google.protobuf import wrappers_pb2 as google_dot_protobuf_dot_wrappers__pb2
|
|
17
17
|
from yandex.cloud.mdb.kafka.v1 import common_pb2 as yandex_dot_cloud_dot_mdb_dot_kafka_dot_v1_dot_common__pb2
|
|
18
18
|
from yandex.cloud.mdb.kafka.v1 import maintenance_pb2 as yandex_dot_cloud_dot_mdb_dot_kafka_dot_v1_dot_maintenance__pb2
|
|
19
|
+
from yandex.cloud import validation_pb2 as yandex_dot_cloud_dot_validation__pb2
|
|
19
20
|
|
|
20
21
|
|
|
21
|
-
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\'yandex/cloud/mdb/kafka/v1/cluster.proto\x12\x19yandex.cloud.mdb.kafka.v1\x1a google/protobuf/descriptor.proto\x1a\x1fgoogle/protobuf/timestamp.proto\x1a\x1egoogle/protobuf/wrappers.proto\x1a&yandex/cloud/mdb/kafka/v1/common.proto\x1a+yandex/cloud/mdb/kafka/v1/maintenance.proto\"\x99\x08\n\x07\x43luster\x12\n\n\x02id\x18\x01 \x01(\t\x12\x11\n\tfolder_id\x18\x02 \x01(\t\x12.\n\ncreated_at\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12\x0c\n\x04name\x18\x04 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x05 \x01(\t\x12>\n\x06labels\x18\x06 \x03(\x0b\x32..yandex.cloud.mdb.kafka.v1.Cluster.LabelsEntry\x12\x43\n\x0b\x65nvironment\x18\x07 \x01(\x0e\x32..yandex.cloud.mdb.kafka.v1.Cluster.Environment\x12\x39\n\nmonitoring\x18\x08 \x03(\x0b\x32%.yandex.cloud.mdb.kafka.v1.Monitoring\x12\x35\n\x06\x63onfig\x18\t \x01(\x0b\x32%.yandex.cloud.mdb.kafka.v1.ConfigSpec\x12\x12\n\nnetwork_id\x18\n \x01(\t\x12\x39\n\x06health\x18\x0b \x01(\x0e\x32).yandex.cloud.mdb.kafka.v1.Cluster.Health\x12\x39\n\x06status\x18\x0c \x01(\x0e\x32).yandex.cloud.mdb.kafka.v1.Cluster.Status\x12\x1a\n\x12security_group_ids\x18\r \x03(\t\x12\x16\n\x0ehost_group_ids\x18\x0e \x03(\t\x12\x1b\n\x13\x64\x65letion_protection\x18\x0f \x01(\x08\x12H\n\x12maintenance_window\x18\x10 \x01(\x0b\x32,.yandex.cloud.mdb.kafka.v1.MaintenanceWindow\x12J\n\x11planned_operation\x18\x11 \x01(\x0b\x32/.yandex.cloud.mdb.kafka.v1.MaintenanceOperation\x1a-\n\x0bLabelsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"I\n\x0b\x45nvironment\x12\x1b\n\x17\x45NVIRONMENT_UNSPECIFIED\x10\x00\x12\x0e\n\nPRODUCTION\x10\x01\x12\r\n\tPRESTABLE\x10\x02\"?\n\x06Health\x12\x12\n\x0eHEALTH_UNKNOWN\x10\x00\x12\t\n\x05\x41LIVE\x10\x01\x12\x08\n\x04\x44\x45\x41\x44\x10\x02\x12\x0c\n\x08\x44\x45GRADED\x10\x03\"y\n\x06Status\x12\x12\n\x0eSTATUS_UNKNOWN\x10\x00\x12\x0c\n\x08\x43REATING\x10\x01\x12\x0b\n\x07RUNNING\x10\x02\x12\t\n\x05\x45RROR\x10\x03\x12\x0c\n\x08UPDATING\x10\x04\x12\x0c\n\x08STOPPING\x10\x05\x12\x0b\n\x07STOPPED\x10\x06\x12\x0c\n\x08STARTING\x10\x07\"=\n\nMonitoring\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x02 \x01(\t\x12\x0c\n\x04link\x18\x03 \x01(\t\"\
|
|
22
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\'yandex/cloud/mdb/kafka/v1/cluster.proto\x12\x19yandex.cloud.mdb.kafka.v1\x1a google/protobuf/descriptor.proto\x1a\x1fgoogle/protobuf/timestamp.proto\x1a\x1egoogle/protobuf/wrappers.proto\x1a&yandex/cloud/mdb/kafka/v1/common.proto\x1a+yandex/cloud/mdb/kafka/v1/maintenance.proto\x1a\x1dyandex/cloud/validation.proto\"\x99\x08\n\x07\x43luster\x12\n\n\x02id\x18\x01 \x01(\t\x12\x11\n\tfolder_id\x18\x02 \x01(\t\x12.\n\ncreated_at\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12\x0c\n\x04name\x18\x04 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x05 \x01(\t\x12>\n\x06labels\x18\x06 \x03(\x0b\x32..yandex.cloud.mdb.kafka.v1.Cluster.LabelsEntry\x12\x43\n\x0b\x65nvironment\x18\x07 \x01(\x0e\x32..yandex.cloud.mdb.kafka.v1.Cluster.Environment\x12\x39\n\nmonitoring\x18\x08 \x03(\x0b\x32%.yandex.cloud.mdb.kafka.v1.Monitoring\x12\x35\n\x06\x63onfig\x18\t \x01(\x0b\x32%.yandex.cloud.mdb.kafka.v1.ConfigSpec\x12\x12\n\nnetwork_id\x18\n \x01(\t\x12\x39\n\x06health\x18\x0b \x01(\x0e\x32).yandex.cloud.mdb.kafka.v1.Cluster.Health\x12\x39\n\x06status\x18\x0c \x01(\x0e\x32).yandex.cloud.mdb.kafka.v1.Cluster.Status\x12\x1a\n\x12security_group_ids\x18\r \x03(\t\x12\x16\n\x0ehost_group_ids\x18\x0e \x03(\t\x12\x1b\n\x13\x64\x65letion_protection\x18\x0f \x01(\x08\x12H\n\x12maintenance_window\x18\x10 \x01(\x0b\x32,.yandex.cloud.mdb.kafka.v1.MaintenanceWindow\x12J\n\x11planned_operation\x18\x11 \x01(\x0b\x32/.yandex.cloud.mdb.kafka.v1.MaintenanceOperation\x1a-\n\x0bLabelsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"I\n\x0b\x45nvironment\x12\x1b\n\x17\x45NVIRONMENT_UNSPECIFIED\x10\x00\x12\x0e\n\nPRODUCTION\x10\x01\x12\r\n\tPRESTABLE\x10\x02\"?\n\x06Health\x12\x12\n\x0eHEALTH_UNKNOWN\x10\x00\x12\t\n\x05\x41LIVE\x10\x01\x12\x08\n\x04\x44\x45\x41\x44\x10\x02\x12\x0c\n\x08\x44\x45GRADED\x10\x03\"y\n\x06Status\x12\x12\n\x0eSTATUS_UNKNOWN\x10\x00\x12\x0c\n\x08\x43REATING\x10\x01\x12\x0b\n\x07RUNNING\x10\x02\x12\t\n\x05\x45RROR\x10\x03\x12\x0c\n\x08UPDATING\x10\x04\x12\x0c\n\x08STOPPING\x10\x05\x12\x0b\n\x07STOPPED\x10\x06\x12\x0c\n\x08STARTING\x10\x07\"=\n\nMonitoring\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x02 \x01(\t\x12\x0c\n\x04link\x18\x03 \x01(\t\"\xee\x06\n\nConfigSpec\x12\x0f\n\x07version\x18\x01 \x01(\t\x12:\n\x05kafka\x18\x02 \x01(\x0b\x32+.yandex.cloud.mdb.kafka.v1.ConfigSpec.Kafka\x12\x42\n\tzookeeper\x18\x03 \x01(\x0b\x32/.yandex.cloud.mdb.kafka.v1.ConfigSpec.Zookeeper\x12\x0f\n\x07zone_id\x18\x04 \x03(\t\x12\x32\n\rbrokers_count\x18\x05 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x18\n\x10\x61ssign_public_ip\x18\x06 \x01(\x08\x12\x1c\n\x10unmanaged_topics\x18\x07 \x01(\x08\x42\x02\x18\x01\x12\x17\n\x0fschema_registry\x18\x08 \x01(\x08\x12\x31\n\x06\x61\x63\x63\x65ss\x18\t \x01(\x0b\x32!.yandex.cloud.mdb.kafka.v1.Access\x12L\n\x0frest_api_config\x18\n \x01(\x0b\x32\x33.yandex.cloud.mdb.kafka.v1.ConfigSpec.RestAPIConfig\x12M\n\x15\x64isk_size_autoscaling\x18\x0b \x01(\x0b\x32..yandex.cloud.mdb.kafka.v1.DiskSizeAutoscaling\x1a\x80\x02\n\x05Kafka\x12\x37\n\tresources\x18\x01 \x01(\x0b\x32$.yandex.cloud.mdb.kafka.v1.Resources\x12V\n\x10kafka_config_2_8\x18\x04 \x01(\x0b\x32).yandex.cloud.mdb.kafka.v1.KafkaConfig2_8H\x00R\x0fkafkaConfig_2_8\x12P\n\x0ekafka_config_3\x18\x05 \x01(\x0b\x32\'.yandex.cloud.mdb.kafka.v1.KafkaConfig3H\x00R\rkafkaConfig_3B\x0e\n\x0ckafka_configJ\x04\x08\x02\x10\x04\x1a\x44\n\tZookeeper\x12\x37\n\tresources\x18\x01 \x01(\x0b\x32$.yandex.cloud.mdb.kafka.v1.Resources\x1a \n\rRestAPIConfig\x12\x0f\n\x07\x65nabled\x18\x01 \x01(\x08\"P\n\tResources\x12\x1a\n\x12resource_preset_id\x18\x01 \x01(\t\x12\x11\n\tdisk_size\x18\x02 \x01(\x03\x12\x14\n\x0c\x64isk_type_id\x18\x03 \x01(\t\"\xc0\t\n\x0eKafkaConfig2_8\x12\x44\n\x10\x63ompression_type\x18\x01 \x01(\x0e\x32*.yandex.cloud.mdb.kafka.v1.CompressionType\x12@\n\x1blog_flush_interval_messages\x18\x02 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12:\n\x15log_flush_interval_ms\x18\x03 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x44\n\x1flog_flush_scheduler_interval_ms\x18\x04 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x38\n\x13log_retention_bytes\x18\x05 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x38\n\x13log_retention_hours\x18\x06 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12:\n\x15log_retention_minutes\x18\x07 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x35\n\x10log_retention_ms\x18\x08 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x36\n\x11log_segment_bytes\x18\t \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x33\n\x0flog_preallocate\x18\n \x01(\x0b\x32\x1a.google.protobuf.BoolValue\x12=\n\x18socket_send_buffer_bytes\x18\x0b \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12@\n\x1bsocket_receive_buffer_bytes\x18\x0c \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12=\n\x19\x61uto_create_topics_enable\x18\r \x01(\x0b\x32\x1a.google.protobuf.BoolValue\x12\x33\n\x0enum_partitions\x18\x0e \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12?\n\x1a\x64\x65\x66\x61ult_replication_factor\x18\x0f \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x36\n\x11message_max_bytes\x18\x10 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12<\n\x17replica_fetch_max_bytes\x18\x11 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x19\n\x11ssl_cipher_suites\x18\x12 \x03(\t\x12>\n\x19offsets_retention_minutes\x18\x13 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12I\n\x17sasl_enabled_mechanisms\x18\x14 \x03(\x0e\x32(.yandex.cloud.mdb.kafka.v1.SaslMechanism\"\xbe\t\n\x0cKafkaConfig3\x12\x44\n\x10\x63ompression_type\x18\x01 \x01(\x0e\x32*.yandex.cloud.mdb.kafka.v1.CompressionType\x12@\n\x1blog_flush_interval_messages\x18\x02 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12:\n\x15log_flush_interval_ms\x18\x03 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x44\n\x1flog_flush_scheduler_interval_ms\x18\x04 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x38\n\x13log_retention_bytes\x18\x05 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x38\n\x13log_retention_hours\x18\x06 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12:\n\x15log_retention_minutes\x18\x07 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x35\n\x10log_retention_ms\x18\x08 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x36\n\x11log_segment_bytes\x18\t \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x33\n\x0flog_preallocate\x18\n \x01(\x0b\x32\x1a.google.protobuf.BoolValue\x12=\n\x18socket_send_buffer_bytes\x18\x0b \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12@\n\x1bsocket_receive_buffer_bytes\x18\x0c \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12=\n\x19\x61uto_create_topics_enable\x18\r \x01(\x0b\x32\x1a.google.protobuf.BoolValue\x12\x33\n\x0enum_partitions\x18\x0e \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12?\n\x1a\x64\x65\x66\x61ult_replication_factor\x18\x0f \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x36\n\x11message_max_bytes\x18\x10 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12<\n\x17replica_fetch_max_bytes\x18\x11 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12\x19\n\x11ssl_cipher_suites\x18\x12 \x03(\t\x12>\n\x19offsets_retention_minutes\x18\x13 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\x12I\n\x17sasl_enabled_mechanisms\x18\x14 \x03(\x0e\x32(.yandex.cloud.mdb.kafka.v1.SaslMechanism\"\x83\x03\n\x04Host\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x12\n\ncluster_id\x18\x02 \x01(\t\x12\x0f\n\x07zone_id\x18\x03 \x01(\t\x12\x32\n\x04role\x18\x04 \x01(\x0e\x32$.yandex.cloud.mdb.kafka.v1.Host.Role\x12\x37\n\tresources\x18\x05 \x01(\x0b\x32$.yandex.cloud.mdb.kafka.v1.Resources\x12\x36\n\x06health\x18\x06 \x01(\x0e\x32&.yandex.cloud.mdb.kafka.v1.Host.Health\x12\x11\n\tsubnet_id\x18\x08 \x01(\t\x12\x18\n\x10\x61ssign_public_ip\x18\t \x01(\x08\"6\n\x04Role\x12\x14\n\x10ROLE_UNSPECIFIED\x10\x00\x12\t\n\x05KAFKA\x10\x01\x12\r\n\tZOOKEEPER\x10\x02\"8\n\x06Health\x12\x0b\n\x07UNKNOWN\x10\x00\x12\t\n\x05\x41LIVE\x10\x01\x12\x08\n\x04\x44\x45\x41\x44\x10\x02\x12\x0c\n\x08\x44\x45GRADED\x10\x03J\x04\x08\x07\x10\x08\"\x1f\n\x06\x41\x63\x63\x65ss\x12\x15\n\rdata_transfer\x18\x01 \x01(\x08\"\x90\x01\n\x13\x44iskSizeAutoscaling\x12.\n\x17planned_usage_threshold\x18\x01 \x01(\x03\x42\r\xe8\xc7\x31\x00\xfa\xc7\x31\x05\x30-100\x12\x30\n\x19\x65mergency_usage_threshold\x18\x02 \x01(\x03\x42\r\xe8\xc7\x31\x00\xfa\xc7\x31\x05\x30-100\x12\x17\n\x0f\x64isk_size_limit\x18\x03 \x01(\x03\x42\x64\n\x1dyandex.cloud.api.mdb.kafka.v1ZCgithub.com/yandex-cloud/go-genproto/yandex/cloud/mdb/kafka/v1;kafkab\x06proto3')
|
|
22
23
|
|
|
23
24
|
_globals = globals()
|
|
24
25
|
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
@@ -30,38 +31,44 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
30
31
|
_CLUSTER_LABELSENTRY._serialized_options = b'8\001'
|
|
31
32
|
_CONFIGSPEC.fields_by_name['unmanaged_topics']._options = None
|
|
32
33
|
_CONFIGSPEC.fields_by_name['unmanaged_topics']._serialized_options = b'\030\001'
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
_globals['
|
|
38
|
-
_globals['
|
|
39
|
-
_globals['
|
|
40
|
-
_globals['
|
|
41
|
-
_globals['
|
|
42
|
-
_globals['
|
|
43
|
-
_globals['
|
|
44
|
-
_globals['
|
|
45
|
-
_globals['
|
|
46
|
-
_globals['
|
|
47
|
-
_globals['
|
|
48
|
-
_globals['
|
|
49
|
-
_globals['
|
|
50
|
-
_globals['
|
|
51
|
-
_globals['
|
|
52
|
-
_globals['
|
|
53
|
-
_globals['
|
|
54
|
-
_globals['
|
|
55
|
-
_globals['
|
|
56
|
-
_globals['
|
|
57
|
-
_globals['
|
|
58
|
-
_globals['
|
|
59
|
-
_globals['
|
|
60
|
-
_globals['
|
|
61
|
-
_globals['
|
|
62
|
-
_globals['
|
|
63
|
-
_globals['
|
|
64
|
-
_globals['
|
|
65
|
-
_globals['
|
|
66
|
-
_globals['
|
|
34
|
+
_DISKSIZEAUTOSCALING.fields_by_name['planned_usage_threshold']._options = None
|
|
35
|
+
_DISKSIZEAUTOSCALING.fields_by_name['planned_usage_threshold']._serialized_options = b'\350\3071\000\372\3071\0050-100'
|
|
36
|
+
_DISKSIZEAUTOSCALING.fields_by_name['emergency_usage_threshold']._options = None
|
|
37
|
+
_DISKSIZEAUTOSCALING.fields_by_name['emergency_usage_threshold']._serialized_options = b'\350\3071\000\372\3071\0050-100'
|
|
38
|
+
_globals['_CLUSTER']._serialized_start=286
|
|
39
|
+
_globals['_CLUSTER']._serialized_end=1335
|
|
40
|
+
_globals['_CLUSTER_LABELSENTRY']._serialized_start=1027
|
|
41
|
+
_globals['_CLUSTER_LABELSENTRY']._serialized_end=1072
|
|
42
|
+
_globals['_CLUSTER_ENVIRONMENT']._serialized_start=1074
|
|
43
|
+
_globals['_CLUSTER_ENVIRONMENT']._serialized_end=1147
|
|
44
|
+
_globals['_CLUSTER_HEALTH']._serialized_start=1149
|
|
45
|
+
_globals['_CLUSTER_HEALTH']._serialized_end=1212
|
|
46
|
+
_globals['_CLUSTER_STATUS']._serialized_start=1214
|
|
47
|
+
_globals['_CLUSTER_STATUS']._serialized_end=1335
|
|
48
|
+
_globals['_MONITORING']._serialized_start=1337
|
|
49
|
+
_globals['_MONITORING']._serialized_end=1398
|
|
50
|
+
_globals['_CONFIGSPEC']._serialized_start=1401
|
|
51
|
+
_globals['_CONFIGSPEC']._serialized_end=2279
|
|
52
|
+
_globals['_CONFIGSPEC_KAFKA']._serialized_start=1919
|
|
53
|
+
_globals['_CONFIGSPEC_KAFKA']._serialized_end=2175
|
|
54
|
+
_globals['_CONFIGSPEC_ZOOKEEPER']._serialized_start=2177
|
|
55
|
+
_globals['_CONFIGSPEC_ZOOKEEPER']._serialized_end=2245
|
|
56
|
+
_globals['_CONFIGSPEC_RESTAPICONFIG']._serialized_start=2247
|
|
57
|
+
_globals['_CONFIGSPEC_RESTAPICONFIG']._serialized_end=2279
|
|
58
|
+
_globals['_RESOURCES']._serialized_start=2281
|
|
59
|
+
_globals['_RESOURCES']._serialized_end=2361
|
|
60
|
+
_globals['_KAFKACONFIG2_8']._serialized_start=2364
|
|
61
|
+
_globals['_KAFKACONFIG2_8']._serialized_end=3580
|
|
62
|
+
_globals['_KAFKACONFIG3']._serialized_start=3583
|
|
63
|
+
_globals['_KAFKACONFIG3']._serialized_end=4797
|
|
64
|
+
_globals['_HOST']._serialized_start=4800
|
|
65
|
+
_globals['_HOST']._serialized_end=5187
|
|
66
|
+
_globals['_HOST_ROLE']._serialized_start=5069
|
|
67
|
+
_globals['_HOST_ROLE']._serialized_end=5123
|
|
68
|
+
_globals['_HOST_HEALTH']._serialized_start=5125
|
|
69
|
+
_globals['_HOST_HEALTH']._serialized_end=5181
|
|
70
|
+
_globals['_ACCESS']._serialized_start=5189
|
|
71
|
+
_globals['_ACCESS']._serialized_end=5220
|
|
72
|
+
_globals['_DISKSIZEAUTOSCALING']._serialized_start=5223
|
|
73
|
+
_globals['_DISKSIZEAUTOSCALING']._serialized_end=5367
|
|
67
74
|
# @@protoc_insertion_point(module_scope)
|
|
@@ -328,6 +328,7 @@ class ConfigSpec(google.protobuf.message.Message):
|
|
|
328
328
|
SCHEMA_REGISTRY_FIELD_NUMBER: builtins.int
|
|
329
329
|
ACCESS_FIELD_NUMBER: builtins.int
|
|
330
330
|
REST_API_CONFIG_FIELD_NUMBER: builtins.int
|
|
331
|
+
DISK_SIZE_AUTOSCALING_FIELD_NUMBER: builtins.int
|
|
331
332
|
version: builtins.str
|
|
332
333
|
"""Version of Apache Kafka® used in the cluster. Possible values: `2.1`, `2.6`."""
|
|
333
334
|
assign_public_ip: builtins.bool
|
|
@@ -364,6 +365,10 @@ class ConfigSpec(google.protobuf.message.Message):
|
|
|
364
365
|
def rest_api_config(self) -> global___ConfigSpec.RestAPIConfig:
|
|
365
366
|
"""Configuration of REST API."""
|
|
366
367
|
|
|
368
|
+
@property
|
|
369
|
+
def disk_size_autoscaling(self) -> global___DiskSizeAutoscaling:
|
|
370
|
+
"""DiskSizeAutoscaling settings"""
|
|
371
|
+
|
|
367
372
|
def __init__(
|
|
368
373
|
self,
|
|
369
374
|
*,
|
|
@@ -377,9 +382,10 @@ class ConfigSpec(google.protobuf.message.Message):
|
|
|
377
382
|
schema_registry: builtins.bool = ...,
|
|
378
383
|
access: global___Access | None = ...,
|
|
379
384
|
rest_api_config: global___ConfigSpec.RestAPIConfig | None = ...,
|
|
385
|
+
disk_size_autoscaling: global___DiskSizeAutoscaling | None = ...,
|
|
380
386
|
) -> None: ...
|
|
381
|
-
def HasField(self, field_name: typing.Literal["access", b"access", "brokers_count", b"brokers_count", "kafka", b"kafka", "rest_api_config", b"rest_api_config", "zookeeper", b"zookeeper"]) -> builtins.bool: ...
|
|
382
|
-
def ClearField(self, field_name: typing.Literal["access", b"access", "assign_public_ip", b"assign_public_ip", "brokers_count", b"brokers_count", "kafka", b"kafka", "rest_api_config", b"rest_api_config", "schema_registry", b"schema_registry", "unmanaged_topics", b"unmanaged_topics", "version", b"version", "zone_id", b"zone_id", "zookeeper", b"zookeeper"]) -> None: ...
|
|
387
|
+
def HasField(self, field_name: typing.Literal["access", b"access", "brokers_count", b"brokers_count", "disk_size_autoscaling", b"disk_size_autoscaling", "kafka", b"kafka", "rest_api_config", b"rest_api_config", "zookeeper", b"zookeeper"]) -> builtins.bool: ...
|
|
388
|
+
def ClearField(self, field_name: typing.Literal["access", b"access", "assign_public_ip", b"assign_public_ip", "brokers_count", b"brokers_count", "disk_size_autoscaling", b"disk_size_autoscaling", "kafka", b"kafka", "rest_api_config", b"rest_api_config", "schema_registry", b"schema_registry", "unmanaged_topics", b"unmanaged_topics", "version", b"version", "zone_id", b"zone_id", "zookeeper", b"zookeeper"]) -> None: ...
|
|
383
389
|
|
|
384
390
|
global___ConfigSpec = ConfigSpec
|
|
385
391
|
|
|
@@ -842,3 +848,27 @@ class Access(google.protobuf.message.Message):
|
|
|
842
848
|
def ClearField(self, field_name: typing.Literal["data_transfer", b"data_transfer"]) -> None: ...
|
|
843
849
|
|
|
844
850
|
global___Access = Access
|
|
851
|
+
|
|
852
|
+
@typing.final
|
|
853
|
+
class DiskSizeAutoscaling(google.protobuf.message.Message):
|
|
854
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
855
|
+
|
|
856
|
+
PLANNED_USAGE_THRESHOLD_FIELD_NUMBER: builtins.int
|
|
857
|
+
EMERGENCY_USAGE_THRESHOLD_FIELD_NUMBER: builtins.int
|
|
858
|
+
DISK_SIZE_LIMIT_FIELD_NUMBER: builtins.int
|
|
859
|
+
planned_usage_threshold: builtins.int
|
|
860
|
+
"""Threshold of storage usage (in percent) that triggers automatic scaling of the storage during the maintenance window. Zero value means disabled threshold."""
|
|
861
|
+
emergency_usage_threshold: builtins.int
|
|
862
|
+
"""Threshold of storage usage (in percent) that triggers immediate automatic scaling of the storage. Zero value means disabled threshold."""
|
|
863
|
+
disk_size_limit: builtins.int
|
|
864
|
+
"""New storage size (in bytes) that is set when one of the thresholds is achieved."""
|
|
865
|
+
def __init__(
|
|
866
|
+
self,
|
|
867
|
+
*,
|
|
868
|
+
planned_usage_threshold: builtins.int = ...,
|
|
869
|
+
emergency_usage_threshold: builtins.int = ...,
|
|
870
|
+
disk_size_limit: builtins.int = ...,
|
|
871
|
+
) -> None: ...
|
|
872
|
+
def ClearField(self, field_name: typing.Literal["disk_size_limit", b"disk_size_limit", "emergency_usage_threshold", b"emergency_usage_threshold", "planned_usage_threshold", b"planned_usage_threshold"]) -> None: ...
|
|
873
|
+
|
|
874
|
+
global___DiskSizeAutoscaling = DiskSizeAutoscaling
|