yandexcloud 0.354.0__py3-none-any.whl → 0.355.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of yandexcloud might be problematic. Click here for more details.

Files changed (76) hide show
  1. yandex/cloud/ai/assistants/v1/common_pb2.py +58 -30
  2. yandex/cloud/ai/assistants/v1/common_pb2.pyi +178 -3
  3. yandex/cloud/airflow/v1/cluster_pb2.py +12 -8
  4. yandex/cloud/airflow/v1/cluster_pb2.pyi +4 -0
  5. yandex/cloud/apploadbalancer/v1/virtual_host_pb2.py +32 -32
  6. yandex/cloud/apploadbalancer/v1/virtual_host_pb2.pyi +5 -1
  7. yandex/cloud/apploadbalancer/v1/virtual_host_service_pb2.py +6 -6
  8. yandex/cloud/apploadbalancer/v1/virtual_host_service_pb2.pyi +5 -1
  9. yandex/cloud/backup/v1/resource_pb2.py +2 -2
  10. yandex/cloud/backup/v1/resource_pb2.pyi +8 -0
  11. yandex/cloud/baremetal/v1alpha/configuration_service_pb2.py +12 -8
  12. yandex/cloud/baremetal/v1alpha/configuration_service_pb2.pyi +16 -2
  13. yandex/cloud/compute/v1/maintenance_service_pb2.py +47 -0
  14. yandex/cloud/compute/v1/maintenance_service_pb2.pyi +8 -0
  15. yandex/cloud/compute/v1/maintenance_service_pb2_grpc.py +190 -0
  16. yandex/cloud/compute/v1/maintenance_service_pb2_grpc.pyi +91 -0
  17. yandex/cloud/compute/v1/reserved_instance_pool_pb2.py +8 -4
  18. yandex/cloud/compute/v1/reserved_instance_pool_pb2.pyi +69 -2
  19. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2.py +44 -20
  20. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2.pyi +138 -2
  21. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2_grpc.py +88 -0
  22. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2_grpc.pyi +40 -0
  23. yandex/cloud/datatransfer/v1/endpoint_service_pb2.pyi +2 -4
  24. yandex/cloud/datatransfer/v1/endpoint_service_pb2_grpc.py +3 -6
  25. yandex/cloud/datatransfer/v1/endpoint_service_pb2_grpc.pyi +3 -9
  26. yandex/cloud/datatransfer/v1/transfer_service_pb2.pyi +2 -4
  27. yandex/cloud/datatransfer/v1/transfer_service_pb2_grpc.py +6 -3
  28. yandex/cloud/datatransfer/v1/transfer_service_pb2_grpc.pyi +6 -0
  29. yandex/cloud/iam/v1/oauth_client_pb2.py +39 -0
  30. yandex/cloud/iam/v1/oauth_client_pb2.pyi +85 -0
  31. yandex/cloud/iam/v1/oauth_client_pb2_grpc.py +24 -0
  32. yandex/cloud/iam/v1/oauth_client_pb2_grpc.pyi +17 -0
  33. yandex/cloud/iam/v1/oauth_client_secret_pb2.py +38 -0
  34. yandex/cloud/iam/v1/oauth_client_secret_pb2.pyi +49 -0
  35. yandex/cloud/iam/v1/oauth_client_secret_pb2_grpc.py +24 -0
  36. yandex/cloud/iam/v1/oauth_client_secret_pb2_grpc.pyi +17 -0
  37. yandex/cloud/iam/v1/oauth_client_secret_service_pb2.py +80 -0
  38. yandex/cloud/iam/v1/oauth_client_secret_service_pb2.pyi +188 -0
  39. yandex/cloud/iam/v1/oauth_client_secret_service_pb2_grpc.py +237 -0
  40. yandex/cloud/iam/v1/oauth_client_secret_service_pb2_grpc.pyi +121 -0
  41. yandex/cloud/iam/v1/oauth_client_service_pb2.py +99 -0
  42. yandex/cloud/iam/v1/oauth_client_service_pb2.pyi +257 -0
  43. yandex/cloud/iam/v1/oauth_client_service_pb2_grpc.py +281 -0
  44. yandex/cloud/iam/v1/oauth_client_service_pb2_grpc.pyi +141 -0
  45. yandex/cloud/iot/broker/v1/broker_service_pb2.py +50 -50
  46. yandex/cloud/iot/devices/v1/device_service_pb2.py +61 -61
  47. yandex/cloud/iot/devices/v1/registry_service_pb2.py +82 -82
  48. yandex/cloud/maintenance/__init__.py +0 -0
  49. yandex/cloud/maintenance/v2/__init__.py +0 -0
  50. yandex/cloud/maintenance/v2/common_pb2.py +73 -0
  51. yandex/cloud/maintenance/v2/common_pb2.pyi +324 -0
  52. yandex/cloud/maintenance/v2/common_pb2_grpc.py +24 -0
  53. yandex/cloud/maintenance/v2/common_pb2_grpc.pyi +17 -0
  54. yandex/cloud/mdb/clickhouse/v1/config/clickhouse_pb2.py +190 -150
  55. yandex/cloud/mdb/clickhouse/v1/config/clickhouse_pb2.pyi +1712 -1025
  56. yandex/cloud/mdb/clickhouse/v1/user_pb2.py +98 -86
  57. yandex/cloud/mdb/clickhouse/v1/user_pb2.pyi +1248 -792
  58. yandex/cloud/mdb/opensearch/v1/backup_pb2.py +14 -14
  59. yandex/cloud/mdb/opensearch/v1/backup_pb2.pyi +13 -1
  60. yandex/cloud/metastore/v1/cluster_pb2.py +18 -16
  61. yandex/cloud/metastore/v1/cluster_pb2.pyi +31 -1
  62. yandex/cloud/searchapi/v2/gen_search_service_pb2.py +10 -10
  63. yandex/cloud/searchapi/v2/gen_search_service_pb2.pyi +7 -1
  64. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_pb2.py +50 -18
  65. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_pb2.pyi +241 -2
  66. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_service_pb2.py +18 -18
  67. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_service_pb2.pyi +22 -2
  68. yandex/cloud/trino/v1/cluster_service_pb2.py +36 -36
  69. yandex/cloud/trino/v1/cluster_service_pb2.pyi +14 -2
  70. yandexcloud/__init__.py +1 -1
  71. {yandexcloud-0.354.0.dist-info → yandexcloud-0.355.0.dist-info}/METADATA +1 -1
  72. {yandexcloud-0.354.0.dist-info → yandexcloud-0.355.0.dist-info}/RECORD +76 -50
  73. {yandexcloud-0.354.0.dist-info → yandexcloud-0.355.0.dist-info}/AUTHORS +0 -0
  74. {yandexcloud-0.354.0.dist-info → yandexcloud-0.355.0.dist-info}/LICENSE +0 -0
  75. {yandexcloud-0.354.0.dist-info → yandexcloud-0.355.0.dist-info}/WHEEL +0 -0
  76. {yandexcloud-0.354.0.dist-info → yandexcloud-0.355.0.dist-info}/top_level.txt +0 -0
@@ -22,10 +22,8 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
22
22
 
23
23
  @typing.final
24
24
  class ClickhouseConfig(google.protobuf.message.Message):
25
- """ClickHouse configuration options. Detailed description for each set of options
26
- is available in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/).
27
-
28
- Any options not listed here are not supported.
25
+ """ClickHouse configuration settings. Supported settings are a subset of settings described
26
+ in [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings).
29
27
  """
30
28
 
31
29
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
@@ -51,9 +49,44 @@ class ClickhouseConfig(google.protobuf.message.Message):
51
49
  WARNING: ClickhouseConfig.LogLevel.ValueType # 4
52
50
  ERROR: ClickhouseConfig.LogLevel.ValueType # 5
53
51
 
52
+ @typing.final
53
+ class AccessControlImprovements(google.protobuf.message.Message):
54
+ """Access control settings.
55
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#access_control_improvements).
56
+ """
57
+
58
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
59
+
60
+ SELECT_FROM_SYSTEM_DB_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
61
+ SELECT_FROM_INFORMATION_SCHEMA_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
62
+ @property
63
+ def select_from_system_db_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
64
+ """Sets whether **SELECT * FROM system.<table>** requires any grants and can be executed by any user.
65
+ If set to true then this query requires **GRANT SELECT ON system.<table>** just as for non-system tables.
66
+
67
+ Default value: **false**.
68
+ """
69
+
70
+ @property
71
+ def select_from_information_schema_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
72
+ """Sets whether **SELECT * FROM information_schema.<table>** requires any grants and can be executed by any user.
73
+ If set to true, then this query requires **GRANT SELECT ON information_schema.<table>**, just as for ordinary tables.
74
+
75
+ Default value: **false**.
76
+ """
77
+
78
+ def __init__(
79
+ self,
80
+ *,
81
+ select_from_system_db_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
82
+ select_from_information_schema_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
83
+ ) -> None: ...
84
+ def HasField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> builtins.bool: ...
85
+ def ClearField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> None: ...
86
+
54
87
  @typing.final
55
88
  class MergeTree(google.protobuf.message.Message):
56
- """Options specific to the MergeTree table engine."""
89
+ """Settings for the MergeTree table engine family."""
57
90
 
58
91
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
59
92
 
@@ -69,7 +102,11 @@ class ClickhouseConfig(google.protobuf.message.Message):
69
102
  DEDUPLICATE_MERGE_PROJECTION_MODE_DROP: ClickhouseConfig.MergeTree._DeduplicateMergeProjectionMode.ValueType # 3
70
103
  DEDUPLICATE_MERGE_PROJECTION_MODE_REBUILD: ClickhouseConfig.MergeTree._DeduplicateMergeProjectionMode.ValueType # 4
71
104
 
72
- class DeduplicateMergeProjectionMode(_DeduplicateMergeProjectionMode, metaclass=_DeduplicateMergeProjectionModeEnumTypeWrapper): ...
105
+ class DeduplicateMergeProjectionMode(_DeduplicateMergeProjectionMode, metaclass=_DeduplicateMergeProjectionModeEnumTypeWrapper):
106
+ """Determines the behavior of background merges for MergeTree tables with projections.
107
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#deduplicate_merge_projection_mode).
108
+ """
109
+
73
110
  DEDUPLICATE_MERGE_PROJECTION_MODE_UNSPECIFIED: ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType # 0
74
111
  DEDUPLICATE_MERGE_PROJECTION_MODE_IGNORE: ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType # 1
75
112
  DEDUPLICATE_MERGE_PROJECTION_MODE_THROW: ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType # 2
@@ -87,488 +124,470 @@ class ClickhouseConfig(google.protobuf.message.Message):
87
124
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_DROP: ClickhouseConfig.MergeTree._LightweightMutationProjectionMode.ValueType # 2
88
125
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_REBUILD: ClickhouseConfig.MergeTree._LightweightMutationProjectionMode.ValueType # 3
89
126
 
90
- class LightweightMutationProjectionMode(_LightweightMutationProjectionMode, metaclass=_LightweightMutationProjectionModeEnumTypeWrapper): ...
127
+ class LightweightMutationProjectionMode(_LightweightMutationProjectionMode, metaclass=_LightweightMutationProjectionModeEnumTypeWrapper):
128
+ """Determines the behavior of lightweight deletes for MergeTree tables with projections."""
129
+
91
130
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_UNSPECIFIED: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 0
92
131
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_THROW: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 1
93
132
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_DROP: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 2
94
133
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_REBUILD: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 3
95
134
 
96
- REPLICATED_DEDUPLICATION_WINDOW_FIELD_NUMBER: builtins.int
97
- REPLICATED_DEDUPLICATION_WINDOW_SECONDS_FIELD_NUMBER: builtins.int
98
135
  PARTS_TO_DELAY_INSERT_FIELD_NUMBER: builtins.int
99
136
  PARTS_TO_THROW_INSERT_FIELD_NUMBER: builtins.int
100
137
  INACTIVE_PARTS_TO_DELAY_INSERT_FIELD_NUMBER: builtins.int
101
138
  INACTIVE_PARTS_TO_THROW_INSERT_FIELD_NUMBER: builtins.int
139
+ MAX_AVG_PART_SIZE_FOR_TOO_MANY_PARTS_FIELD_NUMBER: builtins.int
140
+ MAX_PARTS_IN_TOTAL_FIELD_NUMBER: builtins.int
102
141
  MAX_REPLICATED_MERGES_IN_QUEUE_FIELD_NUMBER: builtins.int
103
142
  NUMBER_OF_FREE_ENTRIES_IN_POOL_TO_LOWER_MAX_SIZE_OF_MERGE_FIELD_NUMBER: builtins.int
143
+ NUMBER_OF_FREE_ENTRIES_IN_POOL_TO_EXECUTE_MUTATION_FIELD_NUMBER: builtins.int
104
144
  MAX_BYTES_TO_MERGE_AT_MIN_SPACE_IN_POOL_FIELD_NUMBER: builtins.int
105
145
  MAX_BYTES_TO_MERGE_AT_MAX_SPACE_IN_POOL_FIELD_NUMBER: builtins.int
106
146
  MIN_BYTES_FOR_WIDE_PART_FIELD_NUMBER: builtins.int
107
147
  MIN_ROWS_FOR_WIDE_PART_FIELD_NUMBER: builtins.int
108
- TTL_ONLY_DROP_PARTS_FIELD_NUMBER: builtins.int
109
- ALLOW_REMOTE_FS_ZERO_COPY_REPLICATION_FIELD_NUMBER: builtins.int
110
- MERGE_WITH_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
111
- MERGE_WITH_RECOMPRESSION_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
112
- MAX_PARTS_IN_TOTAL_FIELD_NUMBER: builtins.int
113
- MAX_NUMBER_OF_MERGES_WITH_TTL_IN_POOL_FIELD_NUMBER: builtins.int
114
148
  CLEANUP_DELAY_PERIOD_FIELD_NUMBER: builtins.int
115
- NUMBER_OF_FREE_ENTRIES_IN_POOL_TO_EXECUTE_MUTATION_FIELD_NUMBER: builtins.int
116
- MAX_AVG_PART_SIZE_FOR_TOO_MANY_PARTS_FIELD_NUMBER: builtins.int
149
+ MAX_CLEANUP_DELAY_PERIOD_FIELD_NUMBER: builtins.int
150
+ MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
151
+ MAX_MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
117
152
  MIN_AGE_TO_FORCE_MERGE_SECONDS_FIELD_NUMBER: builtins.int
118
153
  MIN_AGE_TO_FORCE_MERGE_ON_PARTITION_ONLY_FIELD_NUMBER: builtins.int
119
- MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
120
154
  MERGE_MAX_BLOCK_SIZE_FIELD_NUMBER: builtins.int
121
- CHECK_SAMPLE_COLUMN_IS_CORRECT_FIELD_NUMBER: builtins.int
122
- MAX_MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
123
- MAX_CLEANUP_DELAY_PERIOD_FIELD_NUMBER: builtins.int
124
155
  DEDUPLICATE_MERGE_PROJECTION_MODE_FIELD_NUMBER: builtins.int
125
156
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_FIELD_NUMBER: builtins.int
126
- MATERIALIZE_TTL_RECALCULATE_ONLY_FIELD_NUMBER: builtins.int
157
+ REPLICATED_DEDUPLICATION_WINDOW_FIELD_NUMBER: builtins.int
158
+ REPLICATED_DEDUPLICATION_WINDOW_SECONDS_FIELD_NUMBER: builtins.int
127
159
  FSYNC_AFTER_INSERT_FIELD_NUMBER: builtins.int
128
160
  FSYNC_PART_DIRECTORY_FIELD_NUMBER: builtins.int
129
161
  MIN_COMPRESSED_BYTES_TO_FSYNC_AFTER_FETCH_FIELD_NUMBER: builtins.int
130
162
  MIN_COMPRESSED_BYTES_TO_FSYNC_AFTER_MERGE_FIELD_NUMBER: builtins.int
131
163
  MIN_ROWS_TO_FSYNC_AFTER_MERGE_FIELD_NUMBER: builtins.int
164
+ TTL_ONLY_DROP_PARTS_FIELD_NUMBER: builtins.int
165
+ MERGE_WITH_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
166
+ MERGE_WITH_RECOMPRESSION_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
167
+ MAX_NUMBER_OF_MERGES_WITH_TTL_IN_POOL_FIELD_NUMBER: builtins.int
168
+ MATERIALIZE_TTL_RECALCULATE_ONLY_FIELD_NUMBER: builtins.int
169
+ CHECK_SAMPLE_COLUMN_IS_CORRECT_FIELD_NUMBER: builtins.int
170
+ ALLOW_REMOTE_FS_ZERO_COPY_REPLICATION_FIELD_NUMBER: builtins.int
132
171
  deduplicate_merge_projection_mode: global___ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType
133
172
  """Determines the behavior of background merges for MergeTree tables with projections.
134
- https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#deduplicate_merge_projection_mode
173
+
174
+ Default value: **DEDUPLICATE_MERGE_PROJECTION_MODE_THROW**.
175
+
176
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#deduplicate_merge_projection_mode).
135
177
  """
136
178
  lightweight_mutation_projection_mode: global___ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType
137
- """Determines the behavior of lightweight deletes for MergeTree tables with projections."""
138
- @property
139
- def replicated_deduplication_window(self) -> google.protobuf.wrappers_pb2.Int64Value:
140
- """Number of blocks of hashes to keep in ZooKeeper."""
179
+ """Determines the behavior of lightweight deletes for MergeTree tables with projections.
141
180
 
142
- @property
143
- def replicated_deduplication_window_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
144
- """Period of time to keep blocks of hashes for."""
181
+ Default value: **LIGHTWEIGHT_MUTATION_PROJECTION_MODE_THROW**.
145
182
 
183
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#lightweight_mutation_projection_mode).
184
+ """
146
185
  @property
147
186
  def parts_to_delay_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
148
- """If table contains at least that many active parts in single partition, artificially slow down insert into table."""
187
+ """If the number of active parts in a single partition exceeds the **parts_to_delay_insert** value, an **INSERT** artificially slows down.
188
+
189
+ Default value: **1000** for versions 25.1 and higher, **150** for versions 24.12 and lower.
190
+
191
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#parts_to_delay_insert).
192
+ """
149
193
 
150
194
  @property
151
195
  def parts_to_throw_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
152
- """If more than this number active parts in single partition, throw 'Too many parts ...' exception."""
196
+ """If the number of active parts in a single partition exceeds the **parts_to_throw_insert** value, an **INSERT**
197
+ is interrupted with the error "Too many parts (N). Merges are processing significantly slower than inserts".
198
+
199
+ Default value: **3000** for versions 25.1 and higher, **300** for versions 24.12 and lower.
200
+
201
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#parts_to_throw_insert).
202
+ """
203
+
204
+ @property
205
+ def inactive_parts_to_delay_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
206
+ """If the number of inactive parts in a single partition in the table exceeds the **inactive_parts_to_delay_insert** value,
207
+ an **INSERT** is artificially slowed down.
208
+
209
+ Default value: **0**.
210
+
211
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#inactive_parts_to_delay_insert).
212
+ """
213
+
214
+ @property
215
+ def inactive_parts_to_throw_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
216
+ """If the number of inactive parts in a single partition more than the **inactive_parts_to_throw_insert** value,
217
+ **INSERT** is interrupted with an error.
218
+
219
+ Default value: **0**.
220
+
221
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#inactive_parts_to_throw_insert).
222
+ """
153
223
 
154
224
  @property
155
- def inactive_parts_to_delay_insert(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
225
+ def max_avg_part_size_for_too_many_parts(self) -> google.protobuf.wrappers_pb2.Int64Value:
226
+ """The "Too many parts" check according to **parts_to_delay_insert** and **parts_to_throw_insert** will be active only if the average
227
+ part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold,
228
+ **INSERT** queries will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server
229
+ if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts.
230
+
231
+ Default value: **1073741824** (1 GiB).
232
+
233
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_avg_part_size_for_too_many_parts).
234
+ """
235
+
156
236
  @property
157
- def inactive_parts_to_throw_insert(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
237
+ def max_parts_in_total(self) -> google.protobuf.wrappers_pb2.Int64Value:
238
+ """If the total number of active parts in all partitions of a table exceeds the **max_parts_in_total** value,
239
+ an **INSERT** is interrupted with the error "Too many parts (N)".
240
+
241
+ Default value: **20000** for versions 25.2 and higher, **100000** for versions 25.1 and lower.
242
+
243
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_parts_in_total).
244
+ """
245
+
158
246
  @property
159
247
  def max_replicated_merges_in_queue(self) -> google.protobuf.wrappers_pb2.Int64Value:
160
- """How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue."""
248
+ """How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue.
249
+
250
+ Default value: **16**.
251
+ """
161
252
 
162
253
  @property
163
254
  def number_of_free_entries_in_pool_to_lower_max_size_of_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
164
- """If there is less than specified number of free entries in background pool (or replicated queue), start to lower
165
- maximum size of merge to process.
255
+ """When there is less than the specified number of free entries in pool (or replicated queue), start to lower maximum size of
256
+ merge to process (or to put in queue). This is to allow small merges to process - not filling the pool with long running merges.
257
+
258
+ Default value: **8**.
259
+
260
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#number_of_free_entries_in_pool_to_lower_max_size_of_merge).
261
+ """
262
+
263
+ @property
264
+ def number_of_free_entries_in_pool_to_execute_mutation(self) -> google.protobuf.wrappers_pb2.Int64Value:
265
+ """When there is less than specified number of free entries in pool, do not execute part mutations.
266
+ This is to leave free threads for regular merges and to avoid "Too many parts" errors.
267
+
268
+ Default value: **20**.
269
+
270
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#number_of_free_entries_in_pool_to_execute_mutation).
166
271
  """
167
272
 
168
273
  @property
169
274
  def max_bytes_to_merge_at_min_space_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value:
170
- """Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries
171
- in replication queue).
275
+ """The maximum total part size (in bytes) to be merged into one part, with the minimum available resources in the background pool.
276
+
277
+ Default value: **1048576** (1 MiB).
278
+
279
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_bytes_to_merge_at_min_space_in_pool).
172
280
  """
173
281
 
174
282
  @property
175
- def max_bytes_to_merge_at_max_space_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
283
+ def max_bytes_to_merge_at_max_space_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value:
284
+ """The maximum total parts size (in bytes) to be merged into one part, if there are enough resources available.
285
+ Corresponds roughly to the maximum possible part size created by an automatic background merge. **0** means merges will be disabled.
286
+
287
+ Default value: **161061273600** (150 GiB).
288
+
289
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_bytes_to_merge_at_max_space_in_pool).
290
+ """
291
+
176
292
  @property
177
293
  def min_bytes_for_wide_part(self) -> google.protobuf.wrappers_pb2.Int64Value:
178
- """Minimum number of bytes in a data part that can be stored in **Wide** format.
294
+ """Minimum number of bytes in a data part that can be stored in Wide format.
179
295
 
180
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/engines/table-engines/mergetree-family/mergetree/#min_bytes_for_wide_part).
296
+ Default value: **10485760** (10 MiB).
297
+
298
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_bytes_for_wide_part).
181
299
  """
182
300
 
183
301
  @property
184
302
  def min_rows_for_wide_part(self) -> google.protobuf.wrappers_pb2.Int64Value:
185
- """Minimum number of rows in a data part that can be stored in **Wide** format.
303
+ """Minimum number of rows in a data part that can be stored in Wide format.
304
+
305
+ Default value: **0**.
186
306
 
187
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/engines/table-engines/mergetree-family/mergetree/#min_bytes_for_wide_part).
307
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_rows_for_wide_part).
188
308
  """
189
309
 
190
310
  @property
191
- def ttl_only_drop_parts(self) -> google.protobuf.wrappers_pb2.BoolValue:
192
- """Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables.
311
+ def cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value:
312
+ """Minimum period to clean old queue logs, blocks hashes and parts.
313
+
314
+ Default value: **30**.
193
315
 
194
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/settings/#ttl_only_drop_parts).
316
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#cleanup_delay_period).
195
317
  """
196
318
 
197
319
  @property
198
- def allow_remote_fs_zero_copy_replication(self) -> google.protobuf.wrappers_pb2.BoolValue: ...
199
- @property
200
- def merge_with_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
320
+ def max_cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value:
321
+ """Maximum period to clean old queue logs, blocks hashes and parts.
322
+
323
+ Default value: **300** (5 minutes).
324
+
325
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_cleanup_delay_period).
326
+ """
327
+
201
328
  @property
202
- def merge_with_recompression_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
329
+ def merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
330
+ """Minimum time to wait before trying to select parts to merge again after no parts were selected. A lower setting value will trigger
331
+ selecting tasks in background_schedule_pool frequently which result in large amount of requests to Keeper in large-scale clusters.
332
+
333
+ Default value: **5000** (5 seconds).
334
+
335
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_selecting_sleep_ms).
336
+ """
337
+
203
338
  @property
204
- def max_parts_in_total(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
339
+ def max_merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
340
+ """Maximum time to wait before trying to select parts to merge again after no parts were selected. A lower setting value will trigger
341
+ selecting tasks in background_schedule_pool frequently which result in large amount of requests to Keeper in large-scale clusters.
342
+
343
+ Default value: **60000** (1 minute).
344
+
345
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_merge_selecting_sleep_ms).
346
+ """
347
+
205
348
  @property
206
- def max_number_of_merges_with_ttl_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
349
+ def min_age_to_force_merge_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
350
+ """Merge parts if every part in the range is older than the specified value. **0** means disabled.
351
+
352
+ Default value: **0**.
353
+
354
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_age_to_force_merge_seconds).
355
+ """
356
+
207
357
  @property
208
- def cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
358
+ def min_age_to_force_merge_on_partition_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
359
+ """Whether **min_age_to_force_merge_seconds** should be applied only on the entire partition and not on subset.
360
+
361
+ Default value: **false**.
362
+
363
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_age_to_force_merge_on_partition_only).
364
+ """
365
+
209
366
  @property
210
- def number_of_free_entries_in_pool_to_execute_mutation(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
367
+ def merge_max_block_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
368
+ """The number of rows that are read from the merged parts into memory.
369
+
370
+ Default value: **8192**.
371
+
372
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_max_block_size).
373
+ """
374
+
211
375
  @property
212
- def max_avg_part_size_for_too_many_parts(self) -> google.protobuf.wrappers_pb2.Int64Value:
213
- """The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts.
214
- Default: 1 GiB
215
- Min version: 22.10
216
- See in-depth description in [ClickHouse GitHub](https://github.com/ClickHouse/ClickHouse/blob/f9558345e886876b9132d9c018e357f7fa9b22a3/src/Storages/MergeTree/MergeTreeSettings.h#L80)
376
+ def replicated_deduplication_window(self) -> google.protobuf.wrappers_pb2.Int64Value:
377
+ """The number of most recently inserted blocks for which ClickHouse Keeper stores hash sums to check for duplicates.
378
+
379
+ Default value: **1000** for versions 23.11 and higher, **100** for versions 23.10 and lower.
380
+
381
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#replicated_deduplication_window).
217
382
  """
218
383
 
219
384
  @property
220
- def min_age_to_force_merge_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
221
- """Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds.
222
- Default: 0 - disabled
223
- Min_version: 22.10
224
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#min_age_to_force_merge_seconds)
385
+ def replicated_deduplication_window_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
386
+ """The number of seconds after which the hash sums of the inserted blocks are removed from ClickHouse Keeper.
387
+
388
+ Default value: **604800** (7 days).
389
+
390
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#replicated_deduplication_window_seconds).
225
391
  """
226
392
 
227
393
  @property
228
- def min_age_to_force_merge_on_partition_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
229
- """Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset.
230
- Default: false
231
- Min_version: 22.11
232
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#min_age_to_force_merge_seconds)
394
+ def fsync_after_insert(self) -> google.protobuf.wrappers_pb2.BoolValue:
395
+ """Do fsync for every inserted part. Significantly decreases performance of inserts, not recommended to use with wide parts.
396
+
397
+ Default value: **false**.
398
+
399
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#fsync_after_insert).
233
400
  """
234
401
 
235
402
  @property
236
- def merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
237
- """Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters.
238
- Default: 5000
239
- Min_version: 21.10
240
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/settings#merge_selecting_sleep_ms)
403
+ def fsync_part_directory(self) -> google.protobuf.wrappers_pb2.BoolValue:
404
+ """Do fsync for part directory after all part operations (writes, renames, etc.).
405
+
406
+ Default value: **false**.
407
+
408
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#fsync_part_directory).
241
409
  """
242
410
 
243
411
  @property
244
- def merge_max_block_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
245
- """The number of rows that are read from the merged parts into memory.
246
- Default: 8192
247
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/settings#merge_max_block_size)
412
+ def min_compressed_bytes_to_fsync_after_fetch(self) -> google.protobuf.wrappers_pb2.Int64Value:
413
+ """Minimal number of compressed bytes to do fsync for part after fetch. **0** means disabled.
414
+
415
+ Default value: **0**.
416
+
417
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_compressed_bytes_to_fsync_after_fetch).
248
418
  """
249
419
 
250
420
  @property
251
- def check_sample_column_is_correct(self) -> google.protobuf.wrappers_pb2.BoolValue:
252
- """Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned [integer types](https://clickhouse.com/docs/en/sql-reference/data-types/int-uint): UInt8, UInt16, UInt32, UInt64.
253
- Default: true
254
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#check_sample_column_is_correct)
421
+ def min_compressed_bytes_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
422
+ """Minimal number of compressed bytes to do fsync for part after merge. **0** means disabled.
423
+
424
+ Default value: **0**.
425
+
426
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_compressed_bytes_to_fsync_after_merge).
255
427
  """
256
428
 
257
429
  @property
258
- def max_merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
259
- """Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters.
260
- Default: 60000
261
- Min_version: 23.6
262
- See in-depth description in [ClickHouse GitHub](https://github.com/ClickHouse/ClickHouse/blob/4add9db84859bff7410cf934a3904b0414e36e51/src/Storages/MergeTree/MergeTreeSettings.h#L71)
430
+ def min_rows_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
431
+ """Minimal number of rows to do fsync for part after merge. **0** means disabled.
432
+
433
+ Default value: **0**.
434
+
435
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_rows_to_fsync_after_merge).
263
436
  """
264
437
 
265
438
  @property
266
- def max_cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value:
267
- """Maximum period to clean old queue logs, blocks hashes and parts.
268
- Default: 300
269
- Min_version: 23.6
270
- See in-depth description in [ClickHouse GitHub](https://github.com/ClickHouse/ClickHouse/blob/4add9db84859bff7410cf934a3904b0414e36e51/src/Storages/MergeTree/MergeTreeSettings.h#L142)
439
+ def ttl_only_drop_parts(self) -> google.protobuf.wrappers_pb2.BoolValue:
440
+ """Controls whether data parts are fully dropped in MergeTree tables when all rows in that part have expired according to their **TTL** settings.
441
+ * **true** - the entire part is dropped if all rows in that part have expired according to their **TTL** settings.
442
+ * **false** - only the rows that have expired based on their **TTL** settings are removed.
443
+
444
+ Default value: **false**.
445
+
446
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#ttl_only_drop_parts).
271
447
  """
272
448
 
273
449
  @property
274
- def materialize_ttl_recalculate_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
275
- """Only recalculate ttl info when MATERIALIZE TTL."""
450
+ def merge_with_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
451
+ """Minimum delay in seconds before repeating a merge with delete TTL.
452
+
453
+ Default value: **14400** (4 hours).
454
+
455
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_with_ttl_timeout).
456
+ """
276
457
 
277
458
  @property
278
- def fsync_after_insert(self) -> google.protobuf.wrappers_pb2.BoolValue:
279
- """Do fsync for every inserted part. Significantly decreases performance of inserts, not recommended to use with wide parts."""
459
+ def merge_with_recompression_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
460
+ """Minimum delay in seconds before repeating a merge with recompression TTL.
461
+
462
+ Default value: **14400** (4 hours).
463
+
464
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_with_recompression_ttl_timeout).
465
+ """
280
466
 
281
467
  @property
282
- def fsync_part_directory(self) -> google.protobuf.wrappers_pb2.BoolValue:
283
- """Do fsync for part directory after all part operations (writes, renames, etc.)."""
468
+ def max_number_of_merges_with_ttl_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value:
469
+ """When there is more than specified number of merges with TTL entries in pool, do not assign new merge with TTL.
470
+ This is to leave free threads for regular merges and avoid "Too many parts" errors.
471
+
472
+ Default value: **2**.
473
+
474
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_number_of_merges_with_ttl_in_pool).
475
+ """
284
476
 
285
477
  @property
286
- def min_compressed_bytes_to_fsync_after_fetch(self) -> google.protobuf.wrappers_pb2.Int64Value:
287
- """Minimal number of compressed bytes to do fsync for part after fetch. 0 - disabled."""
478
+ def materialize_ttl_recalculate_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
479
+ """Only recalculate ttl info when **MATERIALIZE TTL**.
480
+
481
+ Default value: **true** for versions 25.2 and higher, **false** for versions 25.1 and lower.
482
+
483
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#materialize_ttl_recalculate_only).
484
+ """
288
485
 
289
486
  @property
290
- def min_compressed_bytes_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
291
- """Minimal number of compressed bytes to do fsync for part after merge. 0 - disabled."""
487
+ def check_sample_column_is_correct(self) -> google.protobuf.wrappers_pb2.BoolValue:
488
+ """Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct.
489
+ The data type must be one of unsigned integer types: UInt8, UInt16, UInt32, UInt64.
490
+
491
+ Default value: **true**.
492
+
493
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#check_sample_column_is_correct).
494
+ """
292
495
 
293
496
  @property
294
- def min_rows_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
295
- """Minimal number of rows to do fsync for part after merge. 0 - disabled."""
497
+ def allow_remote_fs_zero_copy_replication(self) -> google.protobuf.wrappers_pb2.BoolValue:
498
+ """Setting is automatically enabled if cloud storage is enabled, disabled otherwise.
499
+
500
+ Default value: **true**.
501
+ """
296
502
 
297
503
  def __init__(
298
504
  self,
299
505
  *,
300
- replicated_deduplication_window: google.protobuf.wrappers_pb2.Int64Value | None = ...,
301
- replicated_deduplication_window_seconds: google.protobuf.wrappers_pb2.Int64Value | None = ...,
302
506
  parts_to_delay_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
303
507
  parts_to_throw_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
304
508
  inactive_parts_to_delay_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
305
509
  inactive_parts_to_throw_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
510
+ max_avg_part_size_for_too_many_parts: google.protobuf.wrappers_pb2.Int64Value | None = ...,
511
+ max_parts_in_total: google.protobuf.wrappers_pb2.Int64Value | None = ...,
306
512
  max_replicated_merges_in_queue: google.protobuf.wrappers_pb2.Int64Value | None = ...,
307
513
  number_of_free_entries_in_pool_to_lower_max_size_of_merge: google.protobuf.wrappers_pb2.Int64Value | None = ...,
514
+ number_of_free_entries_in_pool_to_execute_mutation: google.protobuf.wrappers_pb2.Int64Value | None = ...,
308
515
  max_bytes_to_merge_at_min_space_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
309
516
  max_bytes_to_merge_at_max_space_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
310
517
  min_bytes_for_wide_part: google.protobuf.wrappers_pb2.Int64Value | None = ...,
311
518
  min_rows_for_wide_part: google.protobuf.wrappers_pb2.Int64Value | None = ...,
312
- ttl_only_drop_parts: google.protobuf.wrappers_pb2.BoolValue | None = ...,
313
- allow_remote_fs_zero_copy_replication: google.protobuf.wrappers_pb2.BoolValue | None = ...,
314
- merge_with_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
315
- merge_with_recompression_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
316
- max_parts_in_total: google.protobuf.wrappers_pb2.Int64Value | None = ...,
317
- max_number_of_merges_with_ttl_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
318
519
  cleanup_delay_period: google.protobuf.wrappers_pb2.Int64Value | None = ...,
319
- number_of_free_entries_in_pool_to_execute_mutation: google.protobuf.wrappers_pb2.Int64Value | None = ...,
320
- max_avg_part_size_for_too_many_parts: google.protobuf.wrappers_pb2.Int64Value | None = ...,
520
+ max_cleanup_delay_period: google.protobuf.wrappers_pb2.Int64Value | None = ...,
521
+ merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
522
+ max_merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
321
523
  min_age_to_force_merge_seconds: google.protobuf.wrappers_pb2.Int64Value | None = ...,
322
524
  min_age_to_force_merge_on_partition_only: google.protobuf.wrappers_pb2.BoolValue | None = ...,
323
- merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
324
525
  merge_max_block_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
325
- check_sample_column_is_correct: google.protobuf.wrappers_pb2.BoolValue | None = ...,
326
- max_merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
327
- max_cleanup_delay_period: google.protobuf.wrappers_pb2.Int64Value | None = ...,
328
526
  deduplicate_merge_projection_mode: global___ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType = ...,
329
527
  lightweight_mutation_projection_mode: global___ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType = ...,
330
- materialize_ttl_recalculate_only: google.protobuf.wrappers_pb2.BoolValue | None = ...,
528
+ replicated_deduplication_window: google.protobuf.wrappers_pb2.Int64Value | None = ...,
529
+ replicated_deduplication_window_seconds: google.protobuf.wrappers_pb2.Int64Value | None = ...,
331
530
  fsync_after_insert: google.protobuf.wrappers_pb2.BoolValue | None = ...,
332
531
  fsync_part_directory: google.protobuf.wrappers_pb2.BoolValue | None = ...,
333
532
  min_compressed_bytes_to_fsync_after_fetch: google.protobuf.wrappers_pb2.Int64Value | None = ...,
334
533
  min_compressed_bytes_to_fsync_after_merge: google.protobuf.wrappers_pb2.Int64Value | None = ...,
335
534
  min_rows_to_fsync_after_merge: google.protobuf.wrappers_pb2.Int64Value | None = ...,
535
+ ttl_only_drop_parts: google.protobuf.wrappers_pb2.BoolValue | None = ...,
536
+ merge_with_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
537
+ merge_with_recompression_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
538
+ max_number_of_merges_with_ttl_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
539
+ materialize_ttl_recalculate_only: google.protobuf.wrappers_pb2.BoolValue | None = ...,
540
+ check_sample_column_is_correct: google.protobuf.wrappers_pb2.BoolValue | None = ...,
541
+ allow_remote_fs_zero_copy_replication: google.protobuf.wrappers_pb2.BoolValue | None = ...,
336
542
  ) -> None: ...
337
543
  def HasField(self, field_name: typing.Literal["allow_remote_fs_zero_copy_replication", b"allow_remote_fs_zero_copy_replication", "check_sample_column_is_correct", b"check_sample_column_is_correct", "cleanup_delay_period", b"cleanup_delay_period", "fsync_after_insert", b"fsync_after_insert", "fsync_part_directory", b"fsync_part_directory", "inactive_parts_to_delay_insert", b"inactive_parts_to_delay_insert", "inactive_parts_to_throw_insert", b"inactive_parts_to_throw_insert", "materialize_ttl_recalculate_only", b"materialize_ttl_recalculate_only", "max_avg_part_size_for_too_many_parts", b"max_avg_part_size_for_too_many_parts", "max_bytes_to_merge_at_max_space_in_pool", b"max_bytes_to_merge_at_max_space_in_pool", "max_bytes_to_merge_at_min_space_in_pool", b"max_bytes_to_merge_at_min_space_in_pool", "max_cleanup_delay_period", b"max_cleanup_delay_period", "max_merge_selecting_sleep_ms", b"max_merge_selecting_sleep_ms", "max_number_of_merges_with_ttl_in_pool", b"max_number_of_merges_with_ttl_in_pool", "max_parts_in_total", b"max_parts_in_total", "max_replicated_merges_in_queue", b"max_replicated_merges_in_queue", "merge_max_block_size", b"merge_max_block_size", "merge_selecting_sleep_ms", b"merge_selecting_sleep_ms", "merge_with_recompression_ttl_timeout", b"merge_with_recompression_ttl_timeout", "merge_with_ttl_timeout", b"merge_with_ttl_timeout", "min_age_to_force_merge_on_partition_only", b"min_age_to_force_merge_on_partition_only", "min_age_to_force_merge_seconds", b"min_age_to_force_merge_seconds", "min_bytes_for_wide_part", b"min_bytes_for_wide_part", "min_compressed_bytes_to_fsync_after_fetch", b"min_compressed_bytes_to_fsync_after_fetch", "min_compressed_bytes_to_fsync_after_merge", b"min_compressed_bytes_to_fsync_after_merge", "min_rows_for_wide_part", b"min_rows_for_wide_part", "min_rows_to_fsync_after_merge", b"min_rows_to_fsync_after_merge", "number_of_free_entries_in_pool_to_execute_mutation", b"number_of_free_entries_in_pool_to_execute_mutation", "number_of_free_entries_in_pool_to_lower_max_size_of_merge", b"number_of_free_entries_in_pool_to_lower_max_size_of_merge", "parts_to_delay_insert", b"parts_to_delay_insert", "parts_to_throw_insert", b"parts_to_throw_insert", "replicated_deduplication_window", b"replicated_deduplication_window", "replicated_deduplication_window_seconds", b"replicated_deduplication_window_seconds", "ttl_only_drop_parts", b"ttl_only_drop_parts"]) -> builtins.bool: ...
338
544
  def ClearField(self, field_name: typing.Literal["allow_remote_fs_zero_copy_replication", b"allow_remote_fs_zero_copy_replication", "check_sample_column_is_correct", b"check_sample_column_is_correct", "cleanup_delay_period", b"cleanup_delay_period", "deduplicate_merge_projection_mode", b"deduplicate_merge_projection_mode", "fsync_after_insert", b"fsync_after_insert", "fsync_part_directory", b"fsync_part_directory", "inactive_parts_to_delay_insert", b"inactive_parts_to_delay_insert", "inactive_parts_to_throw_insert", b"inactive_parts_to_throw_insert", "lightweight_mutation_projection_mode", b"lightweight_mutation_projection_mode", "materialize_ttl_recalculate_only", b"materialize_ttl_recalculate_only", "max_avg_part_size_for_too_many_parts", b"max_avg_part_size_for_too_many_parts", "max_bytes_to_merge_at_max_space_in_pool", b"max_bytes_to_merge_at_max_space_in_pool", "max_bytes_to_merge_at_min_space_in_pool", b"max_bytes_to_merge_at_min_space_in_pool", "max_cleanup_delay_period", b"max_cleanup_delay_period", "max_merge_selecting_sleep_ms", b"max_merge_selecting_sleep_ms", "max_number_of_merges_with_ttl_in_pool", b"max_number_of_merges_with_ttl_in_pool", "max_parts_in_total", b"max_parts_in_total", "max_replicated_merges_in_queue", b"max_replicated_merges_in_queue", "merge_max_block_size", b"merge_max_block_size", "merge_selecting_sleep_ms", b"merge_selecting_sleep_ms", "merge_with_recompression_ttl_timeout", b"merge_with_recompression_ttl_timeout", "merge_with_ttl_timeout", b"merge_with_ttl_timeout", "min_age_to_force_merge_on_partition_only", b"min_age_to_force_merge_on_partition_only", "min_age_to_force_merge_seconds", b"min_age_to_force_merge_seconds", "min_bytes_for_wide_part", b"min_bytes_for_wide_part", "min_compressed_bytes_to_fsync_after_fetch", b"min_compressed_bytes_to_fsync_after_fetch", "min_compressed_bytes_to_fsync_after_merge", b"min_compressed_bytes_to_fsync_after_merge", "min_rows_for_wide_part", b"min_rows_for_wide_part", "min_rows_to_fsync_after_merge", b"min_rows_to_fsync_after_merge", "number_of_free_entries_in_pool_to_execute_mutation", b"number_of_free_entries_in_pool_to_execute_mutation", "number_of_free_entries_in_pool_to_lower_max_size_of_merge", b"number_of_free_entries_in_pool_to_lower_max_size_of_merge", "parts_to_delay_insert", b"parts_to_delay_insert", "parts_to_throw_insert", b"parts_to_throw_insert", "replicated_deduplication_window", b"replicated_deduplication_window", "replicated_deduplication_window_seconds", b"replicated_deduplication_window_seconds", "ttl_only_drop_parts", b"ttl_only_drop_parts"]) -> None: ...
339
545
 
340
546
  @typing.final
341
- class Kafka(google.protobuf.message.Message):
547
+ class Compression(google.protobuf.message.Message):
548
+ """Compression settings.
549
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#compression).
550
+ """
551
+
342
552
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
343
553
 
344
- class _SecurityProtocol:
554
+ class _Method:
345
555
  ValueType = typing.NewType("ValueType", builtins.int)
346
556
  V: typing_extensions.TypeAlias = ValueType
347
557
 
348
- class _SecurityProtocolEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SecurityProtocol.ValueType], builtins.type):
558
+ class _MethodEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Compression._Method.ValueType], builtins.type):
349
559
  DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
350
- SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 0
351
- SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 1
352
- SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 2
353
- SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 3
354
- SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 4
355
-
356
- class SecurityProtocol(_SecurityProtocol, metaclass=_SecurityProtocolEnumTypeWrapper): ...
357
- SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 0
358
- SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 1
359
- SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 2
360
- SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 3
361
- SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 4
362
-
363
- class _SaslMechanism:
364
- ValueType = typing.NewType("ValueType", builtins.int)
365
- V: typing_extensions.TypeAlias = ValueType
366
-
367
- class _SaslMechanismEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SaslMechanism.ValueType], builtins.type):
368
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
369
- SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 0
370
- SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 1
371
- SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 2
372
- SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 3
373
- SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 4
374
-
375
- class SaslMechanism(_SaslMechanism, metaclass=_SaslMechanismEnumTypeWrapper): ...
376
- SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 0
377
- SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 1
378
- SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 2
379
- SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 3
380
- SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 4
381
-
382
- class _Debug:
383
- ValueType = typing.NewType("ValueType", builtins.int)
384
- V: typing_extensions.TypeAlias = ValueType
385
-
386
- class _DebugEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._Debug.ValueType], builtins.type):
387
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
388
- DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka._Debug.ValueType # 0
389
- DEBUG_GENERIC: ClickhouseConfig.Kafka._Debug.ValueType # 1
390
- DEBUG_BROKER: ClickhouseConfig.Kafka._Debug.ValueType # 2
391
- DEBUG_TOPIC: ClickhouseConfig.Kafka._Debug.ValueType # 3
392
- DEBUG_METADATA: ClickhouseConfig.Kafka._Debug.ValueType # 4
393
- DEBUG_FEATURE: ClickhouseConfig.Kafka._Debug.ValueType # 5
394
- DEBUG_QUEUE: ClickhouseConfig.Kafka._Debug.ValueType # 6
395
- DEBUG_MSG: ClickhouseConfig.Kafka._Debug.ValueType # 7
396
- DEBUG_PROTOCOL: ClickhouseConfig.Kafka._Debug.ValueType # 8
397
- DEBUG_CGRP: ClickhouseConfig.Kafka._Debug.ValueType # 9
398
- DEBUG_SECURITY: ClickhouseConfig.Kafka._Debug.ValueType # 10
399
- DEBUG_FETCH: ClickhouseConfig.Kafka._Debug.ValueType # 11
400
- DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka._Debug.ValueType # 12
401
- DEBUG_PLUGIN: ClickhouseConfig.Kafka._Debug.ValueType # 13
402
- DEBUG_CONSUMER: ClickhouseConfig.Kafka._Debug.ValueType # 14
403
- DEBUG_ADMIN: ClickhouseConfig.Kafka._Debug.ValueType # 15
404
- DEBUG_EOS: ClickhouseConfig.Kafka._Debug.ValueType # 16
405
- DEBUG_MOCK: ClickhouseConfig.Kafka._Debug.ValueType # 17
406
- DEBUG_ASSIGNOR: ClickhouseConfig.Kafka._Debug.ValueType # 18
407
- DEBUG_CONF: ClickhouseConfig.Kafka._Debug.ValueType # 19
408
- DEBUG_TELEMETRY: ClickhouseConfig.Kafka._Debug.ValueType # 20
409
- DEBUG_ALL: ClickhouseConfig.Kafka._Debug.ValueType # 21
410
-
411
- class Debug(_Debug, metaclass=_DebugEnumTypeWrapper): ...
412
- DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka.Debug.ValueType # 0
413
- DEBUG_GENERIC: ClickhouseConfig.Kafka.Debug.ValueType # 1
414
- DEBUG_BROKER: ClickhouseConfig.Kafka.Debug.ValueType # 2
415
- DEBUG_TOPIC: ClickhouseConfig.Kafka.Debug.ValueType # 3
416
- DEBUG_METADATA: ClickhouseConfig.Kafka.Debug.ValueType # 4
417
- DEBUG_FEATURE: ClickhouseConfig.Kafka.Debug.ValueType # 5
418
- DEBUG_QUEUE: ClickhouseConfig.Kafka.Debug.ValueType # 6
419
- DEBUG_MSG: ClickhouseConfig.Kafka.Debug.ValueType # 7
420
- DEBUG_PROTOCOL: ClickhouseConfig.Kafka.Debug.ValueType # 8
421
- DEBUG_CGRP: ClickhouseConfig.Kafka.Debug.ValueType # 9
422
- DEBUG_SECURITY: ClickhouseConfig.Kafka.Debug.ValueType # 10
423
- DEBUG_FETCH: ClickhouseConfig.Kafka.Debug.ValueType # 11
424
- DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka.Debug.ValueType # 12
425
- DEBUG_PLUGIN: ClickhouseConfig.Kafka.Debug.ValueType # 13
426
- DEBUG_CONSUMER: ClickhouseConfig.Kafka.Debug.ValueType # 14
427
- DEBUG_ADMIN: ClickhouseConfig.Kafka.Debug.ValueType # 15
428
- DEBUG_EOS: ClickhouseConfig.Kafka.Debug.ValueType # 16
429
- DEBUG_MOCK: ClickhouseConfig.Kafka.Debug.ValueType # 17
430
- DEBUG_ASSIGNOR: ClickhouseConfig.Kafka.Debug.ValueType # 18
431
- DEBUG_CONF: ClickhouseConfig.Kafka.Debug.ValueType # 19
432
- DEBUG_TELEMETRY: ClickhouseConfig.Kafka.Debug.ValueType # 20
433
- DEBUG_ALL: ClickhouseConfig.Kafka.Debug.ValueType # 21
434
-
435
- class _AutoOffsetReset:
436
- ValueType = typing.NewType("ValueType", builtins.int)
437
- V: typing_extensions.TypeAlias = ValueType
438
-
439
- class _AutoOffsetResetEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._AutoOffsetReset.ValueType], builtins.type):
440
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
441
- AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 0
442
- AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 1
443
- AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 2
444
- AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 3
445
- AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 4
446
- AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 5
447
- AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 6
448
- AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 7
449
-
450
- class AutoOffsetReset(_AutoOffsetReset, metaclass=_AutoOffsetResetEnumTypeWrapper): ...
451
- AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 0
452
- AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 1
453
- AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 2
454
- AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 3
455
- AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 4
456
- AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 5
457
- AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 6
458
- AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 7
459
-
460
- SECURITY_PROTOCOL_FIELD_NUMBER: builtins.int
461
- SASL_MECHANISM_FIELD_NUMBER: builtins.int
462
- SASL_USERNAME_FIELD_NUMBER: builtins.int
463
- SASL_PASSWORD_FIELD_NUMBER: builtins.int
464
- ENABLE_SSL_CERTIFICATE_VERIFICATION_FIELD_NUMBER: builtins.int
465
- MAX_POLL_INTERVAL_MS_FIELD_NUMBER: builtins.int
466
- SESSION_TIMEOUT_MS_FIELD_NUMBER: builtins.int
467
- DEBUG_FIELD_NUMBER: builtins.int
468
- AUTO_OFFSET_RESET_FIELD_NUMBER: builtins.int
469
- security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType
470
- sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType
471
- sasl_username: builtins.str
472
- sasl_password: builtins.str
473
- debug: global___ClickhouseConfig.Kafka.Debug.ValueType
474
- auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType
475
- @property
476
- def enable_ssl_certificate_verification(self) -> google.protobuf.wrappers_pb2.BoolValue: ...
477
- @property
478
- def max_poll_interval_ms(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
479
- @property
480
- def session_timeout_ms(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
481
- def __init__(
482
- self,
483
- *,
484
- security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType = ...,
485
- sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType = ...,
486
- sasl_username: builtins.str = ...,
487
- sasl_password: builtins.str = ...,
488
- enable_ssl_certificate_verification: google.protobuf.wrappers_pb2.BoolValue | None = ...,
489
- max_poll_interval_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
490
- session_timeout_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
491
- debug: global___ClickhouseConfig.Kafka.Debug.ValueType = ...,
492
- auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType = ...,
493
- ) -> None: ...
494
- def HasField(self, field_name: typing.Literal["enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "session_timeout_ms", b"session_timeout_ms"]) -> builtins.bool: ...
495
- def ClearField(self, field_name: typing.Literal["auto_offset_reset", b"auto_offset_reset", "debug", b"debug", "enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "sasl_mechanism", b"sasl_mechanism", "sasl_password", b"sasl_password", "sasl_username", b"sasl_username", "security_protocol", b"security_protocol", "session_timeout_ms", b"session_timeout_ms"]) -> None: ...
496
-
497
- @typing.final
498
- class KafkaTopic(google.protobuf.message.Message):
499
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
500
-
501
- NAME_FIELD_NUMBER: builtins.int
502
- SETTINGS_FIELD_NUMBER: builtins.int
503
- name: builtins.str
504
- @property
505
- def settings(self) -> global___ClickhouseConfig.Kafka: ...
506
- def __init__(
507
- self,
508
- *,
509
- name: builtins.str = ...,
510
- settings: global___ClickhouseConfig.Kafka | None = ...,
511
- ) -> None: ...
512
- def HasField(self, field_name: typing.Literal["settings", b"settings"]) -> builtins.bool: ...
513
- def ClearField(self, field_name: typing.Literal["name", b"name", "settings", b"settings"]) -> None: ...
514
-
515
- @typing.final
516
- class Rabbitmq(google.protobuf.message.Message):
517
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
518
-
519
- USERNAME_FIELD_NUMBER: builtins.int
520
- PASSWORD_FIELD_NUMBER: builtins.int
521
- VHOST_FIELD_NUMBER: builtins.int
522
- username: builtins.str
523
- """[RabbitMQ](https://clickhouse.com/docs/en/engines/table-engines/integrations/rabbitmq/) username"""
524
- password: builtins.str
525
- """[RabbitMQ](https://clickhouse.com/docs/en/engines/table-engines/integrations/rabbitmq/) password"""
526
- vhost: builtins.str
527
- """[RabbitMQ](https://clickhouse.com/docs/en/engines/table-engines/integrations/rabbitmq/) virtual host"""
528
- def __init__(
529
- self,
530
- *,
531
- username: builtins.str = ...,
532
- password: builtins.str = ...,
533
- vhost: builtins.str = ...,
534
- ) -> None: ...
535
- def ClearField(self, field_name: typing.Literal["password", b"password", "username", b"username", "vhost", b"vhost"]) -> None: ...
536
-
537
- @typing.final
538
- class Compression(google.protobuf.message.Message):
539
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
540
-
541
- class _Method:
542
- ValueType = typing.NewType("ValueType", builtins.int)
543
- V: typing_extensions.TypeAlias = ValueType
544
-
545
- class _MethodEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Compression._Method.ValueType], builtins.type):
546
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
547
- METHOD_UNSPECIFIED: ClickhouseConfig.Compression._Method.ValueType # 0
548
- LZ4: ClickhouseConfig.Compression._Method.ValueType # 1
549
- """[LZ4 compression algorithm](https://lz4.github.io/lz4/)."""
550
- ZSTD: ClickhouseConfig.Compression._Method.ValueType # 2
551
- """[Zstandard compression algorithm](https://facebook.github.io/zstd/)."""
560
+ METHOD_UNSPECIFIED: ClickhouseConfig.Compression._Method.ValueType # 0
561
+ LZ4: ClickhouseConfig.Compression._Method.ValueType # 1
562
+ """[LZ4 compression algorithm](https://lz4.github.io/lz4)."""
563
+ ZSTD: ClickhouseConfig.Compression._Method.ValueType # 2
564
+ """[ZSTD compression algorithm](https://facebook.github.io/zstd)."""
565
+ LZ4HC: ClickhouseConfig.Compression._Method.ValueType # 3
566
+ """[LZ4 HC (high compression) algorithm](https://clickhouse.com/docs/sql-reference/statements/create/table#lz4hc)."""
552
567
 
553
568
  class Method(_Method, metaclass=_MethodEnumTypeWrapper): ...
554
569
  METHOD_UNSPECIFIED: ClickhouseConfig.Compression.Method.ValueType # 0
555
570
  LZ4: ClickhouseConfig.Compression.Method.ValueType # 1
556
- """[LZ4 compression algorithm](https://lz4.github.io/lz4/)."""
571
+ """[LZ4 compression algorithm](https://lz4.github.io/lz4)."""
557
572
  ZSTD: ClickhouseConfig.Compression.Method.ValueType # 2
558
- """[Zstandard compression algorithm](https://facebook.github.io/zstd/)."""
573
+ """[ZSTD compression algorithm](https://facebook.github.io/zstd)."""
574
+ LZ4HC: ClickhouseConfig.Compression.Method.ValueType # 3
575
+ """[LZ4 HC (high compression) algorithm](https://clickhouse.com/docs/sql-reference/statements/create/table#lz4hc)."""
559
576
 
560
577
  METHOD_FIELD_NUMBER: builtins.int
561
578
  MIN_PART_SIZE_FIELD_NUMBER: builtins.int
562
579
  MIN_PART_SIZE_RATIO_FIELD_NUMBER: builtins.int
563
580
  LEVEL_FIELD_NUMBER: builtins.int
564
581
  method: global___ClickhouseConfig.Compression.Method.ValueType
565
- """Compression method to use for the specified combination of [min_part_size] and [min_part_size_ratio]."""
582
+ """Compression method to use for the specified combination of **min_part_size** and **min_part_size_ratio**."""
566
583
  min_part_size: builtins.int
567
- """Minimum size of a part of a table."""
584
+ """The minimum size of a data part."""
568
585
  min_part_size_ratio: builtins.float
569
- """Minimum ratio of a part relative to the size of all the data in the table."""
586
+ """The ratio of the data part size to the table size."""
570
587
  @property
571
- def level(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
588
+ def level(self) -> google.protobuf.wrappers_pb2.Int64Value:
589
+ """Compression level."""
590
+
572
591
  def __init__(
573
592
  self,
574
593
  *,
@@ -582,305 +601,49 @@ class ClickhouseConfig(google.protobuf.message.Message):
582
601
 
583
602
  @typing.final
584
603
  class ExternalDictionary(google.protobuf.message.Message):
604
+ """External dictionary configuration."""
605
+
585
606
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
586
607
 
587
608
  @typing.final
588
- class HttpSource(google.protobuf.message.Message):
609
+ class Structure(google.protobuf.message.Message):
610
+ """Configuration of external dictionary structure."""
611
+
589
612
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
590
613
 
591
614
  @typing.final
592
- class Header(google.protobuf.message.Message):
615
+ class Id(google.protobuf.message.Message):
616
+ """Numeric key."""
617
+
593
618
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
594
619
 
595
620
  NAME_FIELD_NUMBER: builtins.int
596
- VALUE_FIELD_NUMBER: builtins.int
597
621
  name: builtins.str
598
- value: builtins.str
622
+ """Name of the numeric key."""
599
623
  def __init__(
600
624
  self,
601
625
  *,
602
626
  name: builtins.str = ...,
603
- value: builtins.str = ...,
604
627
  ) -> None: ...
605
- def ClearField(self, field_name: typing.Literal["name", b"name", "value", b"value"]) -> None: ...
606
-
607
- URL_FIELD_NUMBER: builtins.int
608
- FORMAT_FIELD_NUMBER: builtins.int
609
- HEADERS_FIELD_NUMBER: builtins.int
610
- url: builtins.str
611
- """URL of the source dictionary available over HTTP."""
612
- format: builtins.str
613
- """The data format. Valid values are all formats supported by ClickHouse SQL dialect."""
614
- @property
615
- def headers(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header]:
616
- """HTTP headers."""
617
-
618
- def __init__(
619
- self,
620
- *,
621
- url: builtins.str = ...,
622
- format: builtins.str = ...,
623
- headers: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header] | None = ...,
624
- ) -> None: ...
625
- def ClearField(self, field_name: typing.Literal["format", b"format", "headers", b"headers", "url", b"url"]) -> None: ...
626
-
627
- @typing.final
628
- class MysqlSource(google.protobuf.message.Message):
629
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
628
+ def ClearField(self, field_name: typing.Literal["name", b"name"]) -> None: ...
630
629
 
631
630
  @typing.final
632
- class Replica(google.protobuf.message.Message):
631
+ class Key(google.protobuf.message.Message):
632
+ """Complex key."""
633
+
633
634
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
634
635
 
635
- HOST_FIELD_NUMBER: builtins.int
636
- PRIORITY_FIELD_NUMBER: builtins.int
637
- PORT_FIELD_NUMBER: builtins.int
638
- USER_FIELD_NUMBER: builtins.int
639
- PASSWORD_FIELD_NUMBER: builtins.int
640
- host: builtins.str
641
- """MySQL host of the replica."""
642
- priority: builtins.int
643
- """The priority of the replica that ClickHouse takes into account when connecting.
644
- Replica with the highest priority should have this field set to the lowest number.
645
- """
646
- port: builtins.int
647
- """Port to use when connecting to the replica.
648
- If a port is not specified for a replica, ClickHouse uses the port specified for the source.
649
- """
650
- user: builtins.str
651
- """Name of the MySQL database user."""
652
- password: builtins.str
653
- """Password of the MySQL database user."""
636
+ ATTRIBUTES_FIELD_NUMBER: builtins.int
637
+ @property
638
+ def attributes(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute]:
639
+ """Attributes of a complex key."""
640
+
654
641
  def __init__(
655
642
  self,
656
643
  *,
657
- host: builtins.str = ...,
658
- priority: builtins.int = ...,
659
- port: builtins.int = ...,
660
- user: builtins.str = ...,
661
- password: builtins.str = ...,
644
+ attributes: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute] | None = ...,
662
645
  ) -> None: ...
663
- def ClearField(self, field_name: typing.Literal["host", b"host", "password", b"password", "port", b"port", "priority", b"priority", "user", b"user"]) -> None: ...
664
-
665
- DB_FIELD_NUMBER: builtins.int
666
- TABLE_FIELD_NUMBER: builtins.int
667
- PORT_FIELD_NUMBER: builtins.int
668
- USER_FIELD_NUMBER: builtins.int
669
- PASSWORD_FIELD_NUMBER: builtins.int
670
- REPLICAS_FIELD_NUMBER: builtins.int
671
- WHERE_FIELD_NUMBER: builtins.int
672
- INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
673
- CLOSE_CONNECTION_FIELD_NUMBER: builtins.int
674
- SHARE_CONNECTION_FIELD_NUMBER: builtins.int
675
- db: builtins.str
676
- """Name of the MySQL database to connect to."""
677
- table: builtins.str
678
- """Name of the database table to use as a ClickHouse dictionary."""
679
- port: builtins.int
680
- """Default port to use when connecting to a replica of the dictionary source."""
681
- user: builtins.str
682
- """Name of the default user for replicas of the dictionary source."""
683
- password: builtins.str
684
- """Password of the default user for replicas of the dictionary source."""
685
- where: builtins.str
686
- """Selection criteria for the data in the specified MySQL table."""
687
- invalidate_query: builtins.str
688
- """Query for checking the dictionary status, to pull only updated data.
689
- For more details, see [ClickHouse documentation on dictionaries](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_lifetime/).
690
- """
691
- @property
692
- def replicas(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica]:
693
- """List of MySQL replicas of the database used as dictionary source."""
694
-
695
- @property
696
- def close_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
697
- """Should the connection be closed after each request."""
698
-
699
- @property
700
- def share_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
701
- """Should a connection be shared for some requests."""
702
-
703
- def __init__(
704
- self,
705
- *,
706
- db: builtins.str = ...,
707
- table: builtins.str = ...,
708
- port: builtins.int = ...,
709
- user: builtins.str = ...,
710
- password: builtins.str = ...,
711
- replicas: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica] | None = ...,
712
- where: builtins.str = ...,
713
- invalidate_query: builtins.str = ...,
714
- close_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
715
- share_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
716
- ) -> None: ...
717
- def HasField(self, field_name: typing.Literal["close_connection", b"close_connection", "share_connection", b"share_connection"]) -> builtins.bool: ...
718
- def ClearField(self, field_name: typing.Literal["close_connection", b"close_connection", "db", b"db", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "replicas", b"replicas", "share_connection", b"share_connection", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
719
-
720
- @typing.final
721
- class ClickhouseSource(google.protobuf.message.Message):
722
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
723
-
724
- DB_FIELD_NUMBER: builtins.int
725
- TABLE_FIELD_NUMBER: builtins.int
726
- HOST_FIELD_NUMBER: builtins.int
727
- PORT_FIELD_NUMBER: builtins.int
728
- USER_FIELD_NUMBER: builtins.int
729
- PASSWORD_FIELD_NUMBER: builtins.int
730
- WHERE_FIELD_NUMBER: builtins.int
731
- SECURE_FIELD_NUMBER: builtins.int
732
- db: builtins.str
733
- """Name of the ClickHouse database."""
734
- table: builtins.str
735
- """Name of the table in the specified database to be used as the dictionary source."""
736
- host: builtins.str
737
- """ClickHouse host of the specified database."""
738
- port: builtins.int
739
- """Port to use when connecting to the host."""
740
- user: builtins.str
741
- """Name of the ClickHouse database user."""
742
- password: builtins.str
743
- """Password of the ClickHouse database user."""
744
- where: builtins.str
745
- """Selection criteria for the data in the specified ClickHouse table."""
746
- @property
747
- def secure(self) -> google.protobuf.wrappers_pb2.BoolValue:
748
- """Use ssl for connection."""
749
-
750
- def __init__(
751
- self,
752
- *,
753
- db: builtins.str = ...,
754
- table: builtins.str = ...,
755
- host: builtins.str = ...,
756
- port: builtins.int = ...,
757
- user: builtins.str = ...,
758
- password: builtins.str = ...,
759
- where: builtins.str = ...,
760
- secure: google.protobuf.wrappers_pb2.BoolValue | None = ...,
761
- ) -> None: ...
762
- def HasField(self, field_name: typing.Literal["secure", b"secure"]) -> builtins.bool: ...
763
- def ClearField(self, field_name: typing.Literal["db", b"db", "host", b"host", "password", b"password", "port", b"port", "secure", b"secure", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
764
-
765
- @typing.final
766
- class MongodbSource(google.protobuf.message.Message):
767
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
768
-
769
- DB_FIELD_NUMBER: builtins.int
770
- COLLECTION_FIELD_NUMBER: builtins.int
771
- HOST_FIELD_NUMBER: builtins.int
772
- PORT_FIELD_NUMBER: builtins.int
773
- USER_FIELD_NUMBER: builtins.int
774
- PASSWORD_FIELD_NUMBER: builtins.int
775
- OPTIONS_FIELD_NUMBER: builtins.int
776
- db: builtins.str
777
- """Name of the MongoDB database."""
778
- collection: builtins.str
779
- """Name of the collection in the specified database to be used as the dictionary source."""
780
- host: builtins.str
781
- """MongoDB host of the specified database."""
782
- port: builtins.int
783
- """Port to use when connecting to the host."""
784
- user: builtins.str
785
- """Name of the MongoDB database user."""
786
- password: builtins.str
787
- """Password of the MongoDB database user."""
788
- options: builtins.str
789
- def __init__(
790
- self,
791
- *,
792
- db: builtins.str = ...,
793
- collection: builtins.str = ...,
794
- host: builtins.str = ...,
795
- port: builtins.int = ...,
796
- user: builtins.str = ...,
797
- password: builtins.str = ...,
798
- options: builtins.str = ...,
799
- ) -> None: ...
800
- def ClearField(self, field_name: typing.Literal["collection", b"collection", "db", b"db", "host", b"host", "options", b"options", "password", b"password", "port", b"port", "user", b"user"]) -> None: ...
801
-
802
- @typing.final
803
- class PostgresqlSource(google.protobuf.message.Message):
804
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
805
-
806
- class _SslMode:
807
- ValueType = typing.NewType("ValueType", builtins.int)
808
- V: typing_extensions.TypeAlias = ValueType
809
-
810
- class _SslModeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType], builtins.type):
811
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
812
- SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 0
813
- DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 1
814
- """Only try a non-SSL connection."""
815
- ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 2
816
- """First try a non-SSL connection; if that fails, try an SSL connection."""
817
- PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 3
818
- """First try an SSL connection; if that fails, try a non-SSL connection."""
819
- VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 4
820
- """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
821
- VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 5
822
- """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
823
-
824
- class SslMode(_SslMode, metaclass=_SslModeEnumTypeWrapper): ...
825
- SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 0
826
- DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 1
827
- """Only try a non-SSL connection."""
828
- ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 2
829
- """First try a non-SSL connection; if that fails, try an SSL connection."""
830
- PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 3
831
- """First try an SSL connection; if that fails, try a non-SSL connection."""
832
- VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 4
833
- """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
834
- VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 5
835
- """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
836
-
837
- DB_FIELD_NUMBER: builtins.int
838
- TABLE_FIELD_NUMBER: builtins.int
839
- HOSTS_FIELD_NUMBER: builtins.int
840
- PORT_FIELD_NUMBER: builtins.int
841
- USER_FIELD_NUMBER: builtins.int
842
- PASSWORD_FIELD_NUMBER: builtins.int
843
- INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
844
- SSL_MODE_FIELD_NUMBER: builtins.int
845
- db: builtins.str
846
- """Name of the PostrgreSQL database."""
847
- table: builtins.str
848
- """Name of the table in the specified database to be used as the dictionary source."""
849
- port: builtins.int
850
- """Port to use when connecting to the host."""
851
- user: builtins.str
852
- """Name of the PostrgreSQL database user."""
853
- password: builtins.str
854
- """Password of the PostrgreSQL database user."""
855
- invalidate_query: builtins.str
856
- """Query for checking the dictionary status, to pull only updated data.
857
- For more details, see [ClickHouse documentation on dictionaries](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_lifetime/).
858
- """
859
- ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType
860
- """Mode of SSL TCP/IP connection to the PostgreSQL host.
861
- For more details, see [PostgreSQL documentation](https://www.postgresql.org/docs/current/libpq-ssl.html).
862
- """
863
- @property
864
- def hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
865
- """Name of the PostrgreSQL host"""
866
-
867
- def __init__(
868
- self,
869
- *,
870
- db: builtins.str = ...,
871
- table: builtins.str = ...,
872
- hosts: collections.abc.Iterable[builtins.str] | None = ...,
873
- port: builtins.int = ...,
874
- user: builtins.str = ...,
875
- password: builtins.str = ...,
876
- invalidate_query: builtins.str = ...,
877
- ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType = ...,
878
- ) -> None: ...
879
- def ClearField(self, field_name: typing.Literal["db", b"db", "hosts", b"hosts", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "ssl_mode", b"ssl_mode", "table", b"table", "user", b"user"]) -> None: ...
880
-
881
- @typing.final
882
- class Structure(google.protobuf.message.Message):
883
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
646
+ def ClearField(self, field_name: typing.Literal["attributes", b"attributes"]) -> None: ...
884
647
 
885
648
  @typing.final
886
649
  class Attribute(google.protobuf.message.Message):
@@ -902,11 +665,13 @@ class ClickhouseConfig(google.protobuf.message.Message):
902
665
  """Expression, describing the attribute, if applicable."""
903
666
  hierarchical: builtins.bool
904
667
  """Indication of hierarchy support.
905
- Default value: `false`.
668
+
669
+ Default value: **false**.
906
670
  """
907
671
  injective: builtins.bool
908
672
  """Indication of injective mapping "id -> attribute".
909
- Default value: `false`.
673
+
674
+ Default value: **false**.
910
675
  """
911
676
  def __init__(
912
677
  self,
@@ -920,40 +685,6 @@ class ClickhouseConfig(google.protobuf.message.Message):
920
685
  ) -> None: ...
921
686
  def ClearField(self, field_name: typing.Literal["expression", b"expression", "hierarchical", b"hierarchical", "injective", b"injective", "name", b"name", "null_value", b"null_value", "type", b"type"]) -> None: ...
922
687
 
923
- @typing.final
924
- class Id(google.protobuf.message.Message):
925
- """Numeric key."""
926
-
927
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
928
-
929
- NAME_FIELD_NUMBER: builtins.int
930
- name: builtins.str
931
- """Name of the numeric key."""
932
- def __init__(
933
- self,
934
- *,
935
- name: builtins.str = ...,
936
- ) -> None: ...
937
- def ClearField(self, field_name: typing.Literal["name", b"name"]) -> None: ...
938
-
939
- @typing.final
940
- class Key(google.protobuf.message.Message):
941
- """Complex key."""
942
-
943
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
944
-
945
- ATTRIBUTES_FIELD_NUMBER: builtins.int
946
- @property
947
- def attributes(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute]:
948
- """Attributes of a complex key."""
949
-
950
- def __init__(
951
- self,
952
- *,
953
- attributes: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute] | None = ...,
954
- ) -> None: ...
955
- def ClearField(self, field_name: typing.Literal["attributes", b"attributes"]) -> None: ...
956
-
957
688
  ID_FIELD_NUMBER: builtins.int
958
689
  KEY_FIELD_NUMBER: builtins.int
959
690
  RANGE_MIN_FIELD_NUMBER: builtins.int
@@ -966,24 +697,28 @@ class ClickhouseConfig(google.protobuf.message.Message):
966
697
  @property
967
698
  def key(self) -> global___ClickhouseConfig.ExternalDictionary.Structure.Key:
968
699
  """Composite key for the dictionary, containing of one or more key columns.
700
+
969
701
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_structure/#composite-key).
970
702
  """
971
703
 
972
704
  @property
973
705
  def range_min(self) -> global___ClickhouseConfig.ExternalDictionary.Structure.Attribute:
974
- """Field holding the beginning of the range for dictionaries with `RANGE_HASHED` layout.
706
+ """Field holding the beginning of the range for dictionaries with **RANGE_HASHED** layout.
707
+
975
708
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_layout/#range-hashed).
976
709
  """
977
710
 
978
711
  @property
979
712
  def range_max(self) -> global___ClickhouseConfig.ExternalDictionary.Structure.Attribute:
980
- """Field holding the end of the range for dictionaries with `RANGE_HASHED` layout.
713
+ """Field holding the end of the range for dictionaries with **RANGE_HASHED** layout.
714
+
981
715
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_layout/#range-hashed).
982
716
  """
983
717
 
984
718
  @property
985
719
  def attributes(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute]:
986
720
  """Description of the fields available for database queries.
721
+
987
722
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_structure/#attributes).
988
723
  """
989
724
 
@@ -1001,8 +736,6 @@ class ClickhouseConfig(google.protobuf.message.Message):
1001
736
 
1002
737
  @typing.final
1003
738
  class Layout(google.protobuf.message.Message):
1004
- """Layout determining how to store the dictionary in memory."""
1005
-
1006
739
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1007
740
 
1008
741
  class _Type:
@@ -1013,80 +746,106 @@ class ClickhouseConfig(google.protobuf.message.Message):
1013
746
  DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1014
747
  TYPE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 0
1015
748
  FLAT: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 1
1016
- """The entire dictionary is stored in memory in the form of flat arrays.
1017
- Available for all dictionary sources.
749
+ """The dictionary is completely stored in memory in the form of flat arrays.
750
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1018
751
  """
1019
752
  HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 2
1020
- """The entire dictionary is stored in memory in the form of a hash table.
1021
- Available for all dictionary sources.
753
+ """The dictionary is completely stored in memory in the form of a hash table.
754
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1022
755
  """
1023
756
  COMPLEX_KEY_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 3
1024
- """Similar to HASHED, to be used with composite keys.
1025
- Available for all dictionary sources.
757
+ """The dictionary is completely stored in memory in the form of a hash table.
758
+ Applicable for dictionaries with composite keys of arbitrary type.
1026
759
  """
1027
760
  RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 4
1028
- """The entire dictionary is stored in memory in the form of a hash table,
1029
- with an ordered array of ranges and their corresponding values.
1030
- Available for all dictionary sources.
761
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
762
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1031
763
  """
1032
764
  CACHE: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 5
1033
- """The dictionary is stored in a cache with a set number of cells.
1034
- Available for MySQL, ClickHouse and HTTP dictionary sources.
765
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
766
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1035
767
  """
1036
768
  COMPLEX_KEY_CACHE: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 6
1037
- """Similar to CACHE, to be used with composite keys.
1038
- Available for MySQL, ClickHouse and HTTP dictionary sources.
769
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
770
+ Applicable for dictionaries with composite keys of arbitrary type.
1039
771
  """
1040
772
  SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 7
1041
- """Similar to HASHED, but uses less memory in favor of more CPU usage."""
773
+ """The dictionary is completely stored in memory in the form of a hash table.
774
+ It's similar to HASHED layout type but uses less memory in favor of more CPU usage.
775
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
776
+ """
1042
777
  COMPLEX_KEY_SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 8
1043
- """Similar to SPARSE_HASHED, to be used with composite keys."""
778
+ """The dictionary is completely stored in memory in the form of a hash table.
779
+ It's similar to COMPLEX_KEY_HASHED layout type but uses less memory in favor of more CPU usage.
780
+ Applicable for dictionaries with composite keys of arbitrary type.
781
+ """
1044
782
  COMPLEX_KEY_RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 9
1045
- """Similar to RANGE_HASHED, to be used with composite keys."""
783
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
784
+ Applicable for dictionaries with composite keys of arbitrary type.
785
+ """
1046
786
  DIRECT: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 10
1047
- """The dictionary is not stored in memory and directly goes to the source during the processing of a request."""
787
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
788
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
789
+ """
1048
790
  COMPLEX_KEY_DIRECT: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 11
1049
- """Similar to DIRECT, to be used with composite keys."""
791
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
792
+ Applicable for dictionaries with composite keys of arbitrary type.
793
+ """
1050
794
  IP_TRIE: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 12
1051
795
  """The specialized layout type for mapping network prefixes (IP addresses) to metadata such as ASN."""
1052
796
 
1053
- class Type(_Type, metaclass=_TypeEnumTypeWrapper): ...
797
+ class Type(_Type, metaclass=_TypeEnumTypeWrapper):
798
+ """Layout type.
799
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/sql-reference/dictionaries#ways-to-store-dictionaries-in-memory).
800
+ """
801
+
1054
802
  TYPE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 0
1055
803
  FLAT: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 1
1056
- """The entire dictionary is stored in memory in the form of flat arrays.
1057
- Available for all dictionary sources.
804
+ """The dictionary is completely stored in memory in the form of flat arrays.
805
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1058
806
  """
1059
807
  HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 2
1060
- """The entire dictionary is stored in memory in the form of a hash table.
1061
- Available for all dictionary sources.
808
+ """The dictionary is completely stored in memory in the form of a hash table.
809
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1062
810
  """
1063
811
  COMPLEX_KEY_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 3
1064
- """Similar to HASHED, to be used with composite keys.
1065
- Available for all dictionary sources.
812
+ """The dictionary is completely stored in memory in the form of a hash table.
813
+ Applicable for dictionaries with composite keys of arbitrary type.
1066
814
  """
1067
815
  RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 4
1068
- """The entire dictionary is stored in memory in the form of a hash table,
1069
- with an ordered array of ranges and their corresponding values.
1070
- Available for all dictionary sources.
816
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
817
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1071
818
  """
1072
819
  CACHE: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 5
1073
- """The dictionary is stored in a cache with a set number of cells.
1074
- Available for MySQL, ClickHouse and HTTP dictionary sources.
820
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
821
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1075
822
  """
1076
823
  COMPLEX_KEY_CACHE: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 6
1077
- """Similar to CACHE, to be used with composite keys.
1078
- Available for MySQL, ClickHouse and HTTP dictionary sources.
824
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
825
+ Applicable for dictionaries with composite keys of arbitrary type.
1079
826
  """
1080
827
  SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 7
1081
- """Similar to HASHED, but uses less memory in favor of more CPU usage."""
828
+ """The dictionary is completely stored in memory in the form of a hash table.
829
+ It's similar to HASHED layout type but uses less memory in favor of more CPU usage.
830
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
831
+ """
1082
832
  COMPLEX_KEY_SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 8
1083
- """Similar to SPARSE_HASHED, to be used with composite keys."""
833
+ """The dictionary is completely stored in memory in the form of a hash table.
834
+ It's similar to COMPLEX_KEY_HASHED layout type but uses less memory in favor of more CPU usage.
835
+ Applicable for dictionaries with composite keys of arbitrary type.
836
+ """
1084
837
  COMPLEX_KEY_RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 9
1085
- """Similar to RANGE_HASHED, to be used with composite keys."""
838
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
839
+ Applicable for dictionaries with composite keys of arbitrary type.
840
+ """
1086
841
  DIRECT: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 10
1087
- """The dictionary is not stored in memory and directly goes to the source during the processing of a request."""
842
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
843
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
844
+ """
1088
845
  COMPLEX_KEY_DIRECT: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 11
1089
- """Similar to DIRECT, to be used with composite keys."""
846
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
847
+ Applicable for dictionaries with composite keys of arbitrary type.
848
+ """
1090
849
  IP_TRIE: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 12
1091
850
  """The specialized layout type for mapping network prefixes (IP addresses) to metadata such as ASN."""
1092
851
 
@@ -1101,46 +860,83 @@ class ClickhouseConfig(google.protobuf.message.Message):
1101
860
  MAX_ARRAY_SIZE_FIELD_NUMBER: builtins.int
1102
861
  ACCESS_TO_KEY_FROM_ATTRIBUTES_FIELD_NUMBER: builtins.int
1103
862
  type: global___ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType
1104
- """Layout type for an external dictionary."""
863
+ """Layout type.
864
+
865
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#ways-to-store-dictionaries-in-memory).
866
+ """
1105
867
  size_in_cells: builtins.int
1106
868
  """Number of cells in the cache. Rounded up to a power of two.
1107
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
869
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
870
+
871
+ Default value: **1000000000**.
872
+
873
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1108
874
  """
1109
875
  max_update_queue_size: builtins.int
1110
876
  """Max size of update queue.
1111
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
877
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
878
+
879
+ Default value: **100000**.
880
+
881
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1112
882
  """
1113
883
  update_queue_push_timeout_milliseconds: builtins.int
1114
884
  """Max timeout in milliseconds for push update task into queue.
1115
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
885
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
886
+
887
+ Default value: **10**.
888
+
889
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1116
890
  """
1117
891
  query_wait_timeout_milliseconds: builtins.int
1118
892
  """Max wait timeout in milliseconds for update task to complete.
1119
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
893
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
894
+
895
+ Default value: **60000** (1 minute).
896
+
897
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1120
898
  """
1121
899
  max_threads_for_updates: builtins.int
1122
900
  """Max threads for cache dictionary update.
1123
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
901
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
902
+
903
+ Default value: **4**.
904
+
905
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1124
906
  """
1125
907
  initial_array_size: builtins.int
1126
908
  """Initial dictionary key size.
1127
- Applicable only for FLAT layout type.
909
+ Applicable only for **FLAT** layout type.
910
+
911
+ Default value: **1024**.
912
+
913
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#flat).
1128
914
  """
1129
915
  max_array_size: builtins.int
1130
916
  """Maximum dictionary key size.
1131
- Applicable only for FLAT layout type.
917
+ Applicable only for **FLAT** layout type.
918
+
919
+ Default value: **500000**.
920
+
921
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#flat).
1132
922
  """
1133
923
  @property
1134
924
  def allow_read_expired_keys(self) -> google.protobuf.wrappers_pb2.BoolValue:
1135
925
  """Allows to read expired keys.
1136
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
926
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
927
+
928
+ Default value: **false**.
929
+
930
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1137
931
  """
1138
932
 
1139
933
  @property
1140
934
  def access_to_key_from_attributes(self) -> google.protobuf.wrappers_pb2.BoolValue:
1141
- """Allows to retrieve key attribute using dictGetString function.
935
+ """Allows to retrieve key attribute using **dictGetString** function.
1142
936
  Enabling this option increases memory usage.
1143
- Applicable only for IP_TRIE layout type.
937
+ Applicable only for **IP_TRIE** layout type.
938
+
939
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#ip_trie).
1144
940
  """
1145
941
 
1146
942
  def __init__(
@@ -1164,19 +960,318 @@ class ClickhouseConfig(google.protobuf.message.Message):
1164
960
  class Range(google.protobuf.message.Message):
1165
961
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1166
962
 
1167
- MIN_FIELD_NUMBER: builtins.int
1168
- MAX_FIELD_NUMBER: builtins.int
1169
- min: builtins.int
1170
- """Minimum dictionary lifetime."""
1171
- max: builtins.int
1172
- """Maximum dictionary lifetime."""
963
+ MIN_FIELD_NUMBER: builtins.int
964
+ MAX_FIELD_NUMBER: builtins.int
965
+ min: builtins.int
966
+ """Minimum dictionary lifetime."""
967
+ max: builtins.int
968
+ """Maximum dictionary lifetime."""
969
+ def __init__(
970
+ self,
971
+ *,
972
+ min: builtins.int = ...,
973
+ max: builtins.int = ...,
974
+ ) -> None: ...
975
+ def ClearField(self, field_name: typing.Literal["max", b"max", "min", b"min"]) -> None: ...
976
+
977
+ @typing.final
978
+ class HttpSource(google.protobuf.message.Message):
979
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
980
+
981
+ @typing.final
982
+ class Header(google.protobuf.message.Message):
983
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
984
+
985
+ NAME_FIELD_NUMBER: builtins.int
986
+ VALUE_FIELD_NUMBER: builtins.int
987
+ name: builtins.str
988
+ """Header name."""
989
+ value: builtins.str
990
+ """Header value."""
991
+ def __init__(
992
+ self,
993
+ *,
994
+ name: builtins.str = ...,
995
+ value: builtins.str = ...,
996
+ ) -> None: ...
997
+ def ClearField(self, field_name: typing.Literal["name", b"name", "value", b"value"]) -> None: ...
998
+
999
+ URL_FIELD_NUMBER: builtins.int
1000
+ FORMAT_FIELD_NUMBER: builtins.int
1001
+ HEADERS_FIELD_NUMBER: builtins.int
1002
+ url: builtins.str
1003
+ """URL of the source dictionary available over HTTP."""
1004
+ format: builtins.str
1005
+ """The data format. Valid values are all formats [supported by ClickHouse SQL dialect](https://clickhouse.com/docs/en/interfaces/formats/)."""
1006
+ @property
1007
+ def headers(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header]:
1008
+ """HTTP headers."""
1009
+
1010
+ def __init__(
1011
+ self,
1012
+ *,
1013
+ url: builtins.str = ...,
1014
+ format: builtins.str = ...,
1015
+ headers: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header] | None = ...,
1016
+ ) -> None: ...
1017
+ def ClearField(self, field_name: typing.Literal["format", b"format", "headers", b"headers", "url", b"url"]) -> None: ...
1018
+
1019
+ @typing.final
1020
+ class MysqlSource(google.protobuf.message.Message):
1021
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1022
+
1023
+ @typing.final
1024
+ class Replica(google.protobuf.message.Message):
1025
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1026
+
1027
+ HOST_FIELD_NUMBER: builtins.int
1028
+ PRIORITY_FIELD_NUMBER: builtins.int
1029
+ PORT_FIELD_NUMBER: builtins.int
1030
+ USER_FIELD_NUMBER: builtins.int
1031
+ PASSWORD_FIELD_NUMBER: builtins.int
1032
+ host: builtins.str
1033
+ """MySQL host of the replica."""
1034
+ priority: builtins.int
1035
+ """The priority of the replica that ClickHouse takes into account when connecting.
1036
+ Replica with the highest priority should have this field set to the lowest number.
1037
+ """
1038
+ port: builtins.int
1039
+ """Port to use when connecting to the replica.
1040
+ If a port is not specified for a replica, ClickHouse uses the port specified for the source.
1041
+ """
1042
+ user: builtins.str
1043
+ """Name of the MySQL database user.
1044
+ If a user is not specified for a replica, ClickHouse uses the user specified for the source.
1045
+ """
1046
+ password: builtins.str
1047
+ """Password of the MySQL database user.
1048
+ If a password is not specified for a replica, ClickHouse uses the password specified for the source.
1049
+ """
1050
+ def __init__(
1051
+ self,
1052
+ *,
1053
+ host: builtins.str = ...,
1054
+ priority: builtins.int = ...,
1055
+ port: builtins.int = ...,
1056
+ user: builtins.str = ...,
1057
+ password: builtins.str = ...,
1058
+ ) -> None: ...
1059
+ def ClearField(self, field_name: typing.Literal["host", b"host", "password", b"password", "port", b"port", "priority", b"priority", "user", b"user"]) -> None: ...
1060
+
1061
+ DB_FIELD_NUMBER: builtins.int
1062
+ TABLE_FIELD_NUMBER: builtins.int
1063
+ PORT_FIELD_NUMBER: builtins.int
1064
+ USER_FIELD_NUMBER: builtins.int
1065
+ PASSWORD_FIELD_NUMBER: builtins.int
1066
+ REPLICAS_FIELD_NUMBER: builtins.int
1067
+ WHERE_FIELD_NUMBER: builtins.int
1068
+ INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
1069
+ CLOSE_CONNECTION_FIELD_NUMBER: builtins.int
1070
+ SHARE_CONNECTION_FIELD_NUMBER: builtins.int
1071
+ db: builtins.str
1072
+ """Database name."""
1073
+ table: builtins.str
1074
+ """Table name."""
1075
+ port: builtins.int
1076
+ """Port to use when connecting to a replica of the dictionary source."""
1077
+ user: builtins.str
1078
+ """Name of the user for replicas of the dictionary source."""
1079
+ password: builtins.str
1080
+ """Password of the user for replicas of the dictionary source."""
1081
+ where: builtins.str
1082
+ """Selection criteria for the data in the specified MySQL table."""
1083
+ invalidate_query: builtins.str
1084
+ """Query for checking the dictionary status, to pull only updated data."""
1085
+ @property
1086
+ def replicas(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica]:
1087
+ """List of MySQL replicas of the database used as dictionary source."""
1088
+
1089
+ @property
1090
+ def close_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
1091
+ """Should a connection be closed after each request."""
1092
+
1093
+ @property
1094
+ def share_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
1095
+ """Should a connection be shared for some requests."""
1096
+
1097
+ def __init__(
1098
+ self,
1099
+ *,
1100
+ db: builtins.str = ...,
1101
+ table: builtins.str = ...,
1102
+ port: builtins.int = ...,
1103
+ user: builtins.str = ...,
1104
+ password: builtins.str = ...,
1105
+ replicas: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica] | None = ...,
1106
+ where: builtins.str = ...,
1107
+ invalidate_query: builtins.str = ...,
1108
+ close_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1109
+ share_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1110
+ ) -> None: ...
1111
+ def HasField(self, field_name: typing.Literal["close_connection", b"close_connection", "share_connection", b"share_connection"]) -> builtins.bool: ...
1112
+ def ClearField(self, field_name: typing.Literal["close_connection", b"close_connection", "db", b"db", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "replicas", b"replicas", "share_connection", b"share_connection", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
1113
+
1114
+ @typing.final
1115
+ class ClickhouseSource(google.protobuf.message.Message):
1116
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1117
+
1118
+ DB_FIELD_NUMBER: builtins.int
1119
+ TABLE_FIELD_NUMBER: builtins.int
1120
+ HOST_FIELD_NUMBER: builtins.int
1121
+ PORT_FIELD_NUMBER: builtins.int
1122
+ USER_FIELD_NUMBER: builtins.int
1123
+ PASSWORD_FIELD_NUMBER: builtins.int
1124
+ WHERE_FIELD_NUMBER: builtins.int
1125
+ SECURE_FIELD_NUMBER: builtins.int
1126
+ db: builtins.str
1127
+ """Database name."""
1128
+ table: builtins.str
1129
+ """Table name."""
1130
+ host: builtins.str
1131
+ """ClickHouse host."""
1132
+ port: builtins.int
1133
+ """Port to use when connecting to the host."""
1134
+ user: builtins.str
1135
+ """Name of the ClickHouse database user."""
1136
+ password: builtins.str
1137
+ """Password of the ClickHouse database user."""
1138
+ where: builtins.str
1139
+ """Selection criteria for the data in the specified ClickHouse table."""
1140
+ @property
1141
+ def secure(self) -> google.protobuf.wrappers_pb2.BoolValue:
1142
+ """Determines whether to use TLS for connection."""
1143
+
1144
+ def __init__(
1145
+ self,
1146
+ *,
1147
+ db: builtins.str = ...,
1148
+ table: builtins.str = ...,
1149
+ host: builtins.str = ...,
1150
+ port: builtins.int = ...,
1151
+ user: builtins.str = ...,
1152
+ password: builtins.str = ...,
1153
+ where: builtins.str = ...,
1154
+ secure: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1155
+ ) -> None: ...
1156
+ def HasField(self, field_name: typing.Literal["secure", b"secure"]) -> builtins.bool: ...
1157
+ def ClearField(self, field_name: typing.Literal["db", b"db", "host", b"host", "password", b"password", "port", b"port", "secure", b"secure", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
1158
+
1159
+ @typing.final
1160
+ class MongodbSource(google.protobuf.message.Message):
1161
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1162
+
1163
+ DB_FIELD_NUMBER: builtins.int
1164
+ COLLECTION_FIELD_NUMBER: builtins.int
1165
+ HOST_FIELD_NUMBER: builtins.int
1166
+ PORT_FIELD_NUMBER: builtins.int
1167
+ USER_FIELD_NUMBER: builtins.int
1168
+ PASSWORD_FIELD_NUMBER: builtins.int
1169
+ OPTIONS_FIELD_NUMBER: builtins.int
1170
+ db: builtins.str
1171
+ """Database name."""
1172
+ collection: builtins.str
1173
+ """Collection name."""
1174
+ host: builtins.str
1175
+ """MongoDB host."""
1176
+ port: builtins.int
1177
+ """Port to use when connecting to the host."""
1178
+ user: builtins.str
1179
+ """Name of the MongoDB database user."""
1180
+ password: builtins.str
1181
+ """Password of the MongoDB database user."""
1182
+ options: builtins.str
1183
+ """Dictionary source options."""
1184
+ def __init__(
1185
+ self,
1186
+ *,
1187
+ db: builtins.str = ...,
1188
+ collection: builtins.str = ...,
1189
+ host: builtins.str = ...,
1190
+ port: builtins.int = ...,
1191
+ user: builtins.str = ...,
1192
+ password: builtins.str = ...,
1193
+ options: builtins.str = ...,
1194
+ ) -> None: ...
1195
+ def ClearField(self, field_name: typing.Literal["collection", b"collection", "db", b"db", "host", b"host", "options", b"options", "password", b"password", "port", b"port", "user", b"user"]) -> None: ...
1196
+
1197
+ @typing.final
1198
+ class PostgresqlSource(google.protobuf.message.Message):
1199
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1200
+
1201
+ class _SslMode:
1202
+ ValueType = typing.NewType("ValueType", builtins.int)
1203
+ V: typing_extensions.TypeAlias = ValueType
1204
+
1205
+ class _SslModeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType], builtins.type):
1206
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1207
+ SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 0
1208
+ DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 1
1209
+ """Only try a non-SSL connection."""
1210
+ ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 2
1211
+ """First try a non-SSL connection; if that fails, try an SSL connection."""
1212
+ PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 3
1213
+ """First try an SSL connection; if that fails, try a non-SSL connection."""
1214
+ VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 4
1215
+ """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
1216
+ VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 5
1217
+ """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
1218
+
1219
+ class SslMode(_SslMode, metaclass=_SslModeEnumTypeWrapper):
1220
+ """Mode of SSL TCP/IP connection to a PostgreSQL host.
1221
+ For details, see [PostgreSQL documentation](https://www.postgresql.org/docs/current/libpq-ssl.html).
1222
+ """
1223
+
1224
+ SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 0
1225
+ DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 1
1226
+ """Only try a non-SSL connection."""
1227
+ ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 2
1228
+ """First try a non-SSL connection; if that fails, try an SSL connection."""
1229
+ PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 3
1230
+ """First try an SSL connection; if that fails, try a non-SSL connection."""
1231
+ VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 4
1232
+ """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
1233
+ VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 5
1234
+ """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
1235
+
1236
+ DB_FIELD_NUMBER: builtins.int
1237
+ TABLE_FIELD_NUMBER: builtins.int
1238
+ HOSTS_FIELD_NUMBER: builtins.int
1239
+ PORT_FIELD_NUMBER: builtins.int
1240
+ USER_FIELD_NUMBER: builtins.int
1241
+ PASSWORD_FIELD_NUMBER: builtins.int
1242
+ INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
1243
+ SSL_MODE_FIELD_NUMBER: builtins.int
1244
+ db: builtins.str
1245
+ """Database name."""
1246
+ table: builtins.str
1247
+ """Table name."""
1248
+ port: builtins.int
1249
+ """Port to use when connecting to the PostgreSQL hosts."""
1250
+ user: builtins.str
1251
+ """Name of the PostrgreSQL database user."""
1252
+ password: builtins.str
1253
+ """Password of the PostrgreSQL database user."""
1254
+ invalidate_query: builtins.str
1255
+ """Query for checking the dictionary status, to pull only updated data."""
1256
+ ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType
1257
+ """Mode of SSL TCP/IP connection to the PostgreSQL host."""
1258
+ @property
1259
+ def hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
1260
+ """PostgreSQL hosts."""
1261
+
1173
1262
  def __init__(
1174
1263
  self,
1175
1264
  *,
1176
- min: builtins.int = ...,
1177
- max: builtins.int = ...,
1265
+ db: builtins.str = ...,
1266
+ table: builtins.str = ...,
1267
+ hosts: collections.abc.Iterable[builtins.str] | None = ...,
1268
+ port: builtins.int = ...,
1269
+ user: builtins.str = ...,
1270
+ password: builtins.str = ...,
1271
+ invalidate_query: builtins.str = ...,
1272
+ ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType = ...,
1178
1273
  ) -> None: ...
1179
- def ClearField(self, field_name: typing.Literal["max", b"max", "min", b"min"]) -> None: ...
1274
+ def ClearField(self, field_name: typing.Literal["db", b"db", "hosts", b"hosts", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "ssl_mode", b"ssl_mode", "table", b"table", "user", b"user"]) -> None: ...
1180
1275
 
1181
1276
  NAME_FIELD_NUMBER: builtins.int
1182
1277
  STRUCTURE_FIELD_NUMBER: builtins.int
@@ -1194,14 +1289,13 @@ class ClickhouseConfig(google.protobuf.message.Message):
1194
1289
  """Fixed interval between dictionary updates."""
1195
1290
  @property
1196
1291
  def structure(self) -> global___ClickhouseConfig.ExternalDictionary.Structure:
1197
- """Set of attributes for the external dictionary.
1198
- For in-depth description, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_structure/).
1199
- """
1292
+ """Structure of the external dictionary."""
1200
1293
 
1201
1294
  @property
1202
1295
  def layout(self) -> global___ClickhouseConfig.ExternalDictionary.Layout:
1203
- """Layout for storing the dictionary in memory.
1204
- For in-depth description, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_layout/).
1296
+ """Layout determining how to store the dictionary in memory.
1297
+
1298
+ For details, see https://clickhouse.com/docs/sql-reference/dictionaries#ways-to-store-dictionaries-in-memory.
1205
1299
  """
1206
1300
 
1207
1301
  @property
@@ -1249,97 +1343,343 @@ class ClickhouseConfig(google.protobuf.message.Message):
1249
1343
  @typing.overload
1250
1344
  def WhichOneof(self, oneof_group: typing.Literal["source", b"source"]) -> typing.Literal["http_source", "mysql_source", "clickhouse_source", "mongodb_source", "postgresql_source"] | None: ...
1251
1345
 
1252
- @typing.final
1253
- class GraphiteRollup(google.protobuf.message.Message):
1254
- """Rollup settings for the GraphiteMergeTree table engine."""
1346
+ @typing.final
1347
+ class GraphiteRollup(google.protobuf.message.Message):
1348
+ """Rollup settings for the GraphiteMergeTree table engine.
1349
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#graphite-rollup).
1350
+ """
1351
+
1352
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1353
+
1354
+ @typing.final
1355
+ class Pattern(google.protobuf.message.Message):
1356
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1357
+
1358
+ @typing.final
1359
+ class Retention(google.protobuf.message.Message):
1360
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1361
+
1362
+ AGE_FIELD_NUMBER: builtins.int
1363
+ PRECISION_FIELD_NUMBER: builtins.int
1364
+ age: builtins.int
1365
+ """The minimum age of the data in seconds."""
1366
+ precision: builtins.int
1367
+ """Precision of determining the age of the data, in seconds. Should be a divisor for 86400 (seconds in a day)."""
1368
+ def __init__(
1369
+ self,
1370
+ *,
1371
+ age: builtins.int = ...,
1372
+ precision: builtins.int = ...,
1373
+ ) -> None: ...
1374
+ def ClearField(self, field_name: typing.Literal["age", b"age", "precision", b"precision"]) -> None: ...
1375
+
1376
+ REGEXP_FIELD_NUMBER: builtins.int
1377
+ FUNCTION_FIELD_NUMBER: builtins.int
1378
+ RETENTION_FIELD_NUMBER: builtins.int
1379
+ regexp: builtins.str
1380
+ """A pattern for the metric name (a regular or DSL)."""
1381
+ function: builtins.str
1382
+ """The name of the aggregating function to apply to data whose age falls within the range [age, age + precision].
1383
+ Accepted functions: **min**, **max**, **any**, **avg**. The average is calculated imprecisely, like the average of the averages.
1384
+ """
1385
+ @property
1386
+ def retention(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention]:
1387
+ """Retention rules."""
1388
+
1389
+ def __init__(
1390
+ self,
1391
+ *,
1392
+ regexp: builtins.str = ...,
1393
+ function: builtins.str = ...,
1394
+ retention: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention] | None = ...,
1395
+ ) -> None: ...
1396
+ def ClearField(self, field_name: typing.Literal["function", b"function", "regexp", b"regexp", "retention", b"retention"]) -> None: ...
1397
+
1398
+ NAME_FIELD_NUMBER: builtins.int
1399
+ PATTERNS_FIELD_NUMBER: builtins.int
1400
+ PATH_COLUMN_NAME_FIELD_NUMBER: builtins.int
1401
+ TIME_COLUMN_NAME_FIELD_NUMBER: builtins.int
1402
+ VALUE_COLUMN_NAME_FIELD_NUMBER: builtins.int
1403
+ VERSION_COLUMN_NAME_FIELD_NUMBER: builtins.int
1404
+ name: builtins.str
1405
+ """Name for the specified combination of settings for Graphite rollup."""
1406
+ path_column_name: builtins.str
1407
+ """The name of the column storing the metric name (Graphite sensor).
1408
+
1409
+ Default value: **Path**.
1410
+ """
1411
+ time_column_name: builtins.str
1412
+ """The name of the column storing the time of measuring the metric.
1413
+
1414
+ Default value: **Time**.
1415
+ """
1416
+ value_column_name: builtins.str
1417
+ """The name of the column storing the value of the metric at the time set in **time_column_name**.
1418
+
1419
+ Default value: **Value**.
1420
+ """
1421
+ version_column_name: builtins.str
1422
+ """The name of the column storing the version of the metric.
1423
+
1424
+ Default value: **Timestamp**.
1425
+ """
1426
+ @property
1427
+ def patterns(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern]:
1428
+ """Pattern to use for the rollup."""
1429
+
1430
+ def __init__(
1431
+ self,
1432
+ *,
1433
+ name: builtins.str = ...,
1434
+ patterns: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern] | None = ...,
1435
+ path_column_name: builtins.str = ...,
1436
+ time_column_name: builtins.str = ...,
1437
+ value_column_name: builtins.str = ...,
1438
+ version_column_name: builtins.str = ...,
1439
+ ) -> None: ...
1440
+ def ClearField(self, field_name: typing.Literal["name", b"name", "path_column_name", b"path_column_name", "patterns", b"patterns", "time_column_name", b"time_column_name", "value_column_name", b"value_column_name", "version_column_name", b"version_column_name"]) -> None: ...
1441
+
1442
+ @typing.final
1443
+ class Kafka(google.protobuf.message.Message):
1444
+ """Kafka configuration settings.
1445
+ For details, see [librdkafka documentation](https://github.com/confluentinc/librdkafka/blob/master/CONFIGURATION.md).
1446
+ """
1447
+
1448
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1449
+
1450
+ class _SecurityProtocol:
1451
+ ValueType = typing.NewType("ValueType", builtins.int)
1452
+ V: typing_extensions.TypeAlias = ValueType
1453
+
1454
+ class _SecurityProtocolEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SecurityProtocol.ValueType], builtins.type):
1455
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1456
+ SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 0
1457
+ SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 1
1458
+ SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 2
1459
+ SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 3
1460
+ SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 4
1461
+
1462
+ class SecurityProtocol(_SecurityProtocol, metaclass=_SecurityProtocolEnumTypeWrapper): ...
1463
+ SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 0
1464
+ SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 1
1465
+ SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 2
1466
+ SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 3
1467
+ SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 4
1468
+
1469
+ class _SaslMechanism:
1470
+ ValueType = typing.NewType("ValueType", builtins.int)
1471
+ V: typing_extensions.TypeAlias = ValueType
1472
+
1473
+ class _SaslMechanismEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SaslMechanism.ValueType], builtins.type):
1474
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1475
+ SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 0
1476
+ SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 1
1477
+ SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 2
1478
+ SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 3
1479
+ SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 4
1480
+
1481
+ class SaslMechanism(_SaslMechanism, metaclass=_SaslMechanismEnumTypeWrapper): ...
1482
+ SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 0
1483
+ SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 1
1484
+ SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 2
1485
+ SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 3
1486
+ SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 4
1487
+
1488
+ class _Debug:
1489
+ ValueType = typing.NewType("ValueType", builtins.int)
1490
+ V: typing_extensions.TypeAlias = ValueType
1491
+
1492
+ class _DebugEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._Debug.ValueType], builtins.type):
1493
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1494
+ DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka._Debug.ValueType # 0
1495
+ DEBUG_GENERIC: ClickhouseConfig.Kafka._Debug.ValueType # 1
1496
+ DEBUG_BROKER: ClickhouseConfig.Kafka._Debug.ValueType # 2
1497
+ DEBUG_TOPIC: ClickhouseConfig.Kafka._Debug.ValueType # 3
1498
+ DEBUG_METADATA: ClickhouseConfig.Kafka._Debug.ValueType # 4
1499
+ DEBUG_FEATURE: ClickhouseConfig.Kafka._Debug.ValueType # 5
1500
+ DEBUG_QUEUE: ClickhouseConfig.Kafka._Debug.ValueType # 6
1501
+ DEBUG_MSG: ClickhouseConfig.Kafka._Debug.ValueType # 7
1502
+ DEBUG_PROTOCOL: ClickhouseConfig.Kafka._Debug.ValueType # 8
1503
+ DEBUG_CGRP: ClickhouseConfig.Kafka._Debug.ValueType # 9
1504
+ DEBUG_SECURITY: ClickhouseConfig.Kafka._Debug.ValueType # 10
1505
+ DEBUG_FETCH: ClickhouseConfig.Kafka._Debug.ValueType # 11
1506
+ DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka._Debug.ValueType # 12
1507
+ DEBUG_PLUGIN: ClickhouseConfig.Kafka._Debug.ValueType # 13
1508
+ DEBUG_CONSUMER: ClickhouseConfig.Kafka._Debug.ValueType # 14
1509
+ DEBUG_ADMIN: ClickhouseConfig.Kafka._Debug.ValueType # 15
1510
+ DEBUG_EOS: ClickhouseConfig.Kafka._Debug.ValueType # 16
1511
+ DEBUG_MOCK: ClickhouseConfig.Kafka._Debug.ValueType # 17
1512
+ DEBUG_ASSIGNOR: ClickhouseConfig.Kafka._Debug.ValueType # 18
1513
+ DEBUG_CONF: ClickhouseConfig.Kafka._Debug.ValueType # 19
1514
+ DEBUG_TELEMETRY: ClickhouseConfig.Kafka._Debug.ValueType # 20
1515
+ DEBUG_ALL: ClickhouseConfig.Kafka._Debug.ValueType # 21
1516
+
1517
+ class Debug(_Debug, metaclass=_DebugEnumTypeWrapper): ...
1518
+ DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka.Debug.ValueType # 0
1519
+ DEBUG_GENERIC: ClickhouseConfig.Kafka.Debug.ValueType # 1
1520
+ DEBUG_BROKER: ClickhouseConfig.Kafka.Debug.ValueType # 2
1521
+ DEBUG_TOPIC: ClickhouseConfig.Kafka.Debug.ValueType # 3
1522
+ DEBUG_METADATA: ClickhouseConfig.Kafka.Debug.ValueType # 4
1523
+ DEBUG_FEATURE: ClickhouseConfig.Kafka.Debug.ValueType # 5
1524
+ DEBUG_QUEUE: ClickhouseConfig.Kafka.Debug.ValueType # 6
1525
+ DEBUG_MSG: ClickhouseConfig.Kafka.Debug.ValueType # 7
1526
+ DEBUG_PROTOCOL: ClickhouseConfig.Kafka.Debug.ValueType # 8
1527
+ DEBUG_CGRP: ClickhouseConfig.Kafka.Debug.ValueType # 9
1528
+ DEBUG_SECURITY: ClickhouseConfig.Kafka.Debug.ValueType # 10
1529
+ DEBUG_FETCH: ClickhouseConfig.Kafka.Debug.ValueType # 11
1530
+ DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka.Debug.ValueType # 12
1531
+ DEBUG_PLUGIN: ClickhouseConfig.Kafka.Debug.ValueType # 13
1532
+ DEBUG_CONSUMER: ClickhouseConfig.Kafka.Debug.ValueType # 14
1533
+ DEBUG_ADMIN: ClickhouseConfig.Kafka.Debug.ValueType # 15
1534
+ DEBUG_EOS: ClickhouseConfig.Kafka.Debug.ValueType # 16
1535
+ DEBUG_MOCK: ClickhouseConfig.Kafka.Debug.ValueType # 17
1536
+ DEBUG_ASSIGNOR: ClickhouseConfig.Kafka.Debug.ValueType # 18
1537
+ DEBUG_CONF: ClickhouseConfig.Kafka.Debug.ValueType # 19
1538
+ DEBUG_TELEMETRY: ClickhouseConfig.Kafka.Debug.ValueType # 20
1539
+ DEBUG_ALL: ClickhouseConfig.Kafka.Debug.ValueType # 21
1540
+
1541
+ class _AutoOffsetReset:
1542
+ ValueType = typing.NewType("ValueType", builtins.int)
1543
+ V: typing_extensions.TypeAlias = ValueType
1544
+
1545
+ class _AutoOffsetResetEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._AutoOffsetReset.ValueType], builtins.type):
1546
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1547
+ AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 0
1548
+ AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 1
1549
+ AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 2
1550
+ AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 3
1551
+ AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 4
1552
+ AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 5
1553
+ AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 6
1554
+ AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 7
1555
+
1556
+ class AutoOffsetReset(_AutoOffsetReset, metaclass=_AutoOffsetResetEnumTypeWrapper): ...
1557
+ AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 0
1558
+ AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 1
1559
+ AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 2
1560
+ AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 3
1561
+ AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 4
1562
+ AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 5
1563
+ AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 6
1564
+ AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 7
1565
+
1566
+ SECURITY_PROTOCOL_FIELD_NUMBER: builtins.int
1567
+ SASL_MECHANISM_FIELD_NUMBER: builtins.int
1568
+ SASL_USERNAME_FIELD_NUMBER: builtins.int
1569
+ SASL_PASSWORD_FIELD_NUMBER: builtins.int
1570
+ ENABLE_SSL_CERTIFICATE_VERIFICATION_FIELD_NUMBER: builtins.int
1571
+ MAX_POLL_INTERVAL_MS_FIELD_NUMBER: builtins.int
1572
+ SESSION_TIMEOUT_MS_FIELD_NUMBER: builtins.int
1573
+ DEBUG_FIELD_NUMBER: builtins.int
1574
+ AUTO_OFFSET_RESET_FIELD_NUMBER: builtins.int
1575
+ security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType
1576
+ """Protocol used to communicate with brokers.
1577
+
1578
+ Default value: **SECURITY_PROTOCOL_PLAINTEXT**.
1579
+ """
1580
+ sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType
1581
+ """SASL mechanism to use for authentication.
1582
+
1583
+ Default value: **SASL_MECHANISM_GSSAPI**.
1584
+ """
1585
+ sasl_username: builtins.str
1586
+ """SASL username for use with the PLAIN and SASL-SCRAM mechanisms."""
1587
+ sasl_password: builtins.str
1588
+ """SASL password for use with the PLAIN and SASL-SCRAM mechanisms."""
1589
+ debug: global___ClickhouseConfig.Kafka.Debug.ValueType
1590
+ """Debug context to enable."""
1591
+ auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType
1592
+ """Action to take when there is no initial offset in offset store or the desired offset is out of range.
1593
+
1594
+ Default value: **AUTO_OFFSET_RESET_LARGEST**.
1595
+ """
1596
+ @property
1597
+ def enable_ssl_certificate_verification(self) -> google.protobuf.wrappers_pb2.BoolValue:
1598
+ """Enable OpenSSL's builtin broker (server) certificate verification.
1599
+
1600
+ Default value: **true**.
1601
+ """
1255
1602
 
1256
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
1603
+ @property
1604
+ def max_poll_interval_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
1605
+ """Maximum allowed time between calls to consume messages for high-level consumers.
1606
+ If this interval is exceeded the consumer is considered failed and the group will
1607
+ rebalance in order to reassign the partitions to another consumer group member.
1257
1608
 
1258
- @typing.final
1259
- class Pattern(google.protobuf.message.Message):
1260
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
1609
+ Default value: **300000** (5 minutes).
1610
+ """
1261
1611
 
1262
- @typing.final
1263
- class Retention(google.protobuf.message.Message):
1264
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
1612
+ @property
1613
+ def session_timeout_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
1614
+ """Client group session and failure detection timeout. The consumer sends periodic heartbeats (heartbeat.interval.ms)
1615
+ to indicate its liveness to the broker. If no hearts are received by the broker for a group member within
1616
+ the session timeout, the broker will remove the consumer from the group and trigger a rebalance.
1265
1617
 
1266
- AGE_FIELD_NUMBER: builtins.int
1267
- PRECISION_FIELD_NUMBER: builtins.int
1268
- age: builtins.int
1269
- """Minimum age of the data in seconds."""
1270
- precision: builtins.int
1271
- """Precision of determining the age of the data, in seconds."""
1272
- def __init__(
1273
- self,
1274
- *,
1275
- age: builtins.int = ...,
1276
- precision: builtins.int = ...,
1277
- ) -> None: ...
1278
- def ClearField(self, field_name: typing.Literal["age", b"age", "precision", b"precision"]) -> None: ...
1618
+ Default value: **45000** (45 seconds).
1619
+ """
1279
1620
 
1280
- REGEXP_FIELD_NUMBER: builtins.int
1281
- FUNCTION_FIELD_NUMBER: builtins.int
1282
- RETENTION_FIELD_NUMBER: builtins.int
1283
- regexp: builtins.str
1284
- """Pattern for metric names."""
1285
- function: builtins.str
1286
- """Name of the aggregating function to apply to data of the age specified in [retention]."""
1287
- @property
1288
- def retention(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention]:
1289
- """Age of data to use for thinning."""
1621
+ def __init__(
1622
+ self,
1623
+ *,
1624
+ security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType = ...,
1625
+ sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType = ...,
1626
+ sasl_username: builtins.str = ...,
1627
+ sasl_password: builtins.str = ...,
1628
+ enable_ssl_certificate_verification: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1629
+ max_poll_interval_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1630
+ session_timeout_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1631
+ debug: global___ClickhouseConfig.Kafka.Debug.ValueType = ...,
1632
+ auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType = ...,
1633
+ ) -> None: ...
1634
+ def HasField(self, field_name: typing.Literal["enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "session_timeout_ms", b"session_timeout_ms"]) -> builtins.bool: ...
1635
+ def ClearField(self, field_name: typing.Literal["auto_offset_reset", b"auto_offset_reset", "debug", b"debug", "enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "sasl_mechanism", b"sasl_mechanism", "sasl_password", b"sasl_password", "sasl_username", b"sasl_username", "security_protocol", b"security_protocol", "session_timeout_ms", b"session_timeout_ms"]) -> None: ...
1290
1636
 
1291
- def __init__(
1292
- self,
1293
- *,
1294
- regexp: builtins.str = ...,
1295
- function: builtins.str = ...,
1296
- retention: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention] | None = ...,
1297
- ) -> None: ...
1298
- def ClearField(self, field_name: typing.Literal["function", b"function", "regexp", b"regexp", "retention", b"retention"]) -> None: ...
1637
+ @typing.final
1638
+ class KafkaTopic(google.protobuf.message.Message):
1639
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1299
1640
 
1300
1641
  NAME_FIELD_NUMBER: builtins.int
1301
- PATTERNS_FIELD_NUMBER: builtins.int
1302
- PATH_COLUMN_NAME_FIELD_NUMBER: builtins.int
1303
- TIME_COLUMN_NAME_FIELD_NUMBER: builtins.int
1304
- VALUE_COLUMN_NAME_FIELD_NUMBER: builtins.int
1305
- VERSION_COLUMN_NAME_FIELD_NUMBER: builtins.int
1642
+ SETTINGS_FIELD_NUMBER: builtins.int
1306
1643
  name: builtins.str
1307
- """Name for the specified combination of settings for Graphite rollup."""
1308
- path_column_name: builtins.str
1309
- """The name of the column storing the metric name (Graphite sensor).
1310
- Default: Path
1311
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1312
- """
1313
- time_column_name: builtins.str
1314
- """The name of the column storing the time of measuring the metric.
1315
- Default: Time
1316
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1317
- """
1318
- value_column_name: builtins.str
1319
- """The name of the column storing the value of the metric at the time set in time_column_name.
1320
- Default: Value
1321
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1322
- """
1323
- version_column_name: builtins.str
1324
- """The name of the column storing the version of the metric.
1325
- Default: Timestamp
1326
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1327
- """
1644
+ """Kafka topic name."""
1328
1645
  @property
1329
- def patterns(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern]:
1330
- """Pattern to use for the rollup."""
1646
+ def settings(self) -> global___ClickhouseConfig.Kafka:
1647
+ """Kafka topic settings."""
1331
1648
 
1332
1649
  def __init__(
1333
1650
  self,
1334
1651
  *,
1335
1652
  name: builtins.str = ...,
1336
- patterns: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern] | None = ...,
1337
- path_column_name: builtins.str = ...,
1338
- time_column_name: builtins.str = ...,
1339
- value_column_name: builtins.str = ...,
1340
- version_column_name: builtins.str = ...,
1653
+ settings: global___ClickhouseConfig.Kafka | None = ...,
1341
1654
  ) -> None: ...
1342
- def ClearField(self, field_name: typing.Literal["name", b"name", "path_column_name", b"path_column_name", "patterns", b"patterns", "time_column_name", b"time_column_name", "value_column_name", b"value_column_name", "version_column_name", b"version_column_name"]) -> None: ...
1655
+ def HasField(self, field_name: typing.Literal["settings", b"settings"]) -> builtins.bool: ...
1656
+ def ClearField(self, field_name: typing.Literal["name", b"name", "settings", b"settings"]) -> None: ...
1657
+
1658
+ @typing.final
1659
+ class Rabbitmq(google.protobuf.message.Message):
1660
+ """RabbitMQ integration settings.
1661
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/engines/table-engines/integrations/rabbitmq).
1662
+ """
1663
+
1664
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1665
+
1666
+ USERNAME_FIELD_NUMBER: builtins.int
1667
+ PASSWORD_FIELD_NUMBER: builtins.int
1668
+ VHOST_FIELD_NUMBER: builtins.int
1669
+ username: builtins.str
1670
+ """RabbitMQ username."""
1671
+ password: builtins.str
1672
+ """RabbitMQ password."""
1673
+ vhost: builtins.str
1674
+ """RabbitMQ virtual host."""
1675
+ def __init__(
1676
+ self,
1677
+ *,
1678
+ username: builtins.str = ...,
1679
+ password: builtins.str = ...,
1680
+ vhost: builtins.str = ...,
1681
+ ) -> None: ...
1682
+ def ClearField(self, field_name: typing.Literal["password", b"password", "username", b"username", "vhost", b"vhost"]) -> None: ...
1343
1683
 
1344
1684
  @typing.final
1345
1685
  class QueryMaskingRule(google.protobuf.message.Message):
@@ -1351,12 +1691,11 @@ class ClickhouseConfig(google.protobuf.message.Message):
1351
1691
  name: builtins.str
1352
1692
  """Name for the rule."""
1353
1693
  regexp: builtins.str
1354
- """RE2 compatible regular expression.
1355
- Required.
1356
- """
1694
+ """RE2 compatible regular expression."""
1357
1695
  replace: builtins.str
1358
1696
  """Substitution string for sensitive data.
1359
- Default: six asterisks
1697
+
1698
+ Default value: six asterisks.
1360
1699
  """
1361
1700
  def __init__(
1362
1701
  self,
@@ -1369,6 +1708,8 @@ class ClickhouseConfig(google.protobuf.message.Message):
1369
1708
 
1370
1709
  @typing.final
1371
1710
  class QueryCache(google.protobuf.message.Message):
1711
+ """Query cache configuration."""
1712
+
1372
1713
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1373
1714
 
1374
1715
  MAX_SIZE_IN_BYTES_FIELD_NUMBER: builtins.int
@@ -1378,25 +1719,29 @@ class ClickhouseConfig(google.protobuf.message.Message):
1378
1719
  @property
1379
1720
  def max_size_in_bytes(self) -> google.protobuf.wrappers_pb2.Int64Value:
1380
1721
  """The maximum cache size in bytes.
1381
- Default: 1073741824 (1 GiB)
1722
+
1723
+ Default value: **1073741824** (1 GiB).
1382
1724
  """
1383
1725
 
1384
1726
  @property
1385
1727
  def max_entries(self) -> google.protobuf.wrappers_pb2.Int64Value:
1386
- """The maximum number of SELECT query results stored in the cache.
1387
- Default: 1024
1728
+ """The maximum number of **SELECT** query results stored in the cache.
1729
+
1730
+ Default value: **1024**.
1388
1731
  """
1389
1732
 
1390
1733
  @property
1391
1734
  def max_entry_size_in_bytes(self) -> google.protobuf.wrappers_pb2.Int64Value:
1392
- """The maximum size in bytes SELECT query results may have to be saved in the cache.
1393
- Dafault: 1048576 (1 MiB)
1735
+ """The maximum size in bytes **SELECT** query results may have to be saved in the cache.
1736
+
1737
+ Default value: **1048576** (1 MiB).
1394
1738
  """
1395
1739
 
1396
1740
  @property
1397
1741
  def max_entry_size_in_rows(self) -> google.protobuf.wrappers_pb2.Int64Value:
1398
- """The maximum number of rows SELECT query results may have to be saved in the cache.
1399
- Default: 30000000 (30 mil)
1742
+ """The maximum number of rows **SELECT** query results may have to be saved in the cache.
1743
+
1744
+ Default value: **30000000**.
1400
1745
  """
1401
1746
 
1402
1747
  def __init__(
@@ -1412,7 +1757,7 @@ class ClickhouseConfig(google.protobuf.message.Message):
1412
1757
 
1413
1758
  @typing.final
1414
1759
  class JdbcBridge(google.protobuf.message.Message):
1415
- """JDBC bridge for queries to external databases."""
1760
+ """JDBC bridge configuration for queries to external databases."""
1416
1761
 
1417
1762
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1418
1763
 
@@ -1422,7 +1767,10 @@ class ClickhouseConfig(google.protobuf.message.Message):
1422
1767
  """Host of jdbc bridge."""
1423
1768
  @property
1424
1769
  def port(self) -> google.protobuf.wrappers_pb2.Int64Value:
1425
- """Port of jdbc bridge."""
1770
+ """Port of jdbc bridge.
1771
+
1772
+ Default value: **9019**.
1773
+ """
1426
1774
 
1427
1775
  def __init__(
1428
1776
  self,
@@ -1433,54 +1781,17 @@ class ClickhouseConfig(google.protobuf.message.Message):
1433
1781
  def HasField(self, field_name: typing.Literal["port", b"port"]) -> builtins.bool: ...
1434
1782
  def ClearField(self, field_name: typing.Literal["host", b"host", "port", b"port"]) -> None: ...
1435
1783
 
1436
- @typing.final
1437
- class AccessControlImprovements(google.protobuf.message.Message):
1438
- """Access control settings."""
1439
-
1440
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
1441
-
1442
- SELECT_FROM_SYSTEM_DB_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
1443
- SELECT_FROM_INFORMATION_SCHEMA_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
1444
- @property
1445
- def select_from_system_db_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
1446
- """Sets whether SELECT * FROM system.<table> requires any grants and can be executed by any user.
1447
- If set to true then this query requires GRANT SELECT ON system.<table> just as for non-system tables.
1448
- """
1449
-
1450
- @property
1451
- def select_from_information_schema_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
1452
- """Sets whether SELECT * FROM information_schema.<table> requires any grants and can be executed by any user.
1453
- If set to true, then this query requires GRANT SELECT ON information_schema.<table>, just as for ordinary tables.
1454
- """
1455
-
1456
- def __init__(
1457
- self,
1458
- *,
1459
- select_from_system_db_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1460
- select_from_information_schema_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1461
- ) -> None: ...
1462
- def HasField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> builtins.bool: ...
1463
- def ClearField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> None: ...
1464
-
1784
+ BACKGROUND_POOL_SIZE_FIELD_NUMBER: builtins.int
1785
+ BACKGROUND_MERGES_MUTATIONS_CONCURRENCY_RATIO_FIELD_NUMBER: builtins.int
1786
+ BACKGROUND_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1787
+ BACKGROUND_FETCHES_POOL_SIZE_FIELD_NUMBER: builtins.int
1788
+ BACKGROUND_MOVE_POOL_SIZE_FIELD_NUMBER: builtins.int
1789
+ BACKGROUND_DISTRIBUTED_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1790
+ BACKGROUND_BUFFER_FLUSH_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1791
+ BACKGROUND_MESSAGE_BROKER_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1792
+ BACKGROUND_COMMON_POOL_SIZE_FIELD_NUMBER: builtins.int
1793
+ DICTIONARIES_LAZY_LOAD_FIELD_NUMBER: builtins.int
1465
1794
  LOG_LEVEL_FIELD_NUMBER: builtins.int
1466
- MERGE_TREE_FIELD_NUMBER: builtins.int
1467
- COMPRESSION_FIELD_NUMBER: builtins.int
1468
- DICTIONARIES_FIELD_NUMBER: builtins.int
1469
- GRAPHITE_ROLLUP_FIELD_NUMBER: builtins.int
1470
- KAFKA_FIELD_NUMBER: builtins.int
1471
- KAFKA_TOPICS_FIELD_NUMBER: builtins.int
1472
- RABBITMQ_FIELD_NUMBER: builtins.int
1473
- MAX_CONNECTIONS_FIELD_NUMBER: builtins.int
1474
- MAX_CONCURRENT_QUERIES_FIELD_NUMBER: builtins.int
1475
- KEEP_ALIVE_TIMEOUT_FIELD_NUMBER: builtins.int
1476
- UNCOMPRESSED_CACHE_SIZE_FIELD_NUMBER: builtins.int
1477
- MARK_CACHE_SIZE_FIELD_NUMBER: builtins.int
1478
- MAX_TABLE_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1479
- MAX_PARTITION_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1480
- BUILTIN_DICTIONARIES_RELOAD_INTERVAL_FIELD_NUMBER: builtins.int
1481
- TIMEZONE_FIELD_NUMBER: builtins.int
1482
- GEOBASE_ENABLED_FIELD_NUMBER: builtins.int
1483
- GEOBASE_URI_FIELD_NUMBER: builtins.int
1484
1795
  QUERY_LOG_RETENTION_SIZE_FIELD_NUMBER: builtins.int
1485
1796
  QUERY_LOG_RETENTION_TIME_FIELD_NUMBER: builtins.int
1486
1797
  QUERY_THREAD_LOG_ENABLED_FIELD_NUMBER: builtins.int
@@ -1523,312 +1834,524 @@ class ClickhouseConfig(google.protobuf.message.Message):
1523
1834
  ERROR_LOG_RETENTION_SIZE_FIELD_NUMBER: builtins.int
1524
1835
  ERROR_LOG_RETENTION_TIME_FIELD_NUMBER: builtins.int
1525
1836
  ACCESS_CONTROL_IMPROVEMENTS_FIELD_NUMBER: builtins.int
1526
- BACKGROUND_POOL_SIZE_FIELD_NUMBER: builtins.int
1527
- BACKGROUND_MERGES_MUTATIONS_CONCURRENCY_RATIO_FIELD_NUMBER: builtins.int
1528
- BACKGROUND_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1529
- BACKGROUND_FETCHES_POOL_SIZE_FIELD_NUMBER: builtins.int
1530
- BACKGROUND_MOVE_POOL_SIZE_FIELD_NUMBER: builtins.int
1531
- BACKGROUND_DISTRIBUTED_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1532
- BACKGROUND_BUFFER_FLUSH_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1533
- BACKGROUND_MESSAGE_BROKER_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1534
- BACKGROUND_COMMON_POOL_SIZE_FIELD_NUMBER: builtins.int
1837
+ MAX_CONNECTIONS_FIELD_NUMBER: builtins.int
1838
+ MAX_CONCURRENT_QUERIES_FIELD_NUMBER: builtins.int
1839
+ MAX_TABLE_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1840
+ MAX_PARTITION_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1841
+ KEEP_ALIVE_TIMEOUT_FIELD_NUMBER: builtins.int
1842
+ UNCOMPRESSED_CACHE_SIZE_FIELD_NUMBER: builtins.int
1843
+ MARK_CACHE_SIZE_FIELD_NUMBER: builtins.int
1844
+ TIMEZONE_FIELD_NUMBER: builtins.int
1845
+ GEOBASE_ENABLED_FIELD_NUMBER: builtins.int
1846
+ GEOBASE_URI_FIELD_NUMBER: builtins.int
1535
1847
  DEFAULT_DATABASE_FIELD_NUMBER: builtins.int
1536
1848
  TOTAL_MEMORY_PROFILER_STEP_FIELD_NUMBER: builtins.int
1537
1849
  TOTAL_MEMORY_TRACKER_SAMPLE_PROBABILITY_FIELD_NUMBER: builtins.int
1850
+ ASYNC_INSERT_THREADS_FIELD_NUMBER: builtins.int
1851
+ BACKUP_THREADS_FIELD_NUMBER: builtins.int
1852
+ RESTORE_THREADS_FIELD_NUMBER: builtins.int
1853
+ MERGE_TREE_FIELD_NUMBER: builtins.int
1854
+ COMPRESSION_FIELD_NUMBER: builtins.int
1855
+ DICTIONARIES_FIELD_NUMBER: builtins.int
1856
+ GRAPHITE_ROLLUP_FIELD_NUMBER: builtins.int
1857
+ KAFKA_FIELD_NUMBER: builtins.int
1858
+ KAFKA_TOPICS_FIELD_NUMBER: builtins.int
1859
+ RABBITMQ_FIELD_NUMBER: builtins.int
1538
1860
  QUERY_MASKING_RULES_FIELD_NUMBER: builtins.int
1539
- DICTIONARIES_LAZY_LOAD_FIELD_NUMBER: builtins.int
1540
1861
  QUERY_CACHE_FIELD_NUMBER: builtins.int
1541
1862
  JDBC_BRIDGE_FIELD_NUMBER: builtins.int
1863
+ MYSQL_PROTOCOL_FIELD_NUMBER: builtins.int
1864
+ BUILTIN_DICTIONARIES_RELOAD_INTERVAL_FIELD_NUMBER: builtins.int
1542
1865
  log_level: global___ClickhouseConfig.LogLevel.ValueType
1543
- """Logging level for the ClickHouse cluster. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR."""
1866
+ """Logging level."""
1867
+ text_log_level: global___ClickhouseConfig.LogLevel.ValueType
1868
+ """Logging level for text_log system table.
1869
+
1870
+ Default value: **TRACE**.
1871
+
1872
+ Change of the setting is applied with restart.
1873
+ """
1544
1874
  timezone: builtins.str
1545
- """The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier."""
1875
+ """The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier.
1876
+
1877
+ Default value: **Europe/Moscow**.
1878
+
1879
+ Change of the setting is applied with restart.
1880
+
1881
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#timezone).
1882
+ """
1546
1883
  geobase_uri: builtins.str
1547
- """Address of the archive with the user geobase in Object Storage."""
1548
- text_log_level: global___ClickhouseConfig.LogLevel.ValueType
1549
- """Logging level for text_log system table. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR."""
1884
+ """Address of the archive with the user geobase in Object Storage.
1885
+
1886
+ Change of the setting is applied with restart.
1887
+ """
1550
1888
  @property
1551
- def merge_tree(self) -> global___ClickhouseConfig.MergeTree:
1552
- """Settings for the MergeTree engine.
1553
- See description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/#merge_tree).
1889
+ def background_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1890
+ """Sets the number of threads performing background merges and mutations for MergeTree-engine tables.
1891
+
1892
+ Default value: **16**.
1893
+
1894
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1895
+
1896
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_pool_size).
1554
1897
  """
1555
1898
 
1556
1899
  @property
1557
- def compression(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.Compression]:
1558
- """Compression settings for the ClickHouse cluster.
1559
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/#compression).
1900
+ def background_merges_mutations_concurrency_ratio(self) -> google.protobuf.wrappers_pb2.Int64Value:
1901
+ """Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently.
1902
+
1903
+ For example, if the ratio equals to **2** and **background_pool_size** is set to **16** then ClickHouse can execute **32** background merges concurrently.
1904
+ This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority.
1905
+
1906
+ Default value: **2**.
1907
+
1908
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1909
+
1910
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_merges_mutations_concurrency_ratio).
1560
1911
  """
1561
1912
 
1562
1913
  @property
1563
- def dictionaries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary]:
1564
- """Configuration of external dictionaries to be used by the ClickHouse cluster.
1565
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts/).
1914
+ def background_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1915
+ """The maximum number of threads that will be used for constantly executing some lightweight periodic operations
1916
+ for replicated tables, Kafka streaming, and DNS cache updates.
1917
+
1918
+ Default value: **512**.
1919
+
1920
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1921
+
1922
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_schedule_pool_size).
1566
1923
  """
1567
1924
 
1568
1925
  @property
1569
- def graphite_rollup(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup]:
1570
- """Settings for thinning Graphite data.
1571
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/#server_settings-graphite_rollup).
1926
+ def background_fetches_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1927
+ """The maximum number of threads that will be used for fetching data parts from another replica for MergeTree-engine tables in a background.
1928
+
1929
+ Default value: **32** for versions 25.1 and higher, **16** for versions 24.12 and lower.
1930
+
1931
+ Change of the setting is applied with restart.
1932
+
1933
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_fetches_pool_size).
1572
1934
  """
1573
1935
 
1574
1936
  @property
1575
- def kafka(self) -> global___ClickhouseConfig.Kafka: ...
1576
- @property
1577
- def kafka_topics(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.KafkaTopic]: ...
1578
- @property
1579
- def rabbitmq(self) -> global___ClickhouseConfig.Rabbitmq: ...
1580
- @property
1581
- def max_connections(self) -> google.protobuf.wrappers_pb2.Int64Value:
1582
- """Maximum number of inbound connections."""
1937
+ def background_move_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1938
+ """The maximum number of threads that will be used for moving data parts to another disk or volume for MergeTree-engine tables in a background.
1939
+
1940
+ Default value: **8**.
1941
+
1942
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1943
+
1944
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_move_pool_size).
1945
+ """
1583
1946
 
1584
1947
  @property
1585
- def max_concurrent_queries(self) -> google.protobuf.wrappers_pb2.Int64Value:
1586
- """Maximum number of simultaneously processed requests."""
1948
+ def background_distributed_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1949
+ """The maximum number of threads that will be used for executing distributed sends.
1950
+
1951
+ Default value: **16**.
1952
+
1953
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1954
+
1955
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_distributed_schedule_pool_size).
1956
+ """
1587
1957
 
1588
1958
  @property
1589
- def keep_alive_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
1590
- """Number of milliseconds that ClickHouse waits for incoming requests before closing the connection."""
1959
+ def background_buffer_flush_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1960
+ """The maximum number of threads that will be used for performing flush operations for Buffer-engine tables in the background.
1961
+
1962
+ Default value: **16**.
1963
+
1964
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1965
+
1966
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_buffer_flush_schedule_pool_size).
1967
+ """
1591
1968
 
1592
1969
  @property
1593
- def uncompressed_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1594
- """Cache size (in bytes) for uncompressed data used by MergeTree tables."""
1970
+ def background_message_broker_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1971
+ """The maximum number of threads that will be used for executing background operations for message streaming.
1972
+
1973
+ Default value: **16**.
1974
+
1975
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1976
+
1977
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_message_broker_schedule_pool_size).
1978
+ """
1595
1979
 
1596
1980
  @property
1597
- def mark_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1598
- """Approximate size (in bytes) of the cache of "marks" used by MergeTree tables."""
1981
+ def background_common_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1982
+ """The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for MergeTree-engine tables in a background.
1983
+
1984
+ Default value: **8**.
1985
+
1986
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1987
+
1988
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_common_pool_size).
1989
+ """
1599
1990
 
1600
1991
  @property
1601
- def max_table_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
1602
- """Maximum size of the table that can be deleted using a DROP query."""
1992
+ def dictionaries_lazy_load(self) -> google.protobuf.wrappers_pb2.BoolValue:
1993
+ """Lazy loading of dictionaries. If enabled, then each dictionary is loaded on the first use. Otherwise, the server loads all dictionaries at startup.
1603
1994
 
1604
- @property
1605
- def max_partition_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
1606
- """Maximum size of the partition that can be deleted using a DROP query."""
1995
+ Default value: **true** for versions 25.1 and higher, **false** for versions 24.12 and lower.
1607
1996
 
1608
- @property
1609
- def builtin_dictionaries_reload_interval(self) -> google.protobuf.wrappers_pb2.Int64Value:
1610
- """The setting is deprecated and has no effect."""
1997
+ Change of the setting is applied with restart.
1611
1998
 
1612
- @property
1613
- def geobase_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1614
- """Enable or disable geobase."""
1999
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#dictionaries_lazy_load).
2000
+ """
1615
2001
 
1616
2002
  @property
1617
2003
  def query_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1618
- """The maximum size that query_log can grow to before old data will be removed. If set to 0, automatic removal of
1619
- query_log data based on size is disabled.
2004
+ """The maximum size that query_log can grow to before old data will be removed. If set to **0**,
2005
+ automatic removal of query_log data based on size is disabled.
2006
+
2007
+ Default value: **1073741824** (1 GiB).
1620
2008
  """
1621
2009
 
1622
2010
  @property
1623
2011
  def query_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1624
- """The maximum time that query_log records will be retained before removal. If set to 0, automatic removal of
1625
- query_log data based on time is disabled.
2012
+ """The maximum time that query_log records will be retained before removal. If set to **0**, automatic removal of query_log data based on time is disabled.
2013
+
2014
+ Default value: **2592000000** (30 days).
1626
2015
  """
1627
2016
 
1628
2017
  @property
1629
2018
  def query_thread_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1630
- """Whether query_thread_log system table is enabled."""
2019
+ """Enables or disables query_thread_log system table.
2020
+
2021
+ Default value: **true**.
2022
+
2023
+ Change of the setting is applied with restart.
2024
+
2025
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/query_thread_log).
2026
+ """
1631
2027
 
1632
2028
  @property
1633
2029
  def query_thread_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1634
- """The maximum size that query_thread_log can grow to before old data will be removed. If set to 0, automatic removal of
1635
- query_thread_log data based on size is disabled.
2030
+ """The maximum size that query_thread_log can grow to before old data will be removed. If set to **0**,
2031
+ automatic removal of query_thread_log data based on size is disabled.
2032
+
2033
+ Default value: **536870912** (512 MiB).
1636
2034
  """
1637
2035
 
1638
2036
  @property
1639
2037
  def query_thread_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1640
- """The maximum time that query_thread_log records will be retained before removal. If set to 0, automatic removal of
1641
- query_thread_log data based on time is disabled.
2038
+ """The maximum time that query_thread_log records will be retained before removal. If set to **0**,
2039
+ automatic removal of query_thread_log data based on time is disabled.
2040
+
2041
+ Default value: **2592000000** (30 days).
1642
2042
  """
1643
2043
 
1644
2044
  @property
1645
2045
  def part_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1646
- """The maximum size that part_log can grow to before old data will be removed. If set to 0, automatic removal of
1647
- part_log data based on size is disabled.
2046
+ """The maximum size that part_log can grow to before old data will be removed. If set to **0**,
2047
+ automatic removal of part_log data based on size is disabled.
2048
+
2049
+ Default value: **536870912** (512 MiB).
1648
2050
  """
1649
2051
 
1650
2052
  @property
1651
2053
  def part_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1652
- """The maximum time that part_log records will be retained before removal. If set to 0, automatic removal of
1653
- part_log data based on time is disabled.
2054
+ """The maximum time that part_log records will be retained before removal. If set to **0**,
2055
+ automatic removal of part_log data based on time is disabled.
2056
+
2057
+ Default value: **2592000000** (30 days).
1654
2058
  """
1655
2059
 
1656
2060
  @property
1657
2061
  def metric_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1658
- """Whether metric_log system table is enabled."""
2062
+ """Enables or disables metric_log system table.
2063
+
2064
+ Default value: **false** for versions 25.1 and higher, **true** for versions 24.12 and lower.
2065
+
2066
+ Change of the setting is applied with restart.
2067
+
2068
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/metric_log).
2069
+ """
1659
2070
 
1660
2071
  @property
1661
2072
  def metric_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1662
- """The maximum size that metric_log can grow to before old data will be removed. If set to 0, automatic removal of
1663
- metric_log data based on size is disabled.
2073
+ """The maximum size that metric_log can grow to before old data will be removed. If set to **0**,
2074
+ automatic removal of metric_log data based on size is disabled.
2075
+
2076
+ Default value: **536870912** (512 MiB).
1664
2077
  """
1665
2078
 
1666
2079
  @property
1667
2080
  def metric_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1668
- """The maximum time that metric_log records will be retained before removal. If set to 0, automatic removal of
1669
- metric_log data based on time is disabled.
2081
+ """The maximum time that metric_log records will be retained before removal. If set to **0**,
2082
+ automatic removal of metric_log data based on time is disabled.
2083
+
2084
+ Default value: **2592000000** (30 days).
1670
2085
  """
1671
2086
 
1672
2087
  @property
1673
2088
  def trace_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1674
- """Whether trace_log system table is enabled."""
2089
+ """Enables or disables trace_log system table.
2090
+
2091
+ Default value: **true** for versions 25.2 and higher, **false** for versions 25.1 and lower.
2092
+
2093
+ Change of the setting is applied with restart.
2094
+
2095
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/trace_log).
2096
+ """
1675
2097
 
1676
2098
  @property
1677
2099
  def trace_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1678
- """The maximum size that trace_log can grow to before old data will be removed. If set to 0, automatic removal of
1679
- trace_log data based on size is disabled.
2100
+ """The maximum size that trace_log can grow to before old data will be removed. If set to **0**,
2101
+ automatic removal of trace_log data based on size is disabled.
2102
+
2103
+ Default value: **536870912** (512 MiB).
1680
2104
  """
1681
2105
 
1682
2106
  @property
1683
2107
  def trace_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1684
- """The maximum time that trace_log records will be retained before removal. If set to 0, automatic removal of
1685
- trace_log data based on time is disabled.
2108
+ """The maximum time that trace_log records will be retained before removal. If set to **0**,
2109
+ automatic removal of trace_log data based on time is disabled.
2110
+
2111
+ Default value: **2592000000** (30 days).
1686
2112
  """
1687
2113
 
1688
2114
  @property
1689
2115
  def text_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1690
- """Whether text_log system table is enabled."""
2116
+ """Enables or disables text_log system table.
2117
+
2118
+ Default value: **false**.
2119
+
2120
+ Change of the setting is applied with restart.
2121
+
2122
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/text_log).
2123
+ """
1691
2124
 
1692
2125
  @property
1693
2126
  def text_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1694
- """The maximum size that text_log can grow to before old data will be removed. If set to 0, automatic removal of
1695
- text_log data based on size is disabled.
2127
+ """The maximum size that text_log can grow to before old data will be removed. If set to **0**,
2128
+ automatic removal of text_log data based on size is disabled.
2129
+
2130
+ Default value: **536870912** (512 MiB).
1696
2131
  """
1697
2132
 
1698
2133
  @property
1699
2134
  def text_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1700
- """The maximum time that text_log records will be retained before removal. If set to 0, automatic removal of
1701
- text_log data based on time is disabled.
2135
+ """The maximum time that text_log records will be retained before removal. If set to **0**,
2136
+ automatic removal of text_log data based on time is disabled.
2137
+
2138
+ Default value: **2592000000** (30 days).
1702
2139
  """
1703
2140
 
1704
2141
  @property
1705
2142
  def opentelemetry_span_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1706
- """Enable or disable opentelemetry_span_log system table. Default value: false."""
2143
+ """Enables or disables opentelemetry_span_log system table.
2144
+
2145
+ Default value: **false**.
2146
+
2147
+ Change of the setting is applied with restart.
2148
+
2149
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/opentelemetry_span_log).
2150
+ """
1707
2151
 
1708
2152
  @property
1709
2153
  def opentelemetry_span_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1710
- """The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to 0 (default),
2154
+ """The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to **0**,
1711
2155
  automatic removal of opentelemetry_span_log data based on size is disabled.
2156
+
2157
+ Default value: **0**.
1712
2158
  """
1713
2159
 
1714
2160
  @property
1715
2161
  def opentelemetry_span_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1716
- """The maximum time that opentelemetry_span_log records will be retained before removal. If set to 0,
2162
+ """The maximum time that opentelemetry_span_log records will be retained before removal. If set to **0**,
1717
2163
  automatic removal of opentelemetry_span_log data based on time is disabled.
2164
+
2165
+ Default value: **2592000000** (30 days).
1718
2166
  """
1719
2167
 
1720
2168
  @property
1721
2169
  def query_views_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1722
- """Enable or disable query_views_log system table. Default value: false."""
2170
+ """Enables or disables query_views_log system table.
2171
+
2172
+ Default value: **false**.
2173
+
2174
+ Change of the setting is applied with restart.
2175
+
2176
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/query_views_log).
2177
+ """
1723
2178
 
1724
2179
  @property
1725
2180
  def query_views_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1726
- """The maximum size that query_views_log can grow to before old data will be removed. If set to 0 (default),
2181
+ """The maximum size that query_views_log can grow to before old data will be removed. If set to **0**,
1727
2182
  automatic removal of query_views_log data based on size is disabled.
2183
+
2184
+ Default value: **0**.
1728
2185
  """
1729
2186
 
1730
2187
  @property
1731
2188
  def query_views_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1732
- """The maximum time that query_views_log records will be retained before removal. If set to 0,
2189
+ """The maximum time that query_views_log records will be retained before removal. If set to **0**,
1733
2190
  automatic removal of query_views_log data based on time is disabled.
2191
+
2192
+ Default value: **2592000000** (30 days).
1734
2193
  """
1735
2194
 
1736
2195
  @property
1737
2196
  def asynchronous_metric_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1738
- """Enable or disable asynchronous_metric_log system table. Default value: false."""
2197
+ """Enables or disables asynchronous_metric_log system table.
2198
+
2199
+ Default value: **false**.
2200
+
2201
+ Change of the setting is applied with restart.
2202
+
2203
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/asynchronous_metric_log).
2204
+ """
1739
2205
 
1740
2206
  @property
1741
2207
  def asynchronous_metric_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1742
- """The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to 0 (default),
2208
+ """The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to **0**,
1743
2209
  automatic removal of asynchronous_metric_log data based on size is disabled.
2210
+
2211
+ Default value: **0**.
1744
2212
  """
1745
2213
 
1746
2214
  @property
1747
2215
  def asynchronous_metric_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1748
- """The maximum time that asynchronous_metric_log records will be retained before removal. If set to 0,
2216
+ """The maximum time that asynchronous_metric_log records will be retained before removal. If set to **0**,
1749
2217
  automatic removal of asynchronous_metric_log data based on time is disabled.
2218
+
2219
+ Default value: **2592000000** (30 days).
1750
2220
  """
1751
2221
 
1752
2222
  @property
1753
2223
  def session_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1754
- """Enable or disable session_log system table. Default value: false."""
2224
+ """Enables or disables session_log system table.
2225
+
2226
+ Default value: **false**.
2227
+
2228
+ Change of the setting is applied with restart.
2229
+
2230
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/session_log).
2231
+ """
1755
2232
 
1756
2233
  @property
1757
2234
  def session_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1758
- """The maximum size that session_log can grow to before old data will be removed. If set to 0 (default),
2235
+ """The maximum size that session_log can grow to before old data will be removed. If set to **0**,
1759
2236
  automatic removal of session_log data based on size is disabled.
2237
+
2238
+ Default value: **0**.
1760
2239
  """
1761
2240
 
1762
2241
  @property
1763
2242
  def session_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1764
- """The maximum time that session_log records will be retained before removal. If set to 0,
2243
+ """The maximum time that session_log records will be retained before removal. If set to **0**,
1765
2244
  automatic removal of session_log data based on time is disabled.
2245
+
2246
+ Default value: **2592000000** (30 days).
1766
2247
  """
1767
2248
 
1768
2249
  @property
1769
2250
  def zookeeper_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1770
- """Enable or disable zookeeper_log system table. Default value: false."""
2251
+ """Enables or disables zookeeper_log system table.
2252
+
2253
+ Default value: **false**.
2254
+
2255
+ Change of the setting is applied with restart.
2256
+
2257
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/zookeeper_log).
2258
+ """
1771
2259
 
1772
2260
  @property
1773
2261
  def zookeeper_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1774
- """The maximum size that zookeeper_log can grow to before old data will be removed. If set to 0 (default),
2262
+ """The maximum size that zookeeper_log can grow to before old data will be removed. If set to **0**,
1775
2263
  automatic removal of zookeeper_log data based on size is disabled.
2264
+
2265
+ Default value: **0**.
1776
2266
  """
1777
2267
 
1778
2268
  @property
1779
2269
  def zookeeper_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1780
- """The maximum time that zookeeper_log records will be retained before removal. If set to 0,
2270
+ """The maximum time that zookeeper_log records will be retained before removal. If set to **0**,
1781
2271
  automatic removal of zookeeper_log data based on time is disabled.
2272
+
2273
+ Default value: **2592000000** (30 days).
1782
2274
  """
1783
2275
 
1784
2276
  @property
1785
2277
  def asynchronous_insert_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1786
- """Enable or disable asynchronous_insert_log system table. Default value: false.
1787
- Minimal required ClickHouse version: 22.10.
2278
+ """Enables or disables asynchronous_insert_log system table.
2279
+
2280
+ Default value: **false**.
2281
+
2282
+ Change of the setting is applied with restart.
2283
+
2284
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/asynchronous_insert_log).
1788
2285
  """
1789
2286
 
1790
2287
  @property
1791
2288
  def asynchronous_insert_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1792
- """The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to 0 (default),
2289
+ """The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to **0**,
1793
2290
  automatic removal of asynchronous_insert_log data based on size is disabled.
2291
+
2292
+ Default value: **0**.
1794
2293
  """
1795
2294
 
1796
2295
  @property
1797
2296
  def asynchronous_insert_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1798
- """The maximum time that asynchronous_insert_log records will be retained before removal. If set to 0,
2297
+ """The maximum time that asynchronous_insert_log records will be retained before removal. If set to **0**,
1799
2298
  automatic removal of asynchronous_insert_log data based on time is disabled.
2299
+
2300
+ Default value: **2592000000** (30 days).
1800
2301
  """
1801
2302
 
1802
2303
  @property
1803
2304
  def processors_profile_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1804
- """Enable or disable processors_profile_log system table."""
2305
+ """Enables or disables processors_profile_log system table.
2306
+
2307
+ Default value: **true** for versions 25.2 and higher, **false** for versions 25.1 and lower.
2308
+
2309
+ Change of the setting is applied with restart.
2310
+
2311
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/processors_profile_log).
2312
+ """
1805
2313
 
1806
2314
  @property
1807
2315
  def processors_profile_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1808
- """The maximum size that processors_profile_log can grow to before old data will be removed.
1809
- If set to 0 (default), automatic removal of processors_profile_log data based on size is disabled.
2316
+ """The maximum size that processors_profile_log can grow to before old data will be removed. If set to **0**,
2317
+ automatic removal of processors_profile_log data based on size is disabled.
2318
+
2319
+ Default value: **0**.
1810
2320
  """
1811
2321
 
1812
2322
  @property
1813
2323
  def processors_profile_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1814
- """The maximum time that processors_profile_log records will be retained before removal.
1815
- If set to 0, automatic removal of processors_profile_log data based on time is disabled.
2324
+ """The maximum time that processors_profile_log records will be retained before removal. If set to **0**,
2325
+ automatic removal of processors_profile_log data based on time is disabled.
2326
+
2327
+ Default value: **2592000000** (30 days).
1816
2328
  """
1817
2329
 
1818
2330
  @property
1819
2331
  def error_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1820
- """Enable or disable error_log system table."""
2332
+ """Enables or disables error_log system table.
2333
+
2334
+ Default value: **false**.
2335
+
2336
+ Change of the setting is applied with restart.
2337
+
2338
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/error_log).
2339
+ """
1821
2340
 
1822
2341
  @property
1823
2342
  def error_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1824
- """The maximum size that error_log can grow to before old data will be removed.
1825
- If set to 0 (default), automatic removal of error_log data based on size is disabled.
2343
+ """The maximum size that error_log can grow to before old data will be removed. If set to **0**,
2344
+ automatic removal of error_log data based on size is disabled.
2345
+
2346
+ Default value: **0**.
1826
2347
  """
1827
2348
 
1828
2349
  @property
1829
2350
  def error_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1830
- """The maximum time that error_log records will be retained before removal.
1831
- If set to 0, automatic removal of error_log data based on time is disabled.
2351
+ """The maximum time that error_log records will be retained before removal. If set to **0**,
2352
+ automatic removal of error_log data based on time is disabled.
2353
+
2354
+ Default value: **2592000000** (30 days).
1832
2355
  """
1833
2356
 
1834
2357
  @property
@@ -1836,103 +2359,257 @@ class ClickhouseConfig(google.protobuf.message.Message):
1836
2359
  """Access control settings."""
1837
2360
 
1838
2361
  @property
1839
- def background_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
1840
- @property
1841
- def background_merges_mutations_concurrency_ratio(self) -> google.protobuf.wrappers_pb2.Int64Value:
1842
- """Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently. For example, if the ratio equals to 2 and background_pool_size is set to 16 then ClickHouse can execute 32 background merges concurrently. This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority. You can only increase this ratio at runtime. To lower it you have to restart the server. The same as for background_pool_size setting background_merges_mutations_concurrency_ratio could be applied from the default profile for backward compatibility.
1843
- Default: 2
1844
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#background_merges_mutations_concurrency_ratio)
2362
+ def max_connections(self) -> google.protobuf.wrappers_pb2.Int64Value:
2363
+ """Maximum number of inbound connections.
2364
+
2365
+ Default value: **4096**.
2366
+
2367
+ Change of the setting is applied with restart.
2368
+
2369
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_connections).
1845
2370
  """
1846
2371
 
1847
2372
  @property
1848
- def background_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2373
+ def max_concurrent_queries(self) -> google.protobuf.wrappers_pb2.Int64Value:
2374
+ """Maximum number of concurrently executed queries.
2375
+
2376
+ Default value: **500**.
2377
+
2378
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_concurrent_queries).
2379
+ """
2380
+
1849
2381
  @property
1850
- def background_fetches_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1851
- """Sets the number of threads performing background fetches for tables with **ReplicatedMergeTree** engines. Default value: 8.
2382
+ def max_table_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
2383
+ """Maximum size of the table that can be deleted using **DROP** or **TRUNCATE** query.
1852
2384
 
1853
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings/#background_fetches_pool_size).
2385
+ Default value: **50000000000** (48828125 KiB).
2386
+
2387
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_table_size_to_drop).
1854
2388
  """
1855
2389
 
1856
2390
  @property
1857
- def background_move_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2391
+ def max_partition_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
2392
+ """Maximum size of the partition that can be deleted using **DROP** or **TRUNCATE** query.
2393
+
2394
+ Default value: **50000000000** (48828125 KiB).
2395
+
2396
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_partition_size_to_drop).
2397
+ """
2398
+
1858
2399
  @property
1859
- def background_distributed_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2400
+ def keep_alive_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
2401
+ """The number of seconds that ClickHouse waits for incoming requests for HTTP protocol before closing the connection.
2402
+
2403
+ Default value: **30**.
2404
+
2405
+ Change of the setting is applied with restart.
2406
+
2407
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#keep_alive_timeout).
2408
+ """
2409
+
1860
2410
  @property
1861
- def background_buffer_flush_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2411
+ def uncompressed_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
2412
+ """Cache size (in bytes) for uncompressed data used by table engines from the MergeTree family. **0** means disabled.
2413
+
2414
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#uncompressed_cache_size).
2415
+ """
2416
+
1862
2417
  @property
1863
- def background_message_broker_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2418
+ def mark_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
2419
+ """Maximum size (in bytes) of the cache of "marks" used by MergeTree tables.
2420
+
2421
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#mark_cache_size).
2422
+ """
2423
+
1864
2424
  @property
1865
- def background_common_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1866
- """The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for *MergeTree-engine tables in a background.
1867
- Default: 8
1868
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#background_common_pool_size)
2425
+ def geobase_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
2426
+ """Enables or disables geobase.
2427
+
2428
+ Change of the setting is applied with restart.
1869
2429
  """
1870
2430
 
1871
2431
  @property
1872
2432
  def default_database(self) -> google.protobuf.wrappers_pb2.StringValue:
1873
2433
  """The default database.
1874
2434
 
1875
- To get a list of cluster databases, see [Yandex Managed ClickHouse documentation](/docs/managed-clickhouse/operations/databases#list-db).
2435
+ Default value: **default**.
2436
+
2437
+ Change of the setting is applied with restart.
2438
+
2439
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#default_database).
1876
2440
  """
1877
2441
 
1878
2442
  @property
1879
2443
  def total_memory_profiler_step(self) -> google.protobuf.wrappers_pb2.Int64Value:
1880
- """Sets the memory size (in bytes) for a stack trace at every peak allocation step. Default value: **4194304**.
2444
+ """Whenever server memory usage becomes larger than every next step in number of bytes the memory profiler will collect
2445
+ the allocating stack trace. **0** means disabled memory profiler.
2446
+
2447
+ Default value: **0**.
2448
+
2449
+ Change of the setting is applied with restart.
2450
+
2451
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#total_memory_profiler_step).
2452
+ """
2453
+
2454
+ @property
2455
+ def total_memory_tracker_sample_probability(self) -> google.protobuf.wrappers_pb2.DoubleValue:
2456
+ """Allows to collect random allocations and de-allocations and writes them in the system.trace_log system table
2457
+ with trace_type equal to a MemorySample with the specified probability.
2458
+
2459
+ Default value: **0**.
1881
2460
 
1882
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings/#total-memory-profiler-step).
2461
+ Change of the setting is applied with restart.
2462
+
2463
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#total_memory_tracker_sample_probability).
1883
2464
  """
1884
2465
 
1885
2466
  @property
1886
- def total_memory_tracker_sample_probability(self) -> google.protobuf.wrappers_pb2.DoubleValue: ...
2467
+ def async_insert_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
2468
+ """Maximum number of threads to parse and insert data in background. If set to **0**, asynchronous mode is disabled.
2469
+
2470
+ Default value: **16**.
2471
+
2472
+ Change of the setting is applied with restart.
2473
+
2474
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#async_insert_threads).
2475
+ """
2476
+
1887
2477
  @property
1888
- def query_masking_rules(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.QueryMaskingRule]:
1889
- """Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs, system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs.
1890
- Change of these settings is applied with ClickHouse restart
1891
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#query-masking-rules)
2478
+ def backup_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
2479
+ """The maximum number of threads to execute **BACKUP** requests.
2480
+
2481
+ Default value: **16**.
2482
+
2483
+ Change of the setting is applied with restart.
2484
+
2485
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#backup_threads).
1892
2486
  """
1893
2487
 
1894
2488
  @property
1895
- def dictionaries_lazy_load(self) -> google.protobuf.wrappers_pb2.BoolValue:
1896
- """Lazy loading of dictionaries.
1897
- Default: true
1898
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#dictionaries_lazy_load)
2489
+ def restore_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
2490
+ """The maximum number of threads to execute **RESTORE** requests.
2491
+
2492
+ Default value: **16**.
2493
+
2494
+ Change of the setting is applied with restart.
2495
+
2496
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#restore_threads).
2497
+ """
2498
+
2499
+ @property
2500
+ def merge_tree(self) -> global___ClickhouseConfig.MergeTree:
2501
+ """Settings for the MergeTree table engine family.
2502
+
2503
+ Change of the settings of **merge_tree** is applied with restart.
2504
+ """
2505
+
2506
+ @property
2507
+ def compression(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.Compression]:
2508
+ """Data compression settings for MergeTree engine tables.
2509
+
2510
+ Change of the settings of **compression** is applied with restart.
2511
+
2512
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#compression).
2513
+ """
2514
+
2515
+ @property
2516
+ def dictionaries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary]:
2517
+ """Configuration of external dictionaries.
2518
+
2519
+ Change of the settings of **dictionaries** is applied with restart.
2520
+
2521
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries).
2522
+ """
2523
+
2524
+ @property
2525
+ def graphite_rollup(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup]:
2526
+ """Rollup settings for the GraphiteMergeTree engine tables.
2527
+
2528
+ Change of the settings of **graphite_rollup** is applied with restart.
2529
+
2530
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#graphite_rollup).
2531
+ """
2532
+
2533
+ @property
2534
+ def kafka(self) -> global___ClickhouseConfig.Kafka:
2535
+ """Kafka integration settings.
2536
+
2537
+ Change of the settings of **kafka** is applied with restart.
2538
+ """
2539
+
2540
+ @property
2541
+ def kafka_topics(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.KafkaTopic]:
2542
+ """Per-topic Kafka integration settings.
2543
+
2544
+ Change of the settings of **kafka_topics** is applied with restart.
2545
+ """
2546
+
2547
+ @property
2548
+ def rabbitmq(self) -> global___ClickhouseConfig.Rabbitmq:
2549
+ """RabbitMQ integration settings.
2550
+
2551
+ Change of the settings of **rabbitmq** is applied with restart.
2552
+ """
2553
+
2554
+ @property
2555
+ def query_masking_rules(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.QueryMaskingRule]:
2556
+ """Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs,
2557
+ system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing
2558
+ sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs.
2559
+
2560
+ Change of the settings of **query_masking_rules** is applied with restart.
2561
+
2562
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#query_masking_rules).
1899
2563
  """
1900
2564
 
1901
2565
  @property
1902
2566
  def query_cache(self) -> global___ClickhouseConfig.QueryCache:
1903
- """[Query cache](https://clickhouse.com/docs/en/operations/query-cache) configuration.
1904
- Min version: 23.5
1905
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#query_cache)
2567
+ """[Query cache](https://clickhouse.com/docs/operations/query-cache) configuration.
2568
+
2569
+ Change of the settings of **query_cache** is applied with restart.
1906
2570
  """
1907
2571
 
1908
2572
  @property
1909
2573
  def jdbc_bridge(self) -> global___ClickhouseConfig.JdbcBridge:
1910
- """JDBC bridge for queries to external databases.
1911
- https://clickhouse.com/docs/en/integrations/jdbc/jdbc-with-clickhouse
2574
+ """JDBC bridge configuration for queries to external databases.
2575
+
2576
+ Change of the settings of **jdbc_bridge** is applied with restart.
2577
+
2578
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/integrations/jdbc/jdbc-with-clickhouse).
2579
+ """
2580
+
2581
+ @property
2582
+ def mysql_protocol(self) -> google.protobuf.wrappers_pb2.BoolValue:
2583
+ """Enables or disables MySQL interface on ClickHouse server
2584
+
2585
+ Default value: **false**.
2586
+
2587
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/interfaces/mysql).
2588
+ """
2589
+
2590
+ @property
2591
+ def builtin_dictionaries_reload_interval(self) -> google.protobuf.wrappers_pb2.Int64Value:
2592
+ """The interval in seconds before reloading built-in dictionaries.
2593
+
2594
+ Default value: **3600**.
2595
+
2596
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#builtin_dictionaries_reload_interval).
1912
2597
  """
1913
2598
 
1914
2599
  def __init__(
1915
2600
  self,
1916
2601
  *,
2602
+ background_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2603
+ background_merges_mutations_concurrency_ratio: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2604
+ background_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2605
+ background_fetches_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2606
+ background_move_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2607
+ background_distributed_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2608
+ background_buffer_flush_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2609
+ background_message_broker_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2610
+ background_common_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2611
+ dictionaries_lazy_load: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1917
2612
  log_level: global___ClickhouseConfig.LogLevel.ValueType = ...,
1918
- merge_tree: global___ClickhouseConfig.MergeTree | None = ...,
1919
- compression: collections.abc.Iterable[global___ClickhouseConfig.Compression] | None = ...,
1920
- dictionaries: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary] | None = ...,
1921
- graphite_rollup: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup] | None = ...,
1922
- kafka: global___ClickhouseConfig.Kafka | None = ...,
1923
- kafka_topics: collections.abc.Iterable[global___ClickhouseConfig.KafkaTopic] | None = ...,
1924
- rabbitmq: global___ClickhouseConfig.Rabbitmq | None = ...,
1925
- max_connections: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1926
- max_concurrent_queries: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1927
- keep_alive_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1928
- uncompressed_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1929
- mark_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1930
- max_table_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1931
- max_partition_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1932
- builtin_dictionaries_reload_interval: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1933
- timezone: builtins.str = ...,
1934
- geobase_enabled: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1935
- geobase_uri: builtins.str = ...,
1936
2613
  query_log_retention_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1937
2614
  query_log_retention_time: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1938
2615
  query_thread_log_enabled: google.protobuf.wrappers_pb2.BoolValue | None = ...,
@@ -1975,25 +2652,37 @@ class ClickhouseConfig(google.protobuf.message.Message):
1975
2652
  error_log_retention_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1976
2653
  error_log_retention_time: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1977
2654
  access_control_improvements: global___ClickhouseConfig.AccessControlImprovements | None = ...,
1978
- background_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1979
- background_merges_mutations_concurrency_ratio: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1980
- background_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1981
- background_fetches_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1982
- background_move_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1983
- background_distributed_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1984
- background_buffer_flush_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1985
- background_message_broker_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1986
- background_common_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2655
+ max_connections: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2656
+ max_concurrent_queries: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2657
+ max_table_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2658
+ max_partition_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2659
+ keep_alive_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2660
+ uncompressed_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2661
+ mark_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2662
+ timezone: builtins.str = ...,
2663
+ geobase_enabled: google.protobuf.wrappers_pb2.BoolValue | None = ...,
2664
+ geobase_uri: builtins.str = ...,
1987
2665
  default_database: google.protobuf.wrappers_pb2.StringValue | None = ...,
1988
2666
  total_memory_profiler_step: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1989
2667
  total_memory_tracker_sample_probability: google.protobuf.wrappers_pb2.DoubleValue | None = ...,
2668
+ async_insert_threads: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2669
+ backup_threads: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2670
+ restore_threads: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2671
+ merge_tree: global___ClickhouseConfig.MergeTree | None = ...,
2672
+ compression: collections.abc.Iterable[global___ClickhouseConfig.Compression] | None = ...,
2673
+ dictionaries: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary] | None = ...,
2674
+ graphite_rollup: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup] | None = ...,
2675
+ kafka: global___ClickhouseConfig.Kafka | None = ...,
2676
+ kafka_topics: collections.abc.Iterable[global___ClickhouseConfig.KafkaTopic] | None = ...,
2677
+ rabbitmq: global___ClickhouseConfig.Rabbitmq | None = ...,
1990
2678
  query_masking_rules: collections.abc.Iterable[global___ClickhouseConfig.QueryMaskingRule] | None = ...,
1991
- dictionaries_lazy_load: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1992
2679
  query_cache: global___ClickhouseConfig.QueryCache | None = ...,
1993
2680
  jdbc_bridge: global___ClickhouseConfig.JdbcBridge | None = ...,
2681
+ mysql_protocol: google.protobuf.wrappers_pb2.BoolValue | None = ...,
2682
+ builtin_dictionaries_reload_interval: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1994
2683
  ) -> None: ...
1995
- def HasField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "default_database", b"default_database", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "keep_alive_timeout", b"keep_alive_timeout", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> builtins.bool: ...
1996
- def ClearField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "compression", b"compression", "default_database", b"default_database", "dictionaries", b"dictionaries", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "geobase_uri", b"geobase_uri", "graphite_rollup", b"graphite_rollup", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "kafka_topics", b"kafka_topics", "keep_alive_timeout", b"keep_alive_timeout", "log_level", b"log_level", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_masking_rules", b"query_masking_rules", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_level", b"text_log_level", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "timezone", b"timezone", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> None: ...
2684
+ def HasField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "async_insert_threads", b"async_insert_threads", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "backup_threads", b"backup_threads", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "default_database", b"default_database", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "keep_alive_timeout", b"keep_alive_timeout", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "mysql_protocol", b"mysql_protocol", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "restore_threads", b"restore_threads", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> builtins.bool: ...
2685
+ def ClearField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "async_insert_threads", b"async_insert_threads", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "backup_threads", b"backup_threads", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "compression", b"compression", "default_database", b"default_database", "dictionaries", b"dictionaries", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "geobase_uri", b"geobase_uri", "graphite_rollup", b"graphite_rollup", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "kafka_topics", b"kafka_topics", "keep_alive_timeout", b"keep_alive_timeout", "log_level", b"log_level", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "mysql_protocol", b"mysql_protocol", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_masking_rules", b"query_masking_rules", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "restore_threads", b"restore_threads", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_level", b"text_log_level", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "timezone", b"timezone", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> None: ...
1997
2686
 
1998
2687
  global___ClickhouseConfig = ClickhouseConfig
1999
2688
 
@@ -2006,17 +2695,15 @@ class ClickhouseConfigSet(google.protobuf.message.Message):
2006
2695
  DEFAULT_CONFIG_FIELD_NUMBER: builtins.int
2007
2696
  @property
2008
2697
  def effective_config(self) -> global___ClickhouseConfig:
2009
- """Effective settings for a ClickHouse cluster (a combination of settings defined
2010
- in [user_config] and [default_config]).
2011
- """
2698
+ """Effective configuration (a combination of user-defined configuration and default configuration)."""
2012
2699
 
2013
2700
  @property
2014
2701
  def user_config(self) -> global___ClickhouseConfig:
2015
- """User-defined settings for a ClickHouse cluster."""
2702
+ """User-defined configuration."""
2016
2703
 
2017
2704
  @property
2018
2705
  def default_config(self) -> global___ClickhouseConfig:
2019
- """Default configuration for a ClickHouse cluster."""
2706
+ """Default configuration."""
2020
2707
 
2021
2708
  def __init__(
2022
2709
  self,