yandexcloud 0.354.0__py3-none-any.whl → 0.356.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of yandexcloud might be problematic. Click here for more details.

Files changed (102) hide show
  1. yandex/cloud/ai/assistants/v1/common_pb2.py +58 -30
  2. yandex/cloud/ai/assistants/v1/common_pb2.pyi +178 -3
  3. yandex/cloud/airflow/v1/cluster_pb2.py +12 -8
  4. yandex/cloud/airflow/v1/cluster_pb2.pyi +4 -0
  5. yandex/cloud/apploadbalancer/v1/virtual_host_pb2.py +32 -32
  6. yandex/cloud/apploadbalancer/v1/virtual_host_pb2.pyi +5 -1
  7. yandex/cloud/apploadbalancer/v1/virtual_host_service_pb2.py +6 -6
  8. yandex/cloud/apploadbalancer/v1/virtual_host_service_pb2.pyi +5 -1
  9. yandex/cloud/backup/v1/backup_service_pb2.py +46 -40
  10. yandex/cloud/backup/v1/backup_service_pb2.pyi +26 -4
  11. yandex/cloud/backup/v1/resource_pb2.py +2 -2
  12. yandex/cloud/backup/v1/resource_pb2.pyi +8 -0
  13. yandex/cloud/baremetal/v1alpha/configuration_service_pb2.py +18 -9
  14. yandex/cloud/baremetal/v1alpha/configuration_service_pb2.pyi +16 -2
  15. yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2.py +14 -9
  16. yandex/cloud/baremetal/v1alpha/image_service_pb2.py +39 -32
  17. yandex/cloud/baremetal/v1alpha/private_cloud_connection_service_pb2.py +29 -24
  18. yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2.py +43 -36
  19. yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2.py +39 -32
  20. yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2.pyi +3 -2
  21. yandex/cloud/baremetal/v1alpha/server_service_pb2.py +83 -76
  22. yandex/cloud/baremetal/v1alpha/storage_service_pb2.py +16 -11
  23. yandex/cloud/baremetal/v1alpha/vrf_service_pb2.py +39 -32
  24. yandex/cloud/baremetal/v1alpha/zone_service_pb2.py +14 -9
  25. yandex/cloud/compute/v1/maintenance_service_pb2.py +47 -0
  26. yandex/cloud/compute/v1/maintenance_service_pb2.pyi +8 -0
  27. yandex/cloud/compute/v1/maintenance_service_pb2_grpc.py +190 -0
  28. yandex/cloud/compute/v1/maintenance_service_pb2_grpc.pyi +91 -0
  29. yandex/cloud/compute/v1/reserved_instance_pool_pb2.py +8 -4
  30. yandex/cloud/compute/v1/reserved_instance_pool_pb2.pyi +69 -2
  31. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2.py +44 -20
  32. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2.pyi +138 -2
  33. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2_grpc.py +88 -0
  34. yandex/cloud/compute/v1/reserved_instance_pool_service_pb2_grpc.pyi +40 -0
  35. yandex/cloud/datatransfer/v1/transfer_pb2.py +50 -48
  36. yandex/cloud/datatransfer/v1/transfer_pb2.pyi +23 -4
  37. yandex/cloud/datatransfer/v1/transfer_service_pb2.py +32 -32
  38. yandex/cloud/datatransfer/v1/transfer_service_pb2.pyi +12 -4
  39. yandex/cloud/iam/v1/oauth_client_pb2.py +39 -0
  40. yandex/cloud/iam/v1/oauth_client_pb2.pyi +85 -0
  41. yandex/cloud/iam/v1/oauth_client_pb2_grpc.py +24 -0
  42. yandex/cloud/iam/v1/oauth_client_pb2_grpc.pyi +17 -0
  43. yandex/cloud/iam/v1/oauth_client_secret_pb2.py +38 -0
  44. yandex/cloud/iam/v1/oauth_client_secret_pb2.pyi +49 -0
  45. yandex/cloud/iam/v1/oauth_client_secret_pb2_grpc.py +24 -0
  46. yandex/cloud/iam/v1/oauth_client_secret_pb2_grpc.pyi +17 -0
  47. yandex/cloud/iam/v1/oauth_client_secret_service_pb2.py +80 -0
  48. yandex/cloud/iam/v1/oauth_client_secret_service_pb2.pyi +188 -0
  49. yandex/cloud/iam/v1/oauth_client_secret_service_pb2_grpc.py +237 -0
  50. yandex/cloud/iam/v1/oauth_client_secret_service_pb2_grpc.pyi +121 -0
  51. yandex/cloud/iam/v1/oauth_client_service_pb2.py +99 -0
  52. yandex/cloud/iam/v1/oauth_client_service_pb2.pyi +257 -0
  53. yandex/cloud/iam/v1/oauth_client_service_pb2_grpc.py +281 -0
  54. yandex/cloud/iam/v1/oauth_client_service_pb2_grpc.pyi +141 -0
  55. yandex/cloud/iot/broker/v1/broker_service_pb2.py +50 -50
  56. yandex/cloud/iot/devices/v1/device_service_pb2.py +61 -61
  57. yandex/cloud/iot/devices/v1/registry_service_pb2.py +82 -82
  58. yandex/cloud/maintenance/__init__.py +0 -0
  59. yandex/cloud/maintenance/v2/__init__.py +0 -0
  60. yandex/cloud/maintenance/v2/common_pb2.py +73 -0
  61. yandex/cloud/maintenance/v2/common_pb2.pyi +324 -0
  62. yandex/cloud/maintenance/v2/common_pb2_grpc.py +24 -0
  63. yandex/cloud/maintenance/v2/common_pb2_grpc.pyi +17 -0
  64. yandex/cloud/mdb/clickhouse/v1/config/clickhouse_pb2.py +196 -152
  65. yandex/cloud/mdb/clickhouse/v1/config/clickhouse_pb2.pyi +1779 -1066
  66. yandex/cloud/mdb/clickhouse/v1/user_pb2.py +98 -86
  67. yandex/cloud/mdb/clickhouse/v1/user_pb2.pyi +1248 -792
  68. yandex/cloud/mdb/opensearch/v1/backup_pb2.py +14 -14
  69. yandex/cloud/mdb/opensearch/v1/backup_pb2.pyi +13 -1
  70. yandex/cloud/mdb/spqr/v1/cluster_service_pb2.py +76 -76
  71. yandex/cloud/mdb/spqr/v1/cluster_service_pb2.pyi +4 -0
  72. yandex/cloud/metastore/v1/cluster_pb2.py +18 -16
  73. yandex/cloud/metastore/v1/cluster_pb2.pyi +31 -1
  74. yandex/cloud/searchapi/v2/gen_search_service_pb2.py +22 -22
  75. yandex/cloud/searchapi/v2/gen_search_service_pb2.pyi +12 -2
  76. yandex/cloud/serverless/containers/v1/container_pb2.py +46 -44
  77. yandex/cloud/serverless/containers/v1/container_pb2.pyi +25 -3
  78. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_pb2.py +50 -18
  79. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_pb2.pyi +241 -2
  80. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_service_pb2.py +18 -18
  81. yandex/cloud/smartwebsecurity/v1/waf/waf_profile_service_pb2.pyi +22 -2
  82. yandex/cloud/storage/v1/presign_service_pb2.py +47 -0
  83. yandex/cloud/storage/v1/presign_service_pb2.pyi +117 -0
  84. yandex/cloud/storage/v1/presign_service_pb2_grpc.py +101 -0
  85. yandex/cloud/storage/v1/presign_service_pb2_grpc.pyi +50 -0
  86. yandex/cloud/trino/v1/access_control_pb2.py +212 -0
  87. yandex/cloud/trino/v1/access_control_pb2.pyi +926 -0
  88. yandex/cloud/trino/v1/access_control_pb2_grpc.py +24 -0
  89. yandex/cloud/trino/v1/access_control_pb2_grpc.pyi +17 -0
  90. yandex/cloud/trino/v1/cluster_pb2.py +44 -43
  91. yandex/cloud/trino/v1/cluster_pb2.pyi +9 -2
  92. yandex/cloud/trino/v1/cluster_service_pb2.py +50 -49
  93. yandex/cloud/trino/v1/cluster_service_pb2.pyi +29 -4
  94. yandex/cloud/ytsaurus/v1/cluster_pb2.py +40 -37
  95. yandex/cloud/ytsaurus/v1/cluster_pb2.pyi +32 -18
  96. yandexcloud/__init__.py +1 -1
  97. {yandexcloud-0.354.0.dist-info → yandexcloud-0.356.0.dist-info}/METADATA +1 -1
  98. {yandexcloud-0.354.0.dist-info → yandexcloud-0.356.0.dist-info}/RECORD +102 -68
  99. {yandexcloud-0.354.0.dist-info → yandexcloud-0.356.0.dist-info}/AUTHORS +0 -0
  100. {yandexcloud-0.354.0.dist-info → yandexcloud-0.356.0.dist-info}/LICENSE +0 -0
  101. {yandexcloud-0.354.0.dist-info → yandexcloud-0.356.0.dist-info}/WHEEL +0 -0
  102. {yandexcloud-0.354.0.dist-info → yandexcloud-0.356.0.dist-info}/top_level.txt +0 -0
@@ -22,10 +22,8 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
22
22
 
23
23
  @typing.final
24
24
  class ClickhouseConfig(google.protobuf.message.Message):
25
- """ClickHouse configuration options. Detailed description for each set of options
26
- is available in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/).
27
-
28
- Any options not listed here are not supported.
25
+ """ClickHouse configuration settings. Supported settings are a subset of settings described
26
+ in [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings).
29
27
  """
30
28
 
31
29
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
@@ -51,9 +49,44 @@ class ClickhouseConfig(google.protobuf.message.Message):
51
49
  WARNING: ClickhouseConfig.LogLevel.ValueType # 4
52
50
  ERROR: ClickhouseConfig.LogLevel.ValueType # 5
53
51
 
52
+ @typing.final
53
+ class AccessControlImprovements(google.protobuf.message.Message):
54
+ """Access control settings.
55
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#access_control_improvements).
56
+ """
57
+
58
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
59
+
60
+ SELECT_FROM_SYSTEM_DB_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
61
+ SELECT_FROM_INFORMATION_SCHEMA_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
62
+ @property
63
+ def select_from_system_db_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
64
+ """Sets whether **SELECT * FROM system.<table>** requires any grants and can be executed by any user.
65
+ If set to true then this query requires **GRANT SELECT ON system.<table>** just as for non-system tables.
66
+
67
+ Default value: **false**.
68
+ """
69
+
70
+ @property
71
+ def select_from_information_schema_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
72
+ """Sets whether **SELECT * FROM information_schema.<table>** requires any grants and can be executed by any user.
73
+ If set to true, then this query requires **GRANT SELECT ON information_schema.<table>**, just as for ordinary tables.
74
+
75
+ Default value: **false**.
76
+ """
77
+
78
+ def __init__(
79
+ self,
80
+ *,
81
+ select_from_system_db_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
82
+ select_from_information_schema_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
83
+ ) -> None: ...
84
+ def HasField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> builtins.bool: ...
85
+ def ClearField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> None: ...
86
+
54
87
  @typing.final
55
88
  class MergeTree(google.protobuf.message.Message):
56
- """Options specific to the MergeTree table engine."""
89
+ """Settings for the MergeTree table engine family."""
57
90
 
58
91
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
59
92
 
@@ -69,7 +102,11 @@ class ClickhouseConfig(google.protobuf.message.Message):
69
102
  DEDUPLICATE_MERGE_PROJECTION_MODE_DROP: ClickhouseConfig.MergeTree._DeduplicateMergeProjectionMode.ValueType # 3
70
103
  DEDUPLICATE_MERGE_PROJECTION_MODE_REBUILD: ClickhouseConfig.MergeTree._DeduplicateMergeProjectionMode.ValueType # 4
71
104
 
72
- class DeduplicateMergeProjectionMode(_DeduplicateMergeProjectionMode, metaclass=_DeduplicateMergeProjectionModeEnumTypeWrapper): ...
105
+ class DeduplicateMergeProjectionMode(_DeduplicateMergeProjectionMode, metaclass=_DeduplicateMergeProjectionModeEnumTypeWrapper):
106
+ """Determines the behavior of background merges for MergeTree tables with projections.
107
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#deduplicate_merge_projection_mode).
108
+ """
109
+
73
110
  DEDUPLICATE_MERGE_PROJECTION_MODE_UNSPECIFIED: ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType # 0
74
111
  DEDUPLICATE_MERGE_PROJECTION_MODE_IGNORE: ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType # 1
75
112
  DEDUPLICATE_MERGE_PROJECTION_MODE_THROW: ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType # 2
@@ -87,488 +124,470 @@ class ClickhouseConfig(google.protobuf.message.Message):
87
124
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_DROP: ClickhouseConfig.MergeTree._LightweightMutationProjectionMode.ValueType # 2
88
125
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_REBUILD: ClickhouseConfig.MergeTree._LightweightMutationProjectionMode.ValueType # 3
89
126
 
90
- class LightweightMutationProjectionMode(_LightweightMutationProjectionMode, metaclass=_LightweightMutationProjectionModeEnumTypeWrapper): ...
127
+ class LightweightMutationProjectionMode(_LightweightMutationProjectionMode, metaclass=_LightweightMutationProjectionModeEnumTypeWrapper):
128
+ """Determines the behavior of lightweight deletes for MergeTree tables with projections."""
129
+
91
130
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_UNSPECIFIED: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 0
92
131
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_THROW: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 1
93
132
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_DROP: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 2
94
133
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_REBUILD: ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType # 3
95
134
 
96
- REPLICATED_DEDUPLICATION_WINDOW_FIELD_NUMBER: builtins.int
97
- REPLICATED_DEDUPLICATION_WINDOW_SECONDS_FIELD_NUMBER: builtins.int
98
135
  PARTS_TO_DELAY_INSERT_FIELD_NUMBER: builtins.int
99
136
  PARTS_TO_THROW_INSERT_FIELD_NUMBER: builtins.int
100
137
  INACTIVE_PARTS_TO_DELAY_INSERT_FIELD_NUMBER: builtins.int
101
138
  INACTIVE_PARTS_TO_THROW_INSERT_FIELD_NUMBER: builtins.int
139
+ MAX_AVG_PART_SIZE_FOR_TOO_MANY_PARTS_FIELD_NUMBER: builtins.int
140
+ MAX_PARTS_IN_TOTAL_FIELD_NUMBER: builtins.int
102
141
  MAX_REPLICATED_MERGES_IN_QUEUE_FIELD_NUMBER: builtins.int
103
142
  NUMBER_OF_FREE_ENTRIES_IN_POOL_TO_LOWER_MAX_SIZE_OF_MERGE_FIELD_NUMBER: builtins.int
143
+ NUMBER_OF_FREE_ENTRIES_IN_POOL_TO_EXECUTE_MUTATION_FIELD_NUMBER: builtins.int
104
144
  MAX_BYTES_TO_MERGE_AT_MIN_SPACE_IN_POOL_FIELD_NUMBER: builtins.int
105
145
  MAX_BYTES_TO_MERGE_AT_MAX_SPACE_IN_POOL_FIELD_NUMBER: builtins.int
106
146
  MIN_BYTES_FOR_WIDE_PART_FIELD_NUMBER: builtins.int
107
147
  MIN_ROWS_FOR_WIDE_PART_FIELD_NUMBER: builtins.int
108
- TTL_ONLY_DROP_PARTS_FIELD_NUMBER: builtins.int
109
- ALLOW_REMOTE_FS_ZERO_COPY_REPLICATION_FIELD_NUMBER: builtins.int
110
- MERGE_WITH_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
111
- MERGE_WITH_RECOMPRESSION_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
112
- MAX_PARTS_IN_TOTAL_FIELD_NUMBER: builtins.int
113
- MAX_NUMBER_OF_MERGES_WITH_TTL_IN_POOL_FIELD_NUMBER: builtins.int
114
148
  CLEANUP_DELAY_PERIOD_FIELD_NUMBER: builtins.int
115
- NUMBER_OF_FREE_ENTRIES_IN_POOL_TO_EXECUTE_MUTATION_FIELD_NUMBER: builtins.int
116
- MAX_AVG_PART_SIZE_FOR_TOO_MANY_PARTS_FIELD_NUMBER: builtins.int
149
+ MAX_CLEANUP_DELAY_PERIOD_FIELD_NUMBER: builtins.int
150
+ MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
151
+ MAX_MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
117
152
  MIN_AGE_TO_FORCE_MERGE_SECONDS_FIELD_NUMBER: builtins.int
118
153
  MIN_AGE_TO_FORCE_MERGE_ON_PARTITION_ONLY_FIELD_NUMBER: builtins.int
119
- MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
120
154
  MERGE_MAX_BLOCK_SIZE_FIELD_NUMBER: builtins.int
121
- CHECK_SAMPLE_COLUMN_IS_CORRECT_FIELD_NUMBER: builtins.int
122
- MAX_MERGE_SELECTING_SLEEP_MS_FIELD_NUMBER: builtins.int
123
- MAX_CLEANUP_DELAY_PERIOD_FIELD_NUMBER: builtins.int
124
155
  DEDUPLICATE_MERGE_PROJECTION_MODE_FIELD_NUMBER: builtins.int
125
156
  LIGHTWEIGHT_MUTATION_PROJECTION_MODE_FIELD_NUMBER: builtins.int
126
- MATERIALIZE_TTL_RECALCULATE_ONLY_FIELD_NUMBER: builtins.int
157
+ REPLICATED_DEDUPLICATION_WINDOW_FIELD_NUMBER: builtins.int
158
+ REPLICATED_DEDUPLICATION_WINDOW_SECONDS_FIELD_NUMBER: builtins.int
127
159
  FSYNC_AFTER_INSERT_FIELD_NUMBER: builtins.int
128
160
  FSYNC_PART_DIRECTORY_FIELD_NUMBER: builtins.int
129
161
  MIN_COMPRESSED_BYTES_TO_FSYNC_AFTER_FETCH_FIELD_NUMBER: builtins.int
130
162
  MIN_COMPRESSED_BYTES_TO_FSYNC_AFTER_MERGE_FIELD_NUMBER: builtins.int
131
163
  MIN_ROWS_TO_FSYNC_AFTER_MERGE_FIELD_NUMBER: builtins.int
164
+ TTL_ONLY_DROP_PARTS_FIELD_NUMBER: builtins.int
165
+ MERGE_WITH_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
166
+ MERGE_WITH_RECOMPRESSION_TTL_TIMEOUT_FIELD_NUMBER: builtins.int
167
+ MAX_NUMBER_OF_MERGES_WITH_TTL_IN_POOL_FIELD_NUMBER: builtins.int
168
+ MATERIALIZE_TTL_RECALCULATE_ONLY_FIELD_NUMBER: builtins.int
169
+ CHECK_SAMPLE_COLUMN_IS_CORRECT_FIELD_NUMBER: builtins.int
170
+ ALLOW_REMOTE_FS_ZERO_COPY_REPLICATION_FIELD_NUMBER: builtins.int
132
171
  deduplicate_merge_projection_mode: global___ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType
133
172
  """Determines the behavior of background merges for MergeTree tables with projections.
134
- https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#deduplicate_merge_projection_mode
173
+
174
+ Default value: **DEDUPLICATE_MERGE_PROJECTION_MODE_THROW**.
175
+
176
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#deduplicate_merge_projection_mode).
135
177
  """
136
178
  lightweight_mutation_projection_mode: global___ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType
137
- """Determines the behavior of lightweight deletes for MergeTree tables with projections."""
138
- @property
139
- def replicated_deduplication_window(self) -> google.protobuf.wrappers_pb2.Int64Value:
140
- """Number of blocks of hashes to keep in ZooKeeper."""
179
+ """Determines the behavior of lightweight deletes for MergeTree tables with projections.
141
180
 
142
- @property
143
- def replicated_deduplication_window_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
144
- """Period of time to keep blocks of hashes for."""
181
+ Default value: **LIGHTWEIGHT_MUTATION_PROJECTION_MODE_THROW**.
145
182
 
183
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#lightweight_mutation_projection_mode).
184
+ """
146
185
  @property
147
186
  def parts_to_delay_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
148
- """If table contains at least that many active parts in single partition, artificially slow down insert into table."""
187
+ """If the number of active parts in a single partition exceeds the **parts_to_delay_insert** value, an **INSERT** artificially slows down.
188
+
189
+ Default value: **1000** for versions 25.1 and higher, **150** for versions 24.12 and lower.
190
+
191
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#parts_to_delay_insert).
192
+ """
149
193
 
150
194
  @property
151
195
  def parts_to_throw_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
152
- """If more than this number active parts in single partition, throw 'Too many parts ...' exception."""
196
+ """If the number of active parts in a single partition exceeds the **parts_to_throw_insert** value, an **INSERT**
197
+ is interrupted with the error "Too many parts (N). Merges are processing significantly slower than inserts".
198
+
199
+ Default value: **3000** for versions 25.1 and higher, **300** for versions 24.12 and lower.
200
+
201
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#parts_to_throw_insert).
202
+ """
203
+
204
+ @property
205
+ def inactive_parts_to_delay_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
206
+ """If the number of inactive parts in a single partition in the table exceeds the **inactive_parts_to_delay_insert** value,
207
+ an **INSERT** is artificially slowed down.
208
+
209
+ Default value: **0**.
210
+
211
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#inactive_parts_to_delay_insert).
212
+ """
213
+
214
+ @property
215
+ def inactive_parts_to_throw_insert(self) -> google.protobuf.wrappers_pb2.Int64Value:
216
+ """If the number of inactive parts in a single partition more than the **inactive_parts_to_throw_insert** value,
217
+ **INSERT** is interrupted with an error.
218
+
219
+ Default value: **0**.
220
+
221
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#inactive_parts_to_throw_insert).
222
+ """
153
223
 
154
224
  @property
155
- def inactive_parts_to_delay_insert(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
225
+ def max_avg_part_size_for_too_many_parts(self) -> google.protobuf.wrappers_pb2.Int64Value:
226
+ """The "Too many parts" check according to **parts_to_delay_insert** and **parts_to_throw_insert** will be active only if the average
227
+ part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold,
228
+ **INSERT** queries will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server
229
+ if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts.
230
+
231
+ Default value: **1073741824** (1 GiB).
232
+
233
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_avg_part_size_for_too_many_parts).
234
+ """
235
+
156
236
  @property
157
- def inactive_parts_to_throw_insert(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
237
+ def max_parts_in_total(self) -> google.protobuf.wrappers_pb2.Int64Value:
238
+ """If the total number of active parts in all partitions of a table exceeds the **max_parts_in_total** value,
239
+ an **INSERT** is interrupted with the error "Too many parts (N)".
240
+
241
+ Default value: **20000** for versions 25.2 and higher, **100000** for versions 25.1 and lower.
242
+
243
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_parts_in_total).
244
+ """
245
+
158
246
  @property
159
247
  def max_replicated_merges_in_queue(self) -> google.protobuf.wrappers_pb2.Int64Value:
160
- """How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue."""
248
+ """How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue.
249
+
250
+ Default value: **16**.
251
+ """
161
252
 
162
253
  @property
163
254
  def number_of_free_entries_in_pool_to_lower_max_size_of_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
164
- """If there is less than specified number of free entries in background pool (or replicated queue), start to lower
165
- maximum size of merge to process.
255
+ """When there is less than the specified number of free entries in pool (or replicated queue), start to lower maximum size of
256
+ merge to process (or to put in queue). This is to allow small merges to process - not filling the pool with long running merges.
257
+
258
+ Default value: **8**.
259
+
260
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#number_of_free_entries_in_pool_to_lower_max_size_of_merge).
261
+ """
262
+
263
+ @property
264
+ def number_of_free_entries_in_pool_to_execute_mutation(self) -> google.protobuf.wrappers_pb2.Int64Value:
265
+ """When there is less than specified number of free entries in pool, do not execute part mutations.
266
+ This is to leave free threads for regular merges and to avoid "Too many parts" errors.
267
+
268
+ Default value: **20**.
269
+
270
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#number_of_free_entries_in_pool_to_execute_mutation).
166
271
  """
167
272
 
168
273
  @property
169
274
  def max_bytes_to_merge_at_min_space_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value:
170
- """Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries
171
- in replication queue).
275
+ """The maximum total part size (in bytes) to be merged into one part, with the minimum available resources in the background pool.
276
+
277
+ Default value: **1048576** (1 MiB).
278
+
279
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_bytes_to_merge_at_min_space_in_pool).
172
280
  """
173
281
 
174
282
  @property
175
- def max_bytes_to_merge_at_max_space_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
283
+ def max_bytes_to_merge_at_max_space_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value:
284
+ """The maximum total parts size (in bytes) to be merged into one part, if there are enough resources available.
285
+ Corresponds roughly to the maximum possible part size created by an automatic background merge. **0** means merges will be disabled.
286
+
287
+ Default value: **161061273600** (150 GiB).
288
+
289
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_bytes_to_merge_at_max_space_in_pool).
290
+ """
291
+
176
292
  @property
177
293
  def min_bytes_for_wide_part(self) -> google.protobuf.wrappers_pb2.Int64Value:
178
- """Minimum number of bytes in a data part that can be stored in **Wide** format.
294
+ """Minimum number of bytes in a data part that can be stored in Wide format.
179
295
 
180
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/engines/table-engines/mergetree-family/mergetree/#min_bytes_for_wide_part).
296
+ Default value: **10485760** (10 MiB).
297
+
298
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_bytes_for_wide_part).
181
299
  """
182
300
 
183
301
  @property
184
302
  def min_rows_for_wide_part(self) -> google.protobuf.wrappers_pb2.Int64Value:
185
- """Minimum number of rows in a data part that can be stored in **Wide** format.
303
+ """Minimum number of rows in a data part that can be stored in Wide format.
304
+
305
+ Default value: **0**.
186
306
 
187
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/engines/table-engines/mergetree-family/mergetree/#min_bytes_for_wide_part).
307
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_rows_for_wide_part).
188
308
  """
189
309
 
190
310
  @property
191
- def ttl_only_drop_parts(self) -> google.protobuf.wrappers_pb2.BoolValue:
192
- """Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables.
311
+ def cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value:
312
+ """Minimum period to clean old queue logs, blocks hashes and parts.
313
+
314
+ Default value: **30**.
193
315
 
194
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/settings/#ttl_only_drop_parts).
316
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#cleanup_delay_period).
195
317
  """
196
318
 
197
319
  @property
198
- def allow_remote_fs_zero_copy_replication(self) -> google.protobuf.wrappers_pb2.BoolValue: ...
199
- @property
200
- def merge_with_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
320
+ def max_cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value:
321
+ """Maximum period to clean old queue logs, blocks hashes and parts.
322
+
323
+ Default value: **300** (5 minutes).
324
+
325
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_cleanup_delay_period).
326
+ """
327
+
201
328
  @property
202
- def merge_with_recompression_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
329
+ def merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
330
+ """Minimum time to wait before trying to select parts to merge again after no parts were selected. A lower setting value will trigger
331
+ selecting tasks in background_schedule_pool frequently which result in large amount of requests to Keeper in large-scale clusters.
332
+
333
+ Default value: **5000** (5 seconds).
334
+
335
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_selecting_sleep_ms).
336
+ """
337
+
203
338
  @property
204
- def max_parts_in_total(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
339
+ def max_merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
340
+ """Maximum time to wait before trying to select parts to merge again after no parts were selected. A lower setting value will trigger
341
+ selecting tasks in background_schedule_pool frequently which result in large amount of requests to Keeper in large-scale clusters.
342
+
343
+ Default value: **60000** (1 minute).
344
+
345
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_merge_selecting_sleep_ms).
346
+ """
347
+
205
348
  @property
206
- def max_number_of_merges_with_ttl_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
349
+ def min_age_to_force_merge_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
350
+ """Merge parts if every part in the range is older than the specified value. **0** means disabled.
351
+
352
+ Default value: **0**.
353
+
354
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_age_to_force_merge_seconds).
355
+ """
356
+
207
357
  @property
208
- def cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
358
+ def min_age_to_force_merge_on_partition_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
359
+ """Whether **min_age_to_force_merge_seconds** should be applied only on the entire partition and not on subset.
360
+
361
+ Default value: **false**.
362
+
363
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_age_to_force_merge_on_partition_only).
364
+ """
365
+
209
366
  @property
210
- def number_of_free_entries_in_pool_to_execute_mutation(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
367
+ def merge_max_block_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
368
+ """The number of rows that are read from the merged parts into memory.
369
+
370
+ Default value: **8192**.
371
+
372
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_max_block_size).
373
+ """
374
+
211
375
  @property
212
- def max_avg_part_size_for_too_many_parts(self) -> google.protobuf.wrappers_pb2.Int64Value:
213
- """The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts.
214
- Default: 1 GiB
215
- Min version: 22.10
216
- See in-depth description in [ClickHouse GitHub](https://github.com/ClickHouse/ClickHouse/blob/f9558345e886876b9132d9c018e357f7fa9b22a3/src/Storages/MergeTree/MergeTreeSettings.h#L80)
376
+ def replicated_deduplication_window(self) -> google.protobuf.wrappers_pb2.Int64Value:
377
+ """The number of most recently inserted blocks for which ClickHouse Keeper stores hash sums to check for duplicates.
378
+
379
+ Default value: **1000** for versions 23.11 and higher, **100** for versions 23.10 and lower.
380
+
381
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#replicated_deduplication_window).
217
382
  """
218
383
 
219
384
  @property
220
- def min_age_to_force_merge_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
221
- """Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds.
222
- Default: 0 - disabled
223
- Min_version: 22.10
224
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#min_age_to_force_merge_seconds)
385
+ def replicated_deduplication_window_seconds(self) -> google.protobuf.wrappers_pb2.Int64Value:
386
+ """The number of seconds after which the hash sums of the inserted blocks are removed from ClickHouse Keeper.
387
+
388
+ Default value: **604800** (7 days).
389
+
390
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#replicated_deduplication_window_seconds).
225
391
  """
226
392
 
227
393
  @property
228
- def min_age_to_force_merge_on_partition_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
229
- """Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset.
230
- Default: false
231
- Min_version: 22.11
232
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#min_age_to_force_merge_seconds)
394
+ def fsync_after_insert(self) -> google.protobuf.wrappers_pb2.BoolValue:
395
+ """Do fsync for every inserted part. Significantly decreases performance of inserts, not recommended to use with wide parts.
396
+
397
+ Default value: **false**.
398
+
399
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#fsync_after_insert).
233
400
  """
234
401
 
235
402
  @property
236
- def merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
237
- """Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters.
238
- Default: 5000
239
- Min_version: 21.10
240
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/settings#merge_selecting_sleep_ms)
403
+ def fsync_part_directory(self) -> google.protobuf.wrappers_pb2.BoolValue:
404
+ """Do fsync for part directory after all part operations (writes, renames, etc.).
405
+
406
+ Default value: **false**.
407
+
408
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#fsync_part_directory).
241
409
  """
242
410
 
243
411
  @property
244
- def merge_max_block_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
245
- """The number of rows that are read from the merged parts into memory.
246
- Default: 8192
247
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/settings#merge_max_block_size)
412
+ def min_compressed_bytes_to_fsync_after_fetch(self) -> google.protobuf.wrappers_pb2.Int64Value:
413
+ """Minimal number of compressed bytes to do fsync for part after fetch. **0** means disabled.
414
+
415
+ Default value: **0**.
416
+
417
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_compressed_bytes_to_fsync_after_fetch).
248
418
  """
249
419
 
250
420
  @property
251
- def check_sample_column_is_correct(self) -> google.protobuf.wrappers_pb2.BoolValue:
252
- """Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned [integer types](https://clickhouse.com/docs/en/sql-reference/data-types/int-uint): UInt8, UInt16, UInt32, UInt64.
253
- Default: true
254
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/settings/merge-tree-settings#check_sample_column_is_correct)
421
+ def min_compressed_bytes_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
422
+ """Minimal number of compressed bytes to do fsync for part after merge. **0** means disabled.
423
+
424
+ Default value: **0**.
425
+
426
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_compressed_bytes_to_fsync_after_merge).
255
427
  """
256
428
 
257
429
  @property
258
- def max_merge_selecting_sleep_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
259
- """Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters.
260
- Default: 60000
261
- Min_version: 23.6
262
- See in-depth description in [ClickHouse GitHub](https://github.com/ClickHouse/ClickHouse/blob/4add9db84859bff7410cf934a3904b0414e36e51/src/Storages/MergeTree/MergeTreeSettings.h#L71)
430
+ def min_rows_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
431
+ """Minimal number of rows to do fsync for part after merge. **0** means disabled.
432
+
433
+ Default value: **0**.
434
+
435
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#min_rows_to_fsync_after_merge).
263
436
  """
264
437
 
265
438
  @property
266
- def max_cleanup_delay_period(self) -> google.protobuf.wrappers_pb2.Int64Value:
267
- """Maximum period to clean old queue logs, blocks hashes and parts.
268
- Default: 300
269
- Min_version: 23.6
270
- See in-depth description in [ClickHouse GitHub](https://github.com/ClickHouse/ClickHouse/blob/4add9db84859bff7410cf934a3904b0414e36e51/src/Storages/MergeTree/MergeTreeSettings.h#L142)
439
+ def ttl_only_drop_parts(self) -> google.protobuf.wrappers_pb2.BoolValue:
440
+ """Controls whether data parts are fully dropped in MergeTree tables when all rows in that part have expired according to their **TTL** settings.
441
+ * **true** - the entire part is dropped if all rows in that part have expired according to their **TTL** settings.
442
+ * **false** - only the rows that have expired based on their **TTL** settings are removed.
443
+
444
+ Default value: **false**.
445
+
446
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#ttl_only_drop_parts).
271
447
  """
272
448
 
273
449
  @property
274
- def materialize_ttl_recalculate_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
275
- """Only recalculate ttl info when MATERIALIZE TTL."""
450
+ def merge_with_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
451
+ """Minimum delay in seconds before repeating a merge with delete TTL.
452
+
453
+ Default value: **14400** (4 hours).
454
+
455
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_with_ttl_timeout).
456
+ """
276
457
 
277
458
  @property
278
- def fsync_after_insert(self) -> google.protobuf.wrappers_pb2.BoolValue:
279
- """Do fsync for every inserted part. Significantly decreases performance of inserts, not recommended to use with wide parts."""
459
+ def merge_with_recompression_ttl_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
460
+ """Minimum delay in seconds before repeating a merge with recompression TTL.
461
+
462
+ Default value: **14400** (4 hours).
463
+
464
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#merge_with_recompression_ttl_timeout).
465
+ """
280
466
 
281
467
  @property
282
- def fsync_part_directory(self) -> google.protobuf.wrappers_pb2.BoolValue:
283
- """Do fsync for part directory after all part operations (writes, renames, etc.)."""
468
+ def max_number_of_merges_with_ttl_in_pool(self) -> google.protobuf.wrappers_pb2.Int64Value:
469
+ """When there is more than specified number of merges with TTL entries in pool, do not assign new merge with TTL.
470
+ This is to leave free threads for regular merges and avoid "Too many parts" errors.
471
+
472
+ Default value: **2**.
473
+
474
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#max_number_of_merges_with_ttl_in_pool).
475
+ """
284
476
 
285
477
  @property
286
- def min_compressed_bytes_to_fsync_after_fetch(self) -> google.protobuf.wrappers_pb2.Int64Value:
287
- """Minimal number of compressed bytes to do fsync for part after fetch. 0 - disabled."""
478
+ def materialize_ttl_recalculate_only(self) -> google.protobuf.wrappers_pb2.BoolValue:
479
+ """Only recalculate ttl info when **MATERIALIZE TTL**.
480
+
481
+ Default value: **true** for versions 25.2 and higher, **false** for versions 25.1 and lower.
482
+
483
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#materialize_ttl_recalculate_only).
484
+ """
288
485
 
289
486
  @property
290
- def min_compressed_bytes_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
291
- """Minimal number of compressed bytes to do fsync for part after merge. 0 - disabled."""
487
+ def check_sample_column_is_correct(self) -> google.protobuf.wrappers_pb2.BoolValue:
488
+ """Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct.
489
+ The data type must be one of unsigned integer types: UInt8, UInt16, UInt32, UInt64.
490
+
491
+ Default value: **true**.
492
+
493
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/settings/merge-tree-settings#check_sample_column_is_correct).
494
+ """
292
495
 
293
496
  @property
294
- def min_rows_to_fsync_after_merge(self) -> google.protobuf.wrappers_pb2.Int64Value:
295
- """Minimal number of rows to do fsync for part after merge. 0 - disabled."""
497
+ def allow_remote_fs_zero_copy_replication(self) -> google.protobuf.wrappers_pb2.BoolValue:
498
+ """Setting is automatically enabled if cloud storage is enabled, disabled otherwise.
499
+
500
+ Default value: **true**.
501
+ """
296
502
 
297
503
  def __init__(
298
504
  self,
299
505
  *,
300
- replicated_deduplication_window: google.protobuf.wrappers_pb2.Int64Value | None = ...,
301
- replicated_deduplication_window_seconds: google.protobuf.wrappers_pb2.Int64Value | None = ...,
302
506
  parts_to_delay_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
303
507
  parts_to_throw_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
304
508
  inactive_parts_to_delay_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
305
509
  inactive_parts_to_throw_insert: google.protobuf.wrappers_pb2.Int64Value | None = ...,
510
+ max_avg_part_size_for_too_many_parts: google.protobuf.wrappers_pb2.Int64Value | None = ...,
511
+ max_parts_in_total: google.protobuf.wrappers_pb2.Int64Value | None = ...,
306
512
  max_replicated_merges_in_queue: google.protobuf.wrappers_pb2.Int64Value | None = ...,
307
513
  number_of_free_entries_in_pool_to_lower_max_size_of_merge: google.protobuf.wrappers_pb2.Int64Value | None = ...,
514
+ number_of_free_entries_in_pool_to_execute_mutation: google.protobuf.wrappers_pb2.Int64Value | None = ...,
308
515
  max_bytes_to_merge_at_min_space_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
309
516
  max_bytes_to_merge_at_max_space_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
310
517
  min_bytes_for_wide_part: google.protobuf.wrappers_pb2.Int64Value | None = ...,
311
518
  min_rows_for_wide_part: google.protobuf.wrappers_pb2.Int64Value | None = ...,
312
- ttl_only_drop_parts: google.protobuf.wrappers_pb2.BoolValue | None = ...,
313
- allow_remote_fs_zero_copy_replication: google.protobuf.wrappers_pb2.BoolValue | None = ...,
314
- merge_with_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
315
- merge_with_recompression_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
316
- max_parts_in_total: google.protobuf.wrappers_pb2.Int64Value | None = ...,
317
- max_number_of_merges_with_ttl_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
318
519
  cleanup_delay_period: google.protobuf.wrappers_pb2.Int64Value | None = ...,
319
- number_of_free_entries_in_pool_to_execute_mutation: google.protobuf.wrappers_pb2.Int64Value | None = ...,
320
- max_avg_part_size_for_too_many_parts: google.protobuf.wrappers_pb2.Int64Value | None = ...,
520
+ max_cleanup_delay_period: google.protobuf.wrappers_pb2.Int64Value | None = ...,
521
+ merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
522
+ max_merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
321
523
  min_age_to_force_merge_seconds: google.protobuf.wrappers_pb2.Int64Value | None = ...,
322
524
  min_age_to_force_merge_on_partition_only: google.protobuf.wrappers_pb2.BoolValue | None = ...,
323
- merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
324
525
  merge_max_block_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
325
- check_sample_column_is_correct: google.protobuf.wrappers_pb2.BoolValue | None = ...,
326
- max_merge_selecting_sleep_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
327
- max_cleanup_delay_period: google.protobuf.wrappers_pb2.Int64Value | None = ...,
328
526
  deduplicate_merge_projection_mode: global___ClickhouseConfig.MergeTree.DeduplicateMergeProjectionMode.ValueType = ...,
329
527
  lightweight_mutation_projection_mode: global___ClickhouseConfig.MergeTree.LightweightMutationProjectionMode.ValueType = ...,
330
- materialize_ttl_recalculate_only: google.protobuf.wrappers_pb2.BoolValue | None = ...,
528
+ replicated_deduplication_window: google.protobuf.wrappers_pb2.Int64Value | None = ...,
529
+ replicated_deduplication_window_seconds: google.protobuf.wrappers_pb2.Int64Value | None = ...,
331
530
  fsync_after_insert: google.protobuf.wrappers_pb2.BoolValue | None = ...,
332
531
  fsync_part_directory: google.protobuf.wrappers_pb2.BoolValue | None = ...,
333
532
  min_compressed_bytes_to_fsync_after_fetch: google.protobuf.wrappers_pb2.Int64Value | None = ...,
334
533
  min_compressed_bytes_to_fsync_after_merge: google.protobuf.wrappers_pb2.Int64Value | None = ...,
335
534
  min_rows_to_fsync_after_merge: google.protobuf.wrappers_pb2.Int64Value | None = ...,
535
+ ttl_only_drop_parts: google.protobuf.wrappers_pb2.BoolValue | None = ...,
536
+ merge_with_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
537
+ merge_with_recompression_ttl_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
538
+ max_number_of_merges_with_ttl_in_pool: google.protobuf.wrappers_pb2.Int64Value | None = ...,
539
+ materialize_ttl_recalculate_only: google.protobuf.wrappers_pb2.BoolValue | None = ...,
540
+ check_sample_column_is_correct: google.protobuf.wrappers_pb2.BoolValue | None = ...,
541
+ allow_remote_fs_zero_copy_replication: google.protobuf.wrappers_pb2.BoolValue | None = ...,
336
542
  ) -> None: ...
337
543
  def HasField(self, field_name: typing.Literal["allow_remote_fs_zero_copy_replication", b"allow_remote_fs_zero_copy_replication", "check_sample_column_is_correct", b"check_sample_column_is_correct", "cleanup_delay_period", b"cleanup_delay_period", "fsync_after_insert", b"fsync_after_insert", "fsync_part_directory", b"fsync_part_directory", "inactive_parts_to_delay_insert", b"inactive_parts_to_delay_insert", "inactive_parts_to_throw_insert", b"inactive_parts_to_throw_insert", "materialize_ttl_recalculate_only", b"materialize_ttl_recalculate_only", "max_avg_part_size_for_too_many_parts", b"max_avg_part_size_for_too_many_parts", "max_bytes_to_merge_at_max_space_in_pool", b"max_bytes_to_merge_at_max_space_in_pool", "max_bytes_to_merge_at_min_space_in_pool", b"max_bytes_to_merge_at_min_space_in_pool", "max_cleanup_delay_period", b"max_cleanup_delay_period", "max_merge_selecting_sleep_ms", b"max_merge_selecting_sleep_ms", "max_number_of_merges_with_ttl_in_pool", b"max_number_of_merges_with_ttl_in_pool", "max_parts_in_total", b"max_parts_in_total", "max_replicated_merges_in_queue", b"max_replicated_merges_in_queue", "merge_max_block_size", b"merge_max_block_size", "merge_selecting_sleep_ms", b"merge_selecting_sleep_ms", "merge_with_recompression_ttl_timeout", b"merge_with_recompression_ttl_timeout", "merge_with_ttl_timeout", b"merge_with_ttl_timeout", "min_age_to_force_merge_on_partition_only", b"min_age_to_force_merge_on_partition_only", "min_age_to_force_merge_seconds", b"min_age_to_force_merge_seconds", "min_bytes_for_wide_part", b"min_bytes_for_wide_part", "min_compressed_bytes_to_fsync_after_fetch", b"min_compressed_bytes_to_fsync_after_fetch", "min_compressed_bytes_to_fsync_after_merge", b"min_compressed_bytes_to_fsync_after_merge", "min_rows_for_wide_part", b"min_rows_for_wide_part", "min_rows_to_fsync_after_merge", b"min_rows_to_fsync_after_merge", "number_of_free_entries_in_pool_to_execute_mutation", b"number_of_free_entries_in_pool_to_execute_mutation", "number_of_free_entries_in_pool_to_lower_max_size_of_merge", b"number_of_free_entries_in_pool_to_lower_max_size_of_merge", "parts_to_delay_insert", b"parts_to_delay_insert", "parts_to_throw_insert", b"parts_to_throw_insert", "replicated_deduplication_window", b"replicated_deduplication_window", "replicated_deduplication_window_seconds", b"replicated_deduplication_window_seconds", "ttl_only_drop_parts", b"ttl_only_drop_parts"]) -> builtins.bool: ...
338
544
  def ClearField(self, field_name: typing.Literal["allow_remote_fs_zero_copy_replication", b"allow_remote_fs_zero_copy_replication", "check_sample_column_is_correct", b"check_sample_column_is_correct", "cleanup_delay_period", b"cleanup_delay_period", "deduplicate_merge_projection_mode", b"deduplicate_merge_projection_mode", "fsync_after_insert", b"fsync_after_insert", "fsync_part_directory", b"fsync_part_directory", "inactive_parts_to_delay_insert", b"inactive_parts_to_delay_insert", "inactive_parts_to_throw_insert", b"inactive_parts_to_throw_insert", "lightweight_mutation_projection_mode", b"lightweight_mutation_projection_mode", "materialize_ttl_recalculate_only", b"materialize_ttl_recalculate_only", "max_avg_part_size_for_too_many_parts", b"max_avg_part_size_for_too_many_parts", "max_bytes_to_merge_at_max_space_in_pool", b"max_bytes_to_merge_at_max_space_in_pool", "max_bytes_to_merge_at_min_space_in_pool", b"max_bytes_to_merge_at_min_space_in_pool", "max_cleanup_delay_period", b"max_cleanup_delay_period", "max_merge_selecting_sleep_ms", b"max_merge_selecting_sleep_ms", "max_number_of_merges_with_ttl_in_pool", b"max_number_of_merges_with_ttl_in_pool", "max_parts_in_total", b"max_parts_in_total", "max_replicated_merges_in_queue", b"max_replicated_merges_in_queue", "merge_max_block_size", b"merge_max_block_size", "merge_selecting_sleep_ms", b"merge_selecting_sleep_ms", "merge_with_recompression_ttl_timeout", b"merge_with_recompression_ttl_timeout", "merge_with_ttl_timeout", b"merge_with_ttl_timeout", "min_age_to_force_merge_on_partition_only", b"min_age_to_force_merge_on_partition_only", "min_age_to_force_merge_seconds", b"min_age_to_force_merge_seconds", "min_bytes_for_wide_part", b"min_bytes_for_wide_part", "min_compressed_bytes_to_fsync_after_fetch", b"min_compressed_bytes_to_fsync_after_fetch", "min_compressed_bytes_to_fsync_after_merge", b"min_compressed_bytes_to_fsync_after_merge", "min_rows_for_wide_part", b"min_rows_for_wide_part", "min_rows_to_fsync_after_merge", b"min_rows_to_fsync_after_merge", "number_of_free_entries_in_pool_to_execute_mutation", b"number_of_free_entries_in_pool_to_execute_mutation", "number_of_free_entries_in_pool_to_lower_max_size_of_merge", b"number_of_free_entries_in_pool_to_lower_max_size_of_merge", "parts_to_delay_insert", b"parts_to_delay_insert", "parts_to_throw_insert", b"parts_to_throw_insert", "replicated_deduplication_window", b"replicated_deduplication_window", "replicated_deduplication_window_seconds", b"replicated_deduplication_window_seconds", "ttl_only_drop_parts", b"ttl_only_drop_parts"]) -> None: ...
339
545
 
340
546
  @typing.final
341
- class Kafka(google.protobuf.message.Message):
547
+ class Compression(google.protobuf.message.Message):
548
+ """Compression settings.
549
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#compression).
550
+ """
551
+
342
552
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
343
553
 
344
- class _SecurityProtocol:
554
+ class _Method:
345
555
  ValueType = typing.NewType("ValueType", builtins.int)
346
556
  V: typing_extensions.TypeAlias = ValueType
347
557
 
348
- class _SecurityProtocolEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SecurityProtocol.ValueType], builtins.type):
558
+ class _MethodEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Compression._Method.ValueType], builtins.type):
349
559
  DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
350
- SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 0
351
- SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 1
352
- SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 2
353
- SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 3
354
- SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 4
355
-
356
- class SecurityProtocol(_SecurityProtocol, metaclass=_SecurityProtocolEnumTypeWrapper): ...
357
- SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 0
358
- SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 1
359
- SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 2
360
- SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 3
361
- SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 4
362
-
363
- class _SaslMechanism:
364
- ValueType = typing.NewType("ValueType", builtins.int)
365
- V: typing_extensions.TypeAlias = ValueType
366
-
367
- class _SaslMechanismEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SaslMechanism.ValueType], builtins.type):
368
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
369
- SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 0
370
- SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 1
371
- SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 2
372
- SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 3
373
- SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 4
374
-
375
- class SaslMechanism(_SaslMechanism, metaclass=_SaslMechanismEnumTypeWrapper): ...
376
- SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 0
377
- SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 1
378
- SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 2
379
- SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 3
380
- SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 4
381
-
382
- class _Debug:
383
- ValueType = typing.NewType("ValueType", builtins.int)
384
- V: typing_extensions.TypeAlias = ValueType
385
-
386
- class _DebugEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._Debug.ValueType], builtins.type):
387
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
388
- DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka._Debug.ValueType # 0
389
- DEBUG_GENERIC: ClickhouseConfig.Kafka._Debug.ValueType # 1
390
- DEBUG_BROKER: ClickhouseConfig.Kafka._Debug.ValueType # 2
391
- DEBUG_TOPIC: ClickhouseConfig.Kafka._Debug.ValueType # 3
392
- DEBUG_METADATA: ClickhouseConfig.Kafka._Debug.ValueType # 4
393
- DEBUG_FEATURE: ClickhouseConfig.Kafka._Debug.ValueType # 5
394
- DEBUG_QUEUE: ClickhouseConfig.Kafka._Debug.ValueType # 6
395
- DEBUG_MSG: ClickhouseConfig.Kafka._Debug.ValueType # 7
396
- DEBUG_PROTOCOL: ClickhouseConfig.Kafka._Debug.ValueType # 8
397
- DEBUG_CGRP: ClickhouseConfig.Kafka._Debug.ValueType # 9
398
- DEBUG_SECURITY: ClickhouseConfig.Kafka._Debug.ValueType # 10
399
- DEBUG_FETCH: ClickhouseConfig.Kafka._Debug.ValueType # 11
400
- DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka._Debug.ValueType # 12
401
- DEBUG_PLUGIN: ClickhouseConfig.Kafka._Debug.ValueType # 13
402
- DEBUG_CONSUMER: ClickhouseConfig.Kafka._Debug.ValueType # 14
403
- DEBUG_ADMIN: ClickhouseConfig.Kafka._Debug.ValueType # 15
404
- DEBUG_EOS: ClickhouseConfig.Kafka._Debug.ValueType # 16
405
- DEBUG_MOCK: ClickhouseConfig.Kafka._Debug.ValueType # 17
406
- DEBUG_ASSIGNOR: ClickhouseConfig.Kafka._Debug.ValueType # 18
407
- DEBUG_CONF: ClickhouseConfig.Kafka._Debug.ValueType # 19
408
- DEBUG_TELEMETRY: ClickhouseConfig.Kafka._Debug.ValueType # 20
409
- DEBUG_ALL: ClickhouseConfig.Kafka._Debug.ValueType # 21
410
-
411
- class Debug(_Debug, metaclass=_DebugEnumTypeWrapper): ...
412
- DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka.Debug.ValueType # 0
413
- DEBUG_GENERIC: ClickhouseConfig.Kafka.Debug.ValueType # 1
414
- DEBUG_BROKER: ClickhouseConfig.Kafka.Debug.ValueType # 2
415
- DEBUG_TOPIC: ClickhouseConfig.Kafka.Debug.ValueType # 3
416
- DEBUG_METADATA: ClickhouseConfig.Kafka.Debug.ValueType # 4
417
- DEBUG_FEATURE: ClickhouseConfig.Kafka.Debug.ValueType # 5
418
- DEBUG_QUEUE: ClickhouseConfig.Kafka.Debug.ValueType # 6
419
- DEBUG_MSG: ClickhouseConfig.Kafka.Debug.ValueType # 7
420
- DEBUG_PROTOCOL: ClickhouseConfig.Kafka.Debug.ValueType # 8
421
- DEBUG_CGRP: ClickhouseConfig.Kafka.Debug.ValueType # 9
422
- DEBUG_SECURITY: ClickhouseConfig.Kafka.Debug.ValueType # 10
423
- DEBUG_FETCH: ClickhouseConfig.Kafka.Debug.ValueType # 11
424
- DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka.Debug.ValueType # 12
425
- DEBUG_PLUGIN: ClickhouseConfig.Kafka.Debug.ValueType # 13
426
- DEBUG_CONSUMER: ClickhouseConfig.Kafka.Debug.ValueType # 14
427
- DEBUG_ADMIN: ClickhouseConfig.Kafka.Debug.ValueType # 15
428
- DEBUG_EOS: ClickhouseConfig.Kafka.Debug.ValueType # 16
429
- DEBUG_MOCK: ClickhouseConfig.Kafka.Debug.ValueType # 17
430
- DEBUG_ASSIGNOR: ClickhouseConfig.Kafka.Debug.ValueType # 18
431
- DEBUG_CONF: ClickhouseConfig.Kafka.Debug.ValueType # 19
432
- DEBUG_TELEMETRY: ClickhouseConfig.Kafka.Debug.ValueType # 20
433
- DEBUG_ALL: ClickhouseConfig.Kafka.Debug.ValueType # 21
434
-
435
- class _AutoOffsetReset:
436
- ValueType = typing.NewType("ValueType", builtins.int)
437
- V: typing_extensions.TypeAlias = ValueType
438
-
439
- class _AutoOffsetResetEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._AutoOffsetReset.ValueType], builtins.type):
440
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
441
- AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 0
442
- AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 1
443
- AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 2
444
- AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 3
445
- AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 4
446
- AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 5
447
- AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 6
448
- AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 7
449
-
450
- class AutoOffsetReset(_AutoOffsetReset, metaclass=_AutoOffsetResetEnumTypeWrapper): ...
451
- AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 0
452
- AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 1
453
- AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 2
454
- AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 3
455
- AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 4
456
- AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 5
457
- AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 6
458
- AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 7
459
-
460
- SECURITY_PROTOCOL_FIELD_NUMBER: builtins.int
461
- SASL_MECHANISM_FIELD_NUMBER: builtins.int
462
- SASL_USERNAME_FIELD_NUMBER: builtins.int
463
- SASL_PASSWORD_FIELD_NUMBER: builtins.int
464
- ENABLE_SSL_CERTIFICATE_VERIFICATION_FIELD_NUMBER: builtins.int
465
- MAX_POLL_INTERVAL_MS_FIELD_NUMBER: builtins.int
466
- SESSION_TIMEOUT_MS_FIELD_NUMBER: builtins.int
467
- DEBUG_FIELD_NUMBER: builtins.int
468
- AUTO_OFFSET_RESET_FIELD_NUMBER: builtins.int
469
- security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType
470
- sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType
471
- sasl_username: builtins.str
472
- sasl_password: builtins.str
473
- debug: global___ClickhouseConfig.Kafka.Debug.ValueType
474
- auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType
475
- @property
476
- def enable_ssl_certificate_verification(self) -> google.protobuf.wrappers_pb2.BoolValue: ...
477
- @property
478
- def max_poll_interval_ms(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
479
- @property
480
- def session_timeout_ms(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
481
- def __init__(
482
- self,
483
- *,
484
- security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType = ...,
485
- sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType = ...,
486
- sasl_username: builtins.str = ...,
487
- sasl_password: builtins.str = ...,
488
- enable_ssl_certificate_verification: google.protobuf.wrappers_pb2.BoolValue | None = ...,
489
- max_poll_interval_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
490
- session_timeout_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
491
- debug: global___ClickhouseConfig.Kafka.Debug.ValueType = ...,
492
- auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType = ...,
493
- ) -> None: ...
494
- def HasField(self, field_name: typing.Literal["enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "session_timeout_ms", b"session_timeout_ms"]) -> builtins.bool: ...
495
- def ClearField(self, field_name: typing.Literal["auto_offset_reset", b"auto_offset_reset", "debug", b"debug", "enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "sasl_mechanism", b"sasl_mechanism", "sasl_password", b"sasl_password", "sasl_username", b"sasl_username", "security_protocol", b"security_protocol", "session_timeout_ms", b"session_timeout_ms"]) -> None: ...
496
-
497
- @typing.final
498
- class KafkaTopic(google.protobuf.message.Message):
499
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
500
-
501
- NAME_FIELD_NUMBER: builtins.int
502
- SETTINGS_FIELD_NUMBER: builtins.int
503
- name: builtins.str
504
- @property
505
- def settings(self) -> global___ClickhouseConfig.Kafka: ...
506
- def __init__(
507
- self,
508
- *,
509
- name: builtins.str = ...,
510
- settings: global___ClickhouseConfig.Kafka | None = ...,
511
- ) -> None: ...
512
- def HasField(self, field_name: typing.Literal["settings", b"settings"]) -> builtins.bool: ...
513
- def ClearField(self, field_name: typing.Literal["name", b"name", "settings", b"settings"]) -> None: ...
514
-
515
- @typing.final
516
- class Rabbitmq(google.protobuf.message.Message):
517
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
518
-
519
- USERNAME_FIELD_NUMBER: builtins.int
520
- PASSWORD_FIELD_NUMBER: builtins.int
521
- VHOST_FIELD_NUMBER: builtins.int
522
- username: builtins.str
523
- """[RabbitMQ](https://clickhouse.com/docs/en/engines/table-engines/integrations/rabbitmq/) username"""
524
- password: builtins.str
525
- """[RabbitMQ](https://clickhouse.com/docs/en/engines/table-engines/integrations/rabbitmq/) password"""
526
- vhost: builtins.str
527
- """[RabbitMQ](https://clickhouse.com/docs/en/engines/table-engines/integrations/rabbitmq/) virtual host"""
528
- def __init__(
529
- self,
530
- *,
531
- username: builtins.str = ...,
532
- password: builtins.str = ...,
533
- vhost: builtins.str = ...,
534
- ) -> None: ...
535
- def ClearField(self, field_name: typing.Literal["password", b"password", "username", b"username", "vhost", b"vhost"]) -> None: ...
536
-
537
- @typing.final
538
- class Compression(google.protobuf.message.Message):
539
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
540
-
541
- class _Method:
542
- ValueType = typing.NewType("ValueType", builtins.int)
543
- V: typing_extensions.TypeAlias = ValueType
544
-
545
- class _MethodEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Compression._Method.ValueType], builtins.type):
546
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
547
- METHOD_UNSPECIFIED: ClickhouseConfig.Compression._Method.ValueType # 0
548
- LZ4: ClickhouseConfig.Compression._Method.ValueType # 1
549
- """[LZ4 compression algorithm](https://lz4.github.io/lz4/)."""
550
- ZSTD: ClickhouseConfig.Compression._Method.ValueType # 2
551
- """[Zstandard compression algorithm](https://facebook.github.io/zstd/)."""
560
+ METHOD_UNSPECIFIED: ClickhouseConfig.Compression._Method.ValueType # 0
561
+ LZ4: ClickhouseConfig.Compression._Method.ValueType # 1
562
+ """[LZ4 compression algorithm](https://lz4.github.io/lz4)."""
563
+ ZSTD: ClickhouseConfig.Compression._Method.ValueType # 2
564
+ """[ZSTD compression algorithm](https://facebook.github.io/zstd)."""
565
+ LZ4HC: ClickhouseConfig.Compression._Method.ValueType # 3
566
+ """[LZ4 HC (high compression) algorithm](https://clickhouse.com/docs/sql-reference/statements/create/table#lz4hc)."""
552
567
 
553
568
  class Method(_Method, metaclass=_MethodEnumTypeWrapper): ...
554
569
  METHOD_UNSPECIFIED: ClickhouseConfig.Compression.Method.ValueType # 0
555
570
  LZ4: ClickhouseConfig.Compression.Method.ValueType # 1
556
- """[LZ4 compression algorithm](https://lz4.github.io/lz4/)."""
571
+ """[LZ4 compression algorithm](https://lz4.github.io/lz4)."""
557
572
  ZSTD: ClickhouseConfig.Compression.Method.ValueType # 2
558
- """[Zstandard compression algorithm](https://facebook.github.io/zstd/)."""
573
+ """[ZSTD compression algorithm](https://facebook.github.io/zstd)."""
574
+ LZ4HC: ClickhouseConfig.Compression.Method.ValueType # 3
575
+ """[LZ4 HC (high compression) algorithm](https://clickhouse.com/docs/sql-reference/statements/create/table#lz4hc)."""
559
576
 
560
577
  METHOD_FIELD_NUMBER: builtins.int
561
578
  MIN_PART_SIZE_FIELD_NUMBER: builtins.int
562
579
  MIN_PART_SIZE_RATIO_FIELD_NUMBER: builtins.int
563
580
  LEVEL_FIELD_NUMBER: builtins.int
564
581
  method: global___ClickhouseConfig.Compression.Method.ValueType
565
- """Compression method to use for the specified combination of [min_part_size] and [min_part_size_ratio]."""
582
+ """Compression method to use for the specified combination of **min_part_size** and **min_part_size_ratio**."""
566
583
  min_part_size: builtins.int
567
- """Minimum size of a part of a table."""
584
+ """The minimum size of a data part."""
568
585
  min_part_size_ratio: builtins.float
569
- """Minimum ratio of a part relative to the size of all the data in the table."""
586
+ """The ratio of the data part size to the table size."""
570
587
  @property
571
- def level(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
588
+ def level(self) -> google.protobuf.wrappers_pb2.Int64Value:
589
+ """Compression level."""
590
+
572
591
  def __init__(
573
592
  self,
574
593
  *,
@@ -582,305 +601,49 @@ class ClickhouseConfig(google.protobuf.message.Message):
582
601
 
583
602
  @typing.final
584
603
  class ExternalDictionary(google.protobuf.message.Message):
604
+ """External dictionary configuration."""
605
+
585
606
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
586
607
 
587
608
  @typing.final
588
- class HttpSource(google.protobuf.message.Message):
609
+ class Structure(google.protobuf.message.Message):
610
+ """Configuration of external dictionary structure."""
611
+
589
612
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
590
613
 
591
614
  @typing.final
592
- class Header(google.protobuf.message.Message):
615
+ class Id(google.protobuf.message.Message):
616
+ """Numeric key."""
617
+
593
618
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
594
619
 
595
620
  NAME_FIELD_NUMBER: builtins.int
596
- VALUE_FIELD_NUMBER: builtins.int
597
621
  name: builtins.str
598
- value: builtins.str
622
+ """Name of the numeric key."""
599
623
  def __init__(
600
624
  self,
601
625
  *,
602
626
  name: builtins.str = ...,
603
- value: builtins.str = ...,
604
627
  ) -> None: ...
605
- def ClearField(self, field_name: typing.Literal["name", b"name", "value", b"value"]) -> None: ...
606
-
607
- URL_FIELD_NUMBER: builtins.int
608
- FORMAT_FIELD_NUMBER: builtins.int
609
- HEADERS_FIELD_NUMBER: builtins.int
610
- url: builtins.str
611
- """URL of the source dictionary available over HTTP."""
612
- format: builtins.str
613
- """The data format. Valid values are all formats supported by ClickHouse SQL dialect."""
614
- @property
615
- def headers(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header]:
616
- """HTTP headers."""
617
-
618
- def __init__(
619
- self,
620
- *,
621
- url: builtins.str = ...,
622
- format: builtins.str = ...,
623
- headers: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header] | None = ...,
624
- ) -> None: ...
625
- def ClearField(self, field_name: typing.Literal["format", b"format", "headers", b"headers", "url", b"url"]) -> None: ...
626
-
627
- @typing.final
628
- class MysqlSource(google.protobuf.message.Message):
629
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
628
+ def ClearField(self, field_name: typing.Literal["name", b"name"]) -> None: ...
630
629
 
631
630
  @typing.final
632
- class Replica(google.protobuf.message.Message):
631
+ class Key(google.protobuf.message.Message):
632
+ """Complex key."""
633
+
633
634
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
634
635
 
635
- HOST_FIELD_NUMBER: builtins.int
636
- PRIORITY_FIELD_NUMBER: builtins.int
637
- PORT_FIELD_NUMBER: builtins.int
638
- USER_FIELD_NUMBER: builtins.int
639
- PASSWORD_FIELD_NUMBER: builtins.int
640
- host: builtins.str
641
- """MySQL host of the replica."""
642
- priority: builtins.int
643
- """The priority of the replica that ClickHouse takes into account when connecting.
644
- Replica with the highest priority should have this field set to the lowest number.
645
- """
646
- port: builtins.int
647
- """Port to use when connecting to the replica.
648
- If a port is not specified for a replica, ClickHouse uses the port specified for the source.
649
- """
650
- user: builtins.str
651
- """Name of the MySQL database user."""
652
- password: builtins.str
653
- """Password of the MySQL database user."""
636
+ ATTRIBUTES_FIELD_NUMBER: builtins.int
637
+ @property
638
+ def attributes(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute]:
639
+ """Attributes of a complex key."""
640
+
654
641
  def __init__(
655
642
  self,
656
643
  *,
657
- host: builtins.str = ...,
658
- priority: builtins.int = ...,
659
- port: builtins.int = ...,
660
- user: builtins.str = ...,
661
- password: builtins.str = ...,
644
+ attributes: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute] | None = ...,
662
645
  ) -> None: ...
663
- def ClearField(self, field_name: typing.Literal["host", b"host", "password", b"password", "port", b"port", "priority", b"priority", "user", b"user"]) -> None: ...
664
-
665
- DB_FIELD_NUMBER: builtins.int
666
- TABLE_FIELD_NUMBER: builtins.int
667
- PORT_FIELD_NUMBER: builtins.int
668
- USER_FIELD_NUMBER: builtins.int
669
- PASSWORD_FIELD_NUMBER: builtins.int
670
- REPLICAS_FIELD_NUMBER: builtins.int
671
- WHERE_FIELD_NUMBER: builtins.int
672
- INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
673
- CLOSE_CONNECTION_FIELD_NUMBER: builtins.int
674
- SHARE_CONNECTION_FIELD_NUMBER: builtins.int
675
- db: builtins.str
676
- """Name of the MySQL database to connect to."""
677
- table: builtins.str
678
- """Name of the database table to use as a ClickHouse dictionary."""
679
- port: builtins.int
680
- """Default port to use when connecting to a replica of the dictionary source."""
681
- user: builtins.str
682
- """Name of the default user for replicas of the dictionary source."""
683
- password: builtins.str
684
- """Password of the default user for replicas of the dictionary source."""
685
- where: builtins.str
686
- """Selection criteria for the data in the specified MySQL table."""
687
- invalidate_query: builtins.str
688
- """Query for checking the dictionary status, to pull only updated data.
689
- For more details, see [ClickHouse documentation on dictionaries](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_lifetime/).
690
- """
691
- @property
692
- def replicas(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica]:
693
- """List of MySQL replicas of the database used as dictionary source."""
694
-
695
- @property
696
- def close_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
697
- """Should the connection be closed after each request."""
698
-
699
- @property
700
- def share_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
701
- """Should a connection be shared for some requests."""
702
-
703
- def __init__(
704
- self,
705
- *,
706
- db: builtins.str = ...,
707
- table: builtins.str = ...,
708
- port: builtins.int = ...,
709
- user: builtins.str = ...,
710
- password: builtins.str = ...,
711
- replicas: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica] | None = ...,
712
- where: builtins.str = ...,
713
- invalidate_query: builtins.str = ...,
714
- close_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
715
- share_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
716
- ) -> None: ...
717
- def HasField(self, field_name: typing.Literal["close_connection", b"close_connection", "share_connection", b"share_connection"]) -> builtins.bool: ...
718
- def ClearField(self, field_name: typing.Literal["close_connection", b"close_connection", "db", b"db", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "replicas", b"replicas", "share_connection", b"share_connection", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
719
-
720
- @typing.final
721
- class ClickhouseSource(google.protobuf.message.Message):
722
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
723
-
724
- DB_FIELD_NUMBER: builtins.int
725
- TABLE_FIELD_NUMBER: builtins.int
726
- HOST_FIELD_NUMBER: builtins.int
727
- PORT_FIELD_NUMBER: builtins.int
728
- USER_FIELD_NUMBER: builtins.int
729
- PASSWORD_FIELD_NUMBER: builtins.int
730
- WHERE_FIELD_NUMBER: builtins.int
731
- SECURE_FIELD_NUMBER: builtins.int
732
- db: builtins.str
733
- """Name of the ClickHouse database."""
734
- table: builtins.str
735
- """Name of the table in the specified database to be used as the dictionary source."""
736
- host: builtins.str
737
- """ClickHouse host of the specified database."""
738
- port: builtins.int
739
- """Port to use when connecting to the host."""
740
- user: builtins.str
741
- """Name of the ClickHouse database user."""
742
- password: builtins.str
743
- """Password of the ClickHouse database user."""
744
- where: builtins.str
745
- """Selection criteria for the data in the specified ClickHouse table."""
746
- @property
747
- def secure(self) -> google.protobuf.wrappers_pb2.BoolValue:
748
- """Use ssl for connection."""
749
-
750
- def __init__(
751
- self,
752
- *,
753
- db: builtins.str = ...,
754
- table: builtins.str = ...,
755
- host: builtins.str = ...,
756
- port: builtins.int = ...,
757
- user: builtins.str = ...,
758
- password: builtins.str = ...,
759
- where: builtins.str = ...,
760
- secure: google.protobuf.wrappers_pb2.BoolValue | None = ...,
761
- ) -> None: ...
762
- def HasField(self, field_name: typing.Literal["secure", b"secure"]) -> builtins.bool: ...
763
- def ClearField(self, field_name: typing.Literal["db", b"db", "host", b"host", "password", b"password", "port", b"port", "secure", b"secure", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
764
-
765
- @typing.final
766
- class MongodbSource(google.protobuf.message.Message):
767
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
768
-
769
- DB_FIELD_NUMBER: builtins.int
770
- COLLECTION_FIELD_NUMBER: builtins.int
771
- HOST_FIELD_NUMBER: builtins.int
772
- PORT_FIELD_NUMBER: builtins.int
773
- USER_FIELD_NUMBER: builtins.int
774
- PASSWORD_FIELD_NUMBER: builtins.int
775
- OPTIONS_FIELD_NUMBER: builtins.int
776
- db: builtins.str
777
- """Name of the MongoDB database."""
778
- collection: builtins.str
779
- """Name of the collection in the specified database to be used as the dictionary source."""
780
- host: builtins.str
781
- """MongoDB host of the specified database."""
782
- port: builtins.int
783
- """Port to use when connecting to the host."""
784
- user: builtins.str
785
- """Name of the MongoDB database user."""
786
- password: builtins.str
787
- """Password of the MongoDB database user."""
788
- options: builtins.str
789
- def __init__(
790
- self,
791
- *,
792
- db: builtins.str = ...,
793
- collection: builtins.str = ...,
794
- host: builtins.str = ...,
795
- port: builtins.int = ...,
796
- user: builtins.str = ...,
797
- password: builtins.str = ...,
798
- options: builtins.str = ...,
799
- ) -> None: ...
800
- def ClearField(self, field_name: typing.Literal["collection", b"collection", "db", b"db", "host", b"host", "options", b"options", "password", b"password", "port", b"port", "user", b"user"]) -> None: ...
801
-
802
- @typing.final
803
- class PostgresqlSource(google.protobuf.message.Message):
804
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
805
-
806
- class _SslMode:
807
- ValueType = typing.NewType("ValueType", builtins.int)
808
- V: typing_extensions.TypeAlias = ValueType
809
-
810
- class _SslModeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType], builtins.type):
811
- DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
812
- SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 0
813
- DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 1
814
- """Only try a non-SSL connection."""
815
- ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 2
816
- """First try a non-SSL connection; if that fails, try an SSL connection."""
817
- PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 3
818
- """First try an SSL connection; if that fails, try a non-SSL connection."""
819
- VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 4
820
- """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
821
- VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 5
822
- """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
823
-
824
- class SslMode(_SslMode, metaclass=_SslModeEnumTypeWrapper): ...
825
- SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 0
826
- DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 1
827
- """Only try a non-SSL connection."""
828
- ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 2
829
- """First try a non-SSL connection; if that fails, try an SSL connection."""
830
- PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 3
831
- """First try an SSL connection; if that fails, try a non-SSL connection."""
832
- VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 4
833
- """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
834
- VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 5
835
- """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
836
-
837
- DB_FIELD_NUMBER: builtins.int
838
- TABLE_FIELD_NUMBER: builtins.int
839
- HOSTS_FIELD_NUMBER: builtins.int
840
- PORT_FIELD_NUMBER: builtins.int
841
- USER_FIELD_NUMBER: builtins.int
842
- PASSWORD_FIELD_NUMBER: builtins.int
843
- INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
844
- SSL_MODE_FIELD_NUMBER: builtins.int
845
- db: builtins.str
846
- """Name of the PostrgreSQL database."""
847
- table: builtins.str
848
- """Name of the table in the specified database to be used as the dictionary source."""
849
- port: builtins.int
850
- """Port to use when connecting to the host."""
851
- user: builtins.str
852
- """Name of the PostrgreSQL database user."""
853
- password: builtins.str
854
- """Password of the PostrgreSQL database user."""
855
- invalidate_query: builtins.str
856
- """Query for checking the dictionary status, to pull only updated data.
857
- For more details, see [ClickHouse documentation on dictionaries](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_lifetime/).
858
- """
859
- ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType
860
- """Mode of SSL TCP/IP connection to the PostgreSQL host.
861
- For more details, see [PostgreSQL documentation](https://www.postgresql.org/docs/current/libpq-ssl.html).
862
- """
863
- @property
864
- def hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
865
- """Name of the PostrgreSQL host"""
866
-
867
- def __init__(
868
- self,
869
- *,
870
- db: builtins.str = ...,
871
- table: builtins.str = ...,
872
- hosts: collections.abc.Iterable[builtins.str] | None = ...,
873
- port: builtins.int = ...,
874
- user: builtins.str = ...,
875
- password: builtins.str = ...,
876
- invalidate_query: builtins.str = ...,
877
- ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType = ...,
878
- ) -> None: ...
879
- def ClearField(self, field_name: typing.Literal["db", b"db", "hosts", b"hosts", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "ssl_mode", b"ssl_mode", "table", b"table", "user", b"user"]) -> None: ...
880
-
881
- @typing.final
882
- class Structure(google.protobuf.message.Message):
883
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
646
+ def ClearField(self, field_name: typing.Literal["attributes", b"attributes"]) -> None: ...
884
647
 
885
648
  @typing.final
886
649
  class Attribute(google.protobuf.message.Message):
@@ -902,11 +665,13 @@ class ClickhouseConfig(google.protobuf.message.Message):
902
665
  """Expression, describing the attribute, if applicable."""
903
666
  hierarchical: builtins.bool
904
667
  """Indication of hierarchy support.
905
- Default value: `false`.
668
+
669
+ Default value: **false**.
906
670
  """
907
671
  injective: builtins.bool
908
672
  """Indication of injective mapping "id -> attribute".
909
- Default value: `false`.
673
+
674
+ Default value: **false**.
910
675
  """
911
676
  def __init__(
912
677
  self,
@@ -920,40 +685,6 @@ class ClickhouseConfig(google.protobuf.message.Message):
920
685
  ) -> None: ...
921
686
  def ClearField(self, field_name: typing.Literal["expression", b"expression", "hierarchical", b"hierarchical", "injective", b"injective", "name", b"name", "null_value", b"null_value", "type", b"type"]) -> None: ...
922
687
 
923
- @typing.final
924
- class Id(google.protobuf.message.Message):
925
- """Numeric key."""
926
-
927
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
928
-
929
- NAME_FIELD_NUMBER: builtins.int
930
- name: builtins.str
931
- """Name of the numeric key."""
932
- def __init__(
933
- self,
934
- *,
935
- name: builtins.str = ...,
936
- ) -> None: ...
937
- def ClearField(self, field_name: typing.Literal["name", b"name"]) -> None: ...
938
-
939
- @typing.final
940
- class Key(google.protobuf.message.Message):
941
- """Complex key."""
942
-
943
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
944
-
945
- ATTRIBUTES_FIELD_NUMBER: builtins.int
946
- @property
947
- def attributes(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute]:
948
- """Attributes of a complex key."""
949
-
950
- def __init__(
951
- self,
952
- *,
953
- attributes: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute] | None = ...,
954
- ) -> None: ...
955
- def ClearField(self, field_name: typing.Literal["attributes", b"attributes"]) -> None: ...
956
-
957
688
  ID_FIELD_NUMBER: builtins.int
958
689
  KEY_FIELD_NUMBER: builtins.int
959
690
  RANGE_MIN_FIELD_NUMBER: builtins.int
@@ -966,24 +697,28 @@ class ClickhouseConfig(google.protobuf.message.Message):
966
697
  @property
967
698
  def key(self) -> global___ClickhouseConfig.ExternalDictionary.Structure.Key:
968
699
  """Composite key for the dictionary, containing of one or more key columns.
700
+
969
701
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_structure/#composite-key).
970
702
  """
971
703
 
972
704
  @property
973
705
  def range_min(self) -> global___ClickhouseConfig.ExternalDictionary.Structure.Attribute:
974
- """Field holding the beginning of the range for dictionaries with `RANGE_HASHED` layout.
706
+ """Field holding the beginning of the range for dictionaries with **RANGE_HASHED** layout.
707
+
975
708
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_layout/#range-hashed).
976
709
  """
977
710
 
978
711
  @property
979
712
  def range_max(self) -> global___ClickhouseConfig.ExternalDictionary.Structure.Attribute:
980
- """Field holding the end of the range for dictionaries with `RANGE_HASHED` layout.
713
+ """Field holding the end of the range for dictionaries with **RANGE_HASHED** layout.
714
+
981
715
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_layout/#range-hashed).
982
716
  """
983
717
 
984
718
  @property
985
719
  def attributes(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.Structure.Attribute]:
986
720
  """Description of the fields available for database queries.
721
+
987
722
  For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_structure/#attributes).
988
723
  """
989
724
 
@@ -1001,8 +736,6 @@ class ClickhouseConfig(google.protobuf.message.Message):
1001
736
 
1002
737
  @typing.final
1003
738
  class Layout(google.protobuf.message.Message):
1004
- """Layout determining how to store the dictionary in memory."""
1005
-
1006
739
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1007
740
 
1008
741
  class _Type:
@@ -1013,80 +746,106 @@ class ClickhouseConfig(google.protobuf.message.Message):
1013
746
  DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1014
747
  TYPE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 0
1015
748
  FLAT: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 1
1016
- """The entire dictionary is stored in memory in the form of flat arrays.
1017
- Available for all dictionary sources.
749
+ """The dictionary is completely stored in memory in the form of flat arrays.
750
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1018
751
  """
1019
752
  HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 2
1020
- """The entire dictionary is stored in memory in the form of a hash table.
1021
- Available for all dictionary sources.
753
+ """The dictionary is completely stored in memory in the form of a hash table.
754
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1022
755
  """
1023
756
  COMPLEX_KEY_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 3
1024
- """Similar to HASHED, to be used with composite keys.
1025
- Available for all dictionary sources.
757
+ """The dictionary is completely stored in memory in the form of a hash table.
758
+ Applicable for dictionaries with composite keys of arbitrary type.
1026
759
  """
1027
760
  RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 4
1028
- """The entire dictionary is stored in memory in the form of a hash table,
1029
- with an ordered array of ranges and their corresponding values.
1030
- Available for all dictionary sources.
761
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
762
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1031
763
  """
1032
764
  CACHE: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 5
1033
- """The dictionary is stored in a cache with a set number of cells.
1034
- Available for MySQL, ClickHouse and HTTP dictionary sources.
765
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
766
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1035
767
  """
1036
768
  COMPLEX_KEY_CACHE: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 6
1037
- """Similar to CACHE, to be used with composite keys.
1038
- Available for MySQL, ClickHouse and HTTP dictionary sources.
769
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
770
+ Applicable for dictionaries with composite keys of arbitrary type.
1039
771
  """
1040
772
  SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 7
1041
- """Similar to HASHED, but uses less memory in favor of more CPU usage."""
773
+ """The dictionary is completely stored in memory in the form of a hash table.
774
+ It's similar to HASHED layout type but uses less memory in favor of more CPU usage.
775
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
776
+ """
1042
777
  COMPLEX_KEY_SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 8
1043
- """Similar to SPARSE_HASHED, to be used with composite keys."""
778
+ """The dictionary is completely stored in memory in the form of a hash table.
779
+ It's similar to COMPLEX_KEY_HASHED layout type but uses less memory in favor of more CPU usage.
780
+ Applicable for dictionaries with composite keys of arbitrary type.
781
+ """
1044
782
  COMPLEX_KEY_RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 9
1045
- """Similar to RANGE_HASHED, to be used with composite keys."""
783
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
784
+ Applicable for dictionaries with composite keys of arbitrary type.
785
+ """
1046
786
  DIRECT: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 10
1047
- """The dictionary is not stored in memory and directly goes to the source during the processing of a request."""
787
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
788
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
789
+ """
1048
790
  COMPLEX_KEY_DIRECT: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 11
1049
- """Similar to DIRECT, to be used with composite keys."""
791
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
792
+ Applicable for dictionaries with composite keys of arbitrary type.
793
+ """
1050
794
  IP_TRIE: ClickhouseConfig.ExternalDictionary.Layout._Type.ValueType # 12
1051
795
  """The specialized layout type for mapping network prefixes (IP addresses) to metadata such as ASN."""
1052
796
 
1053
- class Type(_Type, metaclass=_TypeEnumTypeWrapper): ...
797
+ class Type(_Type, metaclass=_TypeEnumTypeWrapper):
798
+ """Layout type.
799
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/sql-reference/dictionaries#ways-to-store-dictionaries-in-memory).
800
+ """
801
+
1054
802
  TYPE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 0
1055
803
  FLAT: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 1
1056
- """The entire dictionary is stored in memory in the form of flat arrays.
1057
- Available for all dictionary sources.
804
+ """The dictionary is completely stored in memory in the form of flat arrays.
805
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1058
806
  """
1059
807
  HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 2
1060
- """The entire dictionary is stored in memory in the form of a hash table.
1061
- Available for all dictionary sources.
808
+ """The dictionary is completely stored in memory in the form of a hash table.
809
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1062
810
  """
1063
811
  COMPLEX_KEY_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 3
1064
- """Similar to HASHED, to be used with composite keys.
1065
- Available for all dictionary sources.
812
+ """The dictionary is completely stored in memory in the form of a hash table.
813
+ Applicable for dictionaries with composite keys of arbitrary type.
1066
814
  """
1067
815
  RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 4
1068
- """The entire dictionary is stored in memory in the form of a hash table,
1069
- with an ordered array of ranges and their corresponding values.
1070
- Available for all dictionary sources.
816
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
817
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1071
818
  """
1072
819
  CACHE: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 5
1073
- """The dictionary is stored in a cache with a set number of cells.
1074
- Available for MySQL, ClickHouse and HTTP dictionary sources.
820
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
821
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
1075
822
  """
1076
823
  COMPLEX_KEY_CACHE: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 6
1077
- """Similar to CACHE, to be used with composite keys.
1078
- Available for MySQL, ClickHouse and HTTP dictionary sources.
824
+ """The dictionary is stored in a cache that has a fixed number of cells. These cells contain frequently used elements.
825
+ Applicable for dictionaries with composite keys of arbitrary type.
1079
826
  """
1080
827
  SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 7
1081
- """Similar to HASHED, but uses less memory in favor of more CPU usage."""
828
+ """The dictionary is completely stored in memory in the form of a hash table.
829
+ It's similar to HASHED layout type but uses less memory in favor of more CPU usage.
830
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
831
+ """
1082
832
  COMPLEX_KEY_SPARSE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 8
1083
- """Similar to SPARSE_HASHED, to be used with composite keys."""
833
+ """The dictionary is completely stored in memory in the form of a hash table.
834
+ It's similar to COMPLEX_KEY_HASHED layout type but uses less memory in favor of more CPU usage.
835
+ Applicable for dictionaries with composite keys of arbitrary type.
836
+ """
1084
837
  COMPLEX_KEY_RANGE_HASHED: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 9
1085
- """Similar to RANGE_HASHED, to be used with composite keys."""
838
+ """The dictionary is stored in memory in the form of a hash table with an ordered array of ranges and their corresponding values.
839
+ Applicable for dictionaries with composite keys of arbitrary type.
840
+ """
1086
841
  DIRECT: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 10
1087
- """The dictionary is not stored in memory and directly goes to the source during the processing of a request."""
842
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
843
+ Applicable only for dictionaries with numeric keys of the UInt64 type.
844
+ """
1088
845
  COMPLEX_KEY_DIRECT: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 11
1089
- """Similar to DIRECT, to be used with composite keys."""
846
+ """The dictionary is not stored in memory and directly goes to the source during the processing of a request.
847
+ Applicable for dictionaries with composite keys of arbitrary type.
848
+ """
1090
849
  IP_TRIE: ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType # 12
1091
850
  """The specialized layout type for mapping network prefixes (IP addresses) to metadata such as ASN."""
1092
851
 
@@ -1101,46 +860,83 @@ class ClickhouseConfig(google.protobuf.message.Message):
1101
860
  MAX_ARRAY_SIZE_FIELD_NUMBER: builtins.int
1102
861
  ACCESS_TO_KEY_FROM_ATTRIBUTES_FIELD_NUMBER: builtins.int
1103
862
  type: global___ClickhouseConfig.ExternalDictionary.Layout.Type.ValueType
1104
- """Layout type for an external dictionary."""
863
+ """Layout type.
864
+
865
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#ways-to-store-dictionaries-in-memory).
866
+ """
1105
867
  size_in_cells: builtins.int
1106
868
  """Number of cells in the cache. Rounded up to a power of two.
1107
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
869
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
870
+
871
+ Default value: **1000000000**.
872
+
873
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1108
874
  """
1109
875
  max_update_queue_size: builtins.int
1110
876
  """Max size of update queue.
1111
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
877
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
878
+
879
+ Default value: **100000**.
880
+
881
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1112
882
  """
1113
883
  update_queue_push_timeout_milliseconds: builtins.int
1114
884
  """Max timeout in milliseconds for push update task into queue.
1115
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
885
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
886
+
887
+ Default value: **10**.
888
+
889
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1116
890
  """
1117
891
  query_wait_timeout_milliseconds: builtins.int
1118
892
  """Max wait timeout in milliseconds for update task to complete.
1119
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
893
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
894
+
895
+ Default value: **60000** (1 minute).
896
+
897
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1120
898
  """
1121
899
  max_threads_for_updates: builtins.int
1122
900
  """Max threads for cache dictionary update.
1123
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
901
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
902
+
903
+ Default value: **4**.
904
+
905
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1124
906
  """
1125
907
  initial_array_size: builtins.int
1126
908
  """Initial dictionary key size.
1127
- Applicable only for FLAT layout type.
909
+ Applicable only for **FLAT** layout type.
910
+
911
+ Default value: **1024**.
912
+
913
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#flat).
1128
914
  """
1129
915
  max_array_size: builtins.int
1130
916
  """Maximum dictionary key size.
1131
- Applicable only for FLAT layout type.
917
+ Applicable only for **FLAT** layout type.
918
+
919
+ Default value: **500000**.
920
+
921
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#flat).
1132
922
  """
1133
923
  @property
1134
924
  def allow_read_expired_keys(self) -> google.protobuf.wrappers_pb2.BoolValue:
1135
925
  """Allows to read expired keys.
1136
- Applicable only for CACHE and COMPLEX_KEY_CACHE layout types.
926
+ Applicable only for **CACHE** and **COMPLEX_KEY_CACHE** layout types.
927
+
928
+ Default value: **false**.
929
+
930
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#cache).
1137
931
  """
1138
932
 
1139
933
  @property
1140
934
  def access_to_key_from_attributes(self) -> google.protobuf.wrappers_pb2.BoolValue:
1141
- """Allows to retrieve key attribute using dictGetString function.
935
+ """Allows to retrieve key attribute using **dictGetString** function.
1142
936
  Enabling this option increases memory usage.
1143
- Applicable only for IP_TRIE layout type.
937
+ Applicable only for **IP_TRIE** layout type.
938
+
939
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries#ip_trie).
1144
940
  """
1145
941
 
1146
942
  def __init__(
@@ -1178,168 +974,712 @@ class ClickhouseConfig(google.protobuf.message.Message):
1178
974
  ) -> None: ...
1179
975
  def ClearField(self, field_name: typing.Literal["max", b"max", "min", b"min"]) -> None: ...
1180
976
 
1181
- NAME_FIELD_NUMBER: builtins.int
1182
- STRUCTURE_FIELD_NUMBER: builtins.int
1183
- LAYOUT_FIELD_NUMBER: builtins.int
1184
- FIXED_LIFETIME_FIELD_NUMBER: builtins.int
1185
- LIFETIME_RANGE_FIELD_NUMBER: builtins.int
1186
- HTTP_SOURCE_FIELD_NUMBER: builtins.int
1187
- MYSQL_SOURCE_FIELD_NUMBER: builtins.int
1188
- CLICKHOUSE_SOURCE_FIELD_NUMBER: builtins.int
1189
- MONGODB_SOURCE_FIELD_NUMBER: builtins.int
1190
- POSTGRESQL_SOURCE_FIELD_NUMBER: builtins.int
1191
- name: builtins.str
1192
- """Name of the external dictionary."""
1193
- fixed_lifetime: builtins.int
1194
- """Fixed interval between dictionary updates."""
1195
- @property
1196
- def structure(self) -> global___ClickhouseConfig.ExternalDictionary.Structure:
1197
- """Set of attributes for the external dictionary.
1198
- For in-depth description, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_structure/).
1199
- """
977
+ @typing.final
978
+ class HttpSource(google.protobuf.message.Message):
979
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1200
980
 
1201
- @property
1202
- def layout(self) -> global___ClickhouseConfig.ExternalDictionary.Layout:
1203
- """Layout for storing the dictionary in memory.
1204
- For in-depth description, see [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts_dict_layout/).
1205
- """
981
+ @typing.final
982
+ class Header(google.protobuf.message.Message):
983
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1206
984
 
1207
- @property
1208
- def lifetime_range(self) -> global___ClickhouseConfig.ExternalDictionary.Range:
1209
- """Range of intervals between dictionary updates for ClickHouse to choose from."""
985
+ NAME_FIELD_NUMBER: builtins.int
986
+ VALUE_FIELD_NUMBER: builtins.int
987
+ name: builtins.str
988
+ """Header name."""
989
+ value: builtins.str
990
+ """Header value."""
991
+ def __init__(
992
+ self,
993
+ *,
994
+ name: builtins.str = ...,
995
+ value: builtins.str = ...,
996
+ ) -> None: ...
997
+ def ClearField(self, field_name: typing.Literal["name", b"name", "value", b"value"]) -> None: ...
1210
998
 
1211
- @property
1212
- def http_source(self) -> global___ClickhouseConfig.ExternalDictionary.HttpSource:
1213
- """HTTP source for the dictionary."""
999
+ URL_FIELD_NUMBER: builtins.int
1000
+ FORMAT_FIELD_NUMBER: builtins.int
1001
+ HEADERS_FIELD_NUMBER: builtins.int
1002
+ url: builtins.str
1003
+ """URL of the source dictionary available over HTTP."""
1004
+ format: builtins.str
1005
+ """The data format. Valid values are all formats [supported by ClickHouse SQL dialect](https://clickhouse.com/docs/en/interfaces/formats/)."""
1006
+ @property
1007
+ def headers(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header]:
1008
+ """HTTP headers."""
1214
1009
 
1215
- @property
1216
- def mysql_source(self) -> global___ClickhouseConfig.ExternalDictionary.MysqlSource:
1217
- """MySQL source for the dictionary."""
1010
+ def __init__(
1011
+ self,
1012
+ *,
1013
+ url: builtins.str = ...,
1014
+ format: builtins.str = ...,
1015
+ headers: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.HttpSource.Header] | None = ...,
1016
+ ) -> None: ...
1017
+ def ClearField(self, field_name: typing.Literal["format", b"format", "headers", b"headers", "url", b"url"]) -> None: ...
1218
1018
 
1219
- @property
1220
- def clickhouse_source(self) -> global___ClickhouseConfig.ExternalDictionary.ClickhouseSource:
1221
- """ClickHouse source for the dictionary."""
1019
+ @typing.final
1020
+ class MysqlSource(google.protobuf.message.Message):
1021
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1022
+
1023
+ @typing.final
1024
+ class Replica(google.protobuf.message.Message):
1025
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1026
+
1027
+ HOST_FIELD_NUMBER: builtins.int
1028
+ PRIORITY_FIELD_NUMBER: builtins.int
1029
+ PORT_FIELD_NUMBER: builtins.int
1030
+ USER_FIELD_NUMBER: builtins.int
1031
+ PASSWORD_FIELD_NUMBER: builtins.int
1032
+ host: builtins.str
1033
+ """MySQL host of the replica."""
1034
+ priority: builtins.int
1035
+ """The priority of the replica that ClickHouse takes into account when connecting.
1036
+ Replica with the highest priority should have this field set to the lowest number.
1037
+ """
1038
+ port: builtins.int
1039
+ """Port to use when connecting to the replica.
1040
+ If a port is not specified for a replica, ClickHouse uses the port specified for the source.
1041
+ """
1042
+ user: builtins.str
1043
+ """Name of the MySQL database user.
1044
+ If a user is not specified for a replica, ClickHouse uses the user specified for the source.
1045
+ """
1046
+ password: builtins.str
1047
+ """Password of the MySQL database user.
1048
+ If a password is not specified for a replica, ClickHouse uses the password specified for the source.
1049
+ """
1050
+ def __init__(
1051
+ self,
1052
+ *,
1053
+ host: builtins.str = ...,
1054
+ priority: builtins.int = ...,
1055
+ port: builtins.int = ...,
1056
+ user: builtins.str = ...,
1057
+ password: builtins.str = ...,
1058
+ ) -> None: ...
1059
+ def ClearField(self, field_name: typing.Literal["host", b"host", "password", b"password", "port", b"port", "priority", b"priority", "user", b"user"]) -> None: ...
1060
+
1061
+ DB_FIELD_NUMBER: builtins.int
1062
+ TABLE_FIELD_NUMBER: builtins.int
1063
+ PORT_FIELD_NUMBER: builtins.int
1064
+ USER_FIELD_NUMBER: builtins.int
1065
+ PASSWORD_FIELD_NUMBER: builtins.int
1066
+ REPLICAS_FIELD_NUMBER: builtins.int
1067
+ WHERE_FIELD_NUMBER: builtins.int
1068
+ INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
1069
+ CLOSE_CONNECTION_FIELD_NUMBER: builtins.int
1070
+ SHARE_CONNECTION_FIELD_NUMBER: builtins.int
1071
+ db: builtins.str
1072
+ """Database name."""
1073
+ table: builtins.str
1074
+ """Table name."""
1075
+ port: builtins.int
1076
+ """Port to use when connecting to a replica of the dictionary source."""
1077
+ user: builtins.str
1078
+ """Name of the user for replicas of the dictionary source."""
1079
+ password: builtins.str
1080
+ """Password of the user for replicas of the dictionary source."""
1081
+ where: builtins.str
1082
+ """Selection criteria for the data in the specified MySQL table."""
1083
+ invalidate_query: builtins.str
1084
+ """Query for checking the dictionary status, to pull only updated data."""
1085
+ @property
1086
+ def replicas(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica]:
1087
+ """List of MySQL replicas of the database used as dictionary source."""
1088
+
1089
+ @property
1090
+ def close_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
1091
+ """Should a connection be closed after each request."""
1092
+
1093
+ @property
1094
+ def share_connection(self) -> google.protobuf.wrappers_pb2.BoolValue:
1095
+ """Should a connection be shared for some requests."""
1096
+
1097
+ def __init__(
1098
+ self,
1099
+ *,
1100
+ db: builtins.str = ...,
1101
+ table: builtins.str = ...,
1102
+ port: builtins.int = ...,
1103
+ user: builtins.str = ...,
1104
+ password: builtins.str = ...,
1105
+ replicas: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary.MysqlSource.Replica] | None = ...,
1106
+ where: builtins.str = ...,
1107
+ invalidate_query: builtins.str = ...,
1108
+ close_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1109
+ share_connection: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1110
+ ) -> None: ...
1111
+ def HasField(self, field_name: typing.Literal["close_connection", b"close_connection", "share_connection", b"share_connection"]) -> builtins.bool: ...
1112
+ def ClearField(self, field_name: typing.Literal["close_connection", b"close_connection", "db", b"db", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "replicas", b"replicas", "share_connection", b"share_connection", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
1113
+
1114
+ @typing.final
1115
+ class ClickhouseSource(google.protobuf.message.Message):
1116
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1117
+
1118
+ DB_FIELD_NUMBER: builtins.int
1119
+ TABLE_FIELD_NUMBER: builtins.int
1120
+ HOST_FIELD_NUMBER: builtins.int
1121
+ PORT_FIELD_NUMBER: builtins.int
1122
+ USER_FIELD_NUMBER: builtins.int
1123
+ PASSWORD_FIELD_NUMBER: builtins.int
1124
+ WHERE_FIELD_NUMBER: builtins.int
1125
+ SECURE_FIELD_NUMBER: builtins.int
1126
+ db: builtins.str
1127
+ """Database name."""
1128
+ table: builtins.str
1129
+ """Table name."""
1130
+ host: builtins.str
1131
+ """ClickHouse host."""
1132
+ port: builtins.int
1133
+ """Port to use when connecting to the host."""
1134
+ user: builtins.str
1135
+ """Name of the ClickHouse database user."""
1136
+ password: builtins.str
1137
+ """Password of the ClickHouse database user."""
1138
+ where: builtins.str
1139
+ """Selection criteria for the data in the specified ClickHouse table."""
1140
+ @property
1141
+ def secure(self) -> google.protobuf.wrappers_pb2.BoolValue:
1142
+ """Determines whether to use TLS for connection."""
1143
+
1144
+ def __init__(
1145
+ self,
1146
+ *,
1147
+ db: builtins.str = ...,
1148
+ table: builtins.str = ...,
1149
+ host: builtins.str = ...,
1150
+ port: builtins.int = ...,
1151
+ user: builtins.str = ...,
1152
+ password: builtins.str = ...,
1153
+ where: builtins.str = ...,
1154
+ secure: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1155
+ ) -> None: ...
1156
+ def HasField(self, field_name: typing.Literal["secure", b"secure"]) -> builtins.bool: ...
1157
+ def ClearField(self, field_name: typing.Literal["db", b"db", "host", b"host", "password", b"password", "port", b"port", "secure", b"secure", "table", b"table", "user", b"user", "where", b"where"]) -> None: ...
1158
+
1159
+ @typing.final
1160
+ class MongodbSource(google.protobuf.message.Message):
1161
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1162
+
1163
+ DB_FIELD_NUMBER: builtins.int
1164
+ COLLECTION_FIELD_NUMBER: builtins.int
1165
+ HOST_FIELD_NUMBER: builtins.int
1166
+ PORT_FIELD_NUMBER: builtins.int
1167
+ USER_FIELD_NUMBER: builtins.int
1168
+ PASSWORD_FIELD_NUMBER: builtins.int
1169
+ OPTIONS_FIELD_NUMBER: builtins.int
1170
+ db: builtins.str
1171
+ """Database name."""
1172
+ collection: builtins.str
1173
+ """Collection name."""
1174
+ host: builtins.str
1175
+ """MongoDB host."""
1176
+ port: builtins.int
1177
+ """Port to use when connecting to the host."""
1178
+ user: builtins.str
1179
+ """Name of the MongoDB database user."""
1180
+ password: builtins.str
1181
+ """Password of the MongoDB database user."""
1182
+ options: builtins.str
1183
+ """Dictionary source options."""
1184
+ def __init__(
1185
+ self,
1186
+ *,
1187
+ db: builtins.str = ...,
1188
+ collection: builtins.str = ...,
1189
+ host: builtins.str = ...,
1190
+ port: builtins.int = ...,
1191
+ user: builtins.str = ...,
1192
+ password: builtins.str = ...,
1193
+ options: builtins.str = ...,
1194
+ ) -> None: ...
1195
+ def ClearField(self, field_name: typing.Literal["collection", b"collection", "db", b"db", "host", b"host", "options", b"options", "password", b"password", "port", b"port", "user", b"user"]) -> None: ...
1196
+
1197
+ @typing.final
1198
+ class PostgresqlSource(google.protobuf.message.Message):
1199
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1200
+
1201
+ class _SslMode:
1202
+ ValueType = typing.NewType("ValueType", builtins.int)
1203
+ V: typing_extensions.TypeAlias = ValueType
1204
+
1205
+ class _SslModeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType], builtins.type):
1206
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1207
+ SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 0
1208
+ DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 1
1209
+ """Only try a non-SSL connection."""
1210
+ ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 2
1211
+ """First try a non-SSL connection; if that fails, try an SSL connection."""
1212
+ PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 3
1213
+ """First try an SSL connection; if that fails, try a non-SSL connection."""
1214
+ VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 4
1215
+ """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
1216
+ VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource._SslMode.ValueType # 5
1217
+ """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
1218
+
1219
+ class SslMode(_SslMode, metaclass=_SslModeEnumTypeWrapper):
1220
+ """Mode of SSL TCP/IP connection to a PostgreSQL host.
1221
+ For details, see [PostgreSQL documentation](https://www.postgresql.org/docs/current/libpq-ssl.html).
1222
+ """
1223
+
1224
+ SSL_MODE_UNSPECIFIED: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 0
1225
+ DISABLE: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 1
1226
+ """Only try a non-SSL connection."""
1227
+ ALLOW: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 2
1228
+ """First try a non-SSL connection; if that fails, try an SSL connection."""
1229
+ PREFER: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 3
1230
+ """First try an SSL connection; if that fails, try a non-SSL connection."""
1231
+ VERIFY_CA: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 4
1232
+ """Only try an SSL connection, and verify that the server certificate is issued by a trusted certificate authority (CA)."""
1233
+ VERIFY_FULL: ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType # 5
1234
+ """Only try an SSL connection, verify that the server certificate is issued by a trusted CA and that the requested server host name matches that in the certificate."""
1235
+
1236
+ DB_FIELD_NUMBER: builtins.int
1237
+ TABLE_FIELD_NUMBER: builtins.int
1238
+ HOSTS_FIELD_NUMBER: builtins.int
1239
+ PORT_FIELD_NUMBER: builtins.int
1240
+ USER_FIELD_NUMBER: builtins.int
1241
+ PASSWORD_FIELD_NUMBER: builtins.int
1242
+ INVALIDATE_QUERY_FIELD_NUMBER: builtins.int
1243
+ SSL_MODE_FIELD_NUMBER: builtins.int
1244
+ db: builtins.str
1245
+ """Database name."""
1246
+ table: builtins.str
1247
+ """Table name."""
1248
+ port: builtins.int
1249
+ """Port to use when connecting to the PostgreSQL hosts."""
1250
+ user: builtins.str
1251
+ """Name of the PostrgreSQL database user."""
1252
+ password: builtins.str
1253
+ """Password of the PostrgreSQL database user."""
1254
+ invalidate_query: builtins.str
1255
+ """Query for checking the dictionary status, to pull only updated data."""
1256
+ ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType
1257
+ """Mode of SSL TCP/IP connection to the PostgreSQL host."""
1258
+ @property
1259
+ def hosts(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
1260
+ """PostgreSQL hosts."""
1261
+
1262
+ def __init__(
1263
+ self,
1264
+ *,
1265
+ db: builtins.str = ...,
1266
+ table: builtins.str = ...,
1267
+ hosts: collections.abc.Iterable[builtins.str] | None = ...,
1268
+ port: builtins.int = ...,
1269
+ user: builtins.str = ...,
1270
+ password: builtins.str = ...,
1271
+ invalidate_query: builtins.str = ...,
1272
+ ssl_mode: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource.SslMode.ValueType = ...,
1273
+ ) -> None: ...
1274
+ def ClearField(self, field_name: typing.Literal["db", b"db", "hosts", b"hosts", "invalidate_query", b"invalidate_query", "password", b"password", "port", b"port", "ssl_mode", b"ssl_mode", "table", b"table", "user", b"user"]) -> None: ...
1275
+
1276
+ NAME_FIELD_NUMBER: builtins.int
1277
+ STRUCTURE_FIELD_NUMBER: builtins.int
1278
+ LAYOUT_FIELD_NUMBER: builtins.int
1279
+ FIXED_LIFETIME_FIELD_NUMBER: builtins.int
1280
+ LIFETIME_RANGE_FIELD_NUMBER: builtins.int
1281
+ HTTP_SOURCE_FIELD_NUMBER: builtins.int
1282
+ MYSQL_SOURCE_FIELD_NUMBER: builtins.int
1283
+ CLICKHOUSE_SOURCE_FIELD_NUMBER: builtins.int
1284
+ MONGODB_SOURCE_FIELD_NUMBER: builtins.int
1285
+ POSTGRESQL_SOURCE_FIELD_NUMBER: builtins.int
1286
+ name: builtins.str
1287
+ """Name of the external dictionary."""
1288
+ fixed_lifetime: builtins.int
1289
+ """Fixed interval between dictionary updates."""
1290
+ @property
1291
+ def structure(self) -> global___ClickhouseConfig.ExternalDictionary.Structure:
1292
+ """Structure of the external dictionary."""
1293
+
1294
+ @property
1295
+ def layout(self) -> global___ClickhouseConfig.ExternalDictionary.Layout:
1296
+ """Layout determining how to store the dictionary in memory.
1297
+
1298
+ For details, see https://clickhouse.com/docs/sql-reference/dictionaries#ways-to-store-dictionaries-in-memory.
1299
+ """
1300
+
1301
+ @property
1302
+ def lifetime_range(self) -> global___ClickhouseConfig.ExternalDictionary.Range:
1303
+ """Range of intervals between dictionary updates for ClickHouse to choose from."""
1304
+
1305
+ @property
1306
+ def http_source(self) -> global___ClickhouseConfig.ExternalDictionary.HttpSource:
1307
+ """HTTP source for the dictionary."""
1308
+
1309
+ @property
1310
+ def mysql_source(self) -> global___ClickhouseConfig.ExternalDictionary.MysqlSource:
1311
+ """MySQL source for the dictionary."""
1312
+
1313
+ @property
1314
+ def clickhouse_source(self) -> global___ClickhouseConfig.ExternalDictionary.ClickhouseSource:
1315
+ """ClickHouse source for the dictionary."""
1316
+
1317
+ @property
1318
+ def mongodb_source(self) -> global___ClickhouseConfig.ExternalDictionary.MongodbSource:
1319
+ """MongoDB source for the dictionary."""
1320
+
1321
+ @property
1322
+ def postgresql_source(self) -> global___ClickhouseConfig.ExternalDictionary.PostgresqlSource:
1323
+ """PostgreSQL source for the dictionary."""
1324
+
1325
+ def __init__(
1326
+ self,
1327
+ *,
1328
+ name: builtins.str = ...,
1329
+ structure: global___ClickhouseConfig.ExternalDictionary.Structure | None = ...,
1330
+ layout: global___ClickhouseConfig.ExternalDictionary.Layout | None = ...,
1331
+ fixed_lifetime: builtins.int = ...,
1332
+ lifetime_range: global___ClickhouseConfig.ExternalDictionary.Range | None = ...,
1333
+ http_source: global___ClickhouseConfig.ExternalDictionary.HttpSource | None = ...,
1334
+ mysql_source: global___ClickhouseConfig.ExternalDictionary.MysqlSource | None = ...,
1335
+ clickhouse_source: global___ClickhouseConfig.ExternalDictionary.ClickhouseSource | None = ...,
1336
+ mongodb_source: global___ClickhouseConfig.ExternalDictionary.MongodbSource | None = ...,
1337
+ postgresql_source: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource | None = ...,
1338
+ ) -> None: ...
1339
+ def HasField(self, field_name: typing.Literal["clickhouse_source", b"clickhouse_source", "fixed_lifetime", b"fixed_lifetime", "http_source", b"http_source", "layout", b"layout", "lifetime", b"lifetime", "lifetime_range", b"lifetime_range", "mongodb_source", b"mongodb_source", "mysql_source", b"mysql_source", "postgresql_source", b"postgresql_source", "source", b"source", "structure", b"structure"]) -> builtins.bool: ...
1340
+ def ClearField(self, field_name: typing.Literal["clickhouse_source", b"clickhouse_source", "fixed_lifetime", b"fixed_lifetime", "http_source", b"http_source", "layout", b"layout", "lifetime", b"lifetime", "lifetime_range", b"lifetime_range", "mongodb_source", b"mongodb_source", "mysql_source", b"mysql_source", "name", b"name", "postgresql_source", b"postgresql_source", "source", b"source", "structure", b"structure"]) -> None: ...
1341
+ @typing.overload
1342
+ def WhichOneof(self, oneof_group: typing.Literal["lifetime", b"lifetime"]) -> typing.Literal["fixed_lifetime", "lifetime_range"] | None: ...
1343
+ @typing.overload
1344
+ def WhichOneof(self, oneof_group: typing.Literal["source", b"source"]) -> typing.Literal["http_source", "mysql_source", "clickhouse_source", "mongodb_source", "postgresql_source"] | None: ...
1345
+
1346
+ @typing.final
1347
+ class GraphiteRollup(google.protobuf.message.Message):
1348
+ """Rollup settings for the GraphiteMergeTree table engine.
1349
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#graphite-rollup).
1350
+ """
1351
+
1352
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1353
+
1354
+ @typing.final
1355
+ class Pattern(google.protobuf.message.Message):
1356
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1357
+
1358
+ @typing.final
1359
+ class Retention(google.protobuf.message.Message):
1360
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1361
+
1362
+ AGE_FIELD_NUMBER: builtins.int
1363
+ PRECISION_FIELD_NUMBER: builtins.int
1364
+ age: builtins.int
1365
+ """The minimum age of the data in seconds."""
1366
+ precision: builtins.int
1367
+ """Precision of determining the age of the data, in seconds. Should be a divisor for 86400 (seconds in a day)."""
1368
+ def __init__(
1369
+ self,
1370
+ *,
1371
+ age: builtins.int = ...,
1372
+ precision: builtins.int = ...,
1373
+ ) -> None: ...
1374
+ def ClearField(self, field_name: typing.Literal["age", b"age", "precision", b"precision"]) -> None: ...
1375
+
1376
+ REGEXP_FIELD_NUMBER: builtins.int
1377
+ FUNCTION_FIELD_NUMBER: builtins.int
1378
+ RETENTION_FIELD_NUMBER: builtins.int
1379
+ regexp: builtins.str
1380
+ """A pattern for the metric name (a regular or DSL)."""
1381
+ function: builtins.str
1382
+ """The name of the aggregating function to apply to data whose age falls within the range [age, age + precision].
1383
+ Accepted functions: **min**, **max**, **any**, **avg**. The average is calculated imprecisely, like the average of the averages.
1384
+ """
1385
+ @property
1386
+ def retention(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention]:
1387
+ """Retention rules."""
1388
+
1389
+ def __init__(
1390
+ self,
1391
+ *,
1392
+ regexp: builtins.str = ...,
1393
+ function: builtins.str = ...,
1394
+ retention: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention] | None = ...,
1395
+ ) -> None: ...
1396
+ def ClearField(self, field_name: typing.Literal["function", b"function", "regexp", b"regexp", "retention", b"retention"]) -> None: ...
1397
+
1398
+ NAME_FIELD_NUMBER: builtins.int
1399
+ PATTERNS_FIELD_NUMBER: builtins.int
1400
+ PATH_COLUMN_NAME_FIELD_NUMBER: builtins.int
1401
+ TIME_COLUMN_NAME_FIELD_NUMBER: builtins.int
1402
+ VALUE_COLUMN_NAME_FIELD_NUMBER: builtins.int
1403
+ VERSION_COLUMN_NAME_FIELD_NUMBER: builtins.int
1404
+ name: builtins.str
1405
+ """Name for the specified combination of settings for Graphite rollup."""
1406
+ path_column_name: builtins.str
1407
+ """The name of the column storing the metric name (Graphite sensor).
1408
+
1409
+ Default value: **Path**.
1410
+ """
1411
+ time_column_name: builtins.str
1412
+ """The name of the column storing the time of measuring the metric.
1413
+
1414
+ Default value: **Time**.
1415
+ """
1416
+ value_column_name: builtins.str
1417
+ """The name of the column storing the value of the metric at the time set in **time_column_name**.
1418
+
1419
+ Default value: **Value**.
1420
+ """
1421
+ version_column_name: builtins.str
1422
+ """The name of the column storing the version of the metric.
1423
+
1424
+ Default value: **Timestamp**.
1425
+ """
1426
+ @property
1427
+ def patterns(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern]:
1428
+ """Pattern to use for the rollup."""
1429
+
1430
+ def __init__(
1431
+ self,
1432
+ *,
1433
+ name: builtins.str = ...,
1434
+ patterns: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern] | None = ...,
1435
+ path_column_name: builtins.str = ...,
1436
+ time_column_name: builtins.str = ...,
1437
+ value_column_name: builtins.str = ...,
1438
+ version_column_name: builtins.str = ...,
1439
+ ) -> None: ...
1440
+ def ClearField(self, field_name: typing.Literal["name", b"name", "path_column_name", b"path_column_name", "patterns", b"patterns", "time_column_name", b"time_column_name", "value_column_name", b"value_column_name", "version_column_name", b"version_column_name"]) -> None: ...
1441
+
1442
+ @typing.final
1443
+ class Kafka(google.protobuf.message.Message):
1444
+ """Kafka configuration settings.
1445
+ For details, see [librdkafka documentation](https://github.com/confluentinc/librdkafka/blob/master/CONFIGURATION.md).
1446
+ """
1447
+
1448
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1449
+
1450
+ class _SecurityProtocol:
1451
+ ValueType = typing.NewType("ValueType", builtins.int)
1452
+ V: typing_extensions.TypeAlias = ValueType
1453
+
1454
+ class _SecurityProtocolEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SecurityProtocol.ValueType], builtins.type):
1455
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1456
+ SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 0
1457
+ SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 1
1458
+ SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 2
1459
+ SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 3
1460
+ SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka._SecurityProtocol.ValueType # 4
1461
+
1462
+ class SecurityProtocol(_SecurityProtocol, metaclass=_SecurityProtocolEnumTypeWrapper): ...
1463
+ SECURITY_PROTOCOL_UNSPECIFIED: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 0
1464
+ SECURITY_PROTOCOL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 1
1465
+ SECURITY_PROTOCOL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 2
1466
+ SECURITY_PROTOCOL_SASL_PLAINTEXT: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 3
1467
+ SECURITY_PROTOCOL_SASL_SSL: ClickhouseConfig.Kafka.SecurityProtocol.ValueType # 4
1468
+
1469
+ class _SaslMechanism:
1470
+ ValueType = typing.NewType("ValueType", builtins.int)
1471
+ V: typing_extensions.TypeAlias = ValueType
1472
+
1473
+ class _SaslMechanismEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._SaslMechanism.ValueType], builtins.type):
1474
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1475
+ SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 0
1476
+ SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 1
1477
+ SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 2
1478
+ SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 3
1479
+ SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka._SaslMechanism.ValueType # 4
1480
+
1481
+ class SaslMechanism(_SaslMechanism, metaclass=_SaslMechanismEnumTypeWrapper): ...
1482
+ SASL_MECHANISM_UNSPECIFIED: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 0
1483
+ SASL_MECHANISM_GSSAPI: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 1
1484
+ SASL_MECHANISM_PLAIN: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 2
1485
+ SASL_MECHANISM_SCRAM_SHA_256: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 3
1486
+ SASL_MECHANISM_SCRAM_SHA_512: ClickhouseConfig.Kafka.SaslMechanism.ValueType # 4
1487
+
1488
+ class _Debug:
1489
+ ValueType = typing.NewType("ValueType", builtins.int)
1490
+ V: typing_extensions.TypeAlias = ValueType
1491
+
1492
+ class _DebugEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._Debug.ValueType], builtins.type):
1493
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1494
+ DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka._Debug.ValueType # 0
1495
+ DEBUG_GENERIC: ClickhouseConfig.Kafka._Debug.ValueType # 1
1496
+ DEBUG_BROKER: ClickhouseConfig.Kafka._Debug.ValueType # 2
1497
+ DEBUG_TOPIC: ClickhouseConfig.Kafka._Debug.ValueType # 3
1498
+ DEBUG_METADATA: ClickhouseConfig.Kafka._Debug.ValueType # 4
1499
+ DEBUG_FEATURE: ClickhouseConfig.Kafka._Debug.ValueType # 5
1500
+ DEBUG_QUEUE: ClickhouseConfig.Kafka._Debug.ValueType # 6
1501
+ DEBUG_MSG: ClickhouseConfig.Kafka._Debug.ValueType # 7
1502
+ DEBUG_PROTOCOL: ClickhouseConfig.Kafka._Debug.ValueType # 8
1503
+ DEBUG_CGRP: ClickhouseConfig.Kafka._Debug.ValueType # 9
1504
+ DEBUG_SECURITY: ClickhouseConfig.Kafka._Debug.ValueType # 10
1505
+ DEBUG_FETCH: ClickhouseConfig.Kafka._Debug.ValueType # 11
1506
+ DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka._Debug.ValueType # 12
1507
+ DEBUG_PLUGIN: ClickhouseConfig.Kafka._Debug.ValueType # 13
1508
+ DEBUG_CONSUMER: ClickhouseConfig.Kafka._Debug.ValueType # 14
1509
+ DEBUG_ADMIN: ClickhouseConfig.Kafka._Debug.ValueType # 15
1510
+ DEBUG_EOS: ClickhouseConfig.Kafka._Debug.ValueType # 16
1511
+ DEBUG_MOCK: ClickhouseConfig.Kafka._Debug.ValueType # 17
1512
+ DEBUG_ASSIGNOR: ClickhouseConfig.Kafka._Debug.ValueType # 18
1513
+ DEBUG_CONF: ClickhouseConfig.Kafka._Debug.ValueType # 19
1514
+ DEBUG_TELEMETRY: ClickhouseConfig.Kafka._Debug.ValueType # 20
1515
+ DEBUG_ALL: ClickhouseConfig.Kafka._Debug.ValueType # 21
1516
+
1517
+ class Debug(_Debug, metaclass=_DebugEnumTypeWrapper): ...
1518
+ DEBUG_UNSPECIFIED: ClickhouseConfig.Kafka.Debug.ValueType # 0
1519
+ DEBUG_GENERIC: ClickhouseConfig.Kafka.Debug.ValueType # 1
1520
+ DEBUG_BROKER: ClickhouseConfig.Kafka.Debug.ValueType # 2
1521
+ DEBUG_TOPIC: ClickhouseConfig.Kafka.Debug.ValueType # 3
1522
+ DEBUG_METADATA: ClickhouseConfig.Kafka.Debug.ValueType # 4
1523
+ DEBUG_FEATURE: ClickhouseConfig.Kafka.Debug.ValueType # 5
1524
+ DEBUG_QUEUE: ClickhouseConfig.Kafka.Debug.ValueType # 6
1525
+ DEBUG_MSG: ClickhouseConfig.Kafka.Debug.ValueType # 7
1526
+ DEBUG_PROTOCOL: ClickhouseConfig.Kafka.Debug.ValueType # 8
1527
+ DEBUG_CGRP: ClickhouseConfig.Kafka.Debug.ValueType # 9
1528
+ DEBUG_SECURITY: ClickhouseConfig.Kafka.Debug.ValueType # 10
1529
+ DEBUG_FETCH: ClickhouseConfig.Kafka.Debug.ValueType # 11
1530
+ DEBUG_INTERCEPTOR: ClickhouseConfig.Kafka.Debug.ValueType # 12
1531
+ DEBUG_PLUGIN: ClickhouseConfig.Kafka.Debug.ValueType # 13
1532
+ DEBUG_CONSUMER: ClickhouseConfig.Kafka.Debug.ValueType # 14
1533
+ DEBUG_ADMIN: ClickhouseConfig.Kafka.Debug.ValueType # 15
1534
+ DEBUG_EOS: ClickhouseConfig.Kafka.Debug.ValueType # 16
1535
+ DEBUG_MOCK: ClickhouseConfig.Kafka.Debug.ValueType # 17
1536
+ DEBUG_ASSIGNOR: ClickhouseConfig.Kafka.Debug.ValueType # 18
1537
+ DEBUG_CONF: ClickhouseConfig.Kafka.Debug.ValueType # 19
1538
+ DEBUG_TELEMETRY: ClickhouseConfig.Kafka.Debug.ValueType # 20
1539
+ DEBUG_ALL: ClickhouseConfig.Kafka.Debug.ValueType # 21
1540
+
1541
+ class _AutoOffsetReset:
1542
+ ValueType = typing.NewType("ValueType", builtins.int)
1543
+ V: typing_extensions.TypeAlias = ValueType
1544
+
1545
+ class _AutoOffsetResetEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ClickhouseConfig.Kafka._AutoOffsetReset.ValueType], builtins.type):
1546
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
1547
+ AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 0
1548
+ AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 1
1549
+ AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 2
1550
+ AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 3
1551
+ AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 4
1552
+ AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 5
1553
+ AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 6
1554
+ AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka._AutoOffsetReset.ValueType # 7
1555
+
1556
+ class AutoOffsetReset(_AutoOffsetReset, metaclass=_AutoOffsetResetEnumTypeWrapper): ...
1557
+ AUTO_OFFSET_RESET_UNSPECIFIED: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 0
1558
+ AUTO_OFFSET_RESET_SMALLEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 1
1559
+ AUTO_OFFSET_RESET_EARLIEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 2
1560
+ AUTO_OFFSET_RESET_BEGINNING: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 3
1561
+ AUTO_OFFSET_RESET_LARGEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 4
1562
+ AUTO_OFFSET_RESET_LATEST: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 5
1563
+ AUTO_OFFSET_RESET_END: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 6
1564
+ AUTO_OFFSET_RESET_ERROR: ClickhouseConfig.Kafka.AutoOffsetReset.ValueType # 7
1565
+
1566
+ SECURITY_PROTOCOL_FIELD_NUMBER: builtins.int
1567
+ SASL_MECHANISM_FIELD_NUMBER: builtins.int
1568
+ SASL_USERNAME_FIELD_NUMBER: builtins.int
1569
+ SASL_PASSWORD_FIELD_NUMBER: builtins.int
1570
+ ENABLE_SSL_CERTIFICATE_VERIFICATION_FIELD_NUMBER: builtins.int
1571
+ MAX_POLL_INTERVAL_MS_FIELD_NUMBER: builtins.int
1572
+ SESSION_TIMEOUT_MS_FIELD_NUMBER: builtins.int
1573
+ DEBUG_FIELD_NUMBER: builtins.int
1574
+ AUTO_OFFSET_RESET_FIELD_NUMBER: builtins.int
1575
+ security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType
1576
+ """Protocol used to communicate with brokers.
1577
+
1578
+ Default value: **SECURITY_PROTOCOL_PLAINTEXT**.
1579
+ """
1580
+ sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType
1581
+ """SASL mechanism to use for authentication.
1582
+
1583
+ Default value: **SASL_MECHANISM_GSSAPI**.
1584
+ """
1585
+ sasl_username: builtins.str
1586
+ """SASL username for use with the PLAIN and SASL-SCRAM mechanisms."""
1587
+ sasl_password: builtins.str
1588
+ """SASL password for use with the PLAIN and SASL-SCRAM mechanisms."""
1589
+ debug: global___ClickhouseConfig.Kafka.Debug.ValueType
1590
+ """Debug context to enable."""
1591
+ auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType
1592
+ """Action to take when there is no initial offset in offset store or the desired offset is out of range.
1593
+
1594
+ Default value: **AUTO_OFFSET_RESET_LARGEST**.
1595
+ """
1596
+ @property
1597
+ def enable_ssl_certificate_verification(self) -> google.protobuf.wrappers_pb2.BoolValue:
1598
+ """Enable OpenSSL's builtin broker (server) certificate verification.
1599
+
1600
+ Default value: **true**.
1601
+ """
1602
+
1603
+ @property
1604
+ def max_poll_interval_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
1605
+ """Maximum allowed time between calls to consume messages for high-level consumers.
1606
+ If this interval is exceeded the consumer is considered failed and the group will
1607
+ rebalance in order to reassign the partitions to another consumer group member.
1608
+
1609
+ Default value: **300000** (5 minutes).
1610
+ """
1611
+
1612
+ @property
1613
+ def session_timeout_ms(self) -> google.protobuf.wrappers_pb2.Int64Value:
1614
+ """Client group session and failure detection timeout. The consumer sends periodic heartbeats (heartbeat.interval.ms)
1615
+ to indicate its liveness to the broker. If no hearts are received by the broker for a group member within
1616
+ the session timeout, the broker will remove the consumer from the group and trigger a rebalance.
1617
+
1618
+ Default value: **45000** (45 seconds).
1619
+ """
1620
+
1621
+ def __init__(
1622
+ self,
1623
+ *,
1624
+ security_protocol: global___ClickhouseConfig.Kafka.SecurityProtocol.ValueType = ...,
1625
+ sasl_mechanism: global___ClickhouseConfig.Kafka.SaslMechanism.ValueType = ...,
1626
+ sasl_username: builtins.str = ...,
1627
+ sasl_password: builtins.str = ...,
1628
+ enable_ssl_certificate_verification: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1629
+ max_poll_interval_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1630
+ session_timeout_ms: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1631
+ debug: global___ClickhouseConfig.Kafka.Debug.ValueType = ...,
1632
+ auto_offset_reset: global___ClickhouseConfig.Kafka.AutoOffsetReset.ValueType = ...,
1633
+ ) -> None: ...
1634
+ def HasField(self, field_name: typing.Literal["enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "session_timeout_ms", b"session_timeout_ms"]) -> builtins.bool: ...
1635
+ def ClearField(self, field_name: typing.Literal["auto_offset_reset", b"auto_offset_reset", "debug", b"debug", "enable_ssl_certificate_verification", b"enable_ssl_certificate_verification", "max_poll_interval_ms", b"max_poll_interval_ms", "sasl_mechanism", b"sasl_mechanism", "sasl_password", b"sasl_password", "sasl_username", b"sasl_username", "security_protocol", b"security_protocol", "session_timeout_ms", b"session_timeout_ms"]) -> None: ...
1222
1636
 
1223
- @property
1224
- def mongodb_source(self) -> global___ClickhouseConfig.ExternalDictionary.MongodbSource:
1225
- """MongoDB source for the dictionary."""
1637
+ @typing.final
1638
+ class KafkaTopic(google.protobuf.message.Message):
1639
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1226
1640
 
1641
+ NAME_FIELD_NUMBER: builtins.int
1642
+ SETTINGS_FIELD_NUMBER: builtins.int
1643
+ name: builtins.str
1644
+ """Kafka topic name."""
1227
1645
  @property
1228
- def postgresql_source(self) -> global___ClickhouseConfig.ExternalDictionary.PostgresqlSource:
1229
- """PostgreSQL source for the dictionary."""
1646
+ def settings(self) -> global___ClickhouseConfig.Kafka:
1647
+ """Kafka topic settings."""
1230
1648
 
1231
1649
  def __init__(
1232
1650
  self,
1233
1651
  *,
1234
1652
  name: builtins.str = ...,
1235
- structure: global___ClickhouseConfig.ExternalDictionary.Structure | None = ...,
1236
- layout: global___ClickhouseConfig.ExternalDictionary.Layout | None = ...,
1237
- fixed_lifetime: builtins.int = ...,
1238
- lifetime_range: global___ClickhouseConfig.ExternalDictionary.Range | None = ...,
1239
- http_source: global___ClickhouseConfig.ExternalDictionary.HttpSource | None = ...,
1240
- mysql_source: global___ClickhouseConfig.ExternalDictionary.MysqlSource | None = ...,
1241
- clickhouse_source: global___ClickhouseConfig.ExternalDictionary.ClickhouseSource | None = ...,
1242
- mongodb_source: global___ClickhouseConfig.ExternalDictionary.MongodbSource | None = ...,
1243
- postgresql_source: global___ClickhouseConfig.ExternalDictionary.PostgresqlSource | None = ...,
1653
+ settings: global___ClickhouseConfig.Kafka | None = ...,
1244
1654
  ) -> None: ...
1245
- def HasField(self, field_name: typing.Literal["clickhouse_source", b"clickhouse_source", "fixed_lifetime", b"fixed_lifetime", "http_source", b"http_source", "layout", b"layout", "lifetime", b"lifetime", "lifetime_range", b"lifetime_range", "mongodb_source", b"mongodb_source", "mysql_source", b"mysql_source", "postgresql_source", b"postgresql_source", "source", b"source", "structure", b"structure"]) -> builtins.bool: ...
1246
- def ClearField(self, field_name: typing.Literal["clickhouse_source", b"clickhouse_source", "fixed_lifetime", b"fixed_lifetime", "http_source", b"http_source", "layout", b"layout", "lifetime", b"lifetime", "lifetime_range", b"lifetime_range", "mongodb_source", b"mongodb_source", "mysql_source", b"mysql_source", "name", b"name", "postgresql_source", b"postgresql_source", "source", b"source", "structure", b"structure"]) -> None: ...
1247
- @typing.overload
1248
- def WhichOneof(self, oneof_group: typing.Literal["lifetime", b"lifetime"]) -> typing.Literal["fixed_lifetime", "lifetime_range"] | None: ...
1249
- @typing.overload
1250
- def WhichOneof(self, oneof_group: typing.Literal["source", b"source"]) -> typing.Literal["http_source", "mysql_source", "clickhouse_source", "mongodb_source", "postgresql_source"] | None: ...
1655
+ def HasField(self, field_name: typing.Literal["settings", b"settings"]) -> builtins.bool: ...
1656
+ def ClearField(self, field_name: typing.Literal["name", b"name", "settings", b"settings"]) -> None: ...
1251
1657
 
1252
1658
  @typing.final
1253
- class GraphiteRollup(google.protobuf.message.Message):
1254
- """Rollup settings for the GraphiteMergeTree table engine."""
1659
+ class Rabbitmq(google.protobuf.message.Message):
1660
+ """RabbitMQ integration settings.
1661
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/engines/table-engines/integrations/rabbitmq).
1662
+ """
1255
1663
 
1256
1664
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1257
1665
 
1258
- @typing.final
1259
- class Pattern(google.protobuf.message.Message):
1260
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
1261
-
1262
- @typing.final
1263
- class Retention(google.protobuf.message.Message):
1264
- DESCRIPTOR: google.protobuf.descriptor.Descriptor
1265
-
1266
- AGE_FIELD_NUMBER: builtins.int
1267
- PRECISION_FIELD_NUMBER: builtins.int
1268
- age: builtins.int
1269
- """Minimum age of the data in seconds."""
1270
- precision: builtins.int
1271
- """Precision of determining the age of the data, in seconds."""
1272
- def __init__(
1273
- self,
1274
- *,
1275
- age: builtins.int = ...,
1276
- precision: builtins.int = ...,
1277
- ) -> None: ...
1278
- def ClearField(self, field_name: typing.Literal["age", b"age", "precision", b"precision"]) -> None: ...
1279
-
1280
- REGEXP_FIELD_NUMBER: builtins.int
1281
- FUNCTION_FIELD_NUMBER: builtins.int
1282
- RETENTION_FIELD_NUMBER: builtins.int
1283
- regexp: builtins.str
1284
- """Pattern for metric names."""
1285
- function: builtins.str
1286
- """Name of the aggregating function to apply to data of the age specified in [retention]."""
1287
- @property
1288
- def retention(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention]:
1289
- """Age of data to use for thinning."""
1290
-
1291
- def __init__(
1292
- self,
1293
- *,
1294
- regexp: builtins.str = ...,
1295
- function: builtins.str = ...,
1296
- retention: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern.Retention] | None = ...,
1297
- ) -> None: ...
1298
- def ClearField(self, field_name: typing.Literal["function", b"function", "regexp", b"regexp", "retention", b"retention"]) -> None: ...
1299
-
1300
- NAME_FIELD_NUMBER: builtins.int
1301
- PATTERNS_FIELD_NUMBER: builtins.int
1302
- PATH_COLUMN_NAME_FIELD_NUMBER: builtins.int
1303
- TIME_COLUMN_NAME_FIELD_NUMBER: builtins.int
1304
- VALUE_COLUMN_NAME_FIELD_NUMBER: builtins.int
1305
- VERSION_COLUMN_NAME_FIELD_NUMBER: builtins.int
1306
- name: builtins.str
1307
- """Name for the specified combination of settings for Graphite rollup."""
1308
- path_column_name: builtins.str
1309
- """The name of the column storing the metric name (Graphite sensor).
1310
- Default: Path
1311
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1312
- """
1313
- time_column_name: builtins.str
1314
- """The name of the column storing the time of measuring the metric.
1315
- Default: Time
1316
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1317
- """
1318
- value_column_name: builtins.str
1319
- """The name of the column storing the value of the metric at the time set in time_column_name.
1320
- Default: Value
1321
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1322
- """
1323
- version_column_name: builtins.str
1324
- """The name of the column storing the version of the metric.
1325
- Default: Timestamp
1326
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/ru/engines/table-engines/mergetree-family/graphitemergetree#required-columns)
1327
- """
1328
- @property
1329
- def patterns(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup.Pattern]:
1330
- """Pattern to use for the rollup."""
1331
-
1666
+ USERNAME_FIELD_NUMBER: builtins.int
1667
+ PASSWORD_FIELD_NUMBER: builtins.int
1668
+ VHOST_FIELD_NUMBER: builtins.int
1669
+ username: builtins.str
1670
+ """RabbitMQ username."""
1671
+ password: builtins.str
1672
+ """RabbitMQ password."""
1673
+ vhost: builtins.str
1674
+ """RabbitMQ virtual host."""
1332
1675
  def __init__(
1333
1676
  self,
1334
1677
  *,
1335
- name: builtins.str = ...,
1336
- patterns: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup.Pattern] | None = ...,
1337
- path_column_name: builtins.str = ...,
1338
- time_column_name: builtins.str = ...,
1339
- value_column_name: builtins.str = ...,
1340
- version_column_name: builtins.str = ...,
1678
+ username: builtins.str = ...,
1679
+ password: builtins.str = ...,
1680
+ vhost: builtins.str = ...,
1341
1681
  ) -> None: ...
1342
- def ClearField(self, field_name: typing.Literal["name", b"name", "path_column_name", b"path_column_name", "patterns", b"patterns", "time_column_name", b"time_column_name", "value_column_name", b"value_column_name", "version_column_name", b"version_column_name"]) -> None: ...
1682
+ def ClearField(self, field_name: typing.Literal["password", b"password", "username", b"username", "vhost", b"vhost"]) -> None: ...
1343
1683
 
1344
1684
  @typing.final
1345
1685
  class QueryMaskingRule(google.protobuf.message.Message):
@@ -1351,12 +1691,11 @@ class ClickhouseConfig(google.protobuf.message.Message):
1351
1691
  name: builtins.str
1352
1692
  """Name for the rule."""
1353
1693
  regexp: builtins.str
1354
- """RE2 compatible regular expression.
1355
- Required.
1356
- """
1694
+ """RE2 compatible regular expression."""
1357
1695
  replace: builtins.str
1358
1696
  """Substitution string for sensitive data.
1359
- Default: six asterisks
1697
+
1698
+ Default value: six asterisks.
1360
1699
  """
1361
1700
  def __init__(
1362
1701
  self,
@@ -1369,6 +1708,8 @@ class ClickhouseConfig(google.protobuf.message.Message):
1369
1708
 
1370
1709
  @typing.final
1371
1710
  class QueryCache(google.protobuf.message.Message):
1711
+ """Query cache configuration."""
1712
+
1372
1713
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1373
1714
 
1374
1715
  MAX_SIZE_IN_BYTES_FIELD_NUMBER: builtins.int
@@ -1378,25 +1719,29 @@ class ClickhouseConfig(google.protobuf.message.Message):
1378
1719
  @property
1379
1720
  def max_size_in_bytes(self) -> google.protobuf.wrappers_pb2.Int64Value:
1380
1721
  """The maximum cache size in bytes.
1381
- Default: 1073741824 (1 GiB)
1722
+
1723
+ Default value: **1073741824** (1 GiB).
1382
1724
  """
1383
1725
 
1384
1726
  @property
1385
1727
  def max_entries(self) -> google.protobuf.wrappers_pb2.Int64Value:
1386
- """The maximum number of SELECT query results stored in the cache.
1387
- Default: 1024
1728
+ """The maximum number of **SELECT** query results stored in the cache.
1729
+
1730
+ Default value: **1024**.
1388
1731
  """
1389
1732
 
1390
1733
  @property
1391
1734
  def max_entry_size_in_bytes(self) -> google.protobuf.wrappers_pb2.Int64Value:
1392
- """The maximum size in bytes SELECT query results may have to be saved in the cache.
1393
- Dafault: 1048576 (1 MiB)
1735
+ """The maximum size in bytes **SELECT** query results may have to be saved in the cache.
1736
+
1737
+ Default value: **1048576** (1 MiB).
1394
1738
  """
1395
1739
 
1396
1740
  @property
1397
1741
  def max_entry_size_in_rows(self) -> google.protobuf.wrappers_pb2.Int64Value:
1398
- """The maximum number of rows SELECT query results may have to be saved in the cache.
1399
- Default: 30000000 (30 mil)
1742
+ """The maximum number of rows **SELECT** query results may have to be saved in the cache.
1743
+
1744
+ Default value: **30000000**.
1400
1745
  """
1401
1746
 
1402
1747
  def __init__(
@@ -1412,7 +1757,7 @@ class ClickhouseConfig(google.protobuf.message.Message):
1412
1757
 
1413
1758
  @typing.final
1414
1759
  class JdbcBridge(google.protobuf.message.Message):
1415
- """JDBC bridge for queries to external databases."""
1760
+ """JDBC bridge configuration for queries to external databases."""
1416
1761
 
1417
1762
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1418
1763
 
@@ -1422,7 +1767,10 @@ class ClickhouseConfig(google.protobuf.message.Message):
1422
1767
  """Host of jdbc bridge."""
1423
1768
  @property
1424
1769
  def port(self) -> google.protobuf.wrappers_pb2.Int64Value:
1425
- """Port of jdbc bridge."""
1770
+ """Port of jdbc bridge.
1771
+
1772
+ Default value: **9019**.
1773
+ """
1426
1774
 
1427
1775
  def __init__(
1428
1776
  self,
@@ -1434,53 +1782,36 @@ class ClickhouseConfig(google.protobuf.message.Message):
1434
1782
  def ClearField(self, field_name: typing.Literal["host", b"host", "port", b"port"]) -> None: ...
1435
1783
 
1436
1784
  @typing.final
1437
- class AccessControlImprovements(google.protobuf.message.Message):
1438
- """Access control settings."""
1785
+ class Macro(google.protobuf.message.Message):
1786
+ """ClickHouse macro."""
1439
1787
 
1440
1788
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
1441
1789
 
1442
- SELECT_FROM_SYSTEM_DB_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
1443
- SELECT_FROM_INFORMATION_SCHEMA_REQUIRES_GRANT_FIELD_NUMBER: builtins.int
1444
- @property
1445
- def select_from_system_db_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
1446
- """Sets whether SELECT * FROM system.<table> requires any grants and can be executed by any user.
1447
- If set to true then this query requires GRANT SELECT ON system.<table> just as for non-system tables.
1448
- """
1449
-
1450
- @property
1451
- def select_from_information_schema_requires_grant(self) -> google.protobuf.wrappers_pb2.BoolValue:
1452
- """Sets whether SELECT * FROM information_schema.<table> requires any grants and can be executed by any user.
1453
- If set to true, then this query requires GRANT SELECT ON information_schema.<table>, just as for ordinary tables.
1454
- """
1455
-
1790
+ NAME_FIELD_NUMBER: builtins.int
1791
+ VALUE_FIELD_NUMBER: builtins.int
1792
+ name: builtins.str
1793
+ """Name of the macro."""
1794
+ value: builtins.str
1795
+ """Value of the macro."""
1456
1796
  def __init__(
1457
1797
  self,
1458
1798
  *,
1459
- select_from_system_db_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1460
- select_from_information_schema_requires_grant: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1799
+ name: builtins.str = ...,
1800
+ value: builtins.str = ...,
1461
1801
  ) -> None: ...
1462
- def HasField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> builtins.bool: ...
1463
- def ClearField(self, field_name: typing.Literal["select_from_information_schema_requires_grant", b"select_from_information_schema_requires_grant", "select_from_system_db_requires_grant", b"select_from_system_db_requires_grant"]) -> None: ...
1802
+ def ClearField(self, field_name: typing.Literal["name", b"name", "value", b"value"]) -> None: ...
1464
1803
 
1804
+ BACKGROUND_POOL_SIZE_FIELD_NUMBER: builtins.int
1805
+ BACKGROUND_MERGES_MUTATIONS_CONCURRENCY_RATIO_FIELD_NUMBER: builtins.int
1806
+ BACKGROUND_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1807
+ BACKGROUND_FETCHES_POOL_SIZE_FIELD_NUMBER: builtins.int
1808
+ BACKGROUND_MOVE_POOL_SIZE_FIELD_NUMBER: builtins.int
1809
+ BACKGROUND_DISTRIBUTED_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1810
+ BACKGROUND_BUFFER_FLUSH_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1811
+ BACKGROUND_MESSAGE_BROKER_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1812
+ BACKGROUND_COMMON_POOL_SIZE_FIELD_NUMBER: builtins.int
1813
+ DICTIONARIES_LAZY_LOAD_FIELD_NUMBER: builtins.int
1465
1814
  LOG_LEVEL_FIELD_NUMBER: builtins.int
1466
- MERGE_TREE_FIELD_NUMBER: builtins.int
1467
- COMPRESSION_FIELD_NUMBER: builtins.int
1468
- DICTIONARIES_FIELD_NUMBER: builtins.int
1469
- GRAPHITE_ROLLUP_FIELD_NUMBER: builtins.int
1470
- KAFKA_FIELD_NUMBER: builtins.int
1471
- KAFKA_TOPICS_FIELD_NUMBER: builtins.int
1472
- RABBITMQ_FIELD_NUMBER: builtins.int
1473
- MAX_CONNECTIONS_FIELD_NUMBER: builtins.int
1474
- MAX_CONCURRENT_QUERIES_FIELD_NUMBER: builtins.int
1475
- KEEP_ALIVE_TIMEOUT_FIELD_NUMBER: builtins.int
1476
- UNCOMPRESSED_CACHE_SIZE_FIELD_NUMBER: builtins.int
1477
- MARK_CACHE_SIZE_FIELD_NUMBER: builtins.int
1478
- MAX_TABLE_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1479
- MAX_PARTITION_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1480
- BUILTIN_DICTIONARIES_RELOAD_INTERVAL_FIELD_NUMBER: builtins.int
1481
- TIMEZONE_FIELD_NUMBER: builtins.int
1482
- GEOBASE_ENABLED_FIELD_NUMBER: builtins.int
1483
- GEOBASE_URI_FIELD_NUMBER: builtins.int
1484
1815
  QUERY_LOG_RETENTION_SIZE_FIELD_NUMBER: builtins.int
1485
1816
  QUERY_LOG_RETENTION_TIME_FIELD_NUMBER: builtins.int
1486
1817
  QUERY_THREAD_LOG_ENABLED_FIELD_NUMBER: builtins.int
@@ -1523,312 +1854,525 @@ class ClickhouseConfig(google.protobuf.message.Message):
1523
1854
  ERROR_LOG_RETENTION_SIZE_FIELD_NUMBER: builtins.int
1524
1855
  ERROR_LOG_RETENTION_TIME_FIELD_NUMBER: builtins.int
1525
1856
  ACCESS_CONTROL_IMPROVEMENTS_FIELD_NUMBER: builtins.int
1526
- BACKGROUND_POOL_SIZE_FIELD_NUMBER: builtins.int
1527
- BACKGROUND_MERGES_MUTATIONS_CONCURRENCY_RATIO_FIELD_NUMBER: builtins.int
1528
- BACKGROUND_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1529
- BACKGROUND_FETCHES_POOL_SIZE_FIELD_NUMBER: builtins.int
1530
- BACKGROUND_MOVE_POOL_SIZE_FIELD_NUMBER: builtins.int
1531
- BACKGROUND_DISTRIBUTED_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1532
- BACKGROUND_BUFFER_FLUSH_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1533
- BACKGROUND_MESSAGE_BROKER_SCHEDULE_POOL_SIZE_FIELD_NUMBER: builtins.int
1534
- BACKGROUND_COMMON_POOL_SIZE_FIELD_NUMBER: builtins.int
1857
+ MAX_CONNECTIONS_FIELD_NUMBER: builtins.int
1858
+ MAX_CONCURRENT_QUERIES_FIELD_NUMBER: builtins.int
1859
+ MAX_TABLE_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1860
+ MAX_PARTITION_SIZE_TO_DROP_FIELD_NUMBER: builtins.int
1861
+ KEEP_ALIVE_TIMEOUT_FIELD_NUMBER: builtins.int
1862
+ UNCOMPRESSED_CACHE_SIZE_FIELD_NUMBER: builtins.int
1863
+ MARK_CACHE_SIZE_FIELD_NUMBER: builtins.int
1864
+ TIMEZONE_FIELD_NUMBER: builtins.int
1865
+ GEOBASE_ENABLED_FIELD_NUMBER: builtins.int
1866
+ GEOBASE_URI_FIELD_NUMBER: builtins.int
1535
1867
  DEFAULT_DATABASE_FIELD_NUMBER: builtins.int
1536
1868
  TOTAL_MEMORY_PROFILER_STEP_FIELD_NUMBER: builtins.int
1537
1869
  TOTAL_MEMORY_TRACKER_SAMPLE_PROBABILITY_FIELD_NUMBER: builtins.int
1870
+ ASYNC_INSERT_THREADS_FIELD_NUMBER: builtins.int
1871
+ BACKUP_THREADS_FIELD_NUMBER: builtins.int
1872
+ RESTORE_THREADS_FIELD_NUMBER: builtins.int
1873
+ MERGE_TREE_FIELD_NUMBER: builtins.int
1874
+ COMPRESSION_FIELD_NUMBER: builtins.int
1875
+ DICTIONARIES_FIELD_NUMBER: builtins.int
1876
+ GRAPHITE_ROLLUP_FIELD_NUMBER: builtins.int
1877
+ KAFKA_FIELD_NUMBER: builtins.int
1878
+ KAFKA_TOPICS_FIELD_NUMBER: builtins.int
1879
+ RABBITMQ_FIELD_NUMBER: builtins.int
1538
1880
  QUERY_MASKING_RULES_FIELD_NUMBER: builtins.int
1539
- DICTIONARIES_LAZY_LOAD_FIELD_NUMBER: builtins.int
1540
1881
  QUERY_CACHE_FIELD_NUMBER: builtins.int
1541
1882
  JDBC_BRIDGE_FIELD_NUMBER: builtins.int
1883
+ MYSQL_PROTOCOL_FIELD_NUMBER: builtins.int
1884
+ CUSTOM_MACROS_FIELD_NUMBER: builtins.int
1885
+ BUILTIN_DICTIONARIES_RELOAD_INTERVAL_FIELD_NUMBER: builtins.int
1542
1886
  log_level: global___ClickhouseConfig.LogLevel.ValueType
1543
- """Logging level for the ClickHouse cluster. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR."""
1887
+ """Logging level."""
1888
+ text_log_level: global___ClickhouseConfig.LogLevel.ValueType
1889
+ """Logging level for text_log system table.
1890
+
1891
+ Default value: **TRACE**.
1892
+
1893
+ Change of the setting is applied with restart.
1894
+ """
1544
1895
  timezone: builtins.str
1545
- """The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier."""
1896
+ """The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier.
1897
+
1898
+ Default value: **Europe/Moscow**.
1899
+
1900
+ Change of the setting is applied with restart.
1901
+
1902
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#timezone).
1903
+ """
1546
1904
  geobase_uri: builtins.str
1547
- """Address of the archive with the user geobase in Object Storage."""
1548
- text_log_level: global___ClickhouseConfig.LogLevel.ValueType
1549
- """Logging level for text_log system table. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR."""
1905
+ """Address of the archive with the user geobase in Object Storage.
1906
+
1907
+ Change of the setting is applied with restart.
1908
+ """
1550
1909
  @property
1551
- def merge_tree(self) -> global___ClickhouseConfig.MergeTree:
1552
- """Settings for the MergeTree engine.
1553
- See description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/#merge_tree).
1910
+ def background_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1911
+ """Sets the number of threads performing background merges and mutations for MergeTree-engine tables.
1912
+
1913
+ Default value: **16**.
1914
+
1915
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1916
+
1917
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_pool_size).
1554
1918
  """
1555
1919
 
1556
1920
  @property
1557
- def compression(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.Compression]:
1558
- """Compression settings for the ClickHouse cluster.
1559
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/#compression).
1921
+ def background_merges_mutations_concurrency_ratio(self) -> google.protobuf.wrappers_pb2.Int64Value:
1922
+ """Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently.
1923
+
1924
+ For example, if the ratio equals to **2** and **background_pool_size** is set to **16** then ClickHouse can execute **32** background merges concurrently.
1925
+ This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority.
1926
+
1927
+ Default value: **2**.
1928
+
1929
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1930
+
1931
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_merges_mutations_concurrency_ratio).
1560
1932
  """
1561
1933
 
1562
1934
  @property
1563
- def dictionaries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary]:
1564
- """Configuration of external dictionaries to be used by the ClickHouse cluster.
1565
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/query_language/dicts/external_dicts/).
1935
+ def background_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1936
+ """The maximum number of threads that will be used for constantly executing some lightweight periodic operations
1937
+ for replicated tables, Kafka streaming, and DNS cache updates.
1938
+
1939
+ Default value: **512**.
1940
+
1941
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1942
+
1943
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_schedule_pool_size).
1566
1944
  """
1567
1945
 
1568
1946
  @property
1569
- def graphite_rollup(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup]:
1570
- """Settings for thinning Graphite data.
1571
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server_settings/settings/#server_settings-graphite_rollup).
1947
+ def background_fetches_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1948
+ """The maximum number of threads that will be used for fetching data parts from another replica for MergeTree-engine tables in a background.
1949
+
1950
+ Default value: **32** for versions 25.1 and higher, **16** for versions 24.12 and lower.
1951
+
1952
+ Change of the setting is applied with restart.
1953
+
1954
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_fetches_pool_size).
1572
1955
  """
1573
1956
 
1574
1957
  @property
1575
- def kafka(self) -> global___ClickhouseConfig.Kafka: ...
1576
- @property
1577
- def kafka_topics(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.KafkaTopic]: ...
1578
- @property
1579
- def rabbitmq(self) -> global___ClickhouseConfig.Rabbitmq: ...
1580
- @property
1581
- def max_connections(self) -> google.protobuf.wrappers_pb2.Int64Value:
1582
- """Maximum number of inbound connections."""
1958
+ def background_move_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1959
+ """The maximum number of threads that will be used for moving data parts to another disk or volume for MergeTree-engine tables in a background.
1583
1960
 
1584
- @property
1585
- def max_concurrent_queries(self) -> google.protobuf.wrappers_pb2.Int64Value:
1586
- """Maximum number of simultaneously processed requests."""
1961
+ Default value: **8**.
1962
+
1963
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1964
+
1965
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_move_pool_size).
1966
+ """
1587
1967
 
1588
1968
  @property
1589
- def keep_alive_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
1590
- """Number of milliseconds that ClickHouse waits for incoming requests before closing the connection."""
1969
+ def background_distributed_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1970
+ """The maximum number of threads that will be used for executing distributed sends.
1971
+
1972
+ Default value: **16**.
1973
+
1974
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1975
+
1976
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_distributed_schedule_pool_size).
1977
+ """
1591
1978
 
1592
1979
  @property
1593
- def uncompressed_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1594
- """Cache size (in bytes) for uncompressed data used by MergeTree tables."""
1980
+ def background_buffer_flush_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1981
+ """The maximum number of threads that will be used for performing flush operations for Buffer-engine tables in the background.
1982
+
1983
+ Default value: **16**.
1984
+
1985
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1986
+
1987
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_buffer_flush_schedule_pool_size).
1988
+ """
1595
1989
 
1596
1990
  @property
1597
- def mark_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1598
- """Approximate size (in bytes) of the cache of "marks" used by MergeTree tables."""
1991
+ def background_message_broker_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1992
+ """The maximum number of threads that will be used for executing background operations for message streaming.
1993
+
1994
+ Default value: **16**.
1995
+
1996
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
1997
+
1998
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_message_broker_schedule_pool_size).
1999
+ """
1599
2000
 
1600
2001
  @property
1601
- def max_table_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
1602
- """Maximum size of the table that can be deleted using a DROP query."""
2002
+ def background_common_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
2003
+ """The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for MergeTree-engine tables in a background.
2004
+
2005
+ Default value: **8**.
2006
+
2007
+ Change of the setting is applied with restart on value decrease and without restart on value increase.
2008
+
2009
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#background_common_pool_size).
2010
+ """
1603
2011
 
1604
2012
  @property
1605
- def max_partition_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
1606
- """Maximum size of the partition that can be deleted using a DROP query."""
2013
+ def dictionaries_lazy_load(self) -> google.protobuf.wrappers_pb2.BoolValue:
2014
+ """Lazy loading of dictionaries. If enabled, then each dictionary is loaded on the first use. Otherwise, the server loads all dictionaries at startup.
1607
2015
 
1608
- @property
1609
- def builtin_dictionaries_reload_interval(self) -> google.protobuf.wrappers_pb2.Int64Value:
1610
- """The setting is deprecated and has no effect."""
2016
+ Default value: **true** for versions 25.1 and higher, **false** for versions 24.12 and lower.
1611
2017
 
1612
- @property
1613
- def geobase_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1614
- """Enable or disable geobase."""
2018
+ Change of the setting is applied with restart.
2019
+
2020
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#dictionaries_lazy_load).
2021
+ """
1615
2022
 
1616
2023
  @property
1617
2024
  def query_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1618
- """The maximum size that query_log can grow to before old data will be removed. If set to 0, automatic removal of
1619
- query_log data based on size is disabled.
2025
+ """The maximum size that query_log can grow to before old data will be removed. If set to **0**,
2026
+ automatic removal of query_log data based on size is disabled.
2027
+
2028
+ Default value: **1073741824** (1 GiB).
1620
2029
  """
1621
2030
 
1622
2031
  @property
1623
2032
  def query_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1624
- """The maximum time that query_log records will be retained before removal. If set to 0, automatic removal of
1625
- query_log data based on time is disabled.
2033
+ """The maximum time that query_log records will be retained before removal. If set to **0**, automatic removal of query_log data based on time is disabled.
2034
+
2035
+ Default value: **2592000000** (30 days).
1626
2036
  """
1627
2037
 
1628
2038
  @property
1629
2039
  def query_thread_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1630
- """Whether query_thread_log system table is enabled."""
2040
+ """Enables or disables query_thread_log system table.
2041
+
2042
+ Default value: **true**.
2043
+
2044
+ Change of the setting is applied with restart.
2045
+
2046
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/query_thread_log).
2047
+ """
1631
2048
 
1632
2049
  @property
1633
2050
  def query_thread_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1634
- """The maximum size that query_thread_log can grow to before old data will be removed. If set to 0, automatic removal of
1635
- query_thread_log data based on size is disabled.
2051
+ """The maximum size that query_thread_log can grow to before old data will be removed. If set to **0**,
2052
+ automatic removal of query_thread_log data based on size is disabled.
2053
+
2054
+ Default value: **536870912** (512 MiB).
1636
2055
  """
1637
2056
 
1638
2057
  @property
1639
2058
  def query_thread_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1640
- """The maximum time that query_thread_log records will be retained before removal. If set to 0, automatic removal of
1641
- query_thread_log data based on time is disabled.
2059
+ """The maximum time that query_thread_log records will be retained before removal. If set to **0**,
2060
+ automatic removal of query_thread_log data based on time is disabled.
2061
+
2062
+ Default value: **2592000000** (30 days).
1642
2063
  """
1643
2064
 
1644
2065
  @property
1645
2066
  def part_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1646
- """The maximum size that part_log can grow to before old data will be removed. If set to 0, automatic removal of
1647
- part_log data based on size is disabled.
2067
+ """The maximum size that part_log can grow to before old data will be removed. If set to **0**,
2068
+ automatic removal of part_log data based on size is disabled.
2069
+
2070
+ Default value: **536870912** (512 MiB).
1648
2071
  """
1649
2072
 
1650
2073
  @property
1651
2074
  def part_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1652
- """The maximum time that part_log records will be retained before removal. If set to 0, automatic removal of
1653
- part_log data based on time is disabled.
2075
+ """The maximum time that part_log records will be retained before removal. If set to **0**,
2076
+ automatic removal of part_log data based on time is disabled.
2077
+
2078
+ Default value: **2592000000** (30 days).
1654
2079
  """
1655
2080
 
1656
2081
  @property
1657
2082
  def metric_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1658
- """Whether metric_log system table is enabled."""
2083
+ """Enables or disables metric_log system table.
2084
+
2085
+ Default value: **false** for versions 25.1 and higher, **true** for versions 24.12 and lower.
2086
+
2087
+ Change of the setting is applied with restart.
2088
+
2089
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/metric_log).
2090
+ """
1659
2091
 
1660
2092
  @property
1661
2093
  def metric_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1662
- """The maximum size that metric_log can grow to before old data will be removed. If set to 0, automatic removal of
1663
- metric_log data based on size is disabled.
2094
+ """The maximum size that metric_log can grow to before old data will be removed. If set to **0**,
2095
+ automatic removal of metric_log data based on size is disabled.
2096
+
2097
+ Default value: **536870912** (512 MiB).
1664
2098
  """
1665
2099
 
1666
2100
  @property
1667
2101
  def metric_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1668
- """The maximum time that metric_log records will be retained before removal. If set to 0, automatic removal of
1669
- metric_log data based on time is disabled.
2102
+ """The maximum time that metric_log records will be retained before removal. If set to **0**,
2103
+ automatic removal of metric_log data based on time is disabled.
2104
+
2105
+ Default value: **2592000000** (30 days).
1670
2106
  """
1671
2107
 
1672
2108
  @property
1673
2109
  def trace_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1674
- """Whether trace_log system table is enabled."""
2110
+ """Enables or disables trace_log system table.
2111
+
2112
+ Default value: **true** for versions 25.2 and higher, **false** for versions 25.1 and lower.
2113
+
2114
+ Change of the setting is applied with restart.
2115
+
2116
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/trace_log).
2117
+ """
1675
2118
 
1676
2119
  @property
1677
2120
  def trace_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1678
- """The maximum size that trace_log can grow to before old data will be removed. If set to 0, automatic removal of
1679
- trace_log data based on size is disabled.
2121
+ """The maximum size that trace_log can grow to before old data will be removed. If set to **0**,
2122
+ automatic removal of trace_log data based on size is disabled.
2123
+
2124
+ Default value: **536870912** (512 MiB).
1680
2125
  """
1681
2126
 
1682
2127
  @property
1683
2128
  def trace_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1684
- """The maximum time that trace_log records will be retained before removal. If set to 0, automatic removal of
1685
- trace_log data based on time is disabled.
2129
+ """The maximum time that trace_log records will be retained before removal. If set to **0**,
2130
+ automatic removal of trace_log data based on time is disabled.
2131
+
2132
+ Default value: **2592000000** (30 days).
1686
2133
  """
1687
2134
 
1688
2135
  @property
1689
2136
  def text_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1690
- """Whether text_log system table is enabled."""
2137
+ """Enables or disables text_log system table.
2138
+
2139
+ Default value: **false**.
2140
+
2141
+ Change of the setting is applied with restart.
2142
+
2143
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/text_log).
2144
+ """
1691
2145
 
1692
2146
  @property
1693
2147
  def text_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1694
- """The maximum size that text_log can grow to before old data will be removed. If set to 0, automatic removal of
1695
- text_log data based on size is disabled.
2148
+ """The maximum size that text_log can grow to before old data will be removed. If set to **0**,
2149
+ automatic removal of text_log data based on size is disabled.
2150
+
2151
+ Default value: **536870912** (512 MiB).
1696
2152
  """
1697
2153
 
1698
2154
  @property
1699
2155
  def text_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1700
- """The maximum time that text_log records will be retained before removal. If set to 0, automatic removal of
1701
- text_log data based on time is disabled.
2156
+ """The maximum time that text_log records will be retained before removal. If set to **0**,
2157
+ automatic removal of text_log data based on time is disabled.
2158
+
2159
+ Default value: **2592000000** (30 days).
1702
2160
  """
1703
2161
 
1704
2162
  @property
1705
2163
  def opentelemetry_span_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1706
- """Enable or disable opentelemetry_span_log system table. Default value: false."""
2164
+ """Enables or disables opentelemetry_span_log system table.
2165
+
2166
+ Default value: **false**.
2167
+
2168
+ Change of the setting is applied with restart.
2169
+
2170
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/opentelemetry_span_log).
2171
+ """
1707
2172
 
1708
2173
  @property
1709
2174
  def opentelemetry_span_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1710
- """The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to 0 (default),
2175
+ """The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to **0**,
1711
2176
  automatic removal of opentelemetry_span_log data based on size is disabled.
2177
+
2178
+ Default value: **0**.
1712
2179
  """
1713
2180
 
1714
2181
  @property
1715
2182
  def opentelemetry_span_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1716
- """The maximum time that opentelemetry_span_log records will be retained before removal. If set to 0,
2183
+ """The maximum time that opentelemetry_span_log records will be retained before removal. If set to **0**,
1717
2184
  automatic removal of opentelemetry_span_log data based on time is disabled.
2185
+
2186
+ Default value: **2592000000** (30 days).
1718
2187
  """
1719
2188
 
1720
2189
  @property
1721
2190
  def query_views_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1722
- """Enable or disable query_views_log system table. Default value: false."""
2191
+ """Enables or disables query_views_log system table.
2192
+
2193
+ Default value: **false**.
2194
+
2195
+ Change of the setting is applied with restart.
2196
+
2197
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/query_views_log).
2198
+ """
1723
2199
 
1724
2200
  @property
1725
2201
  def query_views_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1726
- """The maximum size that query_views_log can grow to before old data will be removed. If set to 0 (default),
2202
+ """The maximum size that query_views_log can grow to before old data will be removed. If set to **0**,
1727
2203
  automatic removal of query_views_log data based on size is disabled.
2204
+
2205
+ Default value: **0**.
1728
2206
  """
1729
2207
 
1730
2208
  @property
1731
2209
  def query_views_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1732
- """The maximum time that query_views_log records will be retained before removal. If set to 0,
2210
+ """The maximum time that query_views_log records will be retained before removal. If set to **0**,
1733
2211
  automatic removal of query_views_log data based on time is disabled.
2212
+
2213
+ Default value: **2592000000** (30 days).
1734
2214
  """
1735
2215
 
1736
2216
  @property
1737
2217
  def asynchronous_metric_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1738
- """Enable or disable asynchronous_metric_log system table. Default value: false."""
2218
+ """Enables or disables asynchronous_metric_log system table.
2219
+
2220
+ Default value: **false**.
2221
+
2222
+ Change of the setting is applied with restart.
2223
+
2224
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/asynchronous_metric_log).
2225
+ """
1739
2226
 
1740
2227
  @property
1741
2228
  def asynchronous_metric_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1742
- """The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to 0 (default),
2229
+ """The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to **0**,
1743
2230
  automatic removal of asynchronous_metric_log data based on size is disabled.
2231
+
2232
+ Default value: **0**.
1744
2233
  """
1745
2234
 
1746
2235
  @property
1747
2236
  def asynchronous_metric_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1748
- """The maximum time that asynchronous_metric_log records will be retained before removal. If set to 0,
2237
+ """The maximum time that asynchronous_metric_log records will be retained before removal. If set to **0**,
1749
2238
  automatic removal of asynchronous_metric_log data based on time is disabled.
2239
+
2240
+ Default value: **2592000000** (30 days).
1750
2241
  """
1751
2242
 
1752
2243
  @property
1753
2244
  def session_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1754
- """Enable or disable session_log system table. Default value: false."""
2245
+ """Enables or disables session_log system table.
2246
+
2247
+ Default value: **false**.
2248
+
2249
+ Change of the setting is applied with restart.
2250
+
2251
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/session_log).
2252
+ """
1755
2253
 
1756
2254
  @property
1757
2255
  def session_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1758
- """The maximum size that session_log can grow to before old data will be removed. If set to 0 (default),
2256
+ """The maximum size that session_log can grow to before old data will be removed. If set to **0**,
1759
2257
  automatic removal of session_log data based on size is disabled.
2258
+
2259
+ Default value: **0**.
1760
2260
  """
1761
2261
 
1762
2262
  @property
1763
2263
  def session_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1764
- """The maximum time that session_log records will be retained before removal. If set to 0,
2264
+ """The maximum time that session_log records will be retained before removal. If set to **0**,
1765
2265
  automatic removal of session_log data based on time is disabled.
2266
+
2267
+ Default value: **2592000000** (30 days).
1766
2268
  """
1767
2269
 
1768
2270
  @property
1769
2271
  def zookeeper_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1770
- """Enable or disable zookeeper_log system table. Default value: false."""
2272
+ """Enables or disables zookeeper_log system table.
2273
+
2274
+ Default value: **false**.
2275
+
2276
+ Change of the setting is applied with restart.
2277
+
2278
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/zookeeper_log).
2279
+ """
1771
2280
 
1772
2281
  @property
1773
2282
  def zookeeper_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1774
- """The maximum size that zookeeper_log can grow to before old data will be removed. If set to 0 (default),
2283
+ """The maximum size that zookeeper_log can grow to before old data will be removed. If set to **0**,
1775
2284
  automatic removal of zookeeper_log data based on size is disabled.
2285
+
2286
+ Default value: **0**.
1776
2287
  """
1777
2288
 
1778
2289
  @property
1779
2290
  def zookeeper_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1780
- """The maximum time that zookeeper_log records will be retained before removal. If set to 0,
2291
+ """The maximum time that zookeeper_log records will be retained before removal. If set to **0**,
1781
2292
  automatic removal of zookeeper_log data based on time is disabled.
2293
+
2294
+ Default value: **2592000000** (30 days).
1782
2295
  """
1783
2296
 
1784
2297
  @property
1785
2298
  def asynchronous_insert_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1786
- """Enable or disable asynchronous_insert_log system table. Default value: false.
1787
- Minimal required ClickHouse version: 22.10.
2299
+ """Enables or disables asynchronous_insert_log system table.
2300
+
2301
+ Default value: **false**.
2302
+
2303
+ Change of the setting is applied with restart.
2304
+
2305
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/asynchronous_insert_log).
1788
2306
  """
1789
2307
 
1790
2308
  @property
1791
2309
  def asynchronous_insert_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1792
- """The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to 0 (default),
2310
+ """The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to **0**,
1793
2311
  automatic removal of asynchronous_insert_log data based on size is disabled.
2312
+
2313
+ Default value: **0**.
1794
2314
  """
1795
2315
 
1796
2316
  @property
1797
2317
  def asynchronous_insert_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1798
- """The maximum time that asynchronous_insert_log records will be retained before removal. If set to 0,
2318
+ """The maximum time that asynchronous_insert_log records will be retained before removal. If set to **0**,
1799
2319
  automatic removal of asynchronous_insert_log data based on time is disabled.
2320
+
2321
+ Default value: **2592000000** (30 days).
1800
2322
  """
1801
2323
 
1802
2324
  @property
1803
2325
  def processors_profile_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1804
- """Enable or disable processors_profile_log system table."""
2326
+ """Enables or disables processors_profile_log system table.
2327
+
2328
+ Default value: **true** for versions 25.2 and higher, **false** for versions 25.1 and lower.
2329
+
2330
+ Change of the setting is applied with restart.
2331
+
2332
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/processors_profile_log).
2333
+ """
1805
2334
 
1806
2335
  @property
1807
2336
  def processors_profile_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1808
- """The maximum size that processors_profile_log can grow to before old data will be removed.
1809
- If set to 0 (default), automatic removal of processors_profile_log data based on size is disabled.
2337
+ """The maximum size that processors_profile_log can grow to before old data will be removed. If set to **0**,
2338
+ automatic removal of processors_profile_log data based on size is disabled.
2339
+
2340
+ Default value: **0**.
1810
2341
  """
1811
2342
 
1812
2343
  @property
1813
2344
  def processors_profile_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1814
- """The maximum time that processors_profile_log records will be retained before removal.
1815
- If set to 0, automatic removal of processors_profile_log data based on time is disabled.
2345
+ """The maximum time that processors_profile_log records will be retained before removal. If set to **0**,
2346
+ automatic removal of processors_profile_log data based on time is disabled.
2347
+
2348
+ Default value: **2592000000** (30 days).
1816
2349
  """
1817
2350
 
1818
2351
  @property
1819
2352
  def error_log_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
1820
- """Enable or disable error_log system table."""
2353
+ """Enables or disables error_log system table.
2354
+
2355
+ Default value: **false**.
2356
+
2357
+ Change of the setting is applied with restart.
2358
+
2359
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/system-tables/error_log).
2360
+ """
1821
2361
 
1822
2362
  @property
1823
2363
  def error_log_retention_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1824
- """The maximum size that error_log can grow to before old data will be removed.
1825
- If set to 0 (default), automatic removal of error_log data based on size is disabled.
2364
+ """The maximum size that error_log can grow to before old data will be removed. If set to **0**,
2365
+ automatic removal of error_log data based on size is disabled.
2366
+
2367
+ Default value: **0**.
1826
2368
  """
1827
2369
 
1828
2370
  @property
1829
2371
  def error_log_retention_time(self) -> google.protobuf.wrappers_pb2.Int64Value:
1830
- """The maximum time that error_log records will be retained before removal.
1831
- If set to 0, automatic removal of error_log data based on time is disabled.
2372
+ """The maximum time that error_log records will be retained before removal. If set to **0**,
2373
+ automatic removal of error_log data based on time is disabled.
2374
+
2375
+ Default value: **2592000000** (30 days).
1832
2376
  """
1833
2377
 
1834
2378
  @property
@@ -1836,103 +2380,261 @@ class ClickhouseConfig(google.protobuf.message.Message):
1836
2380
  """Access control settings."""
1837
2381
 
1838
2382
  @property
1839
- def background_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
1840
- @property
1841
- def background_merges_mutations_concurrency_ratio(self) -> google.protobuf.wrappers_pb2.Int64Value:
1842
- """Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently. For example, if the ratio equals to 2 and background_pool_size is set to 16 then ClickHouse can execute 32 background merges concurrently. This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority. You can only increase this ratio at runtime. To lower it you have to restart the server. The same as for background_pool_size setting background_merges_mutations_concurrency_ratio could be applied from the default profile for backward compatibility.
1843
- Default: 2
1844
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#background_merges_mutations_concurrency_ratio)
2383
+ def max_connections(self) -> google.protobuf.wrappers_pb2.Int64Value:
2384
+ """Maximum number of inbound connections.
2385
+
2386
+ Default value: **4096**.
2387
+
2388
+ Change of the setting is applied with restart.
2389
+
2390
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_connections).
1845
2391
  """
1846
2392
 
1847
2393
  @property
1848
- def background_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2394
+ def max_concurrent_queries(self) -> google.protobuf.wrappers_pb2.Int64Value:
2395
+ """Maximum number of concurrently executed queries.
2396
+
2397
+ Default value: **500**.
2398
+
2399
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_concurrent_queries).
2400
+ """
2401
+
1849
2402
  @property
1850
- def background_fetches_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1851
- """Sets the number of threads performing background fetches for tables with **ReplicatedMergeTree** engines. Default value: 8.
2403
+ def max_table_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
2404
+ """Maximum size of the table that can be deleted using **DROP** or **TRUNCATE** query.
1852
2405
 
1853
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings/#background_fetches_pool_size).
2406
+ Default value: **50000000000** (48828125 KiB).
2407
+
2408
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_table_size_to_drop).
1854
2409
  """
1855
2410
 
1856
2411
  @property
1857
- def background_move_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2412
+ def max_partition_size_to_drop(self) -> google.protobuf.wrappers_pb2.Int64Value:
2413
+ """Maximum size of the partition that can be deleted using **DROP** or **TRUNCATE** query.
2414
+
2415
+ Default value: **50000000000** (48828125 KiB).
2416
+
2417
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#max_partition_size_to_drop).
2418
+ """
2419
+
1858
2420
  @property
1859
- def background_distributed_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2421
+ def keep_alive_timeout(self) -> google.protobuf.wrappers_pb2.Int64Value:
2422
+ """The number of seconds that ClickHouse waits for incoming requests for HTTP protocol before closing the connection.
2423
+
2424
+ Default value: **30**.
2425
+
2426
+ Change of the setting is applied with restart.
2427
+
2428
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#keep_alive_timeout).
2429
+ """
2430
+
1860
2431
  @property
1861
- def background_buffer_flush_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2432
+ def uncompressed_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
2433
+ """Cache size (in bytes) for uncompressed data used by table engines from the MergeTree family. **0** means disabled.
2434
+
2435
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#uncompressed_cache_size).
2436
+ """
2437
+
1862
2438
  @property
1863
- def background_message_broker_schedule_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value: ...
2439
+ def mark_cache_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
2440
+ """Maximum size (in bytes) of the cache of "marks" used by MergeTree tables.
2441
+
2442
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#mark_cache_size).
2443
+ """
2444
+
1864
2445
  @property
1865
- def background_common_pool_size(self) -> google.protobuf.wrappers_pb2.Int64Value:
1866
- """The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for *MergeTree-engine tables in a background.
1867
- Default: 8
1868
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#background_common_pool_size)
2446
+ def geobase_enabled(self) -> google.protobuf.wrappers_pb2.BoolValue:
2447
+ """Enables or disables geobase.
2448
+
2449
+ Change of the setting is applied with restart.
1869
2450
  """
1870
2451
 
1871
2452
  @property
1872
2453
  def default_database(self) -> google.protobuf.wrappers_pb2.StringValue:
1873
2454
  """The default database.
1874
2455
 
1875
- To get a list of cluster databases, see [Yandex Managed ClickHouse documentation](/docs/managed-clickhouse/operations/databases#list-db).
2456
+ Default value: **default**.
2457
+
2458
+ Change of the setting is applied with restart.
2459
+
2460
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#default_database).
1876
2461
  """
1877
2462
 
1878
2463
  @property
1879
2464
  def total_memory_profiler_step(self) -> google.protobuf.wrappers_pb2.Int64Value:
1880
- """Sets the memory size (in bytes) for a stack trace at every peak allocation step. Default value: **4194304**.
2465
+ """Whenever server memory usage becomes larger than every next step in number of bytes the memory profiler will collect
2466
+ the allocating stack trace. **0** means disabled memory profiler.
2467
+
2468
+ Default value: **0**.
2469
+
2470
+ Change of the setting is applied with restart.
2471
+
2472
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#total_memory_profiler_step).
2473
+ """
2474
+
2475
+ @property
2476
+ def total_memory_tracker_sample_probability(self) -> google.protobuf.wrappers_pb2.DoubleValue:
2477
+ """Allows to collect random allocations and de-allocations and writes them in the system.trace_log system table
2478
+ with trace_type equal to a MemorySample with the specified probability.
2479
+
2480
+ Default value: **0**.
1881
2481
 
1882
- More info see in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings/#total-memory-profiler-step).
2482
+ Change of the setting is applied with restart.
2483
+
2484
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#total_memory_tracker_sample_probability).
1883
2485
  """
1884
2486
 
1885
2487
  @property
1886
- def total_memory_tracker_sample_probability(self) -> google.protobuf.wrappers_pb2.DoubleValue: ...
2488
+ def async_insert_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
2489
+ """Maximum number of threads to parse and insert data in background. If set to **0**, asynchronous mode is disabled.
2490
+
2491
+ Default value: **16**.
2492
+
2493
+ Change of the setting is applied with restart.
2494
+
2495
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#async_insert_threads).
2496
+ """
2497
+
1887
2498
  @property
1888
- def query_masking_rules(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.QueryMaskingRule]:
1889
- """Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs, system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs.
1890
- Change of these settings is applied with ClickHouse restart
1891
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#query-masking-rules)
2499
+ def backup_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
2500
+ """The maximum number of threads to execute **BACKUP** requests.
2501
+
2502
+ Default value: **16**.
2503
+
2504
+ Change of the setting is applied with restart.
2505
+
2506
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#backup_threads).
1892
2507
  """
1893
2508
 
1894
2509
  @property
1895
- def dictionaries_lazy_load(self) -> google.protobuf.wrappers_pb2.BoolValue:
1896
- """Lazy loading of dictionaries.
1897
- Default: true
1898
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#dictionaries_lazy_load)
2510
+ def restore_threads(self) -> google.protobuf.wrappers_pb2.Int64Value:
2511
+ """The maximum number of threads to execute **RESTORE** requests.
2512
+
2513
+ Default value: **16**.
2514
+
2515
+ Change of the setting is applied with restart.
2516
+
2517
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#restore_threads).
2518
+ """
2519
+
2520
+ @property
2521
+ def merge_tree(self) -> global___ClickhouseConfig.MergeTree:
2522
+ """Settings for the MergeTree table engine family.
2523
+
2524
+ Change of the settings of **merge_tree** is applied with restart.
2525
+ """
2526
+
2527
+ @property
2528
+ def compression(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.Compression]:
2529
+ """Data compression settings for MergeTree engine tables.
2530
+
2531
+ Change of the settings of **compression** is applied with restart.
2532
+
2533
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#compression).
2534
+ """
2535
+
2536
+ @property
2537
+ def dictionaries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.ExternalDictionary]:
2538
+ """Configuration of external dictionaries.
2539
+
2540
+ Change of the settings of **dictionaries** is applied with restart.
2541
+
2542
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/sql-reference/dictionaries).
2543
+ """
2544
+
2545
+ @property
2546
+ def graphite_rollup(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.GraphiteRollup]:
2547
+ """Rollup settings for the GraphiteMergeTree engine tables.
2548
+
2549
+ Change of the settings of **graphite_rollup** is applied with restart.
2550
+
2551
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#graphite_rollup).
2552
+ """
2553
+
2554
+ @property
2555
+ def kafka(self) -> global___ClickhouseConfig.Kafka:
2556
+ """Kafka integration settings.
2557
+
2558
+ Change of the settings of **kafka** is applied with restart.
2559
+ """
2560
+
2561
+ @property
2562
+ def kafka_topics(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.KafkaTopic]:
2563
+ """Per-topic Kafka integration settings.
2564
+
2565
+ Change of the settings of **kafka_topics** is applied with restart.
2566
+ """
2567
+
2568
+ @property
2569
+ def rabbitmq(self) -> global___ClickhouseConfig.Rabbitmq:
2570
+ """RabbitMQ integration settings.
2571
+
2572
+ Change of the settings of **rabbitmq** is applied with restart.
2573
+ """
2574
+
2575
+ @property
2576
+ def query_masking_rules(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.QueryMaskingRule]:
2577
+ """Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs,
2578
+ system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing
2579
+ sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs.
2580
+
2581
+ Change of the settings of **query_masking_rules** is applied with restart.
2582
+
2583
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#query_masking_rules).
1899
2584
  """
1900
2585
 
1901
2586
  @property
1902
2587
  def query_cache(self) -> global___ClickhouseConfig.QueryCache:
1903
- """[Query cache](https://clickhouse.com/docs/en/operations/query-cache) configuration.
1904
- Min version: 23.5
1905
- See in-depth description in [ClickHouse documentation](https://clickhouse.com/docs/en/operations/server-configuration-parameters/settings#query_cache)
2588
+ """[Query cache](https://clickhouse.com/docs/operations/query-cache) configuration.
2589
+
2590
+ Change of the settings of **query_cache** is applied with restart.
1906
2591
  """
1907
2592
 
1908
2593
  @property
1909
2594
  def jdbc_bridge(self) -> global___ClickhouseConfig.JdbcBridge:
1910
- """JDBC bridge for queries to external databases.
1911
- https://clickhouse.com/docs/en/integrations/jdbc/jdbc-with-clickhouse
2595
+ """JDBC bridge configuration for queries to external databases.
2596
+
2597
+ Change of the settings of **jdbc_bridge** is applied with restart.
2598
+
2599
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/en/integrations/jdbc/jdbc-with-clickhouse).
2600
+ """
2601
+
2602
+ @property
2603
+ def mysql_protocol(self) -> google.protobuf.wrappers_pb2.BoolValue:
2604
+ """Enables or disables MySQL interface on ClickHouse server
2605
+
2606
+ Default value: **false**.
2607
+
2608
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/interfaces/mysql).
2609
+ """
2610
+
2611
+ @property
2612
+ def custom_macros(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ClickhouseConfig.Macro]:
2613
+ """Custom ClickHouse macros."""
2614
+
2615
+ @property
2616
+ def builtin_dictionaries_reload_interval(self) -> google.protobuf.wrappers_pb2.Int64Value:
2617
+ """The interval in seconds before reloading built-in dictionaries.
2618
+
2619
+ Default value: **3600**.
2620
+
2621
+ For details, see [ClickHouse documentation](https://clickhouse.com/docs/operations/server-configuration-parameters/settings#builtin_dictionaries_reload_interval).
1912
2622
  """
1913
2623
 
1914
2624
  def __init__(
1915
2625
  self,
1916
2626
  *,
2627
+ background_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2628
+ background_merges_mutations_concurrency_ratio: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2629
+ background_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2630
+ background_fetches_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2631
+ background_move_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2632
+ background_distributed_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2633
+ background_buffer_flush_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2634
+ background_message_broker_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2635
+ background_common_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2636
+ dictionaries_lazy_load: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1917
2637
  log_level: global___ClickhouseConfig.LogLevel.ValueType = ...,
1918
- merge_tree: global___ClickhouseConfig.MergeTree | None = ...,
1919
- compression: collections.abc.Iterable[global___ClickhouseConfig.Compression] | None = ...,
1920
- dictionaries: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary] | None = ...,
1921
- graphite_rollup: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup] | None = ...,
1922
- kafka: global___ClickhouseConfig.Kafka | None = ...,
1923
- kafka_topics: collections.abc.Iterable[global___ClickhouseConfig.KafkaTopic] | None = ...,
1924
- rabbitmq: global___ClickhouseConfig.Rabbitmq | None = ...,
1925
- max_connections: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1926
- max_concurrent_queries: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1927
- keep_alive_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1928
- uncompressed_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1929
- mark_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1930
- max_table_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1931
- max_partition_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1932
- builtin_dictionaries_reload_interval: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1933
- timezone: builtins.str = ...,
1934
- geobase_enabled: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1935
- geobase_uri: builtins.str = ...,
1936
2638
  query_log_retention_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1937
2639
  query_log_retention_time: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1938
2640
  query_thread_log_enabled: google.protobuf.wrappers_pb2.BoolValue | None = ...,
@@ -1975,25 +2677,38 @@ class ClickhouseConfig(google.protobuf.message.Message):
1975
2677
  error_log_retention_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1976
2678
  error_log_retention_time: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1977
2679
  access_control_improvements: global___ClickhouseConfig.AccessControlImprovements | None = ...,
1978
- background_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1979
- background_merges_mutations_concurrency_ratio: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1980
- background_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1981
- background_fetches_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1982
- background_move_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1983
- background_distributed_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1984
- background_buffer_flush_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1985
- background_message_broker_schedule_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1986
- background_common_pool_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2680
+ max_connections: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2681
+ max_concurrent_queries: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2682
+ max_table_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2683
+ max_partition_size_to_drop: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2684
+ keep_alive_timeout: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2685
+ uncompressed_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2686
+ mark_cache_size: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2687
+ timezone: builtins.str = ...,
2688
+ geobase_enabled: google.protobuf.wrappers_pb2.BoolValue | None = ...,
2689
+ geobase_uri: builtins.str = ...,
1987
2690
  default_database: google.protobuf.wrappers_pb2.StringValue | None = ...,
1988
2691
  total_memory_profiler_step: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1989
2692
  total_memory_tracker_sample_probability: google.protobuf.wrappers_pb2.DoubleValue | None = ...,
2693
+ async_insert_threads: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2694
+ backup_threads: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2695
+ restore_threads: google.protobuf.wrappers_pb2.Int64Value | None = ...,
2696
+ merge_tree: global___ClickhouseConfig.MergeTree | None = ...,
2697
+ compression: collections.abc.Iterable[global___ClickhouseConfig.Compression] | None = ...,
2698
+ dictionaries: collections.abc.Iterable[global___ClickhouseConfig.ExternalDictionary] | None = ...,
2699
+ graphite_rollup: collections.abc.Iterable[global___ClickhouseConfig.GraphiteRollup] | None = ...,
2700
+ kafka: global___ClickhouseConfig.Kafka | None = ...,
2701
+ kafka_topics: collections.abc.Iterable[global___ClickhouseConfig.KafkaTopic] | None = ...,
2702
+ rabbitmq: global___ClickhouseConfig.Rabbitmq | None = ...,
1990
2703
  query_masking_rules: collections.abc.Iterable[global___ClickhouseConfig.QueryMaskingRule] | None = ...,
1991
- dictionaries_lazy_load: google.protobuf.wrappers_pb2.BoolValue | None = ...,
1992
2704
  query_cache: global___ClickhouseConfig.QueryCache | None = ...,
1993
2705
  jdbc_bridge: global___ClickhouseConfig.JdbcBridge | None = ...,
2706
+ mysql_protocol: google.protobuf.wrappers_pb2.BoolValue | None = ...,
2707
+ custom_macros: collections.abc.Iterable[global___ClickhouseConfig.Macro] | None = ...,
2708
+ builtin_dictionaries_reload_interval: google.protobuf.wrappers_pb2.Int64Value | None = ...,
1994
2709
  ) -> None: ...
1995
- def HasField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "default_database", b"default_database", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "keep_alive_timeout", b"keep_alive_timeout", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> builtins.bool: ...
1996
- def ClearField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "compression", b"compression", "default_database", b"default_database", "dictionaries", b"dictionaries", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "geobase_uri", b"geobase_uri", "graphite_rollup", b"graphite_rollup", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "kafka_topics", b"kafka_topics", "keep_alive_timeout", b"keep_alive_timeout", "log_level", b"log_level", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_masking_rules", b"query_masking_rules", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_level", b"text_log_level", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "timezone", b"timezone", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> None: ...
2710
+ def HasField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "async_insert_threads", b"async_insert_threads", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "backup_threads", b"backup_threads", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "default_database", b"default_database", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "keep_alive_timeout", b"keep_alive_timeout", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "mysql_protocol", b"mysql_protocol", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "restore_threads", b"restore_threads", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> builtins.bool: ...
2711
+ def ClearField(self, field_name: typing.Literal["access_control_improvements", b"access_control_improvements", "async_insert_threads", b"async_insert_threads", "asynchronous_insert_log_enabled", b"asynchronous_insert_log_enabled", "asynchronous_insert_log_retention_size", b"asynchronous_insert_log_retention_size", "asynchronous_insert_log_retention_time", b"asynchronous_insert_log_retention_time", "asynchronous_metric_log_enabled", b"asynchronous_metric_log_enabled", "asynchronous_metric_log_retention_size", b"asynchronous_metric_log_retention_size", "asynchronous_metric_log_retention_time", b"asynchronous_metric_log_retention_time", "background_buffer_flush_schedule_pool_size", b"background_buffer_flush_schedule_pool_size", "background_common_pool_size", b"background_common_pool_size", "background_distributed_schedule_pool_size", b"background_distributed_schedule_pool_size", "background_fetches_pool_size", b"background_fetches_pool_size", "background_merges_mutations_concurrency_ratio", b"background_merges_mutations_concurrency_ratio", "background_message_broker_schedule_pool_size", b"background_message_broker_schedule_pool_size", "background_move_pool_size", b"background_move_pool_size", "background_pool_size", b"background_pool_size", "background_schedule_pool_size", b"background_schedule_pool_size", "backup_threads", b"backup_threads", "builtin_dictionaries_reload_interval", b"builtin_dictionaries_reload_interval", "compression", b"compression", "custom_macros", b"custom_macros", "default_database", b"default_database", "dictionaries", b"dictionaries", "dictionaries_lazy_load", b"dictionaries_lazy_load", "error_log_enabled", b"error_log_enabled", "error_log_retention_size", b"error_log_retention_size", "error_log_retention_time", b"error_log_retention_time", "geobase_enabled", b"geobase_enabled", "geobase_uri", b"geobase_uri", "graphite_rollup", b"graphite_rollup", "jdbc_bridge", b"jdbc_bridge", "kafka", b"kafka", "kafka_topics", b"kafka_topics", "keep_alive_timeout", b"keep_alive_timeout", "log_level", b"log_level", "mark_cache_size", b"mark_cache_size", "max_concurrent_queries", b"max_concurrent_queries", "max_connections", b"max_connections", "max_partition_size_to_drop", b"max_partition_size_to_drop", "max_table_size_to_drop", b"max_table_size_to_drop", "merge_tree", b"merge_tree", "metric_log_enabled", b"metric_log_enabled", "metric_log_retention_size", b"metric_log_retention_size", "metric_log_retention_time", b"metric_log_retention_time", "mysql_protocol", b"mysql_protocol", "opentelemetry_span_log_enabled", b"opentelemetry_span_log_enabled", "opentelemetry_span_log_retention_size", b"opentelemetry_span_log_retention_size", "opentelemetry_span_log_retention_time", b"opentelemetry_span_log_retention_time", "part_log_retention_size", b"part_log_retention_size", "part_log_retention_time", b"part_log_retention_time", "processors_profile_log_enabled", b"processors_profile_log_enabled", "processors_profile_log_retention_size", b"processors_profile_log_retention_size", "processors_profile_log_retention_time", b"processors_profile_log_retention_time", "query_cache", b"query_cache", "query_log_retention_size", b"query_log_retention_size", "query_log_retention_time", b"query_log_retention_time", "query_masking_rules", b"query_masking_rules", "query_thread_log_enabled", b"query_thread_log_enabled", "query_thread_log_retention_size", b"query_thread_log_retention_size", "query_thread_log_retention_time", b"query_thread_log_retention_time", "query_views_log_enabled", b"query_views_log_enabled", "query_views_log_retention_size", b"query_views_log_retention_size", "query_views_log_retention_time", b"query_views_log_retention_time", "rabbitmq", b"rabbitmq", "restore_threads", b"restore_threads", "session_log_enabled", b"session_log_enabled", "session_log_retention_size", b"session_log_retention_size", "session_log_retention_time", b"session_log_retention_time", "text_log_enabled", b"text_log_enabled", "text_log_level", b"text_log_level", "text_log_retention_size", b"text_log_retention_size", "text_log_retention_time", b"text_log_retention_time", "timezone", b"timezone", "total_memory_profiler_step", b"total_memory_profiler_step", "total_memory_tracker_sample_probability", b"total_memory_tracker_sample_probability", "trace_log_enabled", b"trace_log_enabled", "trace_log_retention_size", b"trace_log_retention_size", "trace_log_retention_time", b"trace_log_retention_time", "uncompressed_cache_size", b"uncompressed_cache_size", "zookeeper_log_enabled", b"zookeeper_log_enabled", "zookeeper_log_retention_size", b"zookeeper_log_retention_size", "zookeeper_log_retention_time", b"zookeeper_log_retention_time"]) -> None: ...
1997
2712
 
1998
2713
  global___ClickhouseConfig = ClickhouseConfig
1999
2714
 
@@ -2006,17 +2721,15 @@ class ClickhouseConfigSet(google.protobuf.message.Message):
2006
2721
  DEFAULT_CONFIG_FIELD_NUMBER: builtins.int
2007
2722
  @property
2008
2723
  def effective_config(self) -> global___ClickhouseConfig:
2009
- """Effective settings for a ClickHouse cluster (a combination of settings defined
2010
- in [user_config] and [default_config]).
2011
- """
2724
+ """Effective configuration (a combination of user-defined configuration and default configuration)."""
2012
2725
 
2013
2726
  @property
2014
2727
  def user_config(self) -> global___ClickhouseConfig:
2015
- """User-defined settings for a ClickHouse cluster."""
2728
+ """User-defined configuration."""
2016
2729
 
2017
2730
  @property
2018
2731
  def default_config(self) -> global___ClickhouseConfig:
2019
- """Default configuration for a ClickHouse cluster."""
2732
+ """Default configuration."""
2020
2733
 
2021
2734
  def __init__(
2022
2735
  self,