pulumi-aiven 6.13.0a1711361534__py3-none-any.whl → 6.14.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pulumi-aiven might be problematic. Click here for more details.
- pulumi_aiven/_inputs.py +200 -262
- pulumi_aiven/account.py +2 -6
- pulumi_aiven/account_authentication.py +2 -2
- pulumi_aiven/account_team.py +6 -6
- pulumi_aiven/account_team_member.py +6 -10
- pulumi_aiven/account_team_project.py +12 -12
- pulumi_aiven/aws_privatelink.py +18 -22
- pulumi_aiven/aws_vpc_peering_connection.py +8 -12
- pulumi_aiven/azure_privatelink.py +18 -22
- pulumi_aiven/azure_privatelink_connection_approval.py +112 -14
- pulumi_aiven/azure_vpc_peering_connection.py +6 -10
- pulumi_aiven/billing_group.py +18 -18
- pulumi_aiven/cassandra.py +25 -57
- pulumi_aiven/cassandra_user.py +18 -22
- pulumi_aiven/clickhouse.py +25 -57
- pulumi_aiven/clickhouse_database.py +24 -26
- pulumi_aiven/clickhouse_grant.py +35 -37
- pulumi_aiven/clickhouse_role.py +16 -20
- pulumi_aiven/clickhouse_user.py +18 -22
- pulumi_aiven/connection_pool.py +24 -28
- pulumi_aiven/dragonfly.py +124 -263
- pulumi_aiven/flink.py +32 -36
- pulumi_aiven/flink_application.py +22 -24
- pulumi_aiven/flink_application_deployment.py +21 -21
- pulumi_aiven/flink_application_version.py +96 -14
- pulumi_aiven/gcp_privatelink.py +21 -21
- pulumi_aiven/gcp_privatelink_connection_approval.py +21 -21
- pulumi_aiven/gcp_vpc_peering_connection.py +4 -8
- pulumi_aiven/get_account.py +0 -4
- pulumi_aiven/get_aws_privatelink.py +10 -14
- pulumi_aiven/get_aws_vpc_peering_connection.py +6 -10
- pulumi_aiven/get_azure_privatelink.py +10 -14
- pulumi_aiven/get_azure_vpc_peering_connection.py +2 -6
- pulumi_aiven/get_billing_group.py +2 -2
- pulumi_aiven/get_cassanda.py +7 -24
- pulumi_aiven/get_cassandra.py +7 -24
- pulumi_aiven/get_cassandra_user.py +10 -14
- pulumi_aiven/get_clickhouse.py +7 -24
- pulumi_aiven/get_clickhouse_database.py +10 -14
- pulumi_aiven/get_clickhouse_user.py +10 -14
- pulumi_aiven/get_connection_pool.py +10 -14
- pulumi_aiven/get_dragonfly.py +27 -20
- pulumi_aiven/get_flink.py +6 -10
- pulumi_aiven/get_flink_application.py +8 -12
- pulumi_aiven/get_flink_application_version.py +8 -12
- pulumi_aiven/get_gcp_vpc_peering_connection.py +2 -6
- pulumi_aiven/get_grafana.py +7 -24
- pulumi_aiven/get_kafka.py +13 -17
- pulumi_aiven/get_kafka_acl.py +10 -14
- pulumi_aiven/get_kafka_connect.py +7 -24
- pulumi_aiven/get_kafka_connector.py +10 -14
- pulumi_aiven/get_kafka_mirror_maker.py +7 -24
- pulumi_aiven/get_kafka_schema.py +10 -14
- pulumi_aiven/get_kafka_schema_configuration.py +10 -14
- pulumi_aiven/get_kafka_schema_registry_acl.py +6 -6
- pulumi_aiven/get_kafka_topic.py +16 -20
- pulumi_aiven/get_kafka_user.py +10 -14
- pulumi_aiven/get_m3_aggregator.py +7 -24
- pulumi_aiven/get_m3_db.py +7 -24
- pulumi_aiven/get_m3db_user.py +10 -14
- pulumi_aiven/get_mirror_maker_replication_flow.py +14 -18
- pulumi_aiven/get_my_sql.py +7 -24
- pulumi_aiven/get_mysql_database.py +10 -14
- pulumi_aiven/get_mysql_user.py +10 -14
- pulumi_aiven/get_open_search.py +6 -10
- pulumi_aiven/get_open_search_acl_config.py +10 -14
- pulumi_aiven/get_open_search_acl_rule.py +10 -14
- pulumi_aiven/get_opensearch_security_plugin_config.py +10 -14
- pulumi_aiven/get_opensearch_user.py +10 -14
- pulumi_aiven/get_organization.py +0 -4
- pulumi_aiven/get_organization_user_group.py +2 -6
- pulumi_aiven/get_pg.py +6 -10
- pulumi_aiven/get_pg_database.py +10 -14
- pulumi_aiven/get_pg_user.py +10 -14
- pulumi_aiven/get_project.py +17 -21
- pulumi_aiven/get_project_user.py +9 -13
- pulumi_aiven/get_project_vpc.py +2 -6
- pulumi_aiven/get_redis.py +7 -24
- pulumi_aiven/get_redis_user.py +10 -14
- pulumi_aiven/get_service_component.py +4 -8
- pulumi_aiven/get_service_integration.py +19 -33
- pulumi_aiven/get_service_integration_endpoint.py +2 -6
- pulumi_aiven/get_transit_gateway_vpc_attachment.py +2 -6
- pulumi_aiven/grafana.py +82 -267
- pulumi_aiven/influx_db.py +21 -21
- pulumi_aiven/influxdb_database.py +21 -21
- pulumi_aiven/influxdb_user.py +21 -21
- pulumi_aiven/kafka.py +50 -54
- pulumi_aiven/kafka_acl.py +20 -24
- pulumi_aiven/kafka_connect.py +80 -267
- pulumi_aiven/kafka_connector.py +26 -30
- pulumi_aiven/kafka_mirror_maker.py +80 -267
- pulumi_aiven/kafka_schema.py +18 -22
- pulumi_aiven/kafka_schema_configuration.py +20 -24
- pulumi_aiven/kafka_schema_registry_acl.py +54 -14
- pulumi_aiven/kafka_topic.py +84 -46
- pulumi_aiven/kafka_user.py +18 -22
- pulumi_aiven/m3_aggregator.py +80 -267
- pulumi_aiven/m3_db.py +80 -267
- pulumi_aiven/m3db_user.py +16 -20
- pulumi_aiven/mirror_maker_replication_flow.py +22 -26
- pulumi_aiven/my_sql.py +80 -267
- pulumi_aiven/mysql_database.py +20 -24
- pulumi_aiven/mysql_user.py +18 -22
- pulumi_aiven/open_search.py +27 -31
- pulumi_aiven/open_search_acl_config.py +28 -36
- pulumi_aiven/open_search_acl_rule.py +36 -40
- pulumi_aiven/opensearch_security_plugin_config.py +28 -32
- pulumi_aiven/opensearch_user.py +18 -22
- pulumi_aiven/organization.py +2 -6
- pulumi_aiven/organization_group_project.py +46 -0
- pulumi_aiven/organization_user.py +2 -20
- pulumi_aiven/organization_user_group.py +4 -6
- pulumi_aiven/organization_user_group_member.py +8 -10
- pulumi_aiven/organizational_unit.py +6 -6
- pulumi_aiven/outputs.py +271 -390
- pulumi_aiven/pg.py +30 -30
- pulumi_aiven/pg_database.py +20 -24
- pulumi_aiven/pg_user.py +18 -22
- pulumi_aiven/project.py +75 -79
- pulumi_aiven/project_user.py +24 -28
- pulumi_aiven/project_vpc.py +31 -7
- pulumi_aiven/redis.py +80 -267
- pulumi_aiven/redis_user.py +18 -22
- pulumi_aiven/service_integration.py +54 -58
- pulumi_aiven/static_ip.py +7 -7
- pulumi_aiven/transit_gateway_vpc_attachment.py +4 -8
- {pulumi_aiven-6.13.0a1711361534.dist-info → pulumi_aiven-6.14.0.dist-info}/METADATA +1 -1
- pulumi_aiven-6.14.0.dist-info/RECORD +150 -0
- pulumi_aiven-6.13.0a1711361534.dist-info/RECORD +0 -150
- {pulumi_aiven-6.13.0a1711361534.dist-info → pulumi_aiven-6.14.0.dist-info}/WHEEL +0 -0
- {pulumi_aiven-6.13.0a1711361534.dist-info → pulumi_aiven-6.14.0.dist-info}/top_level.txt +0 -0
pulumi_aiven/outputs.py
CHANGED
|
@@ -12,7 +12,6 @@ from . import outputs
|
|
|
12
12
|
|
|
13
13
|
__all__ = [
|
|
14
14
|
'AccountAuthenticationSamlFieldMapping',
|
|
15
|
-
'CassandraCassandra',
|
|
16
15
|
'CassandraCassandraUserConfig',
|
|
17
16
|
'CassandraCassandraUserConfigCassandra',
|
|
18
17
|
'CassandraCassandraUserConfigIpFilterObject',
|
|
@@ -22,7 +21,6 @@ __all__ = [
|
|
|
22
21
|
'CassandraServiceIntegration',
|
|
23
22
|
'CassandraTag',
|
|
24
23
|
'CassandraTechEmail',
|
|
25
|
-
'ClickhouseClickhouse',
|
|
26
24
|
'ClickhouseClickhouseUserConfig',
|
|
27
25
|
'ClickhouseClickhouseUserConfigIpFilterObject',
|
|
28
26
|
'ClickhouseClickhouseUserConfigPrivateAccess',
|
|
@@ -35,7 +33,6 @@ __all__ = [
|
|
|
35
33
|
'ClickhouseTag',
|
|
36
34
|
'ClickhouseTechEmail',
|
|
37
35
|
'DragonflyComponent',
|
|
38
|
-
'DragonflyDragonfly',
|
|
39
36
|
'DragonflyDragonflyUserConfig',
|
|
40
37
|
'DragonflyDragonflyUserConfigIpFilterObject',
|
|
41
38
|
'DragonflyDragonflyUserConfigMigration',
|
|
@@ -56,7 +53,6 @@ __all__ = [
|
|
|
56
53
|
'FlinkTag',
|
|
57
54
|
'FlinkTechEmail',
|
|
58
55
|
'GrafanaComponent',
|
|
59
|
-
'GrafanaGrafana',
|
|
60
56
|
'GrafanaGrafanaUserConfig',
|
|
61
57
|
'GrafanaGrafanaUserConfigAuthAzuread',
|
|
62
58
|
'GrafanaGrafanaUserConfigAuthGenericOauth',
|
|
@@ -86,7 +82,6 @@ __all__ = [
|
|
|
86
82
|
'InfluxDbTechEmail',
|
|
87
83
|
'KafkaComponent',
|
|
88
84
|
'KafkaConnectComponent',
|
|
89
|
-
'KafkaConnectKafkaConnect',
|
|
90
85
|
'KafkaConnectKafkaConnectUserConfig',
|
|
91
86
|
'KafkaConnectKafkaConnectUserConfigIpFilterObject',
|
|
92
87
|
'KafkaConnectKafkaConnectUserConfigKafkaConnect',
|
|
@@ -111,7 +106,6 @@ __all__ = [
|
|
|
111
106
|
'KafkaKafkaUserConfigTieredStorage',
|
|
112
107
|
'KafkaKafkaUserConfigTieredStorageLocalCache',
|
|
113
108
|
'KafkaMirrorMakerComponent',
|
|
114
|
-
'KafkaMirrorMakerKafkaMirrormaker',
|
|
115
109
|
'KafkaMirrorMakerKafkaMirrormakerUserConfig',
|
|
116
110
|
'KafkaMirrorMakerKafkaMirrormakerUserConfigIpFilterObject',
|
|
117
111
|
'KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormaker',
|
|
@@ -124,14 +118,12 @@ __all__ = [
|
|
|
124
118
|
'KafkaTopicConfig',
|
|
125
119
|
'KafkaTopicTag',
|
|
126
120
|
'M3AggregatorComponent',
|
|
127
|
-
'M3AggregatorM3aggregator',
|
|
128
121
|
'M3AggregatorM3aggregatorUserConfig',
|
|
129
122
|
'M3AggregatorM3aggregatorUserConfigIpFilterObject',
|
|
130
123
|
'M3AggregatorServiceIntegration',
|
|
131
124
|
'M3AggregatorTag',
|
|
132
125
|
'M3AggregatorTechEmail',
|
|
133
126
|
'M3DbComponent',
|
|
134
|
-
'M3DbM3db',
|
|
135
127
|
'M3DbM3dbUserConfig',
|
|
136
128
|
'M3DbM3dbUserConfigIpFilterObject',
|
|
137
129
|
'M3DbM3dbUserConfigLimits',
|
|
@@ -150,7 +142,6 @@ __all__ = [
|
|
|
150
142
|
'M3DbTag',
|
|
151
143
|
'M3DbTechEmail',
|
|
152
144
|
'MySqlComponent',
|
|
153
|
-
'MySqlMysql',
|
|
154
145
|
'MySqlMysqlUserConfig',
|
|
155
146
|
'MySqlMysqlUserConfigIpFilterObject',
|
|
156
147
|
'MySqlMysqlUserConfigMigration',
|
|
@@ -202,7 +193,6 @@ __all__ = [
|
|
|
202
193
|
'PgTechEmail',
|
|
203
194
|
'ProjectTag',
|
|
204
195
|
'RedisComponent',
|
|
205
|
-
'RedisRedi',
|
|
206
196
|
'RedisRedisUserConfig',
|
|
207
197
|
'RedisRedisUserConfigIpFilterObject',
|
|
208
198
|
'RedisRedisUserConfigMigration',
|
|
@@ -255,7 +245,6 @@ __all__ = [
|
|
|
255
245
|
'ServiceIntegrationPrometheusUserConfigSourceMysql',
|
|
256
246
|
'ServiceIntegrationPrometheusUserConfigSourceMysqlTelegraf',
|
|
257
247
|
'GetAccountAuthenticationSamlFieldMappingResult',
|
|
258
|
-
'GetCassandaCassandraResult',
|
|
259
248
|
'GetCassandaCassandraUserConfigResult',
|
|
260
249
|
'GetCassandaCassandraUserConfigCassandraResult',
|
|
261
250
|
'GetCassandaCassandraUserConfigIpFilterObjectResult',
|
|
@@ -265,7 +254,6 @@ __all__ = [
|
|
|
265
254
|
'GetCassandaServiceIntegrationResult',
|
|
266
255
|
'GetCassandaTagResult',
|
|
267
256
|
'GetCassandaTechEmailResult',
|
|
268
|
-
'GetCassandraCassandraResult',
|
|
269
257
|
'GetCassandraCassandraUserConfigResult',
|
|
270
258
|
'GetCassandraCassandraUserConfigCassandraResult',
|
|
271
259
|
'GetCassandraCassandraUserConfigIpFilterObjectResult',
|
|
@@ -275,7 +263,6 @@ __all__ = [
|
|
|
275
263
|
'GetCassandraServiceIntegrationResult',
|
|
276
264
|
'GetCassandraTagResult',
|
|
277
265
|
'GetCassandraTechEmailResult',
|
|
278
|
-
'GetClickhouseClickhouseResult',
|
|
279
266
|
'GetClickhouseClickhouseUserConfigResult',
|
|
280
267
|
'GetClickhouseClickhouseUserConfigIpFilterObjectResult',
|
|
281
268
|
'GetClickhouseClickhouseUserConfigPrivateAccessResult',
|
|
@@ -286,7 +273,6 @@ __all__ = [
|
|
|
286
273
|
'GetClickhouseTagResult',
|
|
287
274
|
'GetClickhouseTechEmailResult',
|
|
288
275
|
'GetDragonflyComponentResult',
|
|
289
|
-
'GetDragonflyDragonflyResult',
|
|
290
276
|
'GetDragonflyDragonflyUserConfigResult',
|
|
291
277
|
'GetDragonflyDragonflyUserConfigIpFilterObjectResult',
|
|
292
278
|
'GetDragonflyDragonflyUserConfigMigrationResult',
|
|
@@ -307,7 +293,6 @@ __all__ = [
|
|
|
307
293
|
'GetFlinkTagResult',
|
|
308
294
|
'GetFlinkTechEmailResult',
|
|
309
295
|
'GetGrafanaComponentResult',
|
|
310
|
-
'GetGrafanaGrafanaResult',
|
|
311
296
|
'GetGrafanaGrafanaUserConfigResult',
|
|
312
297
|
'GetGrafanaGrafanaUserConfigAuthAzureadResult',
|
|
313
298
|
'GetGrafanaGrafanaUserConfigAuthGenericOauthResult',
|
|
@@ -337,7 +322,6 @@ __all__ = [
|
|
|
337
322
|
'GetInfluxDbTechEmailResult',
|
|
338
323
|
'GetKafkaComponentResult',
|
|
339
324
|
'GetKafkaConnectComponentResult',
|
|
340
|
-
'GetKafkaConnectKafkaConnectResult',
|
|
341
325
|
'GetKafkaConnectKafkaConnectUserConfigResult',
|
|
342
326
|
'GetKafkaConnectKafkaConnectUserConfigIpFilterObjectResult',
|
|
343
327
|
'GetKafkaConnectKafkaConnectUserConfigKafkaConnectResult',
|
|
@@ -362,7 +346,6 @@ __all__ = [
|
|
|
362
346
|
'GetKafkaKafkaUserConfigTieredStorageResult',
|
|
363
347
|
'GetKafkaKafkaUserConfigTieredStorageLocalCacheResult',
|
|
364
348
|
'GetKafkaMirrorMakerComponentResult',
|
|
365
|
-
'GetKafkaMirrorMakerKafkaMirrormakerResult',
|
|
366
349
|
'GetKafkaMirrorMakerKafkaMirrormakerUserConfigResult',
|
|
367
350
|
'GetKafkaMirrorMakerKafkaMirrormakerUserConfigIpFilterObjectResult',
|
|
368
351
|
'GetKafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormakerResult',
|
|
@@ -375,14 +358,12 @@ __all__ = [
|
|
|
375
358
|
'GetKafkaTopicConfigResult',
|
|
376
359
|
'GetKafkaTopicTagResult',
|
|
377
360
|
'GetM3AggregatorComponentResult',
|
|
378
|
-
'GetM3AggregatorM3aggregatorResult',
|
|
379
361
|
'GetM3AggregatorM3aggregatorUserConfigResult',
|
|
380
362
|
'GetM3AggregatorM3aggregatorUserConfigIpFilterObjectResult',
|
|
381
363
|
'GetM3AggregatorServiceIntegrationResult',
|
|
382
364
|
'GetM3AggregatorTagResult',
|
|
383
365
|
'GetM3AggregatorTechEmailResult',
|
|
384
366
|
'GetM3DbComponentResult',
|
|
385
|
-
'GetM3DbM3dbResult',
|
|
386
367
|
'GetM3DbM3dbUserConfigResult',
|
|
387
368
|
'GetM3DbM3dbUserConfigIpFilterObjectResult',
|
|
388
369
|
'GetM3DbM3dbUserConfigLimitsResult',
|
|
@@ -401,7 +382,6 @@ __all__ = [
|
|
|
401
382
|
'GetM3DbTagResult',
|
|
402
383
|
'GetM3DbTechEmailResult',
|
|
403
384
|
'GetMySqlComponentResult',
|
|
404
|
-
'GetMySqlMysqlResult',
|
|
405
385
|
'GetMySqlMysqlUserConfigResult',
|
|
406
386
|
'GetMySqlMysqlUserConfigIpFilterObjectResult',
|
|
407
387
|
'GetMySqlMysqlUserConfigMigrationResult',
|
|
@@ -450,7 +430,6 @@ __all__ = [
|
|
|
450
430
|
'GetPgTechEmailResult',
|
|
451
431
|
'GetProjectTagResult',
|
|
452
432
|
'GetRedisComponentResult',
|
|
453
|
-
'GetRedisRediResult',
|
|
454
433
|
'GetRedisRedisUserConfigResult',
|
|
455
434
|
'GetRedisRedisUserConfigIpFilterObjectResult',
|
|
456
435
|
'GetRedisRedisUserConfigMigrationResult',
|
|
@@ -592,12 +571,6 @@ class AccountAuthenticationSamlFieldMapping(dict):
|
|
|
592
571
|
return pulumi.get(self, "real_name")
|
|
593
572
|
|
|
594
573
|
|
|
595
|
-
@pulumi.output_type
|
|
596
|
-
class CassandraCassandra(dict):
|
|
597
|
-
def __init__(__self__):
|
|
598
|
-
pass
|
|
599
|
-
|
|
600
|
-
|
|
601
574
|
@pulumi.output_type
|
|
602
575
|
class CassandraCassandraUserConfig(dict):
|
|
603
576
|
@staticmethod
|
|
@@ -666,7 +639,7 @@ class CassandraCassandraUserConfig(dict):
|
|
|
666
639
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
667
640
|
:param int backup_hour: The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
668
641
|
:param int backup_minute: The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
669
|
-
:param 'CassandraCassandraUserConfigCassandraArgs' cassandra:
|
|
642
|
+
:param 'CassandraCassandraUserConfigCassandraArgs' cassandra: Cassandra configuration values
|
|
670
643
|
:param str cassandra_version: Cassandra version.
|
|
671
644
|
:param Sequence['CassandraCassandraUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
672
645
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -744,7 +717,7 @@ class CassandraCassandraUserConfig(dict):
|
|
|
744
717
|
@pulumi.getter
|
|
745
718
|
def cassandra(self) -> Optional['outputs.CassandraCassandraUserConfigCassandra']:
|
|
746
719
|
"""
|
|
747
|
-
|
|
720
|
+
Cassandra configuration values
|
|
748
721
|
"""
|
|
749
722
|
return pulumi.get(self, "cassandra")
|
|
750
723
|
|
|
@@ -964,7 +937,7 @@ class CassandraCassandraUserConfigPublicAccess(dict):
|
|
|
964
937
|
def __init__(__self__, *,
|
|
965
938
|
prometheus: Optional[bool] = None):
|
|
966
939
|
"""
|
|
967
|
-
:param bool prometheus: Allow clients to connect to prometheus
|
|
940
|
+
:param bool prometheus: Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
968
941
|
"""
|
|
969
942
|
if prometheus is not None:
|
|
970
943
|
pulumi.set(__self__, "prometheus", prometheus)
|
|
@@ -973,7 +946,7 @@ class CassandraCassandraUserConfigPublicAccess(dict):
|
|
|
973
946
|
@pulumi.getter
|
|
974
947
|
def prometheus(self) -> Optional[bool]:
|
|
975
948
|
"""
|
|
976
|
-
Allow clients to connect to prometheus
|
|
949
|
+
Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
977
950
|
"""
|
|
978
951
|
return pulumi.get(self, "prometheus")
|
|
979
952
|
|
|
@@ -1195,12 +1168,6 @@ class CassandraTechEmail(dict):
|
|
|
1195
1168
|
return pulumi.get(self, "email")
|
|
1196
1169
|
|
|
1197
1170
|
|
|
1198
|
-
@pulumi.output_type
|
|
1199
|
-
class ClickhouseClickhouse(dict):
|
|
1200
|
-
def __init__(__self__):
|
|
1201
|
-
pass
|
|
1202
|
-
|
|
1203
|
-
|
|
1204
1171
|
@pulumi.output_type
|
|
1205
1172
|
class ClickhouseClickhouseUserConfig(dict):
|
|
1206
1173
|
@staticmethod
|
|
@@ -1514,10 +1481,10 @@ class ClickhouseClickhouseUserConfigPrivatelinkAccess(dict):
|
|
|
1514
1481
|
clickhouse_mysql: Optional[bool] = None,
|
|
1515
1482
|
prometheus: Optional[bool] = None):
|
|
1516
1483
|
"""
|
|
1517
|
-
:param bool clickhouse:
|
|
1518
|
-
:param bool clickhouse_https:
|
|
1519
|
-
:param bool clickhouse_mysql:
|
|
1520
|
-
:param bool prometheus:
|
|
1484
|
+
:param bool clickhouse: Enable clickhouse.
|
|
1485
|
+
:param bool clickhouse_https: Enable clickhouse_https.
|
|
1486
|
+
:param bool clickhouse_mysql: Enable clickhouse_mysql.
|
|
1487
|
+
:param bool prometheus: Enable prometheus.
|
|
1521
1488
|
"""
|
|
1522
1489
|
if clickhouse is not None:
|
|
1523
1490
|
pulumi.set(__self__, "clickhouse", clickhouse)
|
|
@@ -1532,7 +1499,7 @@ class ClickhouseClickhouseUserConfigPrivatelinkAccess(dict):
|
|
|
1532
1499
|
@pulumi.getter
|
|
1533
1500
|
def clickhouse(self) -> Optional[bool]:
|
|
1534
1501
|
"""
|
|
1535
|
-
|
|
1502
|
+
Enable clickhouse.
|
|
1536
1503
|
"""
|
|
1537
1504
|
return pulumi.get(self, "clickhouse")
|
|
1538
1505
|
|
|
@@ -1540,7 +1507,7 @@ class ClickhouseClickhouseUserConfigPrivatelinkAccess(dict):
|
|
|
1540
1507
|
@pulumi.getter(name="clickhouseHttps")
|
|
1541
1508
|
def clickhouse_https(self) -> Optional[bool]:
|
|
1542
1509
|
"""
|
|
1543
|
-
|
|
1510
|
+
Enable clickhouse_https.
|
|
1544
1511
|
"""
|
|
1545
1512
|
return pulumi.get(self, "clickhouse_https")
|
|
1546
1513
|
|
|
@@ -1548,7 +1515,7 @@ class ClickhouseClickhouseUserConfigPrivatelinkAccess(dict):
|
|
|
1548
1515
|
@pulumi.getter(name="clickhouseMysql")
|
|
1549
1516
|
def clickhouse_mysql(self) -> Optional[bool]:
|
|
1550
1517
|
"""
|
|
1551
|
-
|
|
1518
|
+
Enable clickhouse_mysql.
|
|
1552
1519
|
"""
|
|
1553
1520
|
return pulumi.get(self, "clickhouse_mysql")
|
|
1554
1521
|
|
|
@@ -1556,7 +1523,7 @@ class ClickhouseClickhouseUserConfigPrivatelinkAccess(dict):
|
|
|
1556
1523
|
@pulumi.getter
|
|
1557
1524
|
def prometheus(self) -> Optional[bool]:
|
|
1558
1525
|
"""
|
|
1559
|
-
|
|
1526
|
+
Enable prometheus.
|
|
1560
1527
|
"""
|
|
1561
1528
|
return pulumi.get(self, "prometheus")
|
|
1562
1529
|
|
|
@@ -1588,10 +1555,10 @@ class ClickhouseClickhouseUserConfigPublicAccess(dict):
|
|
|
1588
1555
|
clickhouse_mysql: Optional[bool] = None,
|
|
1589
1556
|
prometheus: Optional[bool] = None):
|
|
1590
1557
|
"""
|
|
1591
|
-
:param bool clickhouse: Allow clients to connect to clickhouse
|
|
1592
|
-
:param bool clickhouse_https: Allow clients to connect to clickhouse_https
|
|
1593
|
-
:param bool clickhouse_mysql: Allow clients to connect to clickhouse_mysql
|
|
1594
|
-
:param bool prometheus: Allow clients to connect to prometheus
|
|
1558
|
+
:param bool clickhouse: Allow clients to connect to clickhouse from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1559
|
+
:param bool clickhouse_https: Allow clients to connect to clickhouse_https from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1560
|
+
:param bool clickhouse_mysql: Allow clients to connect to clickhouse_mysql from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1561
|
+
:param bool prometheus: Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1595
1562
|
"""
|
|
1596
1563
|
if clickhouse is not None:
|
|
1597
1564
|
pulumi.set(__self__, "clickhouse", clickhouse)
|
|
@@ -1606,7 +1573,7 @@ class ClickhouseClickhouseUserConfigPublicAccess(dict):
|
|
|
1606
1573
|
@pulumi.getter
|
|
1607
1574
|
def clickhouse(self) -> Optional[bool]:
|
|
1608
1575
|
"""
|
|
1609
|
-
Allow clients to connect to clickhouse
|
|
1576
|
+
Allow clients to connect to clickhouse from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1610
1577
|
"""
|
|
1611
1578
|
return pulumi.get(self, "clickhouse")
|
|
1612
1579
|
|
|
@@ -1614,7 +1581,7 @@ class ClickhouseClickhouseUserConfigPublicAccess(dict):
|
|
|
1614
1581
|
@pulumi.getter(name="clickhouseHttps")
|
|
1615
1582
|
def clickhouse_https(self) -> Optional[bool]:
|
|
1616
1583
|
"""
|
|
1617
|
-
Allow clients to connect to clickhouse_https
|
|
1584
|
+
Allow clients to connect to clickhouse_https from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1618
1585
|
"""
|
|
1619
1586
|
return pulumi.get(self, "clickhouse_https")
|
|
1620
1587
|
|
|
@@ -1622,7 +1589,7 @@ class ClickhouseClickhouseUserConfigPublicAccess(dict):
|
|
|
1622
1589
|
@pulumi.getter(name="clickhouseMysql")
|
|
1623
1590
|
def clickhouse_mysql(self) -> Optional[bool]:
|
|
1624
1591
|
"""
|
|
1625
|
-
Allow clients to connect to clickhouse_mysql
|
|
1592
|
+
Allow clients to connect to clickhouse_mysql from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1626
1593
|
"""
|
|
1627
1594
|
return pulumi.get(self, "clickhouse_mysql")
|
|
1628
1595
|
|
|
@@ -1630,7 +1597,7 @@ class ClickhouseClickhouseUserConfigPublicAccess(dict):
|
|
|
1630
1597
|
@pulumi.getter
|
|
1631
1598
|
def prometheus(self) -> Optional[bool]:
|
|
1632
1599
|
"""
|
|
1633
|
-
Allow clients to connect to prometheus
|
|
1600
|
+
Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
1634
1601
|
"""
|
|
1635
1602
|
return pulumi.get(self, "prometheus")
|
|
1636
1603
|
|
|
@@ -2076,12 +2043,6 @@ class DragonflyComponent(dict):
|
|
|
2076
2043
|
return pulumi.get(self, "usage")
|
|
2077
2044
|
|
|
2078
2045
|
|
|
2079
|
-
@pulumi.output_type
|
|
2080
|
-
class DragonflyDragonfly(dict):
|
|
2081
|
-
def __init__(__self__):
|
|
2082
|
-
pass
|
|
2083
|
-
|
|
2084
|
-
|
|
2085
2046
|
@pulumi.output_type
|
|
2086
2047
|
class DragonflyDragonflyUserConfig(dict):
|
|
2087
2048
|
@staticmethod
|
|
@@ -2089,6 +2050,8 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2089
2050
|
suggest = None
|
|
2090
2051
|
if key == "cacheMode":
|
|
2091
2052
|
suggest = "cache_mode"
|
|
2053
|
+
elif key == "dragonflyPersistence":
|
|
2054
|
+
suggest = "dragonfly_persistence"
|
|
2092
2055
|
elif key == "dragonflySsl":
|
|
2093
2056
|
suggest = "dragonfly_ssl"
|
|
2094
2057
|
elif key == "ipFilterObjects":
|
|
@@ -2127,6 +2090,7 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2127
2090
|
|
|
2128
2091
|
def __init__(__self__, *,
|
|
2129
2092
|
cache_mode: Optional[bool] = None,
|
|
2093
|
+
dragonfly_persistence: Optional[str] = None,
|
|
2130
2094
|
dragonfly_ssl: Optional[bool] = None,
|
|
2131
2095
|
ip_filter_objects: Optional[Sequence['outputs.DragonflyDragonflyUserConfigIpFilterObject']] = None,
|
|
2132
2096
|
ip_filter_strings: Optional[Sequence[str]] = None,
|
|
@@ -2142,6 +2106,7 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2142
2106
|
static_ips: Optional[bool] = None):
|
|
2143
2107
|
"""
|
|
2144
2108
|
:param bool cache_mode: Evict entries when getting close to maxmemory limit. The default value is `false`.
|
|
2109
|
+
:param str dragonfly_persistence: When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
2145
2110
|
:param bool dragonfly_ssl: Require SSL to access Dragonfly. The default value is `true`.
|
|
2146
2111
|
:param Sequence['DragonflyDragonflyUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
2147
2112
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -2158,6 +2123,8 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2158
2123
|
"""
|
|
2159
2124
|
if cache_mode is not None:
|
|
2160
2125
|
pulumi.set(__self__, "cache_mode", cache_mode)
|
|
2126
|
+
if dragonfly_persistence is not None:
|
|
2127
|
+
pulumi.set(__self__, "dragonfly_persistence", dragonfly_persistence)
|
|
2161
2128
|
if dragonfly_ssl is not None:
|
|
2162
2129
|
pulumi.set(__self__, "dragonfly_ssl", dragonfly_ssl)
|
|
2163
2130
|
if ip_filter_objects is not None:
|
|
@@ -2193,6 +2160,14 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2193
2160
|
"""
|
|
2194
2161
|
return pulumi.get(self, "cache_mode")
|
|
2195
2162
|
|
|
2163
|
+
@property
|
|
2164
|
+
@pulumi.getter(name="dragonflyPersistence")
|
|
2165
|
+
def dragonfly_persistence(self) -> Optional[str]:
|
|
2166
|
+
"""
|
|
2167
|
+
When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
2168
|
+
"""
|
|
2169
|
+
return pulumi.get(self, "dragonfly_persistence")
|
|
2170
|
+
|
|
2196
2171
|
@property
|
|
2197
2172
|
@pulumi.getter(name="dragonflySsl")
|
|
2198
2173
|
def dragonfly_ssl(self) -> Optional[bool]:
|
|
@@ -3325,12 +3300,6 @@ class GrafanaComponent(dict):
|
|
|
3325
3300
|
return pulumi.get(self, "usage")
|
|
3326
3301
|
|
|
3327
3302
|
|
|
3328
|
-
@pulumi.output_type
|
|
3329
|
-
class GrafanaGrafana(dict):
|
|
3330
|
-
def __init__(__self__):
|
|
3331
|
-
pass
|
|
3332
|
-
|
|
3333
|
-
|
|
3334
3303
|
@pulumi.output_type
|
|
3335
3304
|
class GrafanaGrafanaUserConfig(dict):
|
|
3336
3305
|
@staticmethod
|
|
@@ -3477,7 +3446,7 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3477
3446
|
viewers_can_edit: Optional[bool] = None):
|
|
3478
3447
|
"""
|
|
3479
3448
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
3480
|
-
:param bool alerting_enabled: Enable or disable Grafana legacy alerting functionality. This should not be enabled with
|
|
3449
|
+
:param bool alerting_enabled: Enable or disable Grafana legacy alerting functionality. This should not be enabled with unified*alerting*enabled.
|
|
3481
3450
|
:param str alerting_error_or_timeout: Default error or timeout setting for new alerting rules.
|
|
3482
3451
|
:param int alerting_max_annotations_to_keep: Max number of alert annotations that Grafana stores. 0 (default) keeps all alert annotations.
|
|
3483
3452
|
:param str alerting_nodata_or_nullvalues: Default value for 'no data or null values' for new alerting rules.
|
|
@@ -3514,7 +3483,7 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3514
3483
|
:param str service_to_fork_from: Name of another service to fork from. This has effect only when a new service is being created.
|
|
3515
3484
|
:param 'GrafanaGrafanaUserConfigSmtpServerArgs' smtp_server: SMTP server settings
|
|
3516
3485
|
:param bool static_ips: Use static public IP addresses.
|
|
3517
|
-
:param bool unified_alerting_enabled: Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set
|
|
3486
|
+
:param bool unified_alerting_enabled: Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set unified*alerting*enabled to false and alerting_enabled to true. See https://grafana.com/docs/grafana/latest/alerting/set-up/migrating-alerts/ for more details.
|
|
3518
3487
|
:param bool user_auto_assign_org: Auto-assign new users on signup to main organization. Defaults to false.
|
|
3519
3488
|
:param str user_auto_assign_org_role: Set role for new signups. Defaults to Viewer.
|
|
3520
3489
|
:param bool viewers_can_edit: Users with view-only permission can edit but not save dashboards.
|
|
@@ -3610,16 +3579,13 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3610
3579
|
"""
|
|
3611
3580
|
Additional Cloud Regions for Backup Replication.
|
|
3612
3581
|
"""
|
|
3613
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
3614
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
3615
|
-
|
|
3616
3582
|
return pulumi.get(self, "additional_backup_regions")
|
|
3617
3583
|
|
|
3618
3584
|
@property
|
|
3619
3585
|
@pulumi.getter(name="alertingEnabled")
|
|
3620
3586
|
def alerting_enabled(self) -> Optional[bool]:
|
|
3621
3587
|
"""
|
|
3622
|
-
Enable or disable Grafana legacy alerting functionality. This should not be enabled with
|
|
3588
|
+
Enable or disable Grafana legacy alerting functionality. This should not be enabled with unified*alerting*enabled.
|
|
3623
3589
|
"""
|
|
3624
3590
|
return pulumi.get(self, "alerting_enabled")
|
|
3625
3591
|
|
|
@@ -3918,7 +3884,7 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3918
3884
|
@pulumi.getter(name="unifiedAlertingEnabled")
|
|
3919
3885
|
def unified_alerting_enabled(self) -> Optional[bool]:
|
|
3920
3886
|
"""
|
|
3921
|
-
Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set
|
|
3887
|
+
Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set unified*alerting*enabled to false and alerting_enabled to true. See https://grafana.com/docs/grafana/latest/alerting/set-up/migrating-alerts/ for more details.
|
|
3922
3888
|
"""
|
|
3923
3889
|
return pulumi.get(self, "unified_alerting_enabled")
|
|
3924
3890
|
|
|
@@ -5977,12 +5943,6 @@ class KafkaConnectComponent(dict):
|
|
|
5977
5943
|
return pulumi.get(self, "usage")
|
|
5978
5944
|
|
|
5979
5945
|
|
|
5980
|
-
@pulumi.output_type
|
|
5981
|
-
class KafkaConnectKafkaConnect(dict):
|
|
5982
|
-
def __init__(__self__):
|
|
5983
|
-
pass
|
|
5984
|
-
|
|
5985
|
-
|
|
5986
5946
|
@pulumi.output_type
|
|
5987
5947
|
class KafkaConnectKafkaConnectUserConfig(dict):
|
|
5988
5948
|
@staticmethod
|
|
@@ -6251,8 +6211,8 @@ class KafkaConnectKafkaConnectUserConfigKafkaConnect(dict):
|
|
|
6251
6211
|
:param str connector_client_config_override_policy: Defines what client configurations can be overridden by the connector. Default is None.
|
|
6252
6212
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
6253
6213
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
6254
|
-
:param str consumer_isolation_level: Transaction read isolation level.
|
|
6255
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6214
|
+
:param str consumer_isolation_level: Transaction read isolation level. read*uncommitted is the default, but read*committed can be used if consume-exactly-once behavior is desired.
|
|
6215
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6256
6216
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
6257
6217
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
6258
6218
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -6326,7 +6286,7 @@ class KafkaConnectKafkaConnectUserConfigKafkaConnect(dict):
|
|
|
6326
6286
|
@pulumi.getter(name="consumerIsolationLevel")
|
|
6327
6287
|
def consumer_isolation_level(self) -> Optional[str]:
|
|
6328
6288
|
"""
|
|
6329
|
-
Transaction read isolation level.
|
|
6289
|
+
Transaction read isolation level. read*uncommitted is the default, but read*committed can be used if consume-exactly-once behavior is desired.
|
|
6330
6290
|
"""
|
|
6331
6291
|
return pulumi.get(self, "consumer_isolation_level")
|
|
6332
6292
|
|
|
@@ -6334,7 +6294,7 @@ class KafkaConnectKafkaConnectUserConfigKafkaConnect(dict):
|
|
|
6334
6294
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
6335
6295
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
6336
6296
|
"""
|
|
6337
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6297
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6338
6298
|
"""
|
|
6339
6299
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
6340
6300
|
|
|
@@ -6743,11 +6703,11 @@ class KafkaKafka(dict):
|
|
|
6743
6703
|
rest_uri: Optional[str] = None,
|
|
6744
6704
|
schema_registry_uri: Optional[str] = None):
|
|
6745
6705
|
"""
|
|
6746
|
-
:param str access_cert: The Kafka client certificate
|
|
6747
|
-
:param str access_key: The Kafka client certificate key
|
|
6748
|
-
:param str connect_uri: The Kafka Connect URI
|
|
6749
|
-
:param str rest_uri: The Kafka REST URI
|
|
6750
|
-
:param str schema_registry_uri: The Schema Registry URI
|
|
6706
|
+
:param str access_cert: The Kafka client certificate.
|
|
6707
|
+
:param str access_key: The Kafka client certificate key.
|
|
6708
|
+
:param str connect_uri: The Kafka Connect URI.
|
|
6709
|
+
:param str rest_uri: The Kafka REST URI.
|
|
6710
|
+
:param str schema_registry_uri: The Schema Registry URI.
|
|
6751
6711
|
"""
|
|
6752
6712
|
if access_cert is not None:
|
|
6753
6713
|
pulumi.set(__self__, "access_cert", access_cert)
|
|
@@ -6764,7 +6724,7 @@ class KafkaKafka(dict):
|
|
|
6764
6724
|
@pulumi.getter(name="accessCert")
|
|
6765
6725
|
def access_cert(self) -> Optional[str]:
|
|
6766
6726
|
"""
|
|
6767
|
-
The Kafka client certificate
|
|
6727
|
+
The Kafka client certificate.
|
|
6768
6728
|
"""
|
|
6769
6729
|
return pulumi.get(self, "access_cert")
|
|
6770
6730
|
|
|
@@ -6772,7 +6732,7 @@ class KafkaKafka(dict):
|
|
|
6772
6732
|
@pulumi.getter(name="accessKey")
|
|
6773
6733
|
def access_key(self) -> Optional[str]:
|
|
6774
6734
|
"""
|
|
6775
|
-
The Kafka client certificate key
|
|
6735
|
+
The Kafka client certificate key.
|
|
6776
6736
|
"""
|
|
6777
6737
|
return pulumi.get(self, "access_key")
|
|
6778
6738
|
|
|
@@ -6780,7 +6740,7 @@ class KafkaKafka(dict):
|
|
|
6780
6740
|
@pulumi.getter(name="connectUri")
|
|
6781
6741
|
def connect_uri(self) -> Optional[str]:
|
|
6782
6742
|
"""
|
|
6783
|
-
The Kafka Connect URI
|
|
6743
|
+
The Kafka Connect URI.
|
|
6784
6744
|
"""
|
|
6785
6745
|
return pulumi.get(self, "connect_uri")
|
|
6786
6746
|
|
|
@@ -6788,7 +6748,7 @@ class KafkaKafka(dict):
|
|
|
6788
6748
|
@pulumi.getter(name="restUri")
|
|
6789
6749
|
def rest_uri(self) -> Optional[str]:
|
|
6790
6750
|
"""
|
|
6791
|
-
The Kafka REST URI
|
|
6751
|
+
The Kafka REST URI.
|
|
6792
6752
|
"""
|
|
6793
6753
|
return pulumi.get(self, "rest_uri")
|
|
6794
6754
|
|
|
@@ -6796,7 +6756,7 @@ class KafkaKafka(dict):
|
|
|
6796
6756
|
@pulumi.getter(name="schemaRegistryUri")
|
|
6797
6757
|
def schema_registry_uri(self) -> Optional[str]:
|
|
6798
6758
|
"""
|
|
6799
|
-
The Schema Registry URI
|
|
6759
|
+
The Schema Registry URI.
|
|
6800
6760
|
"""
|
|
6801
6761
|
return pulumi.get(self, "schema_registry_uri")
|
|
6802
6762
|
|
|
@@ -7340,7 +7300,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7340
7300
|
:param int log_index_size_max_bytes: The maximum size in bytes of the offset index.
|
|
7341
7301
|
:param int log_local_retention_bytes: The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value.
|
|
7342
7302
|
:param int log_local_retention_ms: The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value.
|
|
7343
|
-
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7303
|
+
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7344
7304
|
:param int log_message_timestamp_difference_max_ms: The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message.
|
|
7345
7305
|
:param str log_message_timestamp_type: Define whether the timestamp in the message is message create time or log append time.
|
|
7346
7306
|
:param bool log_preallocate: Should pre allocate file when create new segment?
|
|
@@ -7362,7 +7322,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7362
7322
|
:param int replica_fetch_response_max_bytes: Maximum bytes expected for the entire fetch response (defaults to 10485760). Records are fetched in batches, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. As such, this is not an absolute maximum.
|
|
7363
7323
|
:param str sasl_oauthbearer_expected_audience: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences.
|
|
7364
7324
|
:param str sasl_oauthbearer_expected_issuer: Optional setting for the broker to use to verify that the JWT was created by the expected issuer.
|
|
7365
|
-
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7325
|
+
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7366
7326
|
:param str sasl_oauthbearer_sub_claim_name: Name of the scope from which to extract the subject claim from the JWT. Defaults to sub.
|
|
7367
7327
|
:param int socket_request_max_bytes: The maximum number of bytes in a socket request (defaults to 104857600).
|
|
7368
7328
|
:param bool transaction_partition_verification_enable: Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition.
|
|
@@ -7610,7 +7570,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7610
7570
|
@pulumi.getter(name="logMessageDownconversionEnable")
|
|
7611
7571
|
def log_message_downconversion_enable(self) -> Optional[bool]:
|
|
7612
7572
|
"""
|
|
7613
|
-
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7573
|
+
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7614
7574
|
"""
|
|
7615
7575
|
return pulumi.get(self, "log_message_downconversion_enable")
|
|
7616
7576
|
|
|
@@ -7786,7 +7746,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7786
7746
|
@pulumi.getter(name="saslOauthbearerJwksEndpointUrl")
|
|
7787
7747
|
def sasl_oauthbearer_jwks_endpoint_url(self) -> Optional[str]:
|
|
7788
7748
|
"""
|
|
7789
|
-
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7749
|
+
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7790
7750
|
"""
|
|
7791
7751
|
return pulumi.get(self, "sasl_oauthbearer_jwks_endpoint_url")
|
|
7792
7752
|
|
|
@@ -7933,7 +7893,7 @@ class KafkaKafkaUserConfigKafkaConnectConfig(dict):
|
|
|
7933
7893
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
7934
7894
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
7935
7895
|
:param str consumer_isolation_level: Transaction read isolation level. read*uncommitted is the default, but read*committed can be used if consume-exactly-once behavior is desired.
|
|
7936
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
7896
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
7937
7897
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
7938
7898
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
7939
7899
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -8015,7 +7975,7 @@ class KafkaKafkaUserConfigKafkaConnectConfig(dict):
|
|
|
8015
7975
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
8016
7976
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
8017
7977
|
"""
|
|
8018
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
7978
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
8019
7979
|
"""
|
|
8020
7980
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
8021
7981
|
|
|
@@ -8164,8 +8124,8 @@ class KafkaKafkaUserConfigKafkaRestConfig(dict):
|
|
|
8164
8124
|
:param bool name_strategy_validation: If true, validate that given schema is registered under expected subject name by the used name strategy when producing messages. The default value is `true`.
|
|
8165
8125
|
:param str producer_acks: The number of acknowledgments the producer requires the leader to have received before considering a request complete. If set to 'all' or '-1', the leader will wait for the full set of in-sync replicas to acknowledge the record. The default value is `1`.
|
|
8166
8126
|
:param str producer_compression_type: Specify the default compression type for producers. This configuration accepts the standard compression codecs ('gzip', 'snappy', 'lz4', 'zstd'). It additionally accepts 'none' which is the default and equivalent to no compression.
|
|
8167
|
-
:param int producer_linger_ms:
|
|
8168
|
-
:param int producer_max_request_size:
|
|
8127
|
+
:param int producer_linger_ms: Wait for up to the given delay to allow batching records together. The default value is `0`.
|
|
8128
|
+
:param int producer_max_request_size: The maximum size of a request in bytes. Note that Kafka broker can also cap the record batch size. The default value is `1048576`.
|
|
8169
8129
|
:param int simpleconsumer_pool_size_max: Maximum number of SimpleConsumers that can be instantiated per broker. The default value is `25`.
|
|
8170
8130
|
"""
|
|
8171
8131
|
if consumer_enable_auto_commit is not None:
|
|
@@ -8249,7 +8209,7 @@ class KafkaKafkaUserConfigKafkaRestConfig(dict):
|
|
|
8249
8209
|
@pulumi.getter(name="producerLingerMs")
|
|
8250
8210
|
def producer_linger_ms(self) -> Optional[int]:
|
|
8251
8211
|
"""
|
|
8252
|
-
|
|
8212
|
+
Wait for up to the given delay to allow batching records together. The default value is `0`.
|
|
8253
8213
|
"""
|
|
8254
8214
|
return pulumi.get(self, "producer_linger_ms")
|
|
8255
8215
|
|
|
@@ -8257,7 +8217,7 @@ class KafkaKafkaUserConfigKafkaRestConfig(dict):
|
|
|
8257
8217
|
@pulumi.getter(name="producerMaxRequestSize")
|
|
8258
8218
|
def producer_max_request_size(self) -> Optional[int]:
|
|
8259
8219
|
"""
|
|
8260
|
-
|
|
8220
|
+
The maximum size of a request in bytes. Note that Kafka broker can also cap the record batch size. The default value is `1048576`.
|
|
8261
8221
|
"""
|
|
8262
8222
|
return pulumi.get(self, "producer_max_request_size")
|
|
8263
8223
|
|
|
@@ -8300,9 +8260,9 @@ class KafkaKafkaUserConfigPrivateAccess(dict):
|
|
|
8300
8260
|
prometheus: Optional[bool] = None,
|
|
8301
8261
|
schema_registry: Optional[bool] = None):
|
|
8302
8262
|
"""
|
|
8303
|
-
:param bool kafka:
|
|
8304
|
-
:param bool kafka_connect:
|
|
8305
|
-
:param bool kafka_rest:
|
|
8263
|
+
:param bool kafka: Allow clients to connect to kafka with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8264
|
+
:param bool kafka_connect: Allow clients to connect to kafka_connect with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8265
|
+
:param bool kafka_rest: Allow clients to connect to kafka_rest with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8306
8266
|
:param bool prometheus: Allow clients to connect to prometheus with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8307
8267
|
:param bool schema_registry: Allow clients to connect to schema_registry with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8308
8268
|
"""
|
|
@@ -8321,7 +8281,7 @@ class KafkaKafkaUserConfigPrivateAccess(dict):
|
|
|
8321
8281
|
@pulumi.getter
|
|
8322
8282
|
def kafka(self) -> Optional[bool]:
|
|
8323
8283
|
"""
|
|
8324
|
-
|
|
8284
|
+
Allow clients to connect to kafka with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8325
8285
|
"""
|
|
8326
8286
|
return pulumi.get(self, "kafka")
|
|
8327
8287
|
|
|
@@ -8329,7 +8289,7 @@ class KafkaKafkaUserConfigPrivateAccess(dict):
|
|
|
8329
8289
|
@pulumi.getter(name="kafkaConnect")
|
|
8330
8290
|
def kafka_connect(self) -> Optional[bool]:
|
|
8331
8291
|
"""
|
|
8332
|
-
|
|
8292
|
+
Allow clients to connect to kafka_connect with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8333
8293
|
"""
|
|
8334
8294
|
return pulumi.get(self, "kafka_connect")
|
|
8335
8295
|
|
|
@@ -8337,7 +8297,7 @@ class KafkaKafkaUserConfigPrivateAccess(dict):
|
|
|
8337
8297
|
@pulumi.getter(name="kafkaRest")
|
|
8338
8298
|
def kafka_rest(self) -> Optional[bool]:
|
|
8339
8299
|
"""
|
|
8340
|
-
|
|
8300
|
+
Allow clients to connect to kafka_rest with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
8341
8301
|
"""
|
|
8342
8302
|
return pulumi.get(self, "kafka_rest")
|
|
8343
8303
|
|
|
@@ -8390,11 +8350,11 @@ class KafkaKafkaUserConfigPrivatelinkAccess(dict):
|
|
|
8390
8350
|
schema_registry: Optional[bool] = None):
|
|
8391
8351
|
"""
|
|
8392
8352
|
:param bool jolokia: Enable jolokia.
|
|
8393
|
-
:param bool kafka:
|
|
8394
|
-
:param bool kafka_connect: Enable
|
|
8395
|
-
:param bool kafka_rest: Enable
|
|
8396
|
-
:param bool prometheus:
|
|
8397
|
-
:param bool schema_registry:
|
|
8353
|
+
:param bool kafka: Enable kafka.
|
|
8354
|
+
:param bool kafka_connect: Enable kafka_connect.
|
|
8355
|
+
:param bool kafka_rest: Enable kafka_rest.
|
|
8356
|
+
:param bool prometheus: Enable prometheus.
|
|
8357
|
+
:param bool schema_registry: Enable schema_registry.
|
|
8398
8358
|
"""
|
|
8399
8359
|
if jolokia is not None:
|
|
8400
8360
|
pulumi.set(__self__, "jolokia", jolokia)
|
|
@@ -8421,7 +8381,7 @@ class KafkaKafkaUserConfigPrivatelinkAccess(dict):
|
|
|
8421
8381
|
@pulumi.getter
|
|
8422
8382
|
def kafka(self) -> Optional[bool]:
|
|
8423
8383
|
"""
|
|
8424
|
-
|
|
8384
|
+
Enable kafka.
|
|
8425
8385
|
"""
|
|
8426
8386
|
return pulumi.get(self, "kafka")
|
|
8427
8387
|
|
|
@@ -8429,7 +8389,7 @@ class KafkaKafkaUserConfigPrivatelinkAccess(dict):
|
|
|
8429
8389
|
@pulumi.getter(name="kafkaConnect")
|
|
8430
8390
|
def kafka_connect(self) -> Optional[bool]:
|
|
8431
8391
|
"""
|
|
8432
|
-
Enable
|
|
8392
|
+
Enable kafka_connect.
|
|
8433
8393
|
"""
|
|
8434
8394
|
return pulumi.get(self, "kafka_connect")
|
|
8435
8395
|
|
|
@@ -8437,7 +8397,7 @@ class KafkaKafkaUserConfigPrivatelinkAccess(dict):
|
|
|
8437
8397
|
@pulumi.getter(name="kafkaRest")
|
|
8438
8398
|
def kafka_rest(self) -> Optional[bool]:
|
|
8439
8399
|
"""
|
|
8440
|
-
Enable
|
|
8400
|
+
Enable kafka_rest.
|
|
8441
8401
|
"""
|
|
8442
8402
|
return pulumi.get(self, "kafka_rest")
|
|
8443
8403
|
|
|
@@ -8445,7 +8405,7 @@ class KafkaKafkaUserConfigPrivatelinkAccess(dict):
|
|
|
8445
8405
|
@pulumi.getter
|
|
8446
8406
|
def prometheus(self) -> Optional[bool]:
|
|
8447
8407
|
"""
|
|
8448
|
-
|
|
8408
|
+
Enable prometheus.
|
|
8449
8409
|
"""
|
|
8450
8410
|
return pulumi.get(self, "prometheus")
|
|
8451
8411
|
|
|
@@ -8453,7 +8413,7 @@ class KafkaKafkaUserConfigPrivatelinkAccess(dict):
|
|
|
8453
8413
|
@pulumi.getter(name="schemaRegistry")
|
|
8454
8414
|
def schema_registry(self) -> Optional[bool]:
|
|
8455
8415
|
"""
|
|
8456
|
-
|
|
8416
|
+
Enable schema_registry.
|
|
8457
8417
|
"""
|
|
8458
8418
|
return pulumi.get(self, "schema_registry")
|
|
8459
8419
|
|
|
@@ -8488,11 +8448,11 @@ class KafkaKafkaUserConfigPublicAccess(dict):
|
|
|
8488
8448
|
prometheus: Optional[bool] = None,
|
|
8489
8449
|
schema_registry: Optional[bool] = None):
|
|
8490
8450
|
"""
|
|
8491
|
-
:param bool kafka:
|
|
8492
|
-
:param bool kafka_connect:
|
|
8493
|
-
:param bool kafka_rest:
|
|
8494
|
-
:param bool prometheus: Allow clients to connect to prometheus
|
|
8495
|
-
:param bool schema_registry: Allow clients to connect to schema_registry
|
|
8451
|
+
:param bool kafka: Allow clients to connect to kafka from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8452
|
+
:param bool kafka_connect: Allow clients to connect to kafka_connect from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8453
|
+
:param bool kafka_rest: Allow clients to connect to kafka_rest from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8454
|
+
:param bool prometheus: Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8455
|
+
:param bool schema_registry: Allow clients to connect to schema_registry from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8496
8456
|
"""
|
|
8497
8457
|
if kafka is not None:
|
|
8498
8458
|
pulumi.set(__self__, "kafka", kafka)
|
|
@@ -8509,7 +8469,7 @@ class KafkaKafkaUserConfigPublicAccess(dict):
|
|
|
8509
8469
|
@pulumi.getter
|
|
8510
8470
|
def kafka(self) -> Optional[bool]:
|
|
8511
8471
|
"""
|
|
8512
|
-
|
|
8472
|
+
Allow clients to connect to kafka from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8513
8473
|
"""
|
|
8514
8474
|
return pulumi.get(self, "kafka")
|
|
8515
8475
|
|
|
@@ -8517,7 +8477,7 @@ class KafkaKafkaUserConfigPublicAccess(dict):
|
|
|
8517
8477
|
@pulumi.getter(name="kafkaConnect")
|
|
8518
8478
|
def kafka_connect(self) -> Optional[bool]:
|
|
8519
8479
|
"""
|
|
8520
|
-
|
|
8480
|
+
Allow clients to connect to kafka_connect from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8521
8481
|
"""
|
|
8522
8482
|
return pulumi.get(self, "kafka_connect")
|
|
8523
8483
|
|
|
@@ -8525,7 +8485,7 @@ class KafkaKafkaUserConfigPublicAccess(dict):
|
|
|
8525
8485
|
@pulumi.getter(name="kafkaRest")
|
|
8526
8486
|
def kafka_rest(self) -> Optional[bool]:
|
|
8527
8487
|
"""
|
|
8528
|
-
|
|
8488
|
+
Allow clients to connect to kafka_rest from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8529
8489
|
"""
|
|
8530
8490
|
return pulumi.get(self, "kafka_rest")
|
|
8531
8491
|
|
|
@@ -8533,7 +8493,7 @@ class KafkaKafkaUserConfigPublicAccess(dict):
|
|
|
8533
8493
|
@pulumi.getter
|
|
8534
8494
|
def prometheus(self) -> Optional[bool]:
|
|
8535
8495
|
"""
|
|
8536
|
-
Allow clients to connect to prometheus
|
|
8496
|
+
Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8537
8497
|
"""
|
|
8538
8498
|
return pulumi.get(self, "prometheus")
|
|
8539
8499
|
|
|
@@ -8541,7 +8501,7 @@ class KafkaKafkaUserConfigPublicAccess(dict):
|
|
|
8541
8501
|
@pulumi.getter(name="schemaRegistry")
|
|
8542
8502
|
def schema_registry(self) -> Optional[bool]:
|
|
8543
8503
|
"""
|
|
8544
|
-
Allow clients to connect to schema_registry
|
|
8504
|
+
Allow clients to connect to schema_registry from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
8545
8505
|
"""
|
|
8546
8506
|
return pulumi.get(self, "schema_registry")
|
|
8547
8507
|
|
|
@@ -8791,12 +8751,6 @@ class KafkaMirrorMakerComponent(dict):
|
|
|
8791
8751
|
return pulumi.get(self, "usage")
|
|
8792
8752
|
|
|
8793
8753
|
|
|
8794
|
-
@pulumi.output_type
|
|
8795
|
-
class KafkaMirrorMakerKafkaMirrormaker(dict):
|
|
8796
|
-
def __init__(__self__):
|
|
8797
|
-
pass
|
|
8798
|
-
|
|
8799
|
-
|
|
8800
8754
|
@pulumi.output_type
|
|
8801
8755
|
class KafkaMirrorMakerKafkaMirrormakerUserConfig(dict):
|
|
8802
8756
|
@staticmethod
|
|
@@ -9018,7 +8972,7 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormaker(dict):
|
|
|
9018
8972
|
:param int refresh_groups_interval_seconds: Frequency of consumer group refresh in seconds. Defaults to 600 seconds (10 minutes).
|
|
9019
8973
|
:param bool refresh_topics_enabled: Whether to periodically check for new topics and partitions. Defaults to 'true'.
|
|
9020
8974
|
:param int refresh_topics_interval_seconds: Frequency of topic and partitions refresh in seconds. Defaults to 600 seconds (10 minutes).
|
|
9021
|
-
:param bool sync_group_offsets_enabled: Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to
|
|
8975
|
+
:param bool sync_group_offsets_enabled: Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to _*consumer*offsets topic in target cluster, as long as no active consumers in that group are connected to the target cluster.
|
|
9022
8976
|
:param int sync_group_offsets_interval_seconds: Frequency at which consumer group offsets are synced (default: 60, every minute).
|
|
9023
8977
|
:param bool sync_topic_configs_enabled: Whether to periodically configure remote topics to match their corresponding upstream topics.
|
|
9024
8978
|
:param int tasks_max_per_cpu: 'tasks.max' is set to this multiplied by the number of CPUs in the service. The default value is `1`.
|
|
@@ -9126,7 +9080,7 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormaker(dict):
|
|
|
9126
9080
|
@pulumi.getter(name="syncGroupOffsetsEnabled")
|
|
9127
9081
|
def sync_group_offsets_enabled(self) -> Optional[bool]:
|
|
9128
9082
|
"""
|
|
9129
|
-
Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to
|
|
9083
|
+
Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to _*consumer*offsets topic in target cluster, as long as no active consumers in that group are connected to the target cluster.
|
|
9130
9084
|
"""
|
|
9131
9085
|
return pulumi.get(self, "sync_group_offsets_enabled")
|
|
9132
9086
|
|
|
@@ -9898,12 +9852,6 @@ class M3AggregatorComponent(dict):
|
|
|
9898
9852
|
return pulumi.get(self, "usage")
|
|
9899
9853
|
|
|
9900
9854
|
|
|
9901
|
-
@pulumi.output_type
|
|
9902
|
-
class M3AggregatorM3aggregator(dict):
|
|
9903
|
-
def __init__(__self__):
|
|
9904
|
-
pass
|
|
9905
|
-
|
|
9906
|
-
|
|
9907
9855
|
@pulumi.output_type
|
|
9908
9856
|
class M3AggregatorM3aggregatorUserConfig(dict):
|
|
9909
9857
|
@staticmethod
|
|
@@ -10288,12 +10236,6 @@ class M3DbComponent(dict):
|
|
|
10288
10236
|
return pulumi.get(self, "usage")
|
|
10289
10237
|
|
|
10290
10238
|
|
|
10291
|
-
@pulumi.output_type
|
|
10292
|
-
class M3DbM3db(dict):
|
|
10293
|
-
def __init__(__self__):
|
|
10294
|
-
pass
|
|
10295
|
-
|
|
10296
|
-
|
|
10297
10239
|
@pulumi.output_type
|
|
10298
10240
|
class M3DbM3dbUserConfig(dict):
|
|
10299
10241
|
@staticmethod
|
|
@@ -10421,9 +10363,6 @@ class M3DbM3dbUserConfig(dict):
|
|
|
10421
10363
|
"""
|
|
10422
10364
|
Additional Cloud Regions for Backup Replication.
|
|
10423
10365
|
"""
|
|
10424
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
10425
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
10426
|
-
|
|
10427
10366
|
return pulumi.get(self, "additional_backup_regions")
|
|
10428
10367
|
|
|
10429
10368
|
@property
|
|
@@ -10635,7 +10574,7 @@ class M3DbM3dbUserConfigLimits(dict):
|
|
|
10635
10574
|
"""
|
|
10636
10575
|
:param int max_recently_queried_series_blocks: The maximum number of blocks that can be read in a given lookback period.
|
|
10637
10576
|
:param int max_recently_queried_series_disk_bytes_read: The maximum number of disk bytes that can be read in a given lookback period.
|
|
10638
|
-
:param str max_recently_queried_series_lookback: The lookback period for '
|
|
10577
|
+
:param str max_recently_queried_series_lookback: The lookback period for 'max*recently*queried*series*blocks' and 'max*recently*queried*series*disk*bytes*read'.
|
|
10639
10578
|
:param int query_docs: The maximum number of docs fetched in single query.
|
|
10640
10579
|
:param bool query_require_exhaustive: When query limits are exceeded, whether to return error or return partial results.
|
|
10641
10580
|
:param int query_series: The maximum number of series fetched in single query.
|
|
@@ -10673,7 +10612,7 @@ class M3DbM3dbUserConfigLimits(dict):
|
|
|
10673
10612
|
@pulumi.getter(name="maxRecentlyQueriedSeriesLookback")
|
|
10674
10613
|
def max_recently_queried_series_lookback(self) -> Optional[str]:
|
|
10675
10614
|
"""
|
|
10676
|
-
The lookback period for '
|
|
10615
|
+
The lookback period for 'max*recently*queried*series*blocks' and 'max*recently*queried*series*disk*bytes*read'.
|
|
10677
10616
|
"""
|
|
10678
10617
|
return pulumi.get(self, "max_recently_queried_series_lookback")
|
|
10679
10618
|
|
|
@@ -11084,7 +11023,7 @@ class M3DbM3dbUserConfigRulesMapping(dict):
|
|
|
11084
11023
|
namespaces_strings: Optional[Sequence[str]] = None,
|
|
11085
11024
|
tags: Optional[Sequence['outputs.M3DbM3dbUserConfigRulesMappingTag']] = None):
|
|
11086
11025
|
"""
|
|
11087
|
-
:param str filter: Matching metric names with wildcards (using
|
|
11026
|
+
:param str filter: Matching metric names with wildcards (using **name**:wildcard) or matching tags and their (optionally wildcarded) values. For value, ! can be used at start of value for negation, and multiple filters can be supplied using space as separator.
|
|
11088
11027
|
:param Sequence[str] aggregations: List of aggregations to be applied.
|
|
11089
11028
|
:param bool drop: Only store the derived metric (as specified in the roll-up rules), if any.
|
|
11090
11029
|
:param str name: The (optional) name of the rule.
|
|
@@ -11113,7 +11052,7 @@ class M3DbM3dbUserConfigRulesMapping(dict):
|
|
|
11113
11052
|
@pulumi.getter
|
|
11114
11053
|
def filter(self) -> str:
|
|
11115
11054
|
"""
|
|
11116
|
-
Matching metric names with wildcards (using
|
|
11055
|
+
Matching metric names with wildcards (using **name**:wildcard) or matching tags and their (optionally wildcarded) values. For value, ! can be used at start of value for negation, and multiple filters can be supplied using space as separator.
|
|
11117
11056
|
"""
|
|
11118
11057
|
return pulumi.get(self, "filter")
|
|
11119
11058
|
|
|
@@ -11453,12 +11392,6 @@ class MySqlComponent(dict):
|
|
|
11453
11392
|
return pulumi.get(self, "usage")
|
|
11454
11393
|
|
|
11455
11394
|
|
|
11456
|
-
@pulumi.output_type
|
|
11457
|
-
class MySqlMysql(dict):
|
|
11458
|
-
def __init__(__self__):
|
|
11459
|
-
pass
|
|
11460
|
-
|
|
11461
|
-
|
|
11462
11395
|
@pulumi.output_type
|
|
11463
11396
|
class MySqlMysqlUserConfig(dict):
|
|
11464
11397
|
@staticmethod
|
|
@@ -12035,17 +11968,17 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12035
11968
|
:param int innodb_write_io_threads: The number of I/O threads for write operations in InnoDB. Default is 4. Changing this parameter will lead to a restart of the MySQL service.
|
|
12036
11969
|
:param int interactive_timeout: The number of seconds the server waits for activity on an interactive connection before closing it.
|
|
12037
11970
|
:param str internal_tmp_mem_storage_engine: The storage engine for in-memory internal temporary tables.
|
|
12038
|
-
:param float long_query_time: The
|
|
11971
|
+
:param float long_query_time: The slow*query*logs work as SQL statements that take more than long*query*time seconds to execute. Default is 10s.
|
|
12039
11972
|
:param int max_allowed_packet: Size of the largest message in bytes that can be received by the server. Default is 67108864 (64M).
|
|
12040
|
-
:param int max_heap_table_size: Limits the size of internal in-memory tables. Also set
|
|
11973
|
+
:param int max_heap_table_size: Limits the size of internal in-memory tables. Also set tmp*table*size. Default is 16777216 (16M).
|
|
12041
11974
|
:param int net_buffer_length: Start sizes of connection buffer and result buffer. Default is 16384 (16K). Changing this parameter will lead to a restart of the MySQL service.
|
|
12042
11975
|
:param int net_read_timeout: The number of seconds to wait for more data from a connection before aborting the read.
|
|
12043
11976
|
:param int net_write_timeout: The number of seconds to wait for a block to be written to a connection before aborting the write.
|
|
12044
|
-
:param bool slow_query_log: Slow query log enables capturing of slow queries. Setting
|
|
11977
|
+
:param bool slow_query_log: Slow query log enables capturing of slow queries. Setting slow*query*log to false also truncates the mysql.slow_log table. Default is off.
|
|
12045
11978
|
:param int sort_buffer_size: Sort buffer size in bytes for ORDER BY optimization. Default is 262144 (256K).
|
|
12046
11979
|
:param str sql_mode: Global SQL mode. Set to empty to use MySQL server defaults. When creating a new service and not setting this field Aiven default SQL mode (strict, SQL standard compliant) will be assigned.
|
|
12047
11980
|
:param bool sql_require_primary_key: Require primary key to be defined for new tables or old tables modified with ALTER TABLE and fail if missing. It is recommended to always have primary keys because various functionality may break if any large table is missing them.
|
|
12048
|
-
:param int tmp_table_size: Limits the size of internal in-memory tables. Also set
|
|
11981
|
+
:param int tmp_table_size: Limits the size of internal in-memory tables. Also set max*heap*table_size. Default is 16777216 (16M).
|
|
12049
11982
|
:param int wait_timeout: The number of seconds the server waits for activity on a noninteractive connection before closing it.
|
|
12050
11983
|
"""
|
|
12051
11984
|
if connect_timeout is not None:
|
|
@@ -12257,7 +12190,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12257
12190
|
@pulumi.getter(name="longQueryTime")
|
|
12258
12191
|
def long_query_time(self) -> Optional[float]:
|
|
12259
12192
|
"""
|
|
12260
|
-
The
|
|
12193
|
+
The slow*query*logs work as SQL statements that take more than long*query*time seconds to execute. Default is 10s.
|
|
12261
12194
|
"""
|
|
12262
12195
|
return pulumi.get(self, "long_query_time")
|
|
12263
12196
|
|
|
@@ -12273,7 +12206,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12273
12206
|
@pulumi.getter(name="maxHeapTableSize")
|
|
12274
12207
|
def max_heap_table_size(self) -> Optional[int]:
|
|
12275
12208
|
"""
|
|
12276
|
-
Limits the size of internal in-memory tables. Also set
|
|
12209
|
+
Limits the size of internal in-memory tables. Also set tmp*table*size. Default is 16777216 (16M).
|
|
12277
12210
|
"""
|
|
12278
12211
|
return pulumi.get(self, "max_heap_table_size")
|
|
12279
12212
|
|
|
@@ -12305,7 +12238,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12305
12238
|
@pulumi.getter(name="slowQueryLog")
|
|
12306
12239
|
def slow_query_log(self) -> Optional[bool]:
|
|
12307
12240
|
"""
|
|
12308
|
-
Slow query log enables capturing of slow queries. Setting
|
|
12241
|
+
Slow query log enables capturing of slow queries. Setting slow*query*log to false also truncates the mysql.slow_log table. Default is off.
|
|
12309
12242
|
"""
|
|
12310
12243
|
return pulumi.get(self, "slow_query_log")
|
|
12311
12244
|
|
|
@@ -12337,7 +12270,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12337
12270
|
@pulumi.getter(name="tmpTableSize")
|
|
12338
12271
|
def tmp_table_size(self) -> Optional[int]:
|
|
12339
12272
|
"""
|
|
12340
|
-
Limits the size of internal in-memory tables. Also set
|
|
12273
|
+
Limits the size of internal in-memory tables. Also set max*heap*table_size. Default is 16777216 (16M).
|
|
12341
12274
|
"""
|
|
12342
12275
|
return pulumi.get(self, "tmp_table_size")
|
|
12343
12276
|
|
|
@@ -12823,7 +12756,7 @@ class OpenSearchOpensearchUserConfig(dict):
|
|
|
12823
12756
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
12824
12757
|
:param Sequence[str] ip_filters: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
12825
12758
|
:param bool keep_index_refresh_interval: Aiven automation resets index.refresh_interval to default value for every index to be sure that indices are always visible to search. If it doesn't fit your case, you can disable this by setting up this flag to true.
|
|
12826
|
-
:param int max_index_count:
|
|
12759
|
+
:param int max_index_count: Use index_patterns instead. The default value is `0`.
|
|
12827
12760
|
:param 'OpenSearchOpensearchUserConfigOpenidArgs' openid: OpenSearch OpenID Connect Configuration
|
|
12828
12761
|
:param 'OpenSearchOpensearchUserConfigOpensearchArgs' opensearch: OpenSearch settings
|
|
12829
12762
|
:param 'OpenSearchOpensearchUserConfigOpensearchDashboardsArgs' opensearch_dashboards: OpenSearch Dashboards settings
|
|
@@ -12891,9 +12824,6 @@ class OpenSearchOpensearchUserConfig(dict):
|
|
|
12891
12824
|
"""
|
|
12892
12825
|
Additional Cloud Regions for Backup Replication.
|
|
12893
12826
|
"""
|
|
12894
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
12895
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
12896
|
-
|
|
12897
12827
|
return pulumi.get(self, "additional_backup_regions")
|
|
12898
12828
|
|
|
12899
12829
|
@property
|
|
@@ -12967,7 +12897,7 @@ class OpenSearchOpensearchUserConfig(dict):
|
|
|
12967
12897
|
@pulumi.getter(name="maxIndexCount")
|
|
12968
12898
|
def max_index_count(self) -> Optional[int]:
|
|
12969
12899
|
"""
|
|
12970
|
-
|
|
12900
|
+
Use index_patterns instead. The default value is `0`.
|
|
12971
12901
|
"""
|
|
12972
12902
|
return pulumi.get(self, "max_index_count")
|
|
12973
12903
|
|
|
@@ -13568,10 +13498,10 @@ class OpenSearchOpensearchUserConfigOpensearch(dict):
|
|
|
13568
13498
|
:param int http_max_header_size: The max size of allowed headers, in bytes.
|
|
13569
13499
|
:param int http_max_initial_line_length: The max length of an HTTP URL, in bytes.
|
|
13570
13500
|
:param int indices_fielddata_cache_size: Relative amount. Maximum amount of heap memory used for field data cache. This is an expert setting; decreasing the value too much will increase overhead of loading field data; too much memory used for field data cache will decrease amount of heap available for other operations.
|
|
13571
|
-
:param int indices_memory_index_buffer_size: Percentage value. Default is 10
|
|
13501
|
+
:param int indices_memory_index_buffer_size: Percentage value. Default is 10%. Total amount of heap used for indexing buffer, before writing segments to disk. This is an expert setting. Too low value will slow down indexing; too high value will increase indexing performance but causes performance issues for query performance.
|
|
13572
13502
|
:param int indices_memory_max_index_buffer_size: Absolute value. Default is unbound. Doesn't work without indices.memory.index*buffer*size. Maximum amount of heap used for query cache, an absolute indices.memory.index*buffer*size maximum hard limit.
|
|
13573
13503
|
:param int indices_memory_min_index_buffer_size: Absolute value. Default is 48mb. Doesn't work without indices.memory.index*buffer*size. Minimum amount of heap used for query cache, an absolute indices.memory.index*buffer*size minimal hard limit.
|
|
13574
|
-
:param int indices_queries_cache_size: Percentage value. Default is 10
|
|
13504
|
+
:param int indices_queries_cache_size: Percentage value. Default is 10%. Maximum amount of heap used for query cache. This is an expert setting. Too low value will decrease query performance and increase performance for other operations; too high value will cause issues with other OpenSearch functionality.
|
|
13575
13505
|
:param int indices_query_bool_max_clause_count: Maximum number of clauses Lucene BooleanQuery can have. The default value (1024) is relatively high, and increasing it may cause performance issues. Investigate other approaches first before increasing this value.
|
|
13576
13506
|
:param int indices_recovery_max_bytes_per_sec: Limits total inbound and outbound recovery traffic for each node. Applies to both peer recoveries as well as snapshot recoveries (i.e., restores from a snapshot). Defaults to 40mb.
|
|
13577
13507
|
:param int indices_recovery_max_concurrent_file_chunks: Number of file chunks sent in parallel for each recovery. Defaults to 2.
|
|
@@ -13791,7 +13721,7 @@ class OpenSearchOpensearchUserConfigOpensearch(dict):
|
|
|
13791
13721
|
@pulumi.getter(name="indicesMemoryIndexBufferSize")
|
|
13792
13722
|
def indices_memory_index_buffer_size(self) -> Optional[int]:
|
|
13793
13723
|
"""
|
|
13794
|
-
Percentage value. Default is 10
|
|
13724
|
+
Percentage value. Default is 10%. Total amount of heap used for indexing buffer, before writing segments to disk. This is an expert setting. Too low value will slow down indexing; too high value will increase indexing performance but causes performance issues for query performance.
|
|
13795
13725
|
"""
|
|
13796
13726
|
return pulumi.get(self, "indices_memory_index_buffer_size")
|
|
13797
13727
|
|
|
@@ -13815,7 +13745,7 @@ class OpenSearchOpensearchUserConfigOpensearch(dict):
|
|
|
13815
13745
|
@pulumi.getter(name="indicesQueriesCacheSize")
|
|
13816
13746
|
def indices_queries_cache_size(self) -> Optional[int]:
|
|
13817
13747
|
"""
|
|
13818
|
-
Percentage value. Default is 10
|
|
13748
|
+
Percentage value. Default is 10%. Maximum amount of heap used for query cache. This is an expert setting. Too low value will decrease query performance and increase performance for other operations; too high value will cause issues with other OpenSearch functionality.
|
|
13819
13749
|
"""
|
|
13820
13750
|
return pulumi.get(self, "indices_queries_cache_size")
|
|
13821
13751
|
|
|
@@ -14221,10 +14151,10 @@ class OpenSearchOpensearchUserConfigOpensearchAuthFailureListenersIpRateLimiting
|
|
|
14221
14151
|
"""
|
|
14222
14152
|
:param int allowed_tries: The number of login attempts allowed before login is blocked.
|
|
14223
14153
|
:param int block_expiry_seconds: The duration of time that login remains blocked after a failed login.
|
|
14224
|
-
:param int max_blocked_clients:
|
|
14154
|
+
:param int max_blocked_clients: The maximum number of blocked IP addresses.
|
|
14225
14155
|
:param int max_tracked_clients: The maximum number of tracked IP addresses that have failed login.
|
|
14226
14156
|
:param int time_window_seconds: The window of time in which the value for `allowed_tries` is enforced.
|
|
14227
|
-
:param str type:
|
|
14157
|
+
:param str type: The type of rate limiting.
|
|
14228
14158
|
"""
|
|
14229
14159
|
if allowed_tries is not None:
|
|
14230
14160
|
pulumi.set(__self__, "allowed_tries", allowed_tries)
|
|
@@ -14259,7 +14189,7 @@ class OpenSearchOpensearchUserConfigOpensearchAuthFailureListenersIpRateLimiting
|
|
|
14259
14189
|
@pulumi.getter(name="maxBlockedClients")
|
|
14260
14190
|
def max_blocked_clients(self) -> Optional[int]:
|
|
14261
14191
|
"""
|
|
14262
|
-
|
|
14192
|
+
The maximum number of blocked IP addresses.
|
|
14263
14193
|
"""
|
|
14264
14194
|
return pulumi.get(self, "max_blocked_clients")
|
|
14265
14195
|
|
|
@@ -14283,7 +14213,7 @@ class OpenSearchOpensearchUserConfigOpensearchAuthFailureListenersIpRateLimiting
|
|
|
14283
14213
|
@pulumi.getter
|
|
14284
14214
|
def type(self) -> Optional[str]:
|
|
14285
14215
|
"""
|
|
14286
|
-
|
|
14216
|
+
The type of rate limiting.
|
|
14287
14217
|
"""
|
|
14288
14218
|
return pulumi.get(self, "type")
|
|
14289
14219
|
|
|
@@ -14314,7 +14244,7 @@ class OpenSearchOpensearchUserConfigOpensearchDashboards(dict):
|
|
|
14314
14244
|
max_old_space_size: Optional[int] = None,
|
|
14315
14245
|
opensearch_request_timeout: Optional[int] = None):
|
|
14316
14246
|
"""
|
|
14317
|
-
:param bool enabled:
|
|
14247
|
+
:param bool enabled: Enable or disable OpenSearch Dashboards. The default value is `true`.
|
|
14318
14248
|
:param int max_old_space_size: Limits the maximum amount of memory (in MiB) the OpenSearch Dashboards process can use. This sets the max*old*space_size option of the nodejs running the OpenSearch Dashboards. Note: the memory reserved by OpenSearch Dashboards is not available for OpenSearch. The default value is `128`.
|
|
14319
14249
|
:param int opensearch_request_timeout: Timeout in milliseconds for requests made by OpenSearch Dashboards towards OpenSearch. The default value is `30000`.
|
|
14320
14250
|
"""
|
|
@@ -14329,7 +14259,7 @@ class OpenSearchOpensearchUserConfigOpensearchDashboards(dict):
|
|
|
14329
14259
|
@pulumi.getter
|
|
14330
14260
|
def enabled(self) -> Optional[bool]:
|
|
14331
14261
|
"""
|
|
14332
|
-
|
|
14262
|
+
Enable or disable OpenSearch Dashboards. The default value is `true`.
|
|
14333
14263
|
"""
|
|
14334
14264
|
return pulumi.get(self, "enabled")
|
|
14335
14265
|
|
|
@@ -14374,8 +14304,8 @@ class OpenSearchOpensearchUserConfigPrivateAccess(dict):
|
|
|
14374
14304
|
opensearch_dashboards: Optional[bool] = None,
|
|
14375
14305
|
prometheus: Optional[bool] = None):
|
|
14376
14306
|
"""
|
|
14377
|
-
:param bool opensearch:
|
|
14378
|
-
:param bool opensearch_dashboards:
|
|
14307
|
+
:param bool opensearch: Allow clients to connect to opensearch with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
14308
|
+
:param bool opensearch_dashboards: Allow clients to connect to opensearch_dashboards with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
14379
14309
|
:param bool prometheus: Allow clients to connect to prometheus with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
14380
14310
|
"""
|
|
14381
14311
|
if opensearch is not None:
|
|
@@ -14389,7 +14319,7 @@ class OpenSearchOpensearchUserConfigPrivateAccess(dict):
|
|
|
14389
14319
|
@pulumi.getter
|
|
14390
14320
|
def opensearch(self) -> Optional[bool]:
|
|
14391
14321
|
"""
|
|
14392
|
-
|
|
14322
|
+
Allow clients to connect to opensearch with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
14393
14323
|
"""
|
|
14394
14324
|
return pulumi.get(self, "opensearch")
|
|
14395
14325
|
|
|
@@ -14397,7 +14327,7 @@ class OpenSearchOpensearchUserConfigPrivateAccess(dict):
|
|
|
14397
14327
|
@pulumi.getter(name="opensearchDashboards")
|
|
14398
14328
|
def opensearch_dashboards(self) -> Optional[bool]:
|
|
14399
14329
|
"""
|
|
14400
|
-
|
|
14330
|
+
Allow clients to connect to opensearch_dashboards with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
14401
14331
|
"""
|
|
14402
14332
|
return pulumi.get(self, "opensearch_dashboards")
|
|
14403
14333
|
|
|
@@ -14434,9 +14364,9 @@ class OpenSearchOpensearchUserConfigPrivatelinkAccess(dict):
|
|
|
14434
14364
|
opensearch_dashboards: Optional[bool] = None,
|
|
14435
14365
|
prometheus: Optional[bool] = None):
|
|
14436
14366
|
"""
|
|
14437
|
-
:param bool opensearch:
|
|
14438
|
-
:param bool opensearch_dashboards:
|
|
14439
|
-
:param bool prometheus:
|
|
14367
|
+
:param bool opensearch: Enable opensearch.
|
|
14368
|
+
:param bool opensearch_dashboards: Enable opensearch_dashboards.
|
|
14369
|
+
:param bool prometheus: Enable prometheus.
|
|
14440
14370
|
"""
|
|
14441
14371
|
if opensearch is not None:
|
|
14442
14372
|
pulumi.set(__self__, "opensearch", opensearch)
|
|
@@ -14449,7 +14379,7 @@ class OpenSearchOpensearchUserConfigPrivatelinkAccess(dict):
|
|
|
14449
14379
|
@pulumi.getter
|
|
14450
14380
|
def opensearch(self) -> Optional[bool]:
|
|
14451
14381
|
"""
|
|
14452
|
-
|
|
14382
|
+
Enable opensearch.
|
|
14453
14383
|
"""
|
|
14454
14384
|
return pulumi.get(self, "opensearch")
|
|
14455
14385
|
|
|
@@ -14457,7 +14387,7 @@ class OpenSearchOpensearchUserConfigPrivatelinkAccess(dict):
|
|
|
14457
14387
|
@pulumi.getter(name="opensearchDashboards")
|
|
14458
14388
|
def opensearch_dashboards(self) -> Optional[bool]:
|
|
14459
14389
|
"""
|
|
14460
|
-
|
|
14390
|
+
Enable opensearch_dashboards.
|
|
14461
14391
|
"""
|
|
14462
14392
|
return pulumi.get(self, "opensearch_dashboards")
|
|
14463
14393
|
|
|
@@ -14465,7 +14395,7 @@ class OpenSearchOpensearchUserConfigPrivatelinkAccess(dict):
|
|
|
14465
14395
|
@pulumi.getter
|
|
14466
14396
|
def prometheus(self) -> Optional[bool]:
|
|
14467
14397
|
"""
|
|
14468
|
-
|
|
14398
|
+
Enable prometheus.
|
|
14469
14399
|
"""
|
|
14470
14400
|
return pulumi.get(self, "prometheus")
|
|
14471
14401
|
|
|
@@ -14494,9 +14424,9 @@ class OpenSearchOpensearchUserConfigPublicAccess(dict):
|
|
|
14494
14424
|
opensearch_dashboards: Optional[bool] = None,
|
|
14495
14425
|
prometheus: Optional[bool] = None):
|
|
14496
14426
|
"""
|
|
14497
|
-
:param bool opensearch:
|
|
14498
|
-
:param bool opensearch_dashboards:
|
|
14499
|
-
:param bool prometheus: Allow clients to connect to prometheus
|
|
14427
|
+
:param bool opensearch: Allow clients to connect to opensearch from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
14428
|
+
:param bool opensearch_dashboards: Allow clients to connect to opensearch_dashboards from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
14429
|
+
:param bool prometheus: Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
14500
14430
|
"""
|
|
14501
14431
|
if opensearch is not None:
|
|
14502
14432
|
pulumi.set(__self__, "opensearch", opensearch)
|
|
@@ -14509,7 +14439,7 @@ class OpenSearchOpensearchUserConfigPublicAccess(dict):
|
|
|
14509
14439
|
@pulumi.getter
|
|
14510
14440
|
def opensearch(self) -> Optional[bool]:
|
|
14511
14441
|
"""
|
|
14512
|
-
|
|
14442
|
+
Allow clients to connect to opensearch from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
14513
14443
|
"""
|
|
14514
14444
|
return pulumi.get(self, "opensearch")
|
|
14515
14445
|
|
|
@@ -14517,7 +14447,7 @@ class OpenSearchOpensearchUserConfigPublicAccess(dict):
|
|
|
14517
14447
|
@pulumi.getter(name="opensearchDashboards")
|
|
14518
14448
|
def opensearch_dashboards(self) -> Optional[bool]:
|
|
14519
14449
|
"""
|
|
14520
|
-
|
|
14450
|
+
Allow clients to connect to opensearch_dashboards from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
14521
14451
|
"""
|
|
14522
14452
|
return pulumi.get(self, "opensearch_dashboards")
|
|
14523
14453
|
|
|
@@ -14525,7 +14455,7 @@ class OpenSearchOpensearchUserConfigPublicAccess(dict):
|
|
|
14525
14455
|
@pulumi.getter
|
|
14526
14456
|
def prometheus(self) -> Optional[bool]:
|
|
14527
14457
|
"""
|
|
14528
|
-
Allow clients to connect to prometheus
|
|
14458
|
+
Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
14529
14459
|
"""
|
|
14530
14460
|
return pulumi.get(self, "prometheus")
|
|
14531
14461
|
|
|
@@ -14568,13 +14498,13 @@ class OpenSearchOpensearchUserConfigSaml(dict):
|
|
|
14568
14498
|
roles_key: Optional[str] = None,
|
|
14569
14499
|
subject_key: Optional[str] = None):
|
|
14570
14500
|
"""
|
|
14571
|
-
:param bool enabled: Enables or disables
|
|
14501
|
+
:param bool enabled: Enables or disables SAML-based authentication for OpenSearch. When enabled, users can authenticate using SAML with an Identity Provider. The default value is `true`.
|
|
14572
14502
|
:param str idp_entity_id: The unique identifier for the Identity Provider (IdP) entity that is used for SAML authentication. This value is typically provided by the IdP.
|
|
14573
14503
|
:param str idp_metadata_url: The URL of the SAML metadata for the Identity Provider (IdP). This is used to configure SAML-based authentication with the IdP.
|
|
14574
14504
|
:param str sp_entity_id: The unique identifier for the Service Provider (SP) entity that is used for SAML authentication. This value is typically provided by the SP.
|
|
14575
14505
|
:param str idp_pemtrustedcas_content: This parameter specifies the PEM-encoded root certificate authority (CA) content for the SAML identity provider (IdP) server verification. The root CA content is used to verify the SSL/TLS certificate presented by the server.
|
|
14576
|
-
:param str roles_key:
|
|
14577
|
-
:param str subject_key:
|
|
14506
|
+
:param str roles_key: Optional. Specifies the attribute in the SAML response where role information is stored, if available. Role attributes are not required for SAML authentication, but can be included in SAML assertions by most Identity Providers (IdPs) to determine user access levels or permissions.
|
|
14507
|
+
:param str subject_key: Optional. Specifies the attribute in the SAML response where the subject identifier is stored. If not configured, the NameID attribute is used by default.
|
|
14578
14508
|
"""
|
|
14579
14509
|
pulumi.set(__self__, "enabled", enabled)
|
|
14580
14510
|
pulumi.set(__self__, "idp_entity_id", idp_entity_id)
|
|
@@ -14591,7 +14521,7 @@ class OpenSearchOpensearchUserConfigSaml(dict):
|
|
|
14591
14521
|
@pulumi.getter
|
|
14592
14522
|
def enabled(self) -> bool:
|
|
14593
14523
|
"""
|
|
14594
|
-
Enables or disables
|
|
14524
|
+
Enables or disables SAML-based authentication for OpenSearch. When enabled, users can authenticate using SAML with an Identity Provider. The default value is `true`.
|
|
14595
14525
|
"""
|
|
14596
14526
|
return pulumi.get(self, "enabled")
|
|
14597
14527
|
|
|
@@ -14631,7 +14561,7 @@ class OpenSearchOpensearchUserConfigSaml(dict):
|
|
|
14631
14561
|
@pulumi.getter(name="rolesKey")
|
|
14632
14562
|
def roles_key(self) -> Optional[str]:
|
|
14633
14563
|
"""
|
|
14634
|
-
|
|
14564
|
+
Optional. Specifies the attribute in the SAML response where role information is stored, if available. Role attributes are not required for SAML authentication, but can be included in SAML assertions by most Identity Providers (IdPs) to determine user access levels or permissions.
|
|
14635
14565
|
"""
|
|
14636
14566
|
return pulumi.get(self, "roles_key")
|
|
14637
14567
|
|
|
@@ -14639,7 +14569,7 @@ class OpenSearchOpensearchUserConfigSaml(dict):
|
|
|
14639
14569
|
@pulumi.getter(name="subjectKey")
|
|
14640
14570
|
def subject_key(self) -> Optional[str]:
|
|
14641
14571
|
"""
|
|
14642
|
-
|
|
14572
|
+
Optional. Specifies the attribute in the SAML response where the subject identifier is stored. If not configured, the NameID attribute is used by default.
|
|
14643
14573
|
"""
|
|
14644
14574
|
return pulumi.get(self, "subject_key")
|
|
14645
14575
|
|
|
@@ -15287,12 +15217,12 @@ class PgPgUserConfig(dict):
|
|
|
15287
15217
|
:param str recovery_target_time: Recovery target time when forking a service. This has effect only when a new service is being created.
|
|
15288
15218
|
:param bool service_log: Store logs for the service so that they are available in the HTTP API and console.
|
|
15289
15219
|
:param str service_to_fork_from: Name of another service to fork from. This has effect only when a new service is being created.
|
|
15290
|
-
:param float shared_buffers_percentage: Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20%. This setting adjusts the shared_buffers configuration value.
|
|
15220
|
+
:param float shared_buffers_percentage: Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20% - 60%. This setting adjusts the shared_buffers configuration value.
|
|
15291
15221
|
:param bool static_ips: Use static public IP addresses.
|
|
15292
15222
|
:param str synchronous_replication: Synchronous replication type. Note that the service plan also needs to support synchronous replication.
|
|
15293
15223
|
:param 'PgPgUserConfigTimescaledbArgs' timescaledb: System-wide settings for the timescaledb extension
|
|
15294
15224
|
:param str variant: Variant of the PostgreSQL service, may affect the features that are exposed by default.
|
|
15295
|
-
:param int work_mem: Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075
|
|
15225
|
+
:param int work_mem: Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075% of total RAM (up to 32MB).
|
|
15296
15226
|
"""
|
|
15297
15227
|
if additional_backup_regions is not None:
|
|
15298
15228
|
pulumi.set(__self__, "additional_backup_regions", additional_backup_regions)
|
|
@@ -15577,7 +15507,7 @@ class PgPgUserConfig(dict):
|
|
|
15577
15507
|
@pulumi.getter(name="sharedBuffersPercentage")
|
|
15578
15508
|
def shared_buffers_percentage(self) -> Optional[float]:
|
|
15579
15509
|
"""
|
|
15580
|
-
Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20%. This setting adjusts the shared_buffers configuration value.
|
|
15510
|
+
Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20% - 60%. This setting adjusts the shared_buffers configuration value.
|
|
15581
15511
|
"""
|
|
15582
15512
|
return pulumi.get(self, "shared_buffers_percentage")
|
|
15583
15513
|
|
|
@@ -15617,7 +15547,7 @@ class PgPgUserConfig(dict):
|
|
|
15617
15547
|
@pulumi.getter(name="workMem")
|
|
15618
15548
|
def work_mem(self) -> Optional[int]:
|
|
15619
15549
|
"""
|
|
15620
|
-
Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075
|
|
15550
|
+
Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075% of total RAM (up to 32MB).
|
|
15621
15551
|
"""
|
|
15622
15552
|
return pulumi.get(self, "work_mem")
|
|
15623
15553
|
|
|
@@ -15683,10 +15613,10 @@ class PgPgUserConfigMigration(dict):
|
|
|
15683
15613
|
"""
|
|
15684
15614
|
:param str host: Hostname or IP address of the server where to migrate data from.
|
|
15685
15615
|
:param int port: Port number of the server where to migrate data from.
|
|
15686
|
-
:param str dbname:
|
|
15616
|
+
:param str dbname: Database name for bootstrapping the initial connection.
|
|
15687
15617
|
:param str ignore_dbs: Comma-separated list of databases, which should be ignored during migration (supported by MySQL and PostgreSQL only at the moment).
|
|
15688
15618
|
:param str method: The migration method to be used (currently supported only by Redis, Dragonfly, MySQL and PostgreSQL service types).
|
|
15689
|
-
:param str password:
|
|
15619
|
+
:param str password: Password for authentication with the server where to migrate data from.
|
|
15690
15620
|
:param bool ssl: The server where to migrate data from is secured with SSL. The default value is `true`.
|
|
15691
15621
|
:param str username: User name for authentication with the server where to migrate data from.
|
|
15692
15622
|
"""
|
|
@@ -15725,7 +15655,7 @@ class PgPgUserConfigMigration(dict):
|
|
|
15725
15655
|
@pulumi.getter
|
|
15726
15656
|
def dbname(self) -> Optional[str]:
|
|
15727
15657
|
"""
|
|
15728
|
-
|
|
15658
|
+
Database name for bootstrapping the initial connection.
|
|
15729
15659
|
"""
|
|
15730
15660
|
return pulumi.get(self, "dbname")
|
|
15731
15661
|
|
|
@@ -15749,7 +15679,7 @@ class PgPgUserConfigMigration(dict):
|
|
|
15749
15679
|
@pulumi.getter
|
|
15750
15680
|
def password(self) -> Optional[str]:
|
|
15751
15681
|
"""
|
|
15752
|
-
|
|
15682
|
+
Password for authentication with the server where to migrate data from.
|
|
15753
15683
|
"""
|
|
15754
15684
|
return pulumi.get(self, "password")
|
|
15755
15685
|
|
|
@@ -15932,14 +15862,14 @@ class PgPgUserConfigPg(dict):
|
|
|
15932
15862
|
wal_sender_timeout: Optional[int] = None,
|
|
15933
15863
|
wal_writer_delay: Optional[int] = None):
|
|
15934
15864
|
"""
|
|
15935
|
-
:param float autovacuum_analyze_scale_factor: Specifies a fraction of the table size to add to autovacuum*analyze*threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20
|
|
15865
|
+
:param float autovacuum_analyze_scale_factor: Specifies a fraction of the table size to add to autovacuum*analyze*threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20% of table size).
|
|
15936
15866
|
:param int autovacuum_analyze_threshold: Specifies the minimum number of inserted, updated or deleted tuples needed to trigger an ANALYZE in any one table. The default is 50 tuples.
|
|
15937
15867
|
:param int autovacuum_freeze_max_age: Specifies the maximum age (in transactions) that a table's pg_class.relfrozenxid field can attain before a VACUUM operation is forced to prevent transaction ID wraparound within the table. Note that the system will launch autovacuum processes to prevent wraparound even when autovacuum is otherwise disabled. This parameter will cause the server to be restarted.
|
|
15938
15868
|
:param int autovacuum_max_workers: Specifies the maximum number of autovacuum processes (other than the autovacuum launcher) that may be running at any one time. The default is three. This parameter can only be set at server start.
|
|
15939
15869
|
:param int autovacuum_naptime: Specifies the minimum delay between autovacuum runs on any given database. The delay is measured in seconds, and the default is one minute.
|
|
15940
15870
|
:param int autovacuum_vacuum_cost_delay: Specifies the cost delay value that will be used in automatic VACUUM operations. If -1 is specified, the regular vacuum*cost*delay value will be used. The default value is 20 milliseconds.
|
|
15941
15871
|
:param int autovacuum_vacuum_cost_limit: Specifies the cost limit value that will be used in automatic VACUUM operations. If -1 is specified (which is the default), the regular vacuum*cost*limit value will be used.
|
|
15942
|
-
:param float autovacuum_vacuum_scale_factor: Specifies a fraction of the table size to add to autovacuum*vacuum*threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20
|
|
15872
|
+
:param float autovacuum_vacuum_scale_factor: Specifies a fraction of the table size to add to autovacuum*vacuum*threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20% of table size).
|
|
15943
15873
|
:param int autovacuum_vacuum_threshold: Specifies the minimum number of updated or deleted tuples needed to trigger a VACUUM in any one table. The default is 50 tuples.
|
|
15944
15874
|
:param int bgwriter_delay: Specifies the delay between activity rounds for the background writer in milliseconds. Default is 200.
|
|
15945
15875
|
:param int bgwriter_flush_after: Whenever more than bgwriter*flush*after bytes have been written by the background writer, attempt to force the OS to issue these writes to the underlying storage. Specified in kilobytes, default is 512. Setting of 0 disables forced writeback.
|
|
@@ -15971,7 +15901,7 @@ class PgPgUserConfigPg(dict):
|
|
|
15971
15901
|
:param int pg_partman_bgw_dot_interval: Sets the time interval to run pg_partman's scheduled tasks.
|
|
15972
15902
|
:param str pg_partman_bgw_dot_role: Controls which role to use for pg_partman's scheduled background tasks.
|
|
15973
15903
|
:param bool pg_stat_monitor_dot_pgsm_enable_query_plan: Enables or disables query plan monitoring.
|
|
15974
|
-
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets
|
|
15904
|
+
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets.
|
|
15975
15905
|
:param str pg_stat_statements_dot_track: Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top.
|
|
15976
15906
|
:param int temp_file_limit: PostgreSQL temporary file limit in KiB, -1 for unlimited.
|
|
15977
15907
|
:param str timezone: PostgreSQL service timezone.
|
|
@@ -16085,7 +16015,7 @@ class PgPgUserConfigPg(dict):
|
|
|
16085
16015
|
@pulumi.getter(name="autovacuumAnalyzeScaleFactor")
|
|
16086
16016
|
def autovacuum_analyze_scale_factor(self) -> Optional[float]:
|
|
16087
16017
|
"""
|
|
16088
|
-
Specifies a fraction of the table size to add to autovacuum*analyze*threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20
|
|
16018
|
+
Specifies a fraction of the table size to add to autovacuum*analyze*threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20% of table size).
|
|
16089
16019
|
"""
|
|
16090
16020
|
return pulumi.get(self, "autovacuum_analyze_scale_factor")
|
|
16091
16021
|
|
|
@@ -16141,7 +16071,7 @@ class PgPgUserConfigPg(dict):
|
|
|
16141
16071
|
@pulumi.getter(name="autovacuumVacuumScaleFactor")
|
|
16142
16072
|
def autovacuum_vacuum_scale_factor(self) -> Optional[float]:
|
|
16143
16073
|
"""
|
|
16144
|
-
Specifies a fraction of the table size to add to autovacuum*vacuum*threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20
|
|
16074
|
+
Specifies a fraction of the table size to add to autovacuum*vacuum*threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20% of table size).
|
|
16145
16075
|
"""
|
|
16146
16076
|
return pulumi.get(self, "autovacuum_vacuum_scale_factor")
|
|
16147
16077
|
|
|
@@ -16397,7 +16327,7 @@ class PgPgUserConfigPg(dict):
|
|
|
16397
16327
|
@pulumi.getter(name="pgStatMonitorDotPgsmMaxBuckets")
|
|
16398
16328
|
def pg_stat_monitor_dot_pgsm_max_buckets(self) -> Optional[int]:
|
|
16399
16329
|
"""
|
|
16400
|
-
Sets the maximum number of buckets
|
|
16330
|
+
Sets the maximum number of buckets.
|
|
16401
16331
|
"""
|
|
16402
16332
|
return pulumi.get(self, "pg_stat_monitor_dot_pgsm_max_buckets")
|
|
16403
16333
|
|
|
@@ -16984,8 +16914,8 @@ class PgPgUserConfigPrivateAccess(dict):
|
|
|
16984
16914
|
pgbouncer: Optional[bool] = None,
|
|
16985
16915
|
prometheus: Optional[bool] = None):
|
|
16986
16916
|
"""
|
|
16987
|
-
:param bool pg:
|
|
16988
|
-
:param bool pgbouncer:
|
|
16917
|
+
:param bool pg: Allow clients to connect to pg with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
16918
|
+
:param bool pgbouncer: Allow clients to connect to pgbouncer with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
16989
16919
|
:param bool prometheus: Allow clients to connect to prometheus with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
16990
16920
|
"""
|
|
16991
16921
|
if pg is not None:
|
|
@@ -16999,7 +16929,7 @@ class PgPgUserConfigPrivateAccess(dict):
|
|
|
16999
16929
|
@pulumi.getter
|
|
17000
16930
|
def pg(self) -> Optional[bool]:
|
|
17001
16931
|
"""
|
|
17002
|
-
|
|
16932
|
+
Allow clients to connect to pg with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
17003
16933
|
"""
|
|
17004
16934
|
return pulumi.get(self, "pg")
|
|
17005
16935
|
|
|
@@ -17007,7 +16937,7 @@ class PgPgUserConfigPrivateAccess(dict):
|
|
|
17007
16937
|
@pulumi.getter
|
|
17008
16938
|
def pgbouncer(self) -> Optional[bool]:
|
|
17009
16939
|
"""
|
|
17010
|
-
|
|
16940
|
+
Allow clients to connect to pgbouncer with a DNS name that always resolves to the service's private IP addresses. Only available in certain network locations.
|
|
17011
16941
|
"""
|
|
17012
16942
|
return pulumi.get(self, "pgbouncer")
|
|
17013
16943
|
|
|
@@ -17027,9 +16957,9 @@ class PgPgUserConfigPrivatelinkAccess(dict):
|
|
|
17027
16957
|
pgbouncer: Optional[bool] = None,
|
|
17028
16958
|
prometheus: Optional[bool] = None):
|
|
17029
16959
|
"""
|
|
17030
|
-
:param bool pg:
|
|
17031
|
-
:param bool pgbouncer:
|
|
17032
|
-
:param bool prometheus:
|
|
16960
|
+
:param bool pg: Enable pg.
|
|
16961
|
+
:param bool pgbouncer: Enable pgbouncer.
|
|
16962
|
+
:param bool prometheus: Enable prometheus.
|
|
17033
16963
|
"""
|
|
17034
16964
|
if pg is not None:
|
|
17035
16965
|
pulumi.set(__self__, "pg", pg)
|
|
@@ -17042,7 +16972,7 @@ class PgPgUserConfigPrivatelinkAccess(dict):
|
|
|
17042
16972
|
@pulumi.getter
|
|
17043
16973
|
def pg(self) -> Optional[bool]:
|
|
17044
16974
|
"""
|
|
17045
|
-
|
|
16975
|
+
Enable pg.
|
|
17046
16976
|
"""
|
|
17047
16977
|
return pulumi.get(self, "pg")
|
|
17048
16978
|
|
|
@@ -17050,7 +16980,7 @@ class PgPgUserConfigPrivatelinkAccess(dict):
|
|
|
17050
16980
|
@pulumi.getter
|
|
17051
16981
|
def pgbouncer(self) -> Optional[bool]:
|
|
17052
16982
|
"""
|
|
17053
|
-
|
|
16983
|
+
Enable pgbouncer.
|
|
17054
16984
|
"""
|
|
17055
16985
|
return pulumi.get(self, "pgbouncer")
|
|
17056
16986
|
|
|
@@ -17058,7 +16988,7 @@ class PgPgUserConfigPrivatelinkAccess(dict):
|
|
|
17058
16988
|
@pulumi.getter
|
|
17059
16989
|
def prometheus(self) -> Optional[bool]:
|
|
17060
16990
|
"""
|
|
17061
|
-
|
|
16991
|
+
Enable prometheus.
|
|
17062
16992
|
"""
|
|
17063
16993
|
return pulumi.get(self, "prometheus")
|
|
17064
16994
|
|
|
@@ -17070,9 +17000,9 @@ class PgPgUserConfigPublicAccess(dict):
|
|
|
17070
17000
|
pgbouncer: Optional[bool] = None,
|
|
17071
17001
|
prometheus: Optional[bool] = None):
|
|
17072
17002
|
"""
|
|
17073
|
-
:param bool pg:
|
|
17074
|
-
:param bool pgbouncer:
|
|
17075
|
-
:param bool prometheus: Allow clients to connect to prometheus
|
|
17003
|
+
:param bool pg: Allow clients to connect to pg from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
17004
|
+
:param bool pgbouncer: Allow clients to connect to pgbouncer from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
17005
|
+
:param bool prometheus: Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
17076
17006
|
"""
|
|
17077
17007
|
if pg is not None:
|
|
17078
17008
|
pulumi.set(__self__, "pg", pg)
|
|
@@ -17085,7 +17015,7 @@ class PgPgUserConfigPublicAccess(dict):
|
|
|
17085
17015
|
@pulumi.getter
|
|
17086
17016
|
def pg(self) -> Optional[bool]:
|
|
17087
17017
|
"""
|
|
17088
|
-
|
|
17018
|
+
Allow clients to connect to pg from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
17089
17019
|
"""
|
|
17090
17020
|
return pulumi.get(self, "pg")
|
|
17091
17021
|
|
|
@@ -17093,7 +17023,7 @@ class PgPgUserConfigPublicAccess(dict):
|
|
|
17093
17023
|
@pulumi.getter
|
|
17094
17024
|
def pgbouncer(self) -> Optional[bool]:
|
|
17095
17025
|
"""
|
|
17096
|
-
|
|
17026
|
+
Allow clients to connect to pgbouncer from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
17097
17027
|
"""
|
|
17098
17028
|
return pulumi.get(self, "pgbouncer")
|
|
17099
17029
|
|
|
@@ -17101,7 +17031,7 @@ class PgPgUserConfigPublicAccess(dict):
|
|
|
17101
17031
|
@pulumi.getter
|
|
17102
17032
|
def prometheus(self) -> Optional[bool]:
|
|
17103
17033
|
"""
|
|
17104
|
-
Allow clients to connect to prometheus
|
|
17034
|
+
Allow clients to connect to prometheus from the public internet for service nodes that are in a project VPC or another type of private network.
|
|
17105
17035
|
"""
|
|
17106
17036
|
return pulumi.get(self, "prometheus")
|
|
17107
17037
|
|
|
@@ -17243,8 +17173,8 @@ class ProjectTag(dict):
|
|
|
17243
17173
|
key: str,
|
|
17244
17174
|
value: str):
|
|
17245
17175
|
"""
|
|
17246
|
-
:param str key: Project tag key
|
|
17247
|
-
:param str value: Project tag value
|
|
17176
|
+
:param str key: Project tag key.
|
|
17177
|
+
:param str value: Project tag value.
|
|
17248
17178
|
"""
|
|
17249
17179
|
pulumi.set(__self__, "key", key)
|
|
17250
17180
|
pulumi.set(__self__, "value", value)
|
|
@@ -17253,7 +17183,7 @@ class ProjectTag(dict):
|
|
|
17253
17183
|
@pulumi.getter
|
|
17254
17184
|
def key(self) -> str:
|
|
17255
17185
|
"""
|
|
17256
|
-
Project tag key
|
|
17186
|
+
Project tag key.
|
|
17257
17187
|
"""
|
|
17258
17188
|
return pulumi.get(self, "key")
|
|
17259
17189
|
|
|
@@ -17261,7 +17191,7 @@ class ProjectTag(dict):
|
|
|
17261
17191
|
@pulumi.getter
|
|
17262
17192
|
def value(self) -> str:
|
|
17263
17193
|
"""
|
|
17264
|
-
Project tag value
|
|
17194
|
+
Project tag value.
|
|
17265
17195
|
"""
|
|
17266
17196
|
return pulumi.get(self, "value")
|
|
17267
17197
|
|
|
@@ -17388,12 +17318,6 @@ class RedisComponent(dict):
|
|
|
17388
17318
|
return pulumi.get(self, "usage")
|
|
17389
17319
|
|
|
17390
17320
|
|
|
17391
|
-
@pulumi.output_type
|
|
17392
|
-
class RedisRedi(dict):
|
|
17393
|
-
def __init__(__self__):
|
|
17394
|
-
pass
|
|
17395
|
-
|
|
17396
|
-
|
|
17397
17321
|
@pulumi.output_type
|
|
17398
17322
|
class RedisRedisUserConfig(dict):
|
|
17399
17323
|
@staticmethod
|
|
@@ -17569,9 +17493,6 @@ class RedisRedisUserConfig(dict):
|
|
|
17569
17493
|
"""
|
|
17570
17494
|
Additional Cloud Regions for Backup Replication.
|
|
17571
17495
|
"""
|
|
17572
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
17573
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
17574
|
-
|
|
17575
17496
|
return pulumi.get(self, "additional_backup_regions")
|
|
17576
17497
|
|
|
17577
17498
|
@property
|
|
@@ -18180,7 +18101,7 @@ class ServiceIntegrationClickhouseKafkaUserConfigTable(dict):
|
|
|
18180
18101
|
:param Sequence['ServiceIntegrationClickhouseKafkaUserConfigTableColumnArgs'] columns: Table columns
|
|
18181
18102
|
:param str data_format: Message data format. The default value is `JSONEachRow`.
|
|
18182
18103
|
:param str group_name: Kafka consumers group. The default value is `clickhouse`.
|
|
18183
|
-
:param str name:
|
|
18104
|
+
:param str name: Name of the table.
|
|
18184
18105
|
:param Sequence['ServiceIntegrationClickhouseKafkaUserConfigTableTopicArgs'] topics: Kafka topics
|
|
18185
18106
|
:param str auto_offset_reset: Action to take when there is no initial offset in offset store or the desired offset is out of range. The default value is `earliest`.
|
|
18186
18107
|
:param str date_time_input_format: Method to read DateTime from text input formats. The default value is `basic`.
|
|
@@ -18241,7 +18162,7 @@ class ServiceIntegrationClickhouseKafkaUserConfigTable(dict):
|
|
|
18241
18162
|
@pulumi.getter
|
|
18242
18163
|
def name(self) -> str:
|
|
18243
18164
|
"""
|
|
18244
|
-
|
|
18165
|
+
Name of the table.
|
|
18245
18166
|
"""
|
|
18246
18167
|
return pulumi.get(self, "name")
|
|
18247
18168
|
|
|
@@ -18352,7 +18273,7 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableTopic(dict):
|
|
|
18352
18273
|
def __init__(__self__, *,
|
|
18353
18274
|
name: str):
|
|
18354
18275
|
"""
|
|
18355
|
-
:param str name:
|
|
18276
|
+
:param str name: Name of the topic.
|
|
18356
18277
|
"""
|
|
18357
18278
|
pulumi.set(__self__, "name", name)
|
|
18358
18279
|
|
|
@@ -18360,7 +18281,7 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableTopic(dict):
|
|
|
18360
18281
|
@pulumi.getter
|
|
18361
18282
|
def name(self) -> str:
|
|
18362
18283
|
"""
|
|
18363
|
-
|
|
18284
|
+
Name of the topic.
|
|
18364
18285
|
"""
|
|
18365
18286
|
return pulumi.get(self, "name")
|
|
18366
18287
|
|
|
@@ -18608,7 +18529,9 @@ class ServiceIntegrationDatadogUserConfigOpensearch(dict):
|
|
|
18608
18529
|
@staticmethod
|
|
18609
18530
|
def __key_warning(key: str):
|
|
18610
18531
|
suggest = None
|
|
18611
|
-
if key == "
|
|
18532
|
+
if key == "clusterStatsEnabled":
|
|
18533
|
+
suggest = "cluster_stats_enabled"
|
|
18534
|
+
elif key == "indexStatsEnabled":
|
|
18612
18535
|
suggest = "index_stats_enabled"
|
|
18613
18536
|
elif key == "pendingTaskStatsEnabled":
|
|
18614
18537
|
suggest = "pending_task_stats_enabled"
|
|
@@ -18627,14 +18550,18 @@ class ServiceIntegrationDatadogUserConfigOpensearch(dict):
|
|
|
18627
18550
|
return super().get(key, default)
|
|
18628
18551
|
|
|
18629
18552
|
def __init__(__self__, *,
|
|
18553
|
+
cluster_stats_enabled: Optional[bool] = None,
|
|
18630
18554
|
index_stats_enabled: Optional[bool] = None,
|
|
18631
18555
|
pending_task_stats_enabled: Optional[bool] = None,
|
|
18632
18556
|
pshard_stats_enabled: Optional[bool] = None):
|
|
18633
18557
|
"""
|
|
18558
|
+
:param bool cluster_stats_enabled: Enable Datadog Opensearch Cluster Monitoring.
|
|
18634
18559
|
:param bool index_stats_enabled: Enable Datadog Opensearch Index Monitoring.
|
|
18635
18560
|
:param bool pending_task_stats_enabled: Enable Datadog Opensearch Pending Task Monitoring.
|
|
18636
18561
|
:param bool pshard_stats_enabled: Enable Datadog Opensearch Primary Shard Monitoring.
|
|
18637
18562
|
"""
|
|
18563
|
+
if cluster_stats_enabled is not None:
|
|
18564
|
+
pulumi.set(__self__, "cluster_stats_enabled", cluster_stats_enabled)
|
|
18638
18565
|
if index_stats_enabled is not None:
|
|
18639
18566
|
pulumi.set(__self__, "index_stats_enabled", index_stats_enabled)
|
|
18640
18567
|
if pending_task_stats_enabled is not None:
|
|
@@ -18642,6 +18569,14 @@ class ServiceIntegrationDatadogUserConfigOpensearch(dict):
|
|
|
18642
18569
|
if pshard_stats_enabled is not None:
|
|
18643
18570
|
pulumi.set(__self__, "pshard_stats_enabled", pshard_stats_enabled)
|
|
18644
18571
|
|
|
18572
|
+
@property
|
|
18573
|
+
@pulumi.getter(name="clusterStatsEnabled")
|
|
18574
|
+
def cluster_stats_enabled(self) -> Optional[bool]:
|
|
18575
|
+
"""
|
|
18576
|
+
Enable Datadog Opensearch Cluster Monitoring.
|
|
18577
|
+
"""
|
|
18578
|
+
return pulumi.get(self, "cluster_stats_enabled")
|
|
18579
|
+
|
|
18645
18580
|
@property
|
|
18646
18581
|
@pulumi.getter(name="indexStatsEnabled")
|
|
18647
18582
|
def index_stats_enabled(self) -> Optional[bool]:
|
|
@@ -19103,7 +19038,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudBigquery(dict):
|
|
|
19103
19038
|
service_account_credentials: str):
|
|
19104
19039
|
"""
|
|
19105
19040
|
:param str project_id: GCP project id.
|
|
19106
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19041
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19107
19042
|
"""
|
|
19108
19043
|
pulumi.set(__self__, "project_id", project_id)
|
|
19109
19044
|
pulumi.set(__self__, "service_account_credentials", service_account_credentials)
|
|
@@ -19120,7 +19055,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudBigquery(dict):
|
|
|
19120
19055
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
19121
19056
|
def service_account_credentials(self) -> str:
|
|
19122
19057
|
"""
|
|
19123
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19058
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19124
19059
|
"""
|
|
19125
19060
|
return pulumi.get(self, "service_account_credentials")
|
|
19126
19061
|
|
|
@@ -19155,7 +19090,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfig(dict):
|
|
|
19155
19090
|
"""
|
|
19156
19091
|
:param str log_id: Google Cloud Logging log id.
|
|
19157
19092
|
:param str project_id: GCP project id.
|
|
19158
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19093
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19159
19094
|
"""
|
|
19160
19095
|
pulumi.set(__self__, "log_id", log_id)
|
|
19161
19096
|
pulumi.set(__self__, "project_id", project_id)
|
|
@@ -19181,7 +19116,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfig(dict):
|
|
|
19181
19116
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
19182
19117
|
def service_account_credentials(self) -> str:
|
|
19183
19118
|
"""
|
|
19184
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19119
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19185
19120
|
"""
|
|
19186
19121
|
return pulumi.get(self, "service_account_credentials")
|
|
19187
19122
|
|
|
@@ -20633,7 +20568,7 @@ class ServiceIntegrationMetricsUserConfigSourceMysqlTelegraf(dict):
|
|
|
20633
20568
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
20634
20569
|
"""
|
|
20635
20570
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE*SCHEMA.EVENT*WAITS.
|
|
20636
|
-
:param bool gather_file_events_stats:
|
|
20571
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20637
20572
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE*SCHEMA.TABLE*IO*WAITS*SUMMARY*BY*INDEX_USAGE.
|
|
20638
20573
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
20639
20574
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION*SCHEMA.INNODB*METRICS.
|
|
@@ -20688,7 +20623,7 @@ class ServiceIntegrationMetricsUserConfigSourceMysqlTelegraf(dict):
|
|
|
20688
20623
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
20689
20624
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
20690
20625
|
"""
|
|
20691
|
-
|
|
20626
|
+
Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20692
20627
|
"""
|
|
20693
20628
|
return pulumi.get(self, "gather_file_events_stats")
|
|
20694
20629
|
|
|
@@ -20906,7 +20841,7 @@ class ServiceIntegrationPrometheusUserConfigSourceMysqlTelegraf(dict):
|
|
|
20906
20841
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
20907
20842
|
"""
|
|
20908
20843
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE*SCHEMA.EVENT*WAITS.
|
|
20909
|
-
:param bool gather_file_events_stats:
|
|
20844
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20910
20845
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE*SCHEMA.TABLE*IO*WAITS*SUMMARY*BY*INDEX_USAGE.
|
|
20911
20846
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
20912
20847
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION*SCHEMA.INNODB*METRICS.
|
|
@@ -20961,7 +20896,7 @@ class ServiceIntegrationPrometheusUserConfigSourceMysqlTelegraf(dict):
|
|
|
20961
20896
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
20962
20897
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
20963
20898
|
"""
|
|
20964
|
-
|
|
20899
|
+
Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20965
20900
|
"""
|
|
20966
20901
|
return pulumi.get(self, "gather_file_events_stats")
|
|
20967
20902
|
|
|
@@ -21129,12 +21064,6 @@ class GetAccountAuthenticationSamlFieldMappingResult(dict):
|
|
|
21129
21064
|
return pulumi.get(self, "real_name")
|
|
21130
21065
|
|
|
21131
21066
|
|
|
21132
|
-
@pulumi.output_type
|
|
21133
|
-
class GetCassandaCassandraResult(dict):
|
|
21134
|
-
def __init__(__self__):
|
|
21135
|
-
pass
|
|
21136
|
-
|
|
21137
|
-
|
|
21138
21067
|
@pulumi.output_type
|
|
21139
21068
|
class GetCassandaCassandraUserConfigResult(dict):
|
|
21140
21069
|
def __init__(__self__, *,
|
|
@@ -21158,7 +21087,7 @@ class GetCassandaCassandraUserConfigResult(dict):
|
|
|
21158
21087
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
21159
21088
|
:param int backup_hour: The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21160
21089
|
:param int backup_minute: The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21161
|
-
:param 'GetCassandaCassandraUserConfigCassandraArgs' cassandra:
|
|
21090
|
+
:param 'GetCassandaCassandraUserConfigCassandraArgs' cassandra: Cassandra configuration values
|
|
21162
21091
|
:param str cassandra_version: Cassandra version.
|
|
21163
21092
|
:param Sequence['GetCassandaCassandraUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
21164
21093
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -21236,7 +21165,7 @@ class GetCassandaCassandraUserConfigResult(dict):
|
|
|
21236
21165
|
@pulumi.getter
|
|
21237
21166
|
def cassandra(self) -> Optional['outputs.GetCassandaCassandraUserConfigCassandraResult']:
|
|
21238
21167
|
"""
|
|
21239
|
-
|
|
21168
|
+
Cassandra configuration values
|
|
21240
21169
|
"""
|
|
21241
21170
|
return pulumi.get(self, "cassandra")
|
|
21242
21171
|
|
|
@@ -21622,12 +21551,6 @@ class GetCassandaTechEmailResult(dict):
|
|
|
21622
21551
|
return pulumi.get(self, "email")
|
|
21623
21552
|
|
|
21624
21553
|
|
|
21625
|
-
@pulumi.output_type
|
|
21626
|
-
class GetCassandraCassandraResult(dict):
|
|
21627
|
-
def __init__(__self__):
|
|
21628
|
-
pass
|
|
21629
|
-
|
|
21630
|
-
|
|
21631
21554
|
@pulumi.output_type
|
|
21632
21555
|
class GetCassandraCassandraUserConfigResult(dict):
|
|
21633
21556
|
def __init__(__self__, *,
|
|
@@ -21651,7 +21574,7 @@ class GetCassandraCassandraUserConfigResult(dict):
|
|
|
21651
21574
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
21652
21575
|
:param int backup_hour: The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21653
21576
|
:param int backup_minute: The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21654
|
-
:param 'GetCassandraCassandraUserConfigCassandraArgs' cassandra:
|
|
21577
|
+
:param 'GetCassandraCassandraUserConfigCassandraArgs' cassandra: Cassandra configuration values
|
|
21655
21578
|
:param str cassandra_version: Cassandra version.
|
|
21656
21579
|
:param Sequence['GetCassandraCassandraUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
21657
21580
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -21729,7 +21652,7 @@ class GetCassandraCassandraUserConfigResult(dict):
|
|
|
21729
21652
|
@pulumi.getter
|
|
21730
21653
|
def cassandra(self) -> Optional['outputs.GetCassandraCassandraUserConfigCassandraResult']:
|
|
21731
21654
|
"""
|
|
21732
|
-
|
|
21655
|
+
Cassandra configuration values
|
|
21733
21656
|
"""
|
|
21734
21657
|
return pulumi.get(self, "cassandra")
|
|
21735
21658
|
|
|
@@ -22115,12 +22038,6 @@ class GetCassandraTechEmailResult(dict):
|
|
|
22115
22038
|
return pulumi.get(self, "email")
|
|
22116
22039
|
|
|
22117
22040
|
|
|
22118
|
-
@pulumi.output_type
|
|
22119
|
-
class GetClickhouseClickhouseResult(dict):
|
|
22120
|
-
def __init__(__self__):
|
|
22121
|
-
pass
|
|
22122
|
-
|
|
22123
|
-
|
|
22124
22041
|
@pulumi.output_type
|
|
22125
22042
|
class GetClickhouseClickhouseUserConfigResult(dict):
|
|
22126
22043
|
def __init__(__self__, *,
|
|
@@ -22727,16 +22644,11 @@ class GetDragonflyComponentResult(dict):
|
|
|
22727
22644
|
return pulumi.get(self, "usage")
|
|
22728
22645
|
|
|
22729
22646
|
|
|
22730
|
-
@pulumi.output_type
|
|
22731
|
-
class GetDragonflyDragonflyResult(dict):
|
|
22732
|
-
def __init__(__self__):
|
|
22733
|
-
pass
|
|
22734
|
-
|
|
22735
|
-
|
|
22736
22647
|
@pulumi.output_type
|
|
22737
22648
|
class GetDragonflyDragonflyUserConfigResult(dict):
|
|
22738
22649
|
def __init__(__self__, *,
|
|
22739
22650
|
cache_mode: Optional[bool] = None,
|
|
22651
|
+
dragonfly_persistence: Optional[str] = None,
|
|
22740
22652
|
dragonfly_ssl: Optional[bool] = None,
|
|
22741
22653
|
ip_filter_objects: Optional[Sequence['outputs.GetDragonflyDragonflyUserConfigIpFilterObjectResult']] = None,
|
|
22742
22654
|
ip_filter_strings: Optional[Sequence[str]] = None,
|
|
@@ -22752,6 +22664,7 @@ class GetDragonflyDragonflyUserConfigResult(dict):
|
|
|
22752
22664
|
static_ips: Optional[bool] = None):
|
|
22753
22665
|
"""
|
|
22754
22666
|
:param bool cache_mode: Evict entries when getting close to maxmemory limit. The default value is `false`.
|
|
22667
|
+
:param str dragonfly_persistence: When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
22755
22668
|
:param bool dragonfly_ssl: Require SSL to access Dragonfly. The default value is `true`.
|
|
22756
22669
|
:param Sequence['GetDragonflyDragonflyUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
22757
22670
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -22768,6 +22681,8 @@ class GetDragonflyDragonflyUserConfigResult(dict):
|
|
|
22768
22681
|
"""
|
|
22769
22682
|
if cache_mode is not None:
|
|
22770
22683
|
pulumi.set(__self__, "cache_mode", cache_mode)
|
|
22684
|
+
if dragonfly_persistence is not None:
|
|
22685
|
+
pulumi.set(__self__, "dragonfly_persistence", dragonfly_persistence)
|
|
22771
22686
|
if dragonfly_ssl is not None:
|
|
22772
22687
|
pulumi.set(__self__, "dragonfly_ssl", dragonfly_ssl)
|
|
22773
22688
|
if ip_filter_objects is not None:
|
|
@@ -22803,6 +22718,14 @@ class GetDragonflyDragonflyUserConfigResult(dict):
|
|
|
22803
22718
|
"""
|
|
22804
22719
|
return pulumi.get(self, "cache_mode")
|
|
22805
22720
|
|
|
22721
|
+
@property
|
|
22722
|
+
@pulumi.getter(name="dragonflyPersistence")
|
|
22723
|
+
def dragonfly_persistence(self) -> Optional[str]:
|
|
22724
|
+
"""
|
|
22725
|
+
When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
22726
|
+
"""
|
|
22727
|
+
return pulumi.get(self, "dragonfly_persistence")
|
|
22728
|
+
|
|
22806
22729
|
@property
|
|
22807
22730
|
@pulumi.getter(name="dragonflySsl")
|
|
22808
22731
|
def dragonfly_ssl(self) -> Optional[bool]:
|
|
@@ -23737,12 +23660,6 @@ class GetGrafanaComponentResult(dict):
|
|
|
23737
23660
|
return pulumi.get(self, "usage")
|
|
23738
23661
|
|
|
23739
23662
|
|
|
23740
|
-
@pulumi.output_type
|
|
23741
|
-
class GetGrafanaGrafanaResult(dict):
|
|
23742
|
-
def __init__(__self__):
|
|
23743
|
-
pass
|
|
23744
|
-
|
|
23745
|
-
|
|
23746
23663
|
@pulumi.output_type
|
|
23747
23664
|
class GetGrafanaGrafanaUserConfigResult(dict):
|
|
23748
23665
|
def __init__(__self__, *,
|
|
@@ -23923,9 +23840,6 @@ class GetGrafanaGrafanaUserConfigResult(dict):
|
|
|
23923
23840
|
"""
|
|
23924
23841
|
Additional Cloud Regions for Backup Replication.
|
|
23925
23842
|
"""
|
|
23926
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
23927
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
23928
|
-
|
|
23929
23843
|
return pulumi.get(self, "additional_backup_regions")
|
|
23930
23844
|
|
|
23931
23845
|
@property
|
|
@@ -25865,12 +25779,6 @@ class GetKafkaConnectComponentResult(dict):
|
|
|
25865
25779
|
return pulumi.get(self, "usage")
|
|
25866
25780
|
|
|
25867
25781
|
|
|
25868
|
-
@pulumi.output_type
|
|
25869
|
-
class GetKafkaConnectKafkaConnectResult(dict):
|
|
25870
|
-
def __init__(__self__):
|
|
25871
|
-
pass
|
|
25872
|
-
|
|
25873
|
-
|
|
25874
25782
|
@pulumi.output_type
|
|
25875
25783
|
class GetKafkaConnectKafkaConnectUserConfigResult(dict):
|
|
25876
25784
|
def __init__(__self__, *,
|
|
@@ -26058,7 +25966,7 @@ class GetKafkaConnectKafkaConnectUserConfigKafkaConnectResult(dict):
|
|
|
26058
25966
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
26059
25967
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
26060
25968
|
:param str consumer_isolation_level: Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired.
|
|
26061
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
25969
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26062
25970
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
26063
25971
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
26064
25972
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -26140,7 +26048,7 @@ class GetKafkaConnectKafkaConnectUserConfigKafkaConnectResult(dict):
|
|
|
26140
26048
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
26141
26049
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
26142
26050
|
"""
|
|
26143
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26051
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26144
26052
|
"""
|
|
26145
26053
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
26146
26054
|
|
|
@@ -26452,11 +26360,11 @@ class GetKafkaKafkaResult(dict):
|
|
|
26452
26360
|
rest_uri: str,
|
|
26453
26361
|
schema_registry_uri: str):
|
|
26454
26362
|
"""
|
|
26455
|
-
:param str access_cert: The Kafka client certificate
|
|
26456
|
-
:param str access_key: The Kafka client certificate key
|
|
26457
|
-
:param str connect_uri: The Kafka Connect URI
|
|
26458
|
-
:param str rest_uri: The Kafka REST URI
|
|
26459
|
-
:param str schema_registry_uri: The Schema Registry URI
|
|
26363
|
+
:param str access_cert: The Kafka client certificate.
|
|
26364
|
+
:param str access_key: The Kafka client certificate key.
|
|
26365
|
+
:param str connect_uri: The Kafka Connect URI.
|
|
26366
|
+
:param str rest_uri: The Kafka REST URI.
|
|
26367
|
+
:param str schema_registry_uri: The Schema Registry URI.
|
|
26460
26368
|
"""
|
|
26461
26369
|
pulumi.set(__self__, "access_cert", access_cert)
|
|
26462
26370
|
pulumi.set(__self__, "access_key", access_key)
|
|
@@ -26468,7 +26376,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26468
26376
|
@pulumi.getter(name="accessCert")
|
|
26469
26377
|
def access_cert(self) -> str:
|
|
26470
26378
|
"""
|
|
26471
|
-
The Kafka client certificate
|
|
26379
|
+
The Kafka client certificate.
|
|
26472
26380
|
"""
|
|
26473
26381
|
return pulumi.get(self, "access_cert")
|
|
26474
26382
|
|
|
@@ -26476,7 +26384,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26476
26384
|
@pulumi.getter(name="accessKey")
|
|
26477
26385
|
def access_key(self) -> str:
|
|
26478
26386
|
"""
|
|
26479
|
-
The Kafka client certificate key
|
|
26387
|
+
The Kafka client certificate key.
|
|
26480
26388
|
"""
|
|
26481
26389
|
return pulumi.get(self, "access_key")
|
|
26482
26390
|
|
|
@@ -26484,7 +26392,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26484
26392
|
@pulumi.getter(name="connectUri")
|
|
26485
26393
|
def connect_uri(self) -> str:
|
|
26486
26394
|
"""
|
|
26487
|
-
The Kafka Connect URI
|
|
26395
|
+
The Kafka Connect URI.
|
|
26488
26396
|
"""
|
|
26489
26397
|
return pulumi.get(self, "connect_uri")
|
|
26490
26398
|
|
|
@@ -26492,7 +26400,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26492
26400
|
@pulumi.getter(name="restUri")
|
|
26493
26401
|
def rest_uri(self) -> str:
|
|
26494
26402
|
"""
|
|
26495
|
-
The Kafka REST URI
|
|
26403
|
+
The Kafka REST URI.
|
|
26496
26404
|
"""
|
|
26497
26405
|
return pulumi.get(self, "rest_uri")
|
|
26498
26406
|
|
|
@@ -26500,7 +26408,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26500
26408
|
@pulumi.getter(name="schemaRegistryUri")
|
|
26501
26409
|
def schema_registry_uri(self) -> str:
|
|
26502
26410
|
"""
|
|
26503
|
-
The Schema Registry URI
|
|
26411
|
+
The Schema Registry URI.
|
|
26504
26412
|
"""
|
|
26505
26413
|
return pulumi.get(self, "schema_registry_uri")
|
|
26506
26414
|
|
|
@@ -26880,7 +26788,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
26880
26788
|
:param int log_index_size_max_bytes: The maximum size in bytes of the offset index.
|
|
26881
26789
|
:param int log_local_retention_bytes: The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value.
|
|
26882
26790
|
:param int log_local_retention_ms: The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value.
|
|
26883
|
-
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
26791
|
+
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
26884
26792
|
:param int log_message_timestamp_difference_max_ms: The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message.
|
|
26885
26793
|
:param str log_message_timestamp_type: Define whether the timestamp in the message is message create time or log append time.
|
|
26886
26794
|
:param bool log_preallocate: Should pre allocate file when create new segment?
|
|
@@ -26902,7 +26810,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
26902
26810
|
:param int replica_fetch_response_max_bytes: Maximum bytes expected for the entire fetch response (defaults to 10485760). Records are fetched in batches, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. As such, this is not an absolute maximum.
|
|
26903
26811
|
:param str sasl_oauthbearer_expected_audience: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences.
|
|
26904
26812
|
:param str sasl_oauthbearer_expected_issuer: Optional setting for the broker to use to verify that the JWT was created by the expected issuer.
|
|
26905
|
-
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
26813
|
+
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
26906
26814
|
:param str sasl_oauthbearer_sub_claim_name: Name of the scope from which to extract the subject claim from the JWT. Defaults to sub.
|
|
26907
26815
|
:param int socket_request_max_bytes: The maximum number of bytes in a socket request (defaults to 104857600).
|
|
26908
26816
|
:param bool transaction_partition_verification_enable: Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition.
|
|
@@ -27150,7 +27058,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
27150
27058
|
@pulumi.getter(name="logMessageDownconversionEnable")
|
|
27151
27059
|
def log_message_downconversion_enable(self) -> Optional[bool]:
|
|
27152
27060
|
"""
|
|
27153
|
-
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
27061
|
+
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
27154
27062
|
"""
|
|
27155
27063
|
return pulumi.get(self, "log_message_downconversion_enable")
|
|
27156
27064
|
|
|
@@ -27326,7 +27234,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
27326
27234
|
@pulumi.getter(name="saslOauthbearerJwksEndpointUrl")
|
|
27327
27235
|
def sasl_oauthbearer_jwks_endpoint_url(self) -> Optional[str]:
|
|
27328
27236
|
"""
|
|
27329
|
-
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
27237
|
+
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
27330
27238
|
"""
|
|
27331
27239
|
return pulumi.get(self, "sasl_oauthbearer_jwks_endpoint_url")
|
|
27332
27240
|
|
|
@@ -27426,7 +27334,7 @@ class GetKafkaKafkaUserConfigKafkaConnectConfigResult(dict):
|
|
|
27426
27334
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
27427
27335
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
27428
27336
|
:param str consumer_isolation_level: Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired.
|
|
27429
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27337
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27430
27338
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
27431
27339
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
27432
27340
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -27508,7 +27416,7 @@ class GetKafkaKafkaUserConfigKafkaConnectConfigResult(dict):
|
|
|
27508
27416
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
27509
27417
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
27510
27418
|
"""
|
|
27511
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27419
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27512
27420
|
"""
|
|
27513
27421
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
27514
27422
|
|
|
@@ -28123,12 +28031,6 @@ class GetKafkaMirrorMakerComponentResult(dict):
|
|
|
28123
28031
|
return pulumi.get(self, "usage")
|
|
28124
28032
|
|
|
28125
28033
|
|
|
28126
|
-
@pulumi.output_type
|
|
28127
|
-
class GetKafkaMirrorMakerKafkaMirrormakerResult(dict):
|
|
28128
|
-
def __init__(__self__):
|
|
28129
|
-
pass
|
|
28130
|
-
|
|
28131
|
-
|
|
28132
28034
|
@pulumi.output_type
|
|
28133
28035
|
class GetKafkaMirrorMakerKafkaMirrormakerUserConfigResult(dict):
|
|
28134
28036
|
def __init__(__self__, *,
|
|
@@ -29030,12 +28932,6 @@ class GetM3AggregatorComponentResult(dict):
|
|
|
29030
28932
|
return pulumi.get(self, "usage")
|
|
29031
28933
|
|
|
29032
28934
|
|
|
29033
|
-
@pulumi.output_type
|
|
29034
|
-
class GetM3AggregatorM3aggregatorResult(dict):
|
|
29035
|
-
def __init__(__self__):
|
|
29036
|
-
pass
|
|
29037
|
-
|
|
29038
|
-
|
|
29039
28935
|
@pulumi.output_type
|
|
29040
28936
|
class GetM3AggregatorM3aggregatorUserConfigResult(dict):
|
|
29041
28937
|
def __init__(__self__, *,
|
|
@@ -29343,12 +29239,6 @@ class GetM3DbComponentResult(dict):
|
|
|
29343
29239
|
return pulumi.get(self, "usage")
|
|
29344
29240
|
|
|
29345
29241
|
|
|
29346
|
-
@pulumi.output_type
|
|
29347
|
-
class GetM3DbM3dbResult(dict):
|
|
29348
|
-
def __init__(__self__):
|
|
29349
|
-
pass
|
|
29350
|
-
|
|
29351
|
-
|
|
29352
29242
|
@pulumi.output_type
|
|
29353
29243
|
class GetM3DbM3dbUserConfigResult(dict):
|
|
29354
29244
|
def __init__(__self__, *,
|
|
@@ -29433,9 +29323,6 @@ class GetM3DbM3dbUserConfigResult(dict):
|
|
|
29433
29323
|
"""
|
|
29434
29324
|
Additional Cloud Regions for Backup Replication.
|
|
29435
29325
|
"""
|
|
29436
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
29437
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
29438
|
-
|
|
29439
29326
|
return pulumi.get(self, "additional_backup_regions")
|
|
29440
29327
|
|
|
29441
29328
|
@property
|
|
@@ -30291,12 +30178,6 @@ class GetMySqlComponentResult(dict):
|
|
|
30291
30178
|
return pulumi.get(self, "usage")
|
|
30292
30179
|
|
|
30293
30180
|
|
|
30294
|
-
@pulumi.output_type
|
|
30295
|
-
class GetMySqlMysqlResult(dict):
|
|
30296
|
-
def __init__(__self__):
|
|
30297
|
-
pass
|
|
30298
|
-
|
|
30299
|
-
|
|
30300
30181
|
@pulumi.output_type
|
|
30301
30182
|
class GetMySqlMysqlUserConfigResult(dict):
|
|
30302
30183
|
def __init__(__self__, *,
|
|
@@ -31399,7 +31280,7 @@ class GetOpenSearchOpensearchUserConfigResult(dict):
|
|
|
31399
31280
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
31400
31281
|
:param Sequence[str] ip_filters: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
31401
31282
|
:param bool keep_index_refresh_interval: Aiven automation resets index.refresh_interval to default value for every index to be sure that indices are always visible to search. If it doesn't fit your case, you can disable this by setting up this flag to true.
|
|
31402
|
-
:param int max_index_count:
|
|
31283
|
+
:param int max_index_count: Use index_patterns instead. The default value is `0`.
|
|
31403
31284
|
:param 'GetOpenSearchOpensearchUserConfigOpenidArgs' openid: OpenSearch OpenID Connect Configuration
|
|
31404
31285
|
:param 'GetOpenSearchOpensearchUserConfigOpensearchArgs' opensearch: OpenSearch settings
|
|
31405
31286
|
:param 'GetOpenSearchOpensearchUserConfigOpensearchDashboardsArgs' opensearch_dashboards: OpenSearch Dashboards settings
|
|
@@ -31467,9 +31348,6 @@ class GetOpenSearchOpensearchUserConfigResult(dict):
|
|
|
31467
31348
|
"""
|
|
31468
31349
|
Additional Cloud Regions for Backup Replication.
|
|
31469
31350
|
"""
|
|
31470
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
31471
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
31472
|
-
|
|
31473
31351
|
return pulumi.get(self, "additional_backup_regions")
|
|
31474
31352
|
|
|
31475
31353
|
@property
|
|
@@ -31543,7 +31421,7 @@ class GetOpenSearchOpensearchUserConfigResult(dict):
|
|
|
31543
31421
|
@pulumi.getter(name="maxIndexCount")
|
|
31544
31422
|
def max_index_count(self) -> Optional[int]:
|
|
31545
31423
|
"""
|
|
31546
|
-
|
|
31424
|
+
Use index_patterns instead. The default value is `0`.
|
|
31547
31425
|
"""
|
|
31548
31426
|
return pulumi.get(self, "max_index_count")
|
|
31549
31427
|
|
|
@@ -31972,10 +31850,10 @@ class GetOpenSearchOpensearchUserConfigOpensearchResult(dict):
|
|
|
31972
31850
|
:param int http_max_header_size: The max size of allowed headers, in bytes.
|
|
31973
31851
|
:param int http_max_initial_line_length: The max length of an HTTP URL, in bytes.
|
|
31974
31852
|
:param int indices_fielddata_cache_size: Relative amount. Maximum amount of heap memory used for field data cache. This is an expert setting; decreasing the value too much will increase overhead of loading field data; too much memory used for field data cache will decrease amount of heap available for other operations.
|
|
31975
|
-
:param int indices_memory_index_buffer_size: Percentage value. Default is 10
|
|
31853
|
+
:param int indices_memory_index_buffer_size: Percentage value. Default is 10%. Total amount of heap used for indexing buffer, before writing segments to disk. This is an expert setting. Too low value will slow down indexing; too high value will increase indexing performance but causes performance issues for query performance.
|
|
31976
31854
|
:param int indices_memory_max_index_buffer_size: Absolute value. Default is unbound. Doesn't work without indices.memory.index_buffer_size. Maximum amount of heap used for query cache, an absolute indices.memory.index_buffer_size maximum hard limit.
|
|
31977
31855
|
:param int indices_memory_min_index_buffer_size: Absolute value. Default is 48mb. Doesn't work without indices.memory.index_buffer_size. Minimum amount of heap used for query cache, an absolute indices.memory.index_buffer_size minimal hard limit.
|
|
31978
|
-
:param int indices_queries_cache_size: Percentage value. Default is 10
|
|
31856
|
+
:param int indices_queries_cache_size: Percentage value. Default is 10%. Maximum amount of heap used for query cache. This is an expert setting. Too low value will decrease query performance and increase performance for other operations; too high value will cause issues with other OpenSearch functionality.
|
|
31979
31857
|
:param int indices_query_bool_max_clause_count: Maximum number of clauses Lucene BooleanQuery can have. The default value (1024) is relatively high, and increasing it may cause performance issues. Investigate other approaches first before increasing this value.
|
|
31980
31858
|
:param int indices_recovery_max_bytes_per_sec: Limits total inbound and outbound recovery traffic for each node. Applies to both peer recoveries as well as snapshot recoveries (i.e., restores from a snapshot). Defaults to 40mb.
|
|
31981
31859
|
:param int indices_recovery_max_concurrent_file_chunks: Number of file chunks sent in parallel for each recovery. Defaults to 2.
|
|
@@ -32195,7 +32073,7 @@ class GetOpenSearchOpensearchUserConfigOpensearchResult(dict):
|
|
|
32195
32073
|
@pulumi.getter(name="indicesMemoryIndexBufferSize")
|
|
32196
32074
|
def indices_memory_index_buffer_size(self) -> Optional[int]:
|
|
32197
32075
|
"""
|
|
32198
|
-
Percentage value. Default is 10
|
|
32076
|
+
Percentage value. Default is 10%. Total amount of heap used for indexing buffer, before writing segments to disk. This is an expert setting. Too low value will slow down indexing; too high value will increase indexing performance but causes performance issues for query performance.
|
|
32199
32077
|
"""
|
|
32200
32078
|
return pulumi.get(self, "indices_memory_index_buffer_size")
|
|
32201
32079
|
|
|
@@ -32219,7 +32097,7 @@ class GetOpenSearchOpensearchUserConfigOpensearchResult(dict):
|
|
|
32219
32097
|
@pulumi.getter(name="indicesQueriesCacheSize")
|
|
32220
32098
|
def indices_queries_cache_size(self) -> Optional[int]:
|
|
32221
32099
|
"""
|
|
32222
|
-
Percentage value. Default is 10
|
|
32100
|
+
Percentage value. Default is 10%. Maximum amount of heap used for query cache. This is an expert setting. Too low value will decrease query performance and increase performance for other operations; too high value will cause issues with other OpenSearch functionality.
|
|
32223
32101
|
"""
|
|
32224
32102
|
return pulumi.get(self, "indices_queries_cache_size")
|
|
32225
32103
|
|
|
@@ -33219,12 +33097,12 @@ class GetPgPgUserConfigResult(dict):
|
|
|
33219
33097
|
:param str recovery_target_time: Recovery target time when forking a service. This has effect only when a new service is being created.
|
|
33220
33098
|
:param bool service_log: Store logs for the service so that they are available in the HTTP API and console.
|
|
33221
33099
|
:param str service_to_fork_from: Name of another service to fork from. This has effect only when a new service is being created.
|
|
33222
|
-
:param float shared_buffers_percentage: Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20%. This setting adjusts the shared_buffers configuration value.
|
|
33100
|
+
:param float shared_buffers_percentage: Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20% - 60%. This setting adjusts the shared_buffers configuration value.
|
|
33223
33101
|
:param bool static_ips: Use static public IP addresses.
|
|
33224
33102
|
:param str synchronous_replication: Synchronous replication type. Note that the service plan also needs to support synchronous replication.
|
|
33225
33103
|
:param 'GetPgPgUserConfigTimescaledbArgs' timescaledb: System-wide settings for the timescaledb extension
|
|
33226
33104
|
:param str variant: Variant of the PostgreSQL service, may affect the features that are exposed by default.
|
|
33227
|
-
:param int work_mem: Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075
|
|
33105
|
+
:param int work_mem: Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075% of total RAM (up to 32MB).
|
|
33228
33106
|
"""
|
|
33229
33107
|
if additional_backup_regions is not None:
|
|
33230
33108
|
pulumi.set(__self__, "additional_backup_regions", additional_backup_regions)
|
|
@@ -33509,7 +33387,7 @@ class GetPgPgUserConfigResult(dict):
|
|
|
33509
33387
|
@pulumi.getter(name="sharedBuffersPercentage")
|
|
33510
33388
|
def shared_buffers_percentage(self) -> Optional[float]:
|
|
33511
33389
|
"""
|
|
33512
|
-
Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20%. This setting adjusts the shared_buffers configuration value.
|
|
33390
|
+
Percentage of total RAM that the database server uses for shared memory buffers. Valid range is 20-60 (float), which corresponds to 20% - 60%. This setting adjusts the shared_buffers configuration value.
|
|
33513
33391
|
"""
|
|
33514
33392
|
return pulumi.get(self, "shared_buffers_percentage")
|
|
33515
33393
|
|
|
@@ -33549,7 +33427,7 @@ class GetPgPgUserConfigResult(dict):
|
|
|
33549
33427
|
@pulumi.getter(name="workMem")
|
|
33550
33428
|
def work_mem(self) -> Optional[int]:
|
|
33551
33429
|
"""
|
|
33552
|
-
Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075
|
|
33430
|
+
Sets the maximum amount of memory to be used by a query operation (such as a sort or hash table) before writing to temporary disk files, in MB. Default is 1MB + 0.075% of total RAM (up to 32MB).
|
|
33553
33431
|
"""
|
|
33554
33432
|
return pulumi.get(self, "work_mem")
|
|
33555
33433
|
|
|
@@ -33738,14 +33616,14 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
33738
33616
|
wal_sender_timeout: Optional[int] = None,
|
|
33739
33617
|
wal_writer_delay: Optional[int] = None):
|
|
33740
33618
|
"""
|
|
33741
|
-
:param float autovacuum_analyze_scale_factor: Specifies a fraction of the table size to add to autovacuum_analyze_threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20
|
|
33619
|
+
:param float autovacuum_analyze_scale_factor: Specifies a fraction of the table size to add to autovacuum_analyze_threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20% of table size).
|
|
33742
33620
|
:param int autovacuum_analyze_threshold: Specifies the minimum number of inserted, updated or deleted tuples needed to trigger an ANALYZE in any one table. The default is 50 tuples.
|
|
33743
33621
|
:param int autovacuum_freeze_max_age: Specifies the maximum age (in transactions) that a table's pg_class.relfrozenxid field can attain before a VACUUM operation is forced to prevent transaction ID wraparound within the table. Note that the system will launch autovacuum processes to prevent wraparound even when autovacuum is otherwise disabled. This parameter will cause the server to be restarted.
|
|
33744
33622
|
:param int autovacuum_max_workers: Specifies the maximum number of autovacuum processes (other than the autovacuum launcher) that may be running at any one time. The default is three. This parameter can only be set at server start.
|
|
33745
33623
|
:param int autovacuum_naptime: Specifies the minimum delay between autovacuum runs on any given database. The delay is measured in seconds, and the default is one minute.
|
|
33746
33624
|
:param int autovacuum_vacuum_cost_delay: Specifies the cost delay value that will be used in automatic VACUUM operations. If -1 is specified, the regular vacuum_cost_delay value will be used. The default value is 20 milliseconds.
|
|
33747
33625
|
:param int autovacuum_vacuum_cost_limit: Specifies the cost limit value that will be used in automatic VACUUM operations. If -1 is specified (which is the default), the regular vacuum_cost_limit value will be used.
|
|
33748
|
-
:param float autovacuum_vacuum_scale_factor: Specifies a fraction of the table size to add to autovacuum_vacuum_threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20
|
|
33626
|
+
:param float autovacuum_vacuum_scale_factor: Specifies a fraction of the table size to add to autovacuum_vacuum_threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20% of table size).
|
|
33749
33627
|
:param int autovacuum_vacuum_threshold: Specifies the minimum number of updated or deleted tuples needed to trigger a VACUUM in any one table. The default is 50 tuples.
|
|
33750
33628
|
:param int bgwriter_delay: Specifies the delay between activity rounds for the background writer in milliseconds. Default is 200.
|
|
33751
33629
|
:param int bgwriter_flush_after: Whenever more than bgwriter_flush_after bytes have been written by the background writer, attempt to force the OS to issue these writes to the underlying storage. Specified in kilobytes, default is 512. Setting of 0 disables forced writeback.
|
|
@@ -33777,7 +33655,7 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
33777
33655
|
:param int pg_partman_bgw_dot_interval: Sets the time interval to run pg_partman's scheduled tasks.
|
|
33778
33656
|
:param str pg_partman_bgw_dot_role: Controls which role to use for pg_partman's scheduled background tasks.
|
|
33779
33657
|
:param bool pg_stat_monitor_dot_pgsm_enable_query_plan: Enables or disables query plan monitoring.
|
|
33780
|
-
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets
|
|
33658
|
+
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets.
|
|
33781
33659
|
:param str pg_stat_statements_dot_track: Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top.
|
|
33782
33660
|
:param int temp_file_limit: PostgreSQL temporary file limit in KiB, -1 for unlimited.
|
|
33783
33661
|
:param str timezone: PostgreSQL service timezone.
|
|
@@ -33891,7 +33769,7 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
33891
33769
|
@pulumi.getter(name="autovacuumAnalyzeScaleFactor")
|
|
33892
33770
|
def autovacuum_analyze_scale_factor(self) -> Optional[float]:
|
|
33893
33771
|
"""
|
|
33894
|
-
Specifies a fraction of the table size to add to autovacuum_analyze_threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20
|
|
33772
|
+
Specifies a fraction of the table size to add to autovacuum_analyze_threshold when deciding whether to trigger an ANALYZE. The default is 0.2 (20% of table size).
|
|
33895
33773
|
"""
|
|
33896
33774
|
return pulumi.get(self, "autovacuum_analyze_scale_factor")
|
|
33897
33775
|
|
|
@@ -33947,7 +33825,7 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
33947
33825
|
@pulumi.getter(name="autovacuumVacuumScaleFactor")
|
|
33948
33826
|
def autovacuum_vacuum_scale_factor(self) -> Optional[float]:
|
|
33949
33827
|
"""
|
|
33950
|
-
Specifies a fraction of the table size to add to autovacuum_vacuum_threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20
|
|
33828
|
+
Specifies a fraction of the table size to add to autovacuum_vacuum_threshold when deciding whether to trigger a VACUUM. The default is 0.2 (20% of table size).
|
|
33951
33829
|
"""
|
|
33952
33830
|
return pulumi.get(self, "autovacuum_vacuum_scale_factor")
|
|
33953
33831
|
|
|
@@ -34203,7 +34081,7 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
34203
34081
|
@pulumi.getter(name="pgStatMonitorDotPgsmMaxBuckets")
|
|
34204
34082
|
def pg_stat_monitor_dot_pgsm_max_buckets(self) -> Optional[int]:
|
|
34205
34083
|
"""
|
|
34206
|
-
Sets the maximum number of buckets
|
|
34084
|
+
Sets the maximum number of buckets.
|
|
34207
34085
|
"""
|
|
34208
34086
|
return pulumi.get(self, "pg_stat_monitor_dot_pgsm_max_buckets")
|
|
34209
34087
|
|
|
@@ -34901,8 +34779,8 @@ class GetProjectTagResult(dict):
|
|
|
34901
34779
|
key: str,
|
|
34902
34780
|
value: str):
|
|
34903
34781
|
"""
|
|
34904
|
-
:param str key: Project tag key
|
|
34905
|
-
:param str value: Project tag value
|
|
34782
|
+
:param str key: Project tag key.
|
|
34783
|
+
:param str value: Project tag value.
|
|
34906
34784
|
"""
|
|
34907
34785
|
pulumi.set(__self__, "key", key)
|
|
34908
34786
|
pulumi.set(__self__, "value", value)
|
|
@@ -34911,7 +34789,7 @@ class GetProjectTagResult(dict):
|
|
|
34911
34789
|
@pulumi.getter
|
|
34912
34790
|
def key(self) -> str:
|
|
34913
34791
|
"""
|
|
34914
|
-
Project tag key
|
|
34792
|
+
Project tag key.
|
|
34915
34793
|
"""
|
|
34916
34794
|
return pulumi.get(self, "key")
|
|
34917
34795
|
|
|
@@ -34919,7 +34797,7 @@ class GetProjectTagResult(dict):
|
|
|
34919
34797
|
@pulumi.getter
|
|
34920
34798
|
def value(self) -> str:
|
|
34921
34799
|
"""
|
|
34922
|
-
Project tag value
|
|
34800
|
+
Project tag value.
|
|
34923
34801
|
"""
|
|
34924
34802
|
return pulumi.get(self, "value")
|
|
34925
34803
|
|
|
@@ -35019,12 +34897,6 @@ class GetRedisComponentResult(dict):
|
|
|
35019
34897
|
return pulumi.get(self, "usage")
|
|
35020
34898
|
|
|
35021
34899
|
|
|
35022
|
-
@pulumi.output_type
|
|
35023
|
-
class GetRedisRediResult(dict):
|
|
35024
|
-
def __init__(__self__):
|
|
35025
|
-
pass
|
|
35026
|
-
|
|
35027
|
-
|
|
35028
34900
|
@pulumi.output_type
|
|
35029
34901
|
class GetRedisRedisUserConfigResult(dict):
|
|
35030
34902
|
def __init__(__self__, *,
|
|
@@ -35137,9 +35009,6 @@ class GetRedisRedisUserConfigResult(dict):
|
|
|
35137
35009
|
"""
|
|
35138
35010
|
Additional Cloud Regions for Backup Replication.
|
|
35139
35011
|
"""
|
|
35140
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
35141
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
35142
|
-
|
|
35143
35012
|
return pulumi.get(self, "additional_backup_regions")
|
|
35144
35013
|
|
|
35145
35014
|
@property
|
|
@@ -36072,14 +35941,18 @@ class GetServiceIntegrationDatadogUserConfigDatadogTagResult(dict):
|
|
|
36072
35941
|
@pulumi.output_type
|
|
36073
35942
|
class GetServiceIntegrationDatadogUserConfigOpensearchResult(dict):
|
|
36074
35943
|
def __init__(__self__, *,
|
|
35944
|
+
cluster_stats_enabled: Optional[bool] = None,
|
|
36075
35945
|
index_stats_enabled: Optional[bool] = None,
|
|
36076
35946
|
pending_task_stats_enabled: Optional[bool] = None,
|
|
36077
35947
|
pshard_stats_enabled: Optional[bool] = None):
|
|
36078
35948
|
"""
|
|
35949
|
+
:param bool cluster_stats_enabled: Enable Datadog Opensearch Cluster Monitoring.
|
|
36079
35950
|
:param bool index_stats_enabled: Enable Datadog Opensearch Index Monitoring.
|
|
36080
35951
|
:param bool pending_task_stats_enabled: Enable Datadog Opensearch Pending Task Monitoring.
|
|
36081
35952
|
:param bool pshard_stats_enabled: Enable Datadog Opensearch Primary Shard Monitoring.
|
|
36082
35953
|
"""
|
|
35954
|
+
if cluster_stats_enabled is not None:
|
|
35955
|
+
pulumi.set(__self__, "cluster_stats_enabled", cluster_stats_enabled)
|
|
36083
35956
|
if index_stats_enabled is not None:
|
|
36084
35957
|
pulumi.set(__self__, "index_stats_enabled", index_stats_enabled)
|
|
36085
35958
|
if pending_task_stats_enabled is not None:
|
|
@@ -36087,6 +35960,14 @@ class GetServiceIntegrationDatadogUserConfigOpensearchResult(dict):
|
|
|
36087
35960
|
if pshard_stats_enabled is not None:
|
|
36088
35961
|
pulumi.set(__self__, "pshard_stats_enabled", pshard_stats_enabled)
|
|
36089
35962
|
|
|
35963
|
+
@property
|
|
35964
|
+
@pulumi.getter(name="clusterStatsEnabled")
|
|
35965
|
+
def cluster_stats_enabled(self) -> Optional[bool]:
|
|
35966
|
+
"""
|
|
35967
|
+
Enable Datadog Opensearch Cluster Monitoring.
|
|
35968
|
+
"""
|
|
35969
|
+
return pulumi.get(self, "cluster_stats_enabled")
|
|
35970
|
+
|
|
36090
35971
|
@property
|
|
36091
35972
|
@pulumi.getter(name="indexStatsEnabled")
|
|
36092
35973
|
def index_stats_enabled(self) -> Optional[bool]:
|
|
@@ -36426,7 +36307,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudBigqueryResult(dict):
|
|
|
36426
36307
|
service_account_credentials: str):
|
|
36427
36308
|
"""
|
|
36428
36309
|
:param str project_id: GCP project id.
|
|
36429
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36310
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36430
36311
|
"""
|
|
36431
36312
|
pulumi.set(__self__, "project_id", project_id)
|
|
36432
36313
|
pulumi.set(__self__, "service_account_credentials", service_account_credentials)
|
|
@@ -36443,7 +36324,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudBigqueryResult(dict):
|
|
|
36443
36324
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
36444
36325
|
def service_account_credentials(self) -> str:
|
|
36445
36326
|
"""
|
|
36446
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36327
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36447
36328
|
"""
|
|
36448
36329
|
return pulumi.get(self, "service_account_credentials")
|
|
36449
36330
|
|
|
@@ -36457,7 +36338,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfigResult(di
|
|
|
36457
36338
|
"""
|
|
36458
36339
|
:param str log_id: Google Cloud Logging log id.
|
|
36459
36340
|
:param str project_id: GCP project id.
|
|
36460
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36341
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36461
36342
|
"""
|
|
36462
36343
|
pulumi.set(__self__, "log_id", log_id)
|
|
36463
36344
|
pulumi.set(__self__, "project_id", project_id)
|
|
@@ -36483,7 +36364,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfigResult(di
|
|
|
36483
36364
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
36484
36365
|
def service_account_credentials(self) -> str:
|
|
36485
36366
|
"""
|
|
36486
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36367
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36487
36368
|
"""
|
|
36488
36369
|
return pulumi.get(self, "service_account_credentials")
|
|
36489
36370
|
|
|
@@ -37524,7 +37405,7 @@ class GetServiceIntegrationMetricsUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37524
37405
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
37525
37406
|
"""
|
|
37526
37407
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE_SCHEMA.EVENT_WAITS.
|
|
37527
|
-
:param bool gather_file_events_stats:
|
|
37408
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37528
37409
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMARY_BY_INDEX_USAGE.
|
|
37529
37410
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
37530
37411
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION_SCHEMA.INNODB_METRICS.
|
|
@@ -37579,7 +37460,7 @@ class GetServiceIntegrationMetricsUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37579
37460
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
37580
37461
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
37581
37462
|
"""
|
|
37582
|
-
|
|
37463
|
+
Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37583
37464
|
"""
|
|
37584
37465
|
return pulumi.get(self, "gather_file_events_stats")
|
|
37585
37466
|
|
|
@@ -37737,7 +37618,7 @@ class GetServiceIntegrationPrometheusUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37737
37618
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
37738
37619
|
"""
|
|
37739
37620
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE_SCHEMA.EVENT_WAITS.
|
|
37740
|
-
:param bool gather_file_events_stats:
|
|
37621
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37741
37622
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMARY_BY_INDEX_USAGE.
|
|
37742
37623
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
37743
37624
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION_SCHEMA.INNODB_METRICS.
|
|
@@ -37792,7 +37673,7 @@ class GetServiceIntegrationPrometheusUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37792
37673
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
37793
37674
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
37794
37675
|
"""
|
|
37795
|
-
|
|
37676
|
+
Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37796
37677
|
"""
|
|
37797
37678
|
return pulumi.get(self, "gather_file_events_stats")
|
|
37798
37679
|
|