pulumi-aiven 6.14.0a1714562227__py3-none-any.whl → 6.15.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pulumi-aiven might be problematic. Click here for more details.
- pulumi_aiven/__init__.py +19 -0
- pulumi_aiven/_inputs.py +232 -150
- pulumi_aiven/account_authentication.py +2 -2
- pulumi_aiven/account_team_member.py +2 -2
- pulumi_aiven/aws_privatelink.py +14 -14
- pulumi_aiven/azure_privatelink.py +14 -14
- pulumi_aiven/azure_privatelink_connection_approval.py +14 -14
- pulumi_aiven/billing_group.py +18 -16
- pulumi_aiven/cassandra.py +16 -44
- pulumi_aiven/cassandra_user.py +16 -16
- pulumi_aiven/clickhouse.py +16 -44
- pulumi_aiven/clickhouse_database.py +16 -43
- pulumi_aiven/clickhouse_grant.py +14 -14
- pulumi_aiven/clickhouse_role.py +18 -18
- pulumi_aiven/clickhouse_user.py +14 -14
- pulumi_aiven/connection_pool.py +16 -16
- pulumi_aiven/dragonfly.py +124 -263
- pulumi_aiven/flink.py +16 -16
- pulumi_aiven/flink_application.py +16 -16
- pulumi_aiven/flink_application_deployment.py +21 -21
- pulumi_aiven/flink_application_version.py +14 -14
- pulumi_aiven/gcp_privatelink.py +21 -21
- pulumi_aiven/gcp_privatelink_connection_approval.py +21 -21
- pulumi_aiven/gcp_vpc_peering_connection.py +2 -2
- pulumi_aiven/get_aws_privatelink.py +6 -6
- pulumi_aiven/get_azure_privatelink.py +6 -6
- pulumi_aiven/get_billing_group.py +2 -2
- pulumi_aiven/get_cassanda.py +5 -18
- pulumi_aiven/get_cassandra.py +5 -18
- pulumi_aiven/get_cassandra_user.py +6 -6
- pulumi_aiven/get_clickhouse.py +5 -18
- pulumi_aiven/get_clickhouse_database.py +6 -6
- pulumi_aiven/get_clickhouse_user.py +6 -6
- pulumi_aiven/get_connection_pool.py +6 -6
- pulumi_aiven/get_dragonfly.py +27 -20
- pulumi_aiven/get_flink.py +4 -4
- pulumi_aiven/get_flink_application.py +6 -6
- pulumi_aiven/get_flink_application_version.py +6 -6
- pulumi_aiven/get_grafana.py +5 -18
- pulumi_aiven/get_kafka.py +13 -13
- pulumi_aiven/get_kafka_acl.py +6 -6
- pulumi_aiven/get_kafka_connect.py +5 -18
- pulumi_aiven/get_kafka_connector.py +6 -6
- pulumi_aiven/get_kafka_mirror_maker.py +5 -18
- pulumi_aiven/get_kafka_schema.py +6 -6
- pulumi_aiven/get_kafka_schema_configuration.py +6 -6
- pulumi_aiven/get_kafka_schema_registry_acl.py +6 -6
- pulumi_aiven/get_kafka_topic.py +16 -16
- pulumi_aiven/get_kafka_user.py +6 -6
- pulumi_aiven/get_m3_aggregator.py +5 -18
- pulumi_aiven/get_m3_db.py +5 -18
- pulumi_aiven/get_m3db_user.py +6 -6
- pulumi_aiven/get_mirror_maker_replication_flow.py +6 -6
- pulumi_aiven/get_my_sql.py +5 -18
- pulumi_aiven/get_mysql_database.py +6 -6
- pulumi_aiven/get_mysql_user.py +6 -6
- pulumi_aiven/get_open_search.py +4 -4
- pulumi_aiven/get_open_search_acl_config.py +6 -6
- pulumi_aiven/get_open_search_acl_rule.py +6 -6
- pulumi_aiven/get_opensearch_security_plugin_config.py +6 -6
- pulumi_aiven/get_opensearch_user.py +6 -6
- pulumi_aiven/get_organization_application_user.py +156 -0
- pulumi_aiven/get_pg.py +4 -4
- pulumi_aiven/get_pg_database.py +6 -6
- pulumi_aiven/get_pg_user.py +6 -6
- pulumi_aiven/get_project.py +17 -17
- pulumi_aiven/get_project_user.py +7 -7
- pulumi_aiven/get_redis.py +5 -18
- pulumi_aiven/get_redis_user.py +6 -6
- pulumi_aiven/get_service_integration.py +19 -29
- pulumi_aiven/grafana.py +78 -261
- pulumi_aiven/influx_db.py +21 -21
- pulumi_aiven/influxdb_database.py +21 -41
- pulumi_aiven/influxdb_user.py +21 -21
- pulumi_aiven/kafka.py +43 -43
- pulumi_aiven/kafka_acl.py +16 -16
- pulumi_aiven/kafka_connect.py +78 -261
- pulumi_aiven/kafka_connector.py +14 -14
- pulumi_aiven/kafka_mirror_maker.py +78 -261
- pulumi_aiven/kafka_schema.py +14 -14
- pulumi_aiven/kafka_schema_configuration.py +16 -16
- pulumi_aiven/kafka_schema_registry_acl.py +54 -14
- pulumi_aiven/kafka_topic.py +40 -60
- pulumi_aiven/kafka_user.py +16 -16
- pulumi_aiven/m3_aggregator.py +78 -261
- pulumi_aiven/m3_db.py +78 -261
- pulumi_aiven/m3db_user.py +14 -14
- pulumi_aiven/mirror_maker_replication_flow.py +14 -14
- pulumi_aiven/my_sql.py +78 -261
- pulumi_aiven/mysql_database.py +16 -36
- pulumi_aiven/mysql_user.py +16 -16
- pulumi_aiven/open_search.py +14 -14
- pulumi_aiven/open_search_acl_config.py +16 -20
- pulumi_aiven/open_search_acl_rule.py +16 -16
- pulumi_aiven/opensearch_security_plugin_config.py +16 -16
- pulumi_aiven/opensearch_user.py +16 -16
- pulumi_aiven/organization_application_user.py +323 -0
- pulumi_aiven/organization_application_user_token.py +742 -0
- pulumi_aiven/organization_group_project.py +20 -6
- pulumi_aiven/organization_user.py +2 -20
- pulumi_aiven/outputs.py +259 -266
- pulumi_aiven/pg.py +16 -16
- pulumi_aiven/pg_database.py +16 -36
- pulumi_aiven/pg_user.py +16 -16
- pulumi_aiven/project.py +71 -71
- pulumi_aiven/project_user.py +22 -22
- pulumi_aiven/project_vpc.py +7 -7
- pulumi_aiven/redis.py +78 -261
- pulumi_aiven/redis_user.py +16 -16
- pulumi_aiven/service_integration.py +54 -54
- pulumi_aiven/static_ip.py +7 -7
- pulumi_aiven/transit_gateway_vpc_attachment.py +2 -2
- {pulumi_aiven-6.14.0a1714562227.dist-info → pulumi_aiven-6.15.0.dist-info}/METADATA +1 -1
- pulumi_aiven-6.15.0.dist-info/RECORD +153 -0
- pulumi_aiven-6.14.0a1714562227.dist-info/RECORD +0 -150
- {pulumi_aiven-6.14.0a1714562227.dist-info → pulumi_aiven-6.15.0.dist-info}/WHEEL +0 -0
- {pulumi_aiven-6.14.0a1714562227.dist-info → pulumi_aiven-6.15.0.dist-info}/top_level.txt +0 -0
pulumi_aiven/outputs.py
CHANGED
|
@@ -12,7 +12,6 @@ from . import outputs
|
|
|
12
12
|
|
|
13
13
|
__all__ = [
|
|
14
14
|
'AccountAuthenticationSamlFieldMapping',
|
|
15
|
-
'CassandraCassandra',
|
|
16
15
|
'CassandraCassandraUserConfig',
|
|
17
16
|
'CassandraCassandraUserConfigCassandra',
|
|
18
17
|
'CassandraCassandraUserConfigIpFilterObject',
|
|
@@ -22,7 +21,6 @@ __all__ = [
|
|
|
22
21
|
'CassandraServiceIntegration',
|
|
23
22
|
'CassandraTag',
|
|
24
23
|
'CassandraTechEmail',
|
|
25
|
-
'ClickhouseClickhouse',
|
|
26
24
|
'ClickhouseClickhouseUserConfig',
|
|
27
25
|
'ClickhouseClickhouseUserConfigIpFilterObject',
|
|
28
26
|
'ClickhouseClickhouseUserConfigPrivateAccess',
|
|
@@ -35,7 +33,6 @@ __all__ = [
|
|
|
35
33
|
'ClickhouseTag',
|
|
36
34
|
'ClickhouseTechEmail',
|
|
37
35
|
'DragonflyComponent',
|
|
38
|
-
'DragonflyDragonfly',
|
|
39
36
|
'DragonflyDragonflyUserConfig',
|
|
40
37
|
'DragonflyDragonflyUserConfigIpFilterObject',
|
|
41
38
|
'DragonflyDragonflyUserConfigMigration',
|
|
@@ -56,7 +53,6 @@ __all__ = [
|
|
|
56
53
|
'FlinkTag',
|
|
57
54
|
'FlinkTechEmail',
|
|
58
55
|
'GrafanaComponent',
|
|
59
|
-
'GrafanaGrafana',
|
|
60
56
|
'GrafanaGrafanaUserConfig',
|
|
61
57
|
'GrafanaGrafanaUserConfigAuthAzuread',
|
|
62
58
|
'GrafanaGrafanaUserConfigAuthGenericOauth',
|
|
@@ -86,7 +82,6 @@ __all__ = [
|
|
|
86
82
|
'InfluxDbTechEmail',
|
|
87
83
|
'KafkaComponent',
|
|
88
84
|
'KafkaConnectComponent',
|
|
89
|
-
'KafkaConnectKafkaConnect',
|
|
90
85
|
'KafkaConnectKafkaConnectUserConfig',
|
|
91
86
|
'KafkaConnectKafkaConnectUserConfigIpFilterObject',
|
|
92
87
|
'KafkaConnectKafkaConnectUserConfigKafkaConnect',
|
|
@@ -111,7 +106,6 @@ __all__ = [
|
|
|
111
106
|
'KafkaKafkaUserConfigTieredStorage',
|
|
112
107
|
'KafkaKafkaUserConfigTieredStorageLocalCache',
|
|
113
108
|
'KafkaMirrorMakerComponent',
|
|
114
|
-
'KafkaMirrorMakerKafkaMirrormaker',
|
|
115
109
|
'KafkaMirrorMakerKafkaMirrormakerUserConfig',
|
|
116
110
|
'KafkaMirrorMakerKafkaMirrormakerUserConfigIpFilterObject',
|
|
117
111
|
'KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormaker',
|
|
@@ -124,14 +118,12 @@ __all__ = [
|
|
|
124
118
|
'KafkaTopicConfig',
|
|
125
119
|
'KafkaTopicTag',
|
|
126
120
|
'M3AggregatorComponent',
|
|
127
|
-
'M3AggregatorM3aggregator',
|
|
128
121
|
'M3AggregatorM3aggregatorUserConfig',
|
|
129
122
|
'M3AggregatorM3aggregatorUserConfigIpFilterObject',
|
|
130
123
|
'M3AggregatorServiceIntegration',
|
|
131
124
|
'M3AggregatorTag',
|
|
132
125
|
'M3AggregatorTechEmail',
|
|
133
126
|
'M3DbComponent',
|
|
134
|
-
'M3DbM3db',
|
|
135
127
|
'M3DbM3dbUserConfig',
|
|
136
128
|
'M3DbM3dbUserConfigIpFilterObject',
|
|
137
129
|
'M3DbM3dbUserConfigLimits',
|
|
@@ -150,7 +142,6 @@ __all__ = [
|
|
|
150
142
|
'M3DbTag',
|
|
151
143
|
'M3DbTechEmail',
|
|
152
144
|
'MySqlComponent',
|
|
153
|
-
'MySqlMysql',
|
|
154
145
|
'MySqlMysqlUserConfig',
|
|
155
146
|
'MySqlMysqlUserConfigIpFilterObject',
|
|
156
147
|
'MySqlMysqlUserConfigMigration',
|
|
@@ -180,6 +171,8 @@ __all__ = [
|
|
|
180
171
|
'OpenSearchServiceIntegration',
|
|
181
172
|
'OpenSearchTag',
|
|
182
173
|
'OpenSearchTechEmail',
|
|
174
|
+
'OrganizationApplicationUserTimeouts',
|
|
175
|
+
'OrganizationApplicationUserTokenTimeouts',
|
|
183
176
|
'OrganizationGroupProjectTimeouts',
|
|
184
177
|
'OrganizationTimeouts',
|
|
185
178
|
'OrganizationUserGroupMemberTimeouts',
|
|
@@ -202,7 +195,6 @@ __all__ = [
|
|
|
202
195
|
'PgTechEmail',
|
|
203
196
|
'ProjectTag',
|
|
204
197
|
'RedisComponent',
|
|
205
|
-
'RedisRedi',
|
|
206
198
|
'RedisRedisUserConfig',
|
|
207
199
|
'RedisRedisUserConfigIpFilterObject',
|
|
208
200
|
'RedisRedisUserConfigMigration',
|
|
@@ -255,7 +247,6 @@ __all__ = [
|
|
|
255
247
|
'ServiceIntegrationPrometheusUserConfigSourceMysql',
|
|
256
248
|
'ServiceIntegrationPrometheusUserConfigSourceMysqlTelegraf',
|
|
257
249
|
'GetAccountAuthenticationSamlFieldMappingResult',
|
|
258
|
-
'GetCassandaCassandraResult',
|
|
259
250
|
'GetCassandaCassandraUserConfigResult',
|
|
260
251
|
'GetCassandaCassandraUserConfigCassandraResult',
|
|
261
252
|
'GetCassandaCassandraUserConfigIpFilterObjectResult',
|
|
@@ -265,7 +256,6 @@ __all__ = [
|
|
|
265
256
|
'GetCassandaServiceIntegrationResult',
|
|
266
257
|
'GetCassandaTagResult',
|
|
267
258
|
'GetCassandaTechEmailResult',
|
|
268
|
-
'GetCassandraCassandraResult',
|
|
269
259
|
'GetCassandraCassandraUserConfigResult',
|
|
270
260
|
'GetCassandraCassandraUserConfigCassandraResult',
|
|
271
261
|
'GetCassandraCassandraUserConfigIpFilterObjectResult',
|
|
@@ -275,7 +265,6 @@ __all__ = [
|
|
|
275
265
|
'GetCassandraServiceIntegrationResult',
|
|
276
266
|
'GetCassandraTagResult',
|
|
277
267
|
'GetCassandraTechEmailResult',
|
|
278
|
-
'GetClickhouseClickhouseResult',
|
|
279
268
|
'GetClickhouseClickhouseUserConfigResult',
|
|
280
269
|
'GetClickhouseClickhouseUserConfigIpFilterObjectResult',
|
|
281
270
|
'GetClickhouseClickhouseUserConfigPrivateAccessResult',
|
|
@@ -286,7 +275,6 @@ __all__ = [
|
|
|
286
275
|
'GetClickhouseTagResult',
|
|
287
276
|
'GetClickhouseTechEmailResult',
|
|
288
277
|
'GetDragonflyComponentResult',
|
|
289
|
-
'GetDragonflyDragonflyResult',
|
|
290
278
|
'GetDragonflyDragonflyUserConfigResult',
|
|
291
279
|
'GetDragonflyDragonflyUserConfigIpFilterObjectResult',
|
|
292
280
|
'GetDragonflyDragonflyUserConfigMigrationResult',
|
|
@@ -307,7 +295,6 @@ __all__ = [
|
|
|
307
295
|
'GetFlinkTagResult',
|
|
308
296
|
'GetFlinkTechEmailResult',
|
|
309
297
|
'GetGrafanaComponentResult',
|
|
310
|
-
'GetGrafanaGrafanaResult',
|
|
311
298
|
'GetGrafanaGrafanaUserConfigResult',
|
|
312
299
|
'GetGrafanaGrafanaUserConfigAuthAzureadResult',
|
|
313
300
|
'GetGrafanaGrafanaUserConfigAuthGenericOauthResult',
|
|
@@ -337,7 +324,6 @@ __all__ = [
|
|
|
337
324
|
'GetInfluxDbTechEmailResult',
|
|
338
325
|
'GetKafkaComponentResult',
|
|
339
326
|
'GetKafkaConnectComponentResult',
|
|
340
|
-
'GetKafkaConnectKafkaConnectResult',
|
|
341
327
|
'GetKafkaConnectKafkaConnectUserConfigResult',
|
|
342
328
|
'GetKafkaConnectKafkaConnectUserConfigIpFilterObjectResult',
|
|
343
329
|
'GetKafkaConnectKafkaConnectUserConfigKafkaConnectResult',
|
|
@@ -362,7 +348,6 @@ __all__ = [
|
|
|
362
348
|
'GetKafkaKafkaUserConfigTieredStorageResult',
|
|
363
349
|
'GetKafkaKafkaUserConfigTieredStorageLocalCacheResult',
|
|
364
350
|
'GetKafkaMirrorMakerComponentResult',
|
|
365
|
-
'GetKafkaMirrorMakerKafkaMirrormakerResult',
|
|
366
351
|
'GetKafkaMirrorMakerKafkaMirrormakerUserConfigResult',
|
|
367
352
|
'GetKafkaMirrorMakerKafkaMirrormakerUserConfigIpFilterObjectResult',
|
|
368
353
|
'GetKafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormakerResult',
|
|
@@ -375,14 +360,12 @@ __all__ = [
|
|
|
375
360
|
'GetKafkaTopicConfigResult',
|
|
376
361
|
'GetKafkaTopicTagResult',
|
|
377
362
|
'GetM3AggregatorComponentResult',
|
|
378
|
-
'GetM3AggregatorM3aggregatorResult',
|
|
379
363
|
'GetM3AggregatorM3aggregatorUserConfigResult',
|
|
380
364
|
'GetM3AggregatorM3aggregatorUserConfigIpFilterObjectResult',
|
|
381
365
|
'GetM3AggregatorServiceIntegrationResult',
|
|
382
366
|
'GetM3AggregatorTagResult',
|
|
383
367
|
'GetM3AggregatorTechEmailResult',
|
|
384
368
|
'GetM3DbComponentResult',
|
|
385
|
-
'GetM3DbM3dbResult',
|
|
386
369
|
'GetM3DbM3dbUserConfigResult',
|
|
387
370
|
'GetM3DbM3dbUserConfigIpFilterObjectResult',
|
|
388
371
|
'GetM3DbM3dbUserConfigLimitsResult',
|
|
@@ -401,7 +384,6 @@ __all__ = [
|
|
|
401
384
|
'GetM3DbTagResult',
|
|
402
385
|
'GetM3DbTechEmailResult',
|
|
403
386
|
'GetMySqlComponentResult',
|
|
404
|
-
'GetMySqlMysqlResult',
|
|
405
387
|
'GetMySqlMysqlUserConfigResult',
|
|
406
388
|
'GetMySqlMysqlUserConfigIpFilterObjectResult',
|
|
407
389
|
'GetMySqlMysqlUserConfigMigrationResult',
|
|
@@ -450,7 +432,6 @@ __all__ = [
|
|
|
450
432
|
'GetPgTechEmailResult',
|
|
451
433
|
'GetProjectTagResult',
|
|
452
434
|
'GetRedisComponentResult',
|
|
453
|
-
'GetRedisRediResult',
|
|
454
435
|
'GetRedisRedisUserConfigResult',
|
|
455
436
|
'GetRedisRedisUserConfigIpFilterObjectResult',
|
|
456
437
|
'GetRedisRedisUserConfigMigrationResult',
|
|
@@ -592,12 +573,6 @@ class AccountAuthenticationSamlFieldMapping(dict):
|
|
|
592
573
|
return pulumi.get(self, "real_name")
|
|
593
574
|
|
|
594
575
|
|
|
595
|
-
@pulumi.output_type
|
|
596
|
-
class CassandraCassandra(dict):
|
|
597
|
-
def __init__(__self__):
|
|
598
|
-
pass
|
|
599
|
-
|
|
600
|
-
|
|
601
576
|
@pulumi.output_type
|
|
602
577
|
class CassandraCassandraUserConfig(dict):
|
|
603
578
|
@staticmethod
|
|
@@ -666,7 +641,7 @@ class CassandraCassandraUserConfig(dict):
|
|
|
666
641
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
667
642
|
:param int backup_hour: The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
668
643
|
:param int backup_minute: The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
669
|
-
:param 'CassandraCassandraUserConfigCassandraArgs' cassandra:
|
|
644
|
+
:param 'CassandraCassandraUserConfigCassandraArgs' cassandra: Cassandra configuration values
|
|
670
645
|
:param str cassandra_version: Cassandra version.
|
|
671
646
|
:param Sequence['CassandraCassandraUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
672
647
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -744,7 +719,7 @@ class CassandraCassandraUserConfig(dict):
|
|
|
744
719
|
@pulumi.getter
|
|
745
720
|
def cassandra(self) -> Optional['outputs.CassandraCassandraUserConfigCassandra']:
|
|
746
721
|
"""
|
|
747
|
-
|
|
722
|
+
Cassandra configuration values
|
|
748
723
|
"""
|
|
749
724
|
return pulumi.get(self, "cassandra")
|
|
750
725
|
|
|
@@ -1195,12 +1170,6 @@ class CassandraTechEmail(dict):
|
|
|
1195
1170
|
return pulumi.get(self, "email")
|
|
1196
1171
|
|
|
1197
1172
|
|
|
1198
|
-
@pulumi.output_type
|
|
1199
|
-
class ClickhouseClickhouse(dict):
|
|
1200
|
-
def __init__(__self__):
|
|
1201
|
-
pass
|
|
1202
|
-
|
|
1203
|
-
|
|
1204
1173
|
@pulumi.output_type
|
|
1205
1174
|
class ClickhouseClickhouseUserConfig(dict):
|
|
1206
1175
|
@staticmethod
|
|
@@ -2076,12 +2045,6 @@ class DragonflyComponent(dict):
|
|
|
2076
2045
|
return pulumi.get(self, "usage")
|
|
2077
2046
|
|
|
2078
2047
|
|
|
2079
|
-
@pulumi.output_type
|
|
2080
|
-
class DragonflyDragonfly(dict):
|
|
2081
|
-
def __init__(__self__):
|
|
2082
|
-
pass
|
|
2083
|
-
|
|
2084
|
-
|
|
2085
2048
|
@pulumi.output_type
|
|
2086
2049
|
class DragonflyDragonflyUserConfig(dict):
|
|
2087
2050
|
@staticmethod
|
|
@@ -2089,6 +2052,8 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2089
2052
|
suggest = None
|
|
2090
2053
|
if key == "cacheMode":
|
|
2091
2054
|
suggest = "cache_mode"
|
|
2055
|
+
elif key == "dragonflyPersistence":
|
|
2056
|
+
suggest = "dragonfly_persistence"
|
|
2092
2057
|
elif key == "dragonflySsl":
|
|
2093
2058
|
suggest = "dragonfly_ssl"
|
|
2094
2059
|
elif key == "ipFilterObjects":
|
|
@@ -2127,6 +2092,7 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2127
2092
|
|
|
2128
2093
|
def __init__(__self__, *,
|
|
2129
2094
|
cache_mode: Optional[bool] = None,
|
|
2095
|
+
dragonfly_persistence: Optional[str] = None,
|
|
2130
2096
|
dragonfly_ssl: Optional[bool] = None,
|
|
2131
2097
|
ip_filter_objects: Optional[Sequence['outputs.DragonflyDragonflyUserConfigIpFilterObject']] = None,
|
|
2132
2098
|
ip_filter_strings: Optional[Sequence[str]] = None,
|
|
@@ -2142,6 +2108,7 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2142
2108
|
static_ips: Optional[bool] = None):
|
|
2143
2109
|
"""
|
|
2144
2110
|
:param bool cache_mode: Evict entries when getting close to maxmemory limit. The default value is `false`.
|
|
2111
|
+
:param str dragonfly_persistence: When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
2145
2112
|
:param bool dragonfly_ssl: Require SSL to access Dragonfly. The default value is `true`.
|
|
2146
2113
|
:param Sequence['DragonflyDragonflyUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
2147
2114
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -2158,6 +2125,8 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2158
2125
|
"""
|
|
2159
2126
|
if cache_mode is not None:
|
|
2160
2127
|
pulumi.set(__self__, "cache_mode", cache_mode)
|
|
2128
|
+
if dragonfly_persistence is not None:
|
|
2129
|
+
pulumi.set(__self__, "dragonfly_persistence", dragonfly_persistence)
|
|
2161
2130
|
if dragonfly_ssl is not None:
|
|
2162
2131
|
pulumi.set(__self__, "dragonfly_ssl", dragonfly_ssl)
|
|
2163
2132
|
if ip_filter_objects is not None:
|
|
@@ -2193,6 +2162,14 @@ class DragonflyDragonflyUserConfig(dict):
|
|
|
2193
2162
|
"""
|
|
2194
2163
|
return pulumi.get(self, "cache_mode")
|
|
2195
2164
|
|
|
2165
|
+
@property
|
|
2166
|
+
@pulumi.getter(name="dragonflyPersistence")
|
|
2167
|
+
def dragonfly_persistence(self) -> Optional[str]:
|
|
2168
|
+
"""
|
|
2169
|
+
When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
2170
|
+
"""
|
|
2171
|
+
return pulumi.get(self, "dragonfly_persistence")
|
|
2172
|
+
|
|
2196
2173
|
@property
|
|
2197
2174
|
@pulumi.getter(name="dragonflySsl")
|
|
2198
2175
|
def dragonfly_ssl(self) -> Optional[bool]:
|
|
@@ -3325,12 +3302,6 @@ class GrafanaComponent(dict):
|
|
|
3325
3302
|
return pulumi.get(self, "usage")
|
|
3326
3303
|
|
|
3327
3304
|
|
|
3328
|
-
@pulumi.output_type
|
|
3329
|
-
class GrafanaGrafana(dict):
|
|
3330
|
-
def __init__(__self__):
|
|
3331
|
-
pass
|
|
3332
|
-
|
|
3333
|
-
|
|
3334
3305
|
@pulumi.output_type
|
|
3335
3306
|
class GrafanaGrafanaUserConfig(dict):
|
|
3336
3307
|
@staticmethod
|
|
@@ -3477,7 +3448,7 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3477
3448
|
viewers_can_edit: Optional[bool] = None):
|
|
3478
3449
|
"""
|
|
3479
3450
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
3480
|
-
:param bool alerting_enabled: Enable or disable Grafana legacy alerting functionality. This should not be enabled with
|
|
3451
|
+
:param bool alerting_enabled: Enable or disable Grafana legacy alerting functionality. This should not be enabled with unified*alerting*enabled.
|
|
3481
3452
|
:param str alerting_error_or_timeout: Default error or timeout setting for new alerting rules.
|
|
3482
3453
|
:param int alerting_max_annotations_to_keep: Max number of alert annotations that Grafana stores. 0 (default) keeps all alert annotations.
|
|
3483
3454
|
:param str alerting_nodata_or_nullvalues: Default value for 'no data or null values' for new alerting rules.
|
|
@@ -3514,7 +3485,7 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3514
3485
|
:param str service_to_fork_from: Name of another service to fork from. This has effect only when a new service is being created.
|
|
3515
3486
|
:param 'GrafanaGrafanaUserConfigSmtpServerArgs' smtp_server: SMTP server settings
|
|
3516
3487
|
:param bool static_ips: Use static public IP addresses.
|
|
3517
|
-
:param bool unified_alerting_enabled: Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set
|
|
3488
|
+
:param bool unified_alerting_enabled: Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set unified*alerting*enabled to false and alerting_enabled to true. See https://grafana.com/docs/grafana/latest/alerting/set-up/migrating-alerts/ for more details.
|
|
3518
3489
|
:param bool user_auto_assign_org: Auto-assign new users on signup to main organization. Defaults to false.
|
|
3519
3490
|
:param str user_auto_assign_org_role: Set role for new signups. Defaults to Viewer.
|
|
3520
3491
|
:param bool viewers_can_edit: Users with view-only permission can edit but not save dashboards.
|
|
@@ -3610,16 +3581,13 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3610
3581
|
"""
|
|
3611
3582
|
Additional Cloud Regions for Backup Replication.
|
|
3612
3583
|
"""
|
|
3613
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
3614
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
3615
|
-
|
|
3616
3584
|
return pulumi.get(self, "additional_backup_regions")
|
|
3617
3585
|
|
|
3618
3586
|
@property
|
|
3619
3587
|
@pulumi.getter(name="alertingEnabled")
|
|
3620
3588
|
def alerting_enabled(self) -> Optional[bool]:
|
|
3621
3589
|
"""
|
|
3622
|
-
Enable or disable Grafana legacy alerting functionality. This should not be enabled with
|
|
3590
|
+
Enable or disable Grafana legacy alerting functionality. This should not be enabled with unified*alerting*enabled.
|
|
3623
3591
|
"""
|
|
3624
3592
|
return pulumi.get(self, "alerting_enabled")
|
|
3625
3593
|
|
|
@@ -3918,7 +3886,7 @@ class GrafanaGrafanaUserConfig(dict):
|
|
|
3918
3886
|
@pulumi.getter(name="unifiedAlertingEnabled")
|
|
3919
3887
|
def unified_alerting_enabled(self) -> Optional[bool]:
|
|
3920
3888
|
"""
|
|
3921
|
-
Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set
|
|
3889
|
+
Enable or disable Grafana unified alerting functionality. By default this is enabled and any legacy alerts will be migrated on upgrade to Grafana 9+. To stay on legacy alerting, set unified*alerting*enabled to false and alerting_enabled to true. See https://grafana.com/docs/grafana/latest/alerting/set-up/migrating-alerts/ for more details.
|
|
3922
3890
|
"""
|
|
3923
3891
|
return pulumi.get(self, "unified_alerting_enabled")
|
|
3924
3892
|
|
|
@@ -5977,12 +5945,6 @@ class KafkaConnectComponent(dict):
|
|
|
5977
5945
|
return pulumi.get(self, "usage")
|
|
5978
5946
|
|
|
5979
5947
|
|
|
5980
|
-
@pulumi.output_type
|
|
5981
|
-
class KafkaConnectKafkaConnect(dict):
|
|
5982
|
-
def __init__(__self__):
|
|
5983
|
-
pass
|
|
5984
|
-
|
|
5985
|
-
|
|
5986
5948
|
@pulumi.output_type
|
|
5987
5949
|
class KafkaConnectKafkaConnectUserConfig(dict):
|
|
5988
5950
|
@staticmethod
|
|
@@ -6251,8 +6213,8 @@ class KafkaConnectKafkaConnectUserConfigKafkaConnect(dict):
|
|
|
6251
6213
|
:param str connector_client_config_override_policy: Defines what client configurations can be overridden by the connector. Default is None.
|
|
6252
6214
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
6253
6215
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
6254
|
-
:param str consumer_isolation_level: Transaction read isolation level.
|
|
6255
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6216
|
+
:param str consumer_isolation_level: Transaction read isolation level. read*uncommitted is the default, but read*committed can be used if consume-exactly-once behavior is desired.
|
|
6217
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6256
6218
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
6257
6219
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
6258
6220
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -6326,7 +6288,7 @@ class KafkaConnectKafkaConnectUserConfigKafkaConnect(dict):
|
|
|
6326
6288
|
@pulumi.getter(name="consumerIsolationLevel")
|
|
6327
6289
|
def consumer_isolation_level(self) -> Optional[str]:
|
|
6328
6290
|
"""
|
|
6329
|
-
Transaction read isolation level.
|
|
6291
|
+
Transaction read isolation level. read*uncommitted is the default, but read*committed can be used if consume-exactly-once behavior is desired.
|
|
6330
6292
|
"""
|
|
6331
6293
|
return pulumi.get(self, "consumer_isolation_level")
|
|
6332
6294
|
|
|
@@ -6334,7 +6296,7 @@ class KafkaConnectKafkaConnectUserConfigKafkaConnect(dict):
|
|
|
6334
6296
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
6335
6297
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
6336
6298
|
"""
|
|
6337
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6299
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
6338
6300
|
"""
|
|
6339
6301
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
6340
6302
|
|
|
@@ -6743,11 +6705,11 @@ class KafkaKafka(dict):
|
|
|
6743
6705
|
rest_uri: Optional[str] = None,
|
|
6744
6706
|
schema_registry_uri: Optional[str] = None):
|
|
6745
6707
|
"""
|
|
6746
|
-
:param str access_cert: The Kafka client certificate
|
|
6747
|
-
:param str access_key: The Kafka client certificate key
|
|
6748
|
-
:param str connect_uri: The Kafka Connect URI
|
|
6749
|
-
:param str rest_uri: The Kafka REST URI
|
|
6750
|
-
:param str schema_registry_uri: The Schema Registry URI
|
|
6708
|
+
:param str access_cert: The Kafka client certificate.
|
|
6709
|
+
:param str access_key: The Kafka client certificate key.
|
|
6710
|
+
:param str connect_uri: The Kafka Connect URI.
|
|
6711
|
+
:param str rest_uri: The Kafka REST URI.
|
|
6712
|
+
:param str schema_registry_uri: The Schema Registry URI.
|
|
6751
6713
|
"""
|
|
6752
6714
|
if access_cert is not None:
|
|
6753
6715
|
pulumi.set(__self__, "access_cert", access_cert)
|
|
@@ -6764,7 +6726,7 @@ class KafkaKafka(dict):
|
|
|
6764
6726
|
@pulumi.getter(name="accessCert")
|
|
6765
6727
|
def access_cert(self) -> Optional[str]:
|
|
6766
6728
|
"""
|
|
6767
|
-
The Kafka client certificate
|
|
6729
|
+
The Kafka client certificate.
|
|
6768
6730
|
"""
|
|
6769
6731
|
return pulumi.get(self, "access_cert")
|
|
6770
6732
|
|
|
@@ -6772,7 +6734,7 @@ class KafkaKafka(dict):
|
|
|
6772
6734
|
@pulumi.getter(name="accessKey")
|
|
6773
6735
|
def access_key(self) -> Optional[str]:
|
|
6774
6736
|
"""
|
|
6775
|
-
The Kafka client certificate key
|
|
6737
|
+
The Kafka client certificate key.
|
|
6776
6738
|
"""
|
|
6777
6739
|
return pulumi.get(self, "access_key")
|
|
6778
6740
|
|
|
@@ -6780,7 +6742,7 @@ class KafkaKafka(dict):
|
|
|
6780
6742
|
@pulumi.getter(name="connectUri")
|
|
6781
6743
|
def connect_uri(self) -> Optional[str]:
|
|
6782
6744
|
"""
|
|
6783
|
-
The Kafka Connect URI
|
|
6745
|
+
The Kafka Connect URI.
|
|
6784
6746
|
"""
|
|
6785
6747
|
return pulumi.get(self, "connect_uri")
|
|
6786
6748
|
|
|
@@ -6788,7 +6750,7 @@ class KafkaKafka(dict):
|
|
|
6788
6750
|
@pulumi.getter(name="restUri")
|
|
6789
6751
|
def rest_uri(self) -> Optional[str]:
|
|
6790
6752
|
"""
|
|
6791
|
-
The Kafka REST URI
|
|
6753
|
+
The Kafka REST URI.
|
|
6792
6754
|
"""
|
|
6793
6755
|
return pulumi.get(self, "rest_uri")
|
|
6794
6756
|
|
|
@@ -6796,7 +6758,7 @@ class KafkaKafka(dict):
|
|
|
6796
6758
|
@pulumi.getter(name="schemaRegistryUri")
|
|
6797
6759
|
def schema_registry_uri(self) -> Optional[str]:
|
|
6798
6760
|
"""
|
|
6799
|
-
The Schema Registry URI
|
|
6761
|
+
The Schema Registry URI.
|
|
6800
6762
|
"""
|
|
6801
6763
|
return pulumi.get(self, "schema_registry_uri")
|
|
6802
6764
|
|
|
@@ -7340,7 +7302,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7340
7302
|
:param int log_index_size_max_bytes: The maximum size in bytes of the offset index.
|
|
7341
7303
|
:param int log_local_retention_bytes: The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value.
|
|
7342
7304
|
:param int log_local_retention_ms: The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value.
|
|
7343
|
-
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7305
|
+
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7344
7306
|
:param int log_message_timestamp_difference_max_ms: The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message.
|
|
7345
7307
|
:param str log_message_timestamp_type: Define whether the timestamp in the message is message create time or log append time.
|
|
7346
7308
|
:param bool log_preallocate: Should pre allocate file when create new segment?
|
|
@@ -7362,7 +7324,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7362
7324
|
:param int replica_fetch_response_max_bytes: Maximum bytes expected for the entire fetch response (defaults to 10485760). Records are fetched in batches, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. As such, this is not an absolute maximum.
|
|
7363
7325
|
:param str sasl_oauthbearer_expected_audience: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences.
|
|
7364
7326
|
:param str sasl_oauthbearer_expected_issuer: Optional setting for the broker to use to verify that the JWT was created by the expected issuer.
|
|
7365
|
-
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7327
|
+
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7366
7328
|
:param str sasl_oauthbearer_sub_claim_name: Name of the scope from which to extract the subject claim from the JWT. Defaults to sub.
|
|
7367
7329
|
:param int socket_request_max_bytes: The maximum number of bytes in a socket request (defaults to 104857600).
|
|
7368
7330
|
:param bool transaction_partition_verification_enable: Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition.
|
|
@@ -7610,7 +7572,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7610
7572
|
@pulumi.getter(name="logMessageDownconversionEnable")
|
|
7611
7573
|
def log_message_downconversion_enable(self) -> Optional[bool]:
|
|
7612
7574
|
"""
|
|
7613
|
-
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7575
|
+
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7614
7576
|
"""
|
|
7615
7577
|
return pulumi.get(self, "log_message_downconversion_enable")
|
|
7616
7578
|
|
|
@@ -7786,7 +7748,7 @@ class KafkaKafkaUserConfigKafka(dict):
|
|
|
7786
7748
|
@pulumi.getter(name="saslOauthbearerJwksEndpointUrl")
|
|
7787
7749
|
def sasl_oauthbearer_jwks_endpoint_url(self) -> Optional[str]:
|
|
7788
7750
|
"""
|
|
7789
|
-
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7751
|
+
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7790
7752
|
"""
|
|
7791
7753
|
return pulumi.get(self, "sasl_oauthbearer_jwks_endpoint_url")
|
|
7792
7754
|
|
|
@@ -7933,7 +7895,7 @@ class KafkaKafkaUserConfigKafkaConnectConfig(dict):
|
|
|
7933
7895
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
7934
7896
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
7935
7897
|
:param str consumer_isolation_level: Transaction read isolation level. read*uncommitted is the default, but read*committed can be used if consume-exactly-once behavior is desired.
|
|
7936
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
7898
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
7937
7899
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
7938
7900
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
7939
7901
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -8015,7 +7977,7 @@ class KafkaKafkaUserConfigKafkaConnectConfig(dict):
|
|
|
8015
7977
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
8016
7978
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
8017
7979
|
"""
|
|
8018
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
7980
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
8019
7981
|
"""
|
|
8020
7982
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
8021
7983
|
|
|
@@ -8791,12 +8753,6 @@ class KafkaMirrorMakerComponent(dict):
|
|
|
8791
8753
|
return pulumi.get(self, "usage")
|
|
8792
8754
|
|
|
8793
8755
|
|
|
8794
|
-
@pulumi.output_type
|
|
8795
|
-
class KafkaMirrorMakerKafkaMirrormaker(dict):
|
|
8796
|
-
def __init__(__self__):
|
|
8797
|
-
pass
|
|
8798
|
-
|
|
8799
|
-
|
|
8800
8756
|
@pulumi.output_type
|
|
8801
8757
|
class KafkaMirrorMakerKafkaMirrormakerUserConfig(dict):
|
|
8802
8758
|
@staticmethod
|
|
@@ -9018,7 +8974,7 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormaker(dict):
|
|
|
9018
8974
|
:param int refresh_groups_interval_seconds: Frequency of consumer group refresh in seconds. Defaults to 600 seconds (10 minutes).
|
|
9019
8975
|
:param bool refresh_topics_enabled: Whether to periodically check for new topics and partitions. Defaults to 'true'.
|
|
9020
8976
|
:param int refresh_topics_interval_seconds: Frequency of topic and partitions refresh in seconds. Defaults to 600 seconds (10 minutes).
|
|
9021
|
-
:param bool sync_group_offsets_enabled: Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to
|
|
8977
|
+
:param bool sync_group_offsets_enabled: Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to _*consumer*offsets topic in target cluster, as long as no active consumers in that group are connected to the target cluster.
|
|
9022
8978
|
:param int sync_group_offsets_interval_seconds: Frequency at which consumer group offsets are synced (default: 60, every minute).
|
|
9023
8979
|
:param bool sync_topic_configs_enabled: Whether to periodically configure remote topics to match their corresponding upstream topics.
|
|
9024
8980
|
:param int tasks_max_per_cpu: 'tasks.max' is set to this multiplied by the number of CPUs in the service. The default value is `1`.
|
|
@@ -9126,7 +9082,7 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormaker(dict):
|
|
|
9126
9082
|
@pulumi.getter(name="syncGroupOffsetsEnabled")
|
|
9127
9083
|
def sync_group_offsets_enabled(self) -> Optional[bool]:
|
|
9128
9084
|
"""
|
|
9129
|
-
Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to
|
|
9085
|
+
Whether to periodically write the translated offsets of replicated consumer groups (in the source cluster) to _*consumer*offsets topic in target cluster, as long as no active consumers in that group are connected to the target cluster.
|
|
9130
9086
|
"""
|
|
9131
9087
|
return pulumi.get(self, "sync_group_offsets_enabled")
|
|
9132
9088
|
|
|
@@ -9898,12 +9854,6 @@ class M3AggregatorComponent(dict):
|
|
|
9898
9854
|
return pulumi.get(self, "usage")
|
|
9899
9855
|
|
|
9900
9856
|
|
|
9901
|
-
@pulumi.output_type
|
|
9902
|
-
class M3AggregatorM3aggregator(dict):
|
|
9903
|
-
def __init__(__self__):
|
|
9904
|
-
pass
|
|
9905
|
-
|
|
9906
|
-
|
|
9907
9857
|
@pulumi.output_type
|
|
9908
9858
|
class M3AggregatorM3aggregatorUserConfig(dict):
|
|
9909
9859
|
@staticmethod
|
|
@@ -10288,12 +10238,6 @@ class M3DbComponent(dict):
|
|
|
10288
10238
|
return pulumi.get(self, "usage")
|
|
10289
10239
|
|
|
10290
10240
|
|
|
10291
|
-
@pulumi.output_type
|
|
10292
|
-
class M3DbM3db(dict):
|
|
10293
|
-
def __init__(__self__):
|
|
10294
|
-
pass
|
|
10295
|
-
|
|
10296
|
-
|
|
10297
10241
|
@pulumi.output_type
|
|
10298
10242
|
class M3DbM3dbUserConfig(dict):
|
|
10299
10243
|
@staticmethod
|
|
@@ -10421,9 +10365,6 @@ class M3DbM3dbUserConfig(dict):
|
|
|
10421
10365
|
"""
|
|
10422
10366
|
Additional Cloud Regions for Backup Replication.
|
|
10423
10367
|
"""
|
|
10424
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
10425
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
10426
|
-
|
|
10427
10368
|
return pulumi.get(self, "additional_backup_regions")
|
|
10428
10369
|
|
|
10429
10370
|
@property
|
|
@@ -10635,7 +10576,7 @@ class M3DbM3dbUserConfigLimits(dict):
|
|
|
10635
10576
|
"""
|
|
10636
10577
|
:param int max_recently_queried_series_blocks: The maximum number of blocks that can be read in a given lookback period.
|
|
10637
10578
|
:param int max_recently_queried_series_disk_bytes_read: The maximum number of disk bytes that can be read in a given lookback period.
|
|
10638
|
-
:param str max_recently_queried_series_lookback: The lookback period for '
|
|
10579
|
+
:param str max_recently_queried_series_lookback: The lookback period for 'max*recently*queried*series*blocks' and 'max*recently*queried*series*disk*bytes*read'.
|
|
10639
10580
|
:param int query_docs: The maximum number of docs fetched in single query.
|
|
10640
10581
|
:param bool query_require_exhaustive: When query limits are exceeded, whether to return error or return partial results.
|
|
10641
10582
|
:param int query_series: The maximum number of series fetched in single query.
|
|
@@ -10673,7 +10614,7 @@ class M3DbM3dbUserConfigLimits(dict):
|
|
|
10673
10614
|
@pulumi.getter(name="maxRecentlyQueriedSeriesLookback")
|
|
10674
10615
|
def max_recently_queried_series_lookback(self) -> Optional[str]:
|
|
10675
10616
|
"""
|
|
10676
|
-
The lookback period for '
|
|
10617
|
+
The lookback period for 'max*recently*queried*series*blocks' and 'max*recently*queried*series*disk*bytes*read'.
|
|
10677
10618
|
"""
|
|
10678
10619
|
return pulumi.get(self, "max_recently_queried_series_lookback")
|
|
10679
10620
|
|
|
@@ -11084,7 +11025,7 @@ class M3DbM3dbUserConfigRulesMapping(dict):
|
|
|
11084
11025
|
namespaces_strings: Optional[Sequence[str]] = None,
|
|
11085
11026
|
tags: Optional[Sequence['outputs.M3DbM3dbUserConfigRulesMappingTag']] = None):
|
|
11086
11027
|
"""
|
|
11087
|
-
:param str filter: Matching metric names with wildcards (using
|
|
11028
|
+
:param str filter: Matching metric names with wildcards (using **name**:wildcard) or matching tags and their (optionally wildcarded) values. For value, ! can be used at start of value for negation, and multiple filters can be supplied using space as separator.
|
|
11088
11029
|
:param Sequence[str] aggregations: List of aggregations to be applied.
|
|
11089
11030
|
:param bool drop: Only store the derived metric (as specified in the roll-up rules), if any.
|
|
11090
11031
|
:param str name: The (optional) name of the rule.
|
|
@@ -11113,7 +11054,7 @@ class M3DbM3dbUserConfigRulesMapping(dict):
|
|
|
11113
11054
|
@pulumi.getter
|
|
11114
11055
|
def filter(self) -> str:
|
|
11115
11056
|
"""
|
|
11116
|
-
Matching metric names with wildcards (using
|
|
11057
|
+
Matching metric names with wildcards (using **name**:wildcard) or matching tags and their (optionally wildcarded) values. For value, ! can be used at start of value for negation, and multiple filters can be supplied using space as separator.
|
|
11117
11058
|
"""
|
|
11118
11059
|
return pulumi.get(self, "filter")
|
|
11119
11060
|
|
|
@@ -11453,12 +11394,6 @@ class MySqlComponent(dict):
|
|
|
11453
11394
|
return pulumi.get(self, "usage")
|
|
11454
11395
|
|
|
11455
11396
|
|
|
11456
|
-
@pulumi.output_type
|
|
11457
|
-
class MySqlMysql(dict):
|
|
11458
|
-
def __init__(__self__):
|
|
11459
|
-
pass
|
|
11460
|
-
|
|
11461
|
-
|
|
11462
11397
|
@pulumi.output_type
|
|
11463
11398
|
class MySqlMysqlUserConfig(dict):
|
|
11464
11399
|
@staticmethod
|
|
@@ -12035,17 +11970,17 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12035
11970
|
:param int innodb_write_io_threads: The number of I/O threads for write operations in InnoDB. Default is 4. Changing this parameter will lead to a restart of the MySQL service.
|
|
12036
11971
|
:param int interactive_timeout: The number of seconds the server waits for activity on an interactive connection before closing it.
|
|
12037
11972
|
:param str internal_tmp_mem_storage_engine: The storage engine for in-memory internal temporary tables.
|
|
12038
|
-
:param float long_query_time: The
|
|
11973
|
+
:param float long_query_time: The slow*query*logs work as SQL statements that take more than long*query*time seconds to execute. Default is 10s.
|
|
12039
11974
|
:param int max_allowed_packet: Size of the largest message in bytes that can be received by the server. Default is 67108864 (64M).
|
|
12040
|
-
:param int max_heap_table_size: Limits the size of internal in-memory tables. Also set
|
|
11975
|
+
:param int max_heap_table_size: Limits the size of internal in-memory tables. Also set tmp*table*size. Default is 16777216 (16M).
|
|
12041
11976
|
:param int net_buffer_length: Start sizes of connection buffer and result buffer. Default is 16384 (16K). Changing this parameter will lead to a restart of the MySQL service.
|
|
12042
11977
|
:param int net_read_timeout: The number of seconds to wait for more data from a connection before aborting the read.
|
|
12043
11978
|
:param int net_write_timeout: The number of seconds to wait for a block to be written to a connection before aborting the write.
|
|
12044
|
-
:param bool slow_query_log: Slow query log enables capturing of slow queries. Setting
|
|
11979
|
+
:param bool slow_query_log: Slow query log enables capturing of slow queries. Setting slow*query*log to false also truncates the mysql.slow_log table. Default is off.
|
|
12045
11980
|
:param int sort_buffer_size: Sort buffer size in bytes for ORDER BY optimization. Default is 262144 (256K).
|
|
12046
11981
|
:param str sql_mode: Global SQL mode. Set to empty to use MySQL server defaults. When creating a new service and not setting this field Aiven default SQL mode (strict, SQL standard compliant) will be assigned.
|
|
12047
11982
|
:param bool sql_require_primary_key: Require primary key to be defined for new tables or old tables modified with ALTER TABLE and fail if missing. It is recommended to always have primary keys because various functionality may break if any large table is missing them.
|
|
12048
|
-
:param int tmp_table_size: Limits the size of internal in-memory tables. Also set
|
|
11983
|
+
:param int tmp_table_size: Limits the size of internal in-memory tables. Also set max*heap*table_size. Default is 16777216 (16M).
|
|
12049
11984
|
:param int wait_timeout: The number of seconds the server waits for activity on a noninteractive connection before closing it.
|
|
12050
11985
|
"""
|
|
12051
11986
|
if connect_timeout is not None:
|
|
@@ -12257,7 +12192,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12257
12192
|
@pulumi.getter(name="longQueryTime")
|
|
12258
12193
|
def long_query_time(self) -> Optional[float]:
|
|
12259
12194
|
"""
|
|
12260
|
-
The
|
|
12195
|
+
The slow*query*logs work as SQL statements that take more than long*query*time seconds to execute. Default is 10s.
|
|
12261
12196
|
"""
|
|
12262
12197
|
return pulumi.get(self, "long_query_time")
|
|
12263
12198
|
|
|
@@ -12273,7 +12208,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12273
12208
|
@pulumi.getter(name="maxHeapTableSize")
|
|
12274
12209
|
def max_heap_table_size(self) -> Optional[int]:
|
|
12275
12210
|
"""
|
|
12276
|
-
Limits the size of internal in-memory tables. Also set
|
|
12211
|
+
Limits the size of internal in-memory tables. Also set tmp*table*size. Default is 16777216 (16M).
|
|
12277
12212
|
"""
|
|
12278
12213
|
return pulumi.get(self, "max_heap_table_size")
|
|
12279
12214
|
|
|
@@ -12305,7 +12240,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12305
12240
|
@pulumi.getter(name="slowQueryLog")
|
|
12306
12241
|
def slow_query_log(self) -> Optional[bool]:
|
|
12307
12242
|
"""
|
|
12308
|
-
Slow query log enables capturing of slow queries. Setting
|
|
12243
|
+
Slow query log enables capturing of slow queries. Setting slow*query*log to false also truncates the mysql.slow_log table. Default is off.
|
|
12309
12244
|
"""
|
|
12310
12245
|
return pulumi.get(self, "slow_query_log")
|
|
12311
12246
|
|
|
@@ -12337,7 +12272,7 @@ class MySqlMysqlUserConfigMysql(dict):
|
|
|
12337
12272
|
@pulumi.getter(name="tmpTableSize")
|
|
12338
12273
|
def tmp_table_size(self) -> Optional[int]:
|
|
12339
12274
|
"""
|
|
12340
|
-
Limits the size of internal in-memory tables. Also set
|
|
12275
|
+
Limits the size of internal in-memory tables. Also set max*heap*table_size. Default is 16777216 (16M).
|
|
12341
12276
|
"""
|
|
12342
12277
|
return pulumi.get(self, "tmp_table_size")
|
|
12343
12278
|
|
|
@@ -12823,7 +12758,7 @@ class OpenSearchOpensearchUserConfig(dict):
|
|
|
12823
12758
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
12824
12759
|
:param Sequence[str] ip_filters: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
12825
12760
|
:param bool keep_index_refresh_interval: Aiven automation resets index.refresh_interval to default value for every index to be sure that indices are always visible to search. If it doesn't fit your case, you can disable this by setting up this flag to true.
|
|
12826
|
-
:param int max_index_count:
|
|
12761
|
+
:param int max_index_count: Use index_patterns instead. The default value is `0`.
|
|
12827
12762
|
:param 'OpenSearchOpensearchUserConfigOpenidArgs' openid: OpenSearch OpenID Connect Configuration
|
|
12828
12763
|
:param 'OpenSearchOpensearchUserConfigOpensearchArgs' opensearch: OpenSearch settings
|
|
12829
12764
|
:param 'OpenSearchOpensearchUserConfigOpensearchDashboardsArgs' opensearch_dashboards: OpenSearch Dashboards settings
|
|
@@ -12891,9 +12826,6 @@ class OpenSearchOpensearchUserConfig(dict):
|
|
|
12891
12826
|
"""
|
|
12892
12827
|
Additional Cloud Regions for Backup Replication.
|
|
12893
12828
|
"""
|
|
12894
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
12895
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
12896
|
-
|
|
12897
12829
|
return pulumi.get(self, "additional_backup_regions")
|
|
12898
12830
|
|
|
12899
12831
|
@property
|
|
@@ -12967,7 +12899,7 @@ class OpenSearchOpensearchUserConfig(dict):
|
|
|
12967
12899
|
@pulumi.getter(name="maxIndexCount")
|
|
12968
12900
|
def max_index_count(self) -> Optional[int]:
|
|
12969
12901
|
"""
|
|
12970
|
-
|
|
12902
|
+
Use index_patterns instead. The default value is `0`.
|
|
12971
12903
|
"""
|
|
12972
12904
|
return pulumi.get(self, "max_index_count")
|
|
12973
12905
|
|
|
@@ -14739,6 +14671,116 @@ class OpenSearchTechEmail(dict):
|
|
|
14739
14671
|
return pulumi.get(self, "email")
|
|
14740
14672
|
|
|
14741
14673
|
|
|
14674
|
+
@pulumi.output_type
|
|
14675
|
+
class OrganizationApplicationUserTimeouts(dict):
|
|
14676
|
+
def __init__(__self__, *,
|
|
14677
|
+
create: Optional[str] = None,
|
|
14678
|
+
delete: Optional[str] = None,
|
|
14679
|
+
read: Optional[str] = None,
|
|
14680
|
+
update: Optional[str] = None):
|
|
14681
|
+
"""
|
|
14682
|
+
:param str create: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14683
|
+
:param str delete: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Setting a timeout for a Delete operation is only applicable if changes are saved into state before the destroy operation occurs.
|
|
14684
|
+
:param str read: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Read operations occur during any refresh or planning operation when refresh is enabled.
|
|
14685
|
+
:param str update: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14686
|
+
"""
|
|
14687
|
+
if create is not None:
|
|
14688
|
+
pulumi.set(__self__, "create", create)
|
|
14689
|
+
if delete is not None:
|
|
14690
|
+
pulumi.set(__self__, "delete", delete)
|
|
14691
|
+
if read is not None:
|
|
14692
|
+
pulumi.set(__self__, "read", read)
|
|
14693
|
+
if update is not None:
|
|
14694
|
+
pulumi.set(__self__, "update", update)
|
|
14695
|
+
|
|
14696
|
+
@property
|
|
14697
|
+
@pulumi.getter
|
|
14698
|
+
def create(self) -> Optional[str]:
|
|
14699
|
+
"""
|
|
14700
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14701
|
+
"""
|
|
14702
|
+
return pulumi.get(self, "create")
|
|
14703
|
+
|
|
14704
|
+
@property
|
|
14705
|
+
@pulumi.getter
|
|
14706
|
+
def delete(self) -> Optional[str]:
|
|
14707
|
+
"""
|
|
14708
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Setting a timeout for a Delete operation is only applicable if changes are saved into state before the destroy operation occurs.
|
|
14709
|
+
"""
|
|
14710
|
+
return pulumi.get(self, "delete")
|
|
14711
|
+
|
|
14712
|
+
@property
|
|
14713
|
+
@pulumi.getter
|
|
14714
|
+
def read(self) -> Optional[str]:
|
|
14715
|
+
"""
|
|
14716
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Read operations occur during any refresh or planning operation when refresh is enabled.
|
|
14717
|
+
"""
|
|
14718
|
+
return pulumi.get(self, "read")
|
|
14719
|
+
|
|
14720
|
+
@property
|
|
14721
|
+
@pulumi.getter
|
|
14722
|
+
def update(self) -> Optional[str]:
|
|
14723
|
+
"""
|
|
14724
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14725
|
+
"""
|
|
14726
|
+
return pulumi.get(self, "update")
|
|
14727
|
+
|
|
14728
|
+
|
|
14729
|
+
@pulumi.output_type
|
|
14730
|
+
class OrganizationApplicationUserTokenTimeouts(dict):
|
|
14731
|
+
def __init__(__self__, *,
|
|
14732
|
+
create: Optional[str] = None,
|
|
14733
|
+
delete: Optional[str] = None,
|
|
14734
|
+
read: Optional[str] = None,
|
|
14735
|
+
update: Optional[str] = None):
|
|
14736
|
+
"""
|
|
14737
|
+
:param str create: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14738
|
+
:param str delete: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Setting a timeout for a Delete operation is only applicable if changes are saved into state before the destroy operation occurs.
|
|
14739
|
+
:param str read: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Read operations occur during any refresh or planning operation when refresh is enabled.
|
|
14740
|
+
:param str update: A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14741
|
+
"""
|
|
14742
|
+
if create is not None:
|
|
14743
|
+
pulumi.set(__self__, "create", create)
|
|
14744
|
+
if delete is not None:
|
|
14745
|
+
pulumi.set(__self__, "delete", delete)
|
|
14746
|
+
if read is not None:
|
|
14747
|
+
pulumi.set(__self__, "read", read)
|
|
14748
|
+
if update is not None:
|
|
14749
|
+
pulumi.set(__self__, "update", update)
|
|
14750
|
+
|
|
14751
|
+
@property
|
|
14752
|
+
@pulumi.getter
|
|
14753
|
+
def create(self) -> Optional[str]:
|
|
14754
|
+
"""
|
|
14755
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14756
|
+
"""
|
|
14757
|
+
return pulumi.get(self, "create")
|
|
14758
|
+
|
|
14759
|
+
@property
|
|
14760
|
+
@pulumi.getter
|
|
14761
|
+
def delete(self) -> Optional[str]:
|
|
14762
|
+
"""
|
|
14763
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Setting a timeout for a Delete operation is only applicable if changes are saved into state before the destroy operation occurs.
|
|
14764
|
+
"""
|
|
14765
|
+
return pulumi.get(self, "delete")
|
|
14766
|
+
|
|
14767
|
+
@property
|
|
14768
|
+
@pulumi.getter
|
|
14769
|
+
def read(self) -> Optional[str]:
|
|
14770
|
+
"""
|
|
14771
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Read operations occur during any refresh or planning operation when refresh is enabled.
|
|
14772
|
+
"""
|
|
14773
|
+
return pulumi.get(self, "read")
|
|
14774
|
+
|
|
14775
|
+
@property
|
|
14776
|
+
@pulumi.getter
|
|
14777
|
+
def update(self) -> Optional[str]:
|
|
14778
|
+
"""
|
|
14779
|
+
A string that can be [parsed as a duration](https://pkg.go.dev/time#ParseDuration) consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours).
|
|
14780
|
+
"""
|
|
14781
|
+
return pulumi.get(self, "update")
|
|
14782
|
+
|
|
14783
|
+
|
|
14742
14784
|
@pulumi.output_type
|
|
14743
14785
|
class OrganizationGroupProjectTimeouts(dict):
|
|
14744
14786
|
def __init__(__self__, *,
|
|
@@ -15971,7 +16013,7 @@ class PgPgUserConfigPg(dict):
|
|
|
15971
16013
|
:param int pg_partman_bgw_dot_interval: Sets the time interval to run pg_partman's scheduled tasks.
|
|
15972
16014
|
:param str pg_partman_bgw_dot_role: Controls which role to use for pg_partman's scheduled background tasks.
|
|
15973
16015
|
:param bool pg_stat_monitor_dot_pgsm_enable_query_plan: Enables or disables query plan monitoring.
|
|
15974
|
-
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets
|
|
16016
|
+
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets.
|
|
15975
16017
|
:param str pg_stat_statements_dot_track: Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top.
|
|
15976
16018
|
:param int temp_file_limit: PostgreSQL temporary file limit in KiB, -1 for unlimited.
|
|
15977
16019
|
:param str timezone: PostgreSQL service timezone.
|
|
@@ -16397,7 +16439,7 @@ class PgPgUserConfigPg(dict):
|
|
|
16397
16439
|
@pulumi.getter(name="pgStatMonitorDotPgsmMaxBuckets")
|
|
16398
16440
|
def pg_stat_monitor_dot_pgsm_max_buckets(self) -> Optional[int]:
|
|
16399
16441
|
"""
|
|
16400
|
-
Sets the maximum number of buckets
|
|
16442
|
+
Sets the maximum number of buckets.
|
|
16401
16443
|
"""
|
|
16402
16444
|
return pulumi.get(self, "pg_stat_monitor_dot_pgsm_max_buckets")
|
|
16403
16445
|
|
|
@@ -17243,8 +17285,8 @@ class ProjectTag(dict):
|
|
|
17243
17285
|
key: str,
|
|
17244
17286
|
value: str):
|
|
17245
17287
|
"""
|
|
17246
|
-
:param str key: Project tag key
|
|
17247
|
-
:param str value: Project tag value
|
|
17288
|
+
:param str key: Project tag key.
|
|
17289
|
+
:param str value: Project tag value.
|
|
17248
17290
|
"""
|
|
17249
17291
|
pulumi.set(__self__, "key", key)
|
|
17250
17292
|
pulumi.set(__self__, "value", value)
|
|
@@ -17253,7 +17295,7 @@ class ProjectTag(dict):
|
|
|
17253
17295
|
@pulumi.getter
|
|
17254
17296
|
def key(self) -> str:
|
|
17255
17297
|
"""
|
|
17256
|
-
Project tag key
|
|
17298
|
+
Project tag key.
|
|
17257
17299
|
"""
|
|
17258
17300
|
return pulumi.get(self, "key")
|
|
17259
17301
|
|
|
@@ -17261,7 +17303,7 @@ class ProjectTag(dict):
|
|
|
17261
17303
|
@pulumi.getter
|
|
17262
17304
|
def value(self) -> str:
|
|
17263
17305
|
"""
|
|
17264
|
-
Project tag value
|
|
17306
|
+
Project tag value.
|
|
17265
17307
|
"""
|
|
17266
17308
|
return pulumi.get(self, "value")
|
|
17267
17309
|
|
|
@@ -17388,12 +17430,6 @@ class RedisComponent(dict):
|
|
|
17388
17430
|
return pulumi.get(self, "usage")
|
|
17389
17431
|
|
|
17390
17432
|
|
|
17391
|
-
@pulumi.output_type
|
|
17392
|
-
class RedisRedi(dict):
|
|
17393
|
-
def __init__(__self__):
|
|
17394
|
-
pass
|
|
17395
|
-
|
|
17396
|
-
|
|
17397
17433
|
@pulumi.output_type
|
|
17398
17434
|
class RedisRedisUserConfig(dict):
|
|
17399
17435
|
@staticmethod
|
|
@@ -17569,9 +17605,6 @@ class RedisRedisUserConfig(dict):
|
|
|
17569
17605
|
"""
|
|
17570
17606
|
Additional Cloud Regions for Backup Replication.
|
|
17571
17607
|
"""
|
|
17572
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
17573
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
17574
|
-
|
|
17575
17608
|
return pulumi.get(self, "additional_backup_regions")
|
|
17576
17609
|
|
|
17577
17610
|
@property
|
|
@@ -18608,7 +18641,9 @@ class ServiceIntegrationDatadogUserConfigOpensearch(dict):
|
|
|
18608
18641
|
@staticmethod
|
|
18609
18642
|
def __key_warning(key: str):
|
|
18610
18643
|
suggest = None
|
|
18611
|
-
if key == "
|
|
18644
|
+
if key == "clusterStatsEnabled":
|
|
18645
|
+
suggest = "cluster_stats_enabled"
|
|
18646
|
+
elif key == "indexStatsEnabled":
|
|
18612
18647
|
suggest = "index_stats_enabled"
|
|
18613
18648
|
elif key == "pendingTaskStatsEnabled":
|
|
18614
18649
|
suggest = "pending_task_stats_enabled"
|
|
@@ -18627,14 +18662,18 @@ class ServiceIntegrationDatadogUserConfigOpensearch(dict):
|
|
|
18627
18662
|
return super().get(key, default)
|
|
18628
18663
|
|
|
18629
18664
|
def __init__(__self__, *,
|
|
18665
|
+
cluster_stats_enabled: Optional[bool] = None,
|
|
18630
18666
|
index_stats_enabled: Optional[bool] = None,
|
|
18631
18667
|
pending_task_stats_enabled: Optional[bool] = None,
|
|
18632
18668
|
pshard_stats_enabled: Optional[bool] = None):
|
|
18633
18669
|
"""
|
|
18670
|
+
:param bool cluster_stats_enabled: Enable Datadog Opensearch Cluster Monitoring.
|
|
18634
18671
|
:param bool index_stats_enabled: Enable Datadog Opensearch Index Monitoring.
|
|
18635
18672
|
:param bool pending_task_stats_enabled: Enable Datadog Opensearch Pending Task Monitoring.
|
|
18636
18673
|
:param bool pshard_stats_enabled: Enable Datadog Opensearch Primary Shard Monitoring.
|
|
18637
18674
|
"""
|
|
18675
|
+
if cluster_stats_enabled is not None:
|
|
18676
|
+
pulumi.set(__self__, "cluster_stats_enabled", cluster_stats_enabled)
|
|
18638
18677
|
if index_stats_enabled is not None:
|
|
18639
18678
|
pulumi.set(__self__, "index_stats_enabled", index_stats_enabled)
|
|
18640
18679
|
if pending_task_stats_enabled is not None:
|
|
@@ -18642,6 +18681,14 @@ class ServiceIntegrationDatadogUserConfigOpensearch(dict):
|
|
|
18642
18681
|
if pshard_stats_enabled is not None:
|
|
18643
18682
|
pulumi.set(__self__, "pshard_stats_enabled", pshard_stats_enabled)
|
|
18644
18683
|
|
|
18684
|
+
@property
|
|
18685
|
+
@pulumi.getter(name="clusterStatsEnabled")
|
|
18686
|
+
def cluster_stats_enabled(self) -> Optional[bool]:
|
|
18687
|
+
"""
|
|
18688
|
+
Enable Datadog Opensearch Cluster Monitoring.
|
|
18689
|
+
"""
|
|
18690
|
+
return pulumi.get(self, "cluster_stats_enabled")
|
|
18691
|
+
|
|
18645
18692
|
@property
|
|
18646
18693
|
@pulumi.getter(name="indexStatsEnabled")
|
|
18647
18694
|
def index_stats_enabled(self) -> Optional[bool]:
|
|
@@ -19103,7 +19150,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudBigquery(dict):
|
|
|
19103
19150
|
service_account_credentials: str):
|
|
19104
19151
|
"""
|
|
19105
19152
|
:param str project_id: GCP project id.
|
|
19106
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19153
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19107
19154
|
"""
|
|
19108
19155
|
pulumi.set(__self__, "project_id", project_id)
|
|
19109
19156
|
pulumi.set(__self__, "service_account_credentials", service_account_credentials)
|
|
@@ -19120,7 +19167,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudBigquery(dict):
|
|
|
19120
19167
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
19121
19168
|
def service_account_credentials(self) -> str:
|
|
19122
19169
|
"""
|
|
19123
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19170
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19124
19171
|
"""
|
|
19125
19172
|
return pulumi.get(self, "service_account_credentials")
|
|
19126
19173
|
|
|
@@ -19155,7 +19202,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfig(dict):
|
|
|
19155
19202
|
"""
|
|
19156
19203
|
:param str log_id: Google Cloud Logging log id.
|
|
19157
19204
|
:param str project_id: GCP project id.
|
|
19158
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19205
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19159
19206
|
"""
|
|
19160
19207
|
pulumi.set(__self__, "log_id", log_id)
|
|
19161
19208
|
pulumi.set(__self__, "project_id", project_id)
|
|
@@ -19181,7 +19228,7 @@ class ServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfig(dict):
|
|
|
19181
19228
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
19182
19229
|
def service_account_credentials(self) -> str:
|
|
19183
19230
|
"""
|
|
19184
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
19231
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
19185
19232
|
"""
|
|
19186
19233
|
return pulumi.get(self, "service_account_credentials")
|
|
19187
19234
|
|
|
@@ -20633,7 +20680,7 @@ class ServiceIntegrationMetricsUserConfigSourceMysqlTelegraf(dict):
|
|
|
20633
20680
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
20634
20681
|
"""
|
|
20635
20682
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE*SCHEMA.EVENT*WAITS.
|
|
20636
|
-
:param bool gather_file_events_stats:
|
|
20683
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20637
20684
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE*SCHEMA.TABLE*IO*WAITS*SUMMARY*BY*INDEX_USAGE.
|
|
20638
20685
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
20639
20686
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION*SCHEMA.INNODB*METRICS.
|
|
@@ -20688,7 +20735,7 @@ class ServiceIntegrationMetricsUserConfigSourceMysqlTelegraf(dict):
|
|
|
20688
20735
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
20689
20736
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
20690
20737
|
"""
|
|
20691
|
-
|
|
20738
|
+
Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20692
20739
|
"""
|
|
20693
20740
|
return pulumi.get(self, "gather_file_events_stats")
|
|
20694
20741
|
|
|
@@ -20906,7 +20953,7 @@ class ServiceIntegrationPrometheusUserConfigSourceMysqlTelegraf(dict):
|
|
|
20906
20953
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
20907
20954
|
"""
|
|
20908
20955
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE*SCHEMA.EVENT*WAITS.
|
|
20909
|
-
:param bool gather_file_events_stats:
|
|
20956
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20910
20957
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE*SCHEMA.TABLE*IO*WAITS*SUMMARY*BY*INDEX_USAGE.
|
|
20911
20958
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
20912
20959
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION*SCHEMA.INNODB*METRICS.
|
|
@@ -20961,7 +21008,7 @@ class ServiceIntegrationPrometheusUserConfigSourceMysqlTelegraf(dict):
|
|
|
20961
21008
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
20962
21009
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
20963
21010
|
"""
|
|
20964
|
-
|
|
21011
|
+
Gather metrics from PERFORMANCE*SCHEMA.FILE*SUMMARY*BY*EVENT_NAME.
|
|
20965
21012
|
"""
|
|
20966
21013
|
return pulumi.get(self, "gather_file_events_stats")
|
|
20967
21014
|
|
|
@@ -21129,12 +21176,6 @@ class GetAccountAuthenticationSamlFieldMappingResult(dict):
|
|
|
21129
21176
|
return pulumi.get(self, "real_name")
|
|
21130
21177
|
|
|
21131
21178
|
|
|
21132
|
-
@pulumi.output_type
|
|
21133
|
-
class GetCassandaCassandraResult(dict):
|
|
21134
|
-
def __init__(__self__):
|
|
21135
|
-
pass
|
|
21136
|
-
|
|
21137
|
-
|
|
21138
21179
|
@pulumi.output_type
|
|
21139
21180
|
class GetCassandaCassandraUserConfigResult(dict):
|
|
21140
21181
|
def __init__(__self__, *,
|
|
@@ -21158,7 +21199,7 @@ class GetCassandaCassandraUserConfigResult(dict):
|
|
|
21158
21199
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
21159
21200
|
:param int backup_hour: The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21160
21201
|
:param int backup_minute: The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21161
|
-
:param 'GetCassandaCassandraUserConfigCassandraArgs' cassandra:
|
|
21202
|
+
:param 'GetCassandaCassandraUserConfigCassandraArgs' cassandra: Cassandra configuration values
|
|
21162
21203
|
:param str cassandra_version: Cassandra version.
|
|
21163
21204
|
:param Sequence['GetCassandaCassandraUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
21164
21205
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -21236,7 +21277,7 @@ class GetCassandaCassandraUserConfigResult(dict):
|
|
|
21236
21277
|
@pulumi.getter
|
|
21237
21278
|
def cassandra(self) -> Optional['outputs.GetCassandaCassandraUserConfigCassandraResult']:
|
|
21238
21279
|
"""
|
|
21239
|
-
|
|
21280
|
+
Cassandra configuration values
|
|
21240
21281
|
"""
|
|
21241
21282
|
return pulumi.get(self, "cassandra")
|
|
21242
21283
|
|
|
@@ -21622,12 +21663,6 @@ class GetCassandaTechEmailResult(dict):
|
|
|
21622
21663
|
return pulumi.get(self, "email")
|
|
21623
21664
|
|
|
21624
21665
|
|
|
21625
|
-
@pulumi.output_type
|
|
21626
|
-
class GetCassandraCassandraResult(dict):
|
|
21627
|
-
def __init__(__self__):
|
|
21628
|
-
pass
|
|
21629
|
-
|
|
21630
|
-
|
|
21631
21666
|
@pulumi.output_type
|
|
21632
21667
|
class GetCassandraCassandraUserConfigResult(dict):
|
|
21633
21668
|
def __init__(__self__, *,
|
|
@@ -21651,7 +21686,7 @@ class GetCassandraCassandraUserConfigResult(dict):
|
|
|
21651
21686
|
:param str additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
21652
21687
|
:param int backup_hour: The hour of day (in UTC) when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21653
21688
|
:param int backup_minute: The minute of an hour when backup for the service is started. New backup is only started if previous backup has already completed.
|
|
21654
|
-
:param 'GetCassandraCassandraUserConfigCassandraArgs' cassandra:
|
|
21689
|
+
:param 'GetCassandraCassandraUserConfigCassandraArgs' cassandra: Cassandra configuration values
|
|
21655
21690
|
:param str cassandra_version: Cassandra version.
|
|
21656
21691
|
:param Sequence['GetCassandraCassandraUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
21657
21692
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -21729,7 +21764,7 @@ class GetCassandraCassandraUserConfigResult(dict):
|
|
|
21729
21764
|
@pulumi.getter
|
|
21730
21765
|
def cassandra(self) -> Optional['outputs.GetCassandraCassandraUserConfigCassandraResult']:
|
|
21731
21766
|
"""
|
|
21732
|
-
|
|
21767
|
+
Cassandra configuration values
|
|
21733
21768
|
"""
|
|
21734
21769
|
return pulumi.get(self, "cassandra")
|
|
21735
21770
|
|
|
@@ -22115,12 +22150,6 @@ class GetCassandraTechEmailResult(dict):
|
|
|
22115
22150
|
return pulumi.get(self, "email")
|
|
22116
22151
|
|
|
22117
22152
|
|
|
22118
|
-
@pulumi.output_type
|
|
22119
|
-
class GetClickhouseClickhouseResult(dict):
|
|
22120
|
-
def __init__(__self__):
|
|
22121
|
-
pass
|
|
22122
|
-
|
|
22123
|
-
|
|
22124
22153
|
@pulumi.output_type
|
|
22125
22154
|
class GetClickhouseClickhouseUserConfigResult(dict):
|
|
22126
22155
|
def __init__(__self__, *,
|
|
@@ -22727,16 +22756,11 @@ class GetDragonflyComponentResult(dict):
|
|
|
22727
22756
|
return pulumi.get(self, "usage")
|
|
22728
22757
|
|
|
22729
22758
|
|
|
22730
|
-
@pulumi.output_type
|
|
22731
|
-
class GetDragonflyDragonflyResult(dict):
|
|
22732
|
-
def __init__(__self__):
|
|
22733
|
-
pass
|
|
22734
|
-
|
|
22735
|
-
|
|
22736
22759
|
@pulumi.output_type
|
|
22737
22760
|
class GetDragonflyDragonflyUserConfigResult(dict):
|
|
22738
22761
|
def __init__(__self__, *,
|
|
22739
22762
|
cache_mode: Optional[bool] = None,
|
|
22763
|
+
dragonfly_persistence: Optional[str] = None,
|
|
22740
22764
|
dragonfly_ssl: Optional[bool] = None,
|
|
22741
22765
|
ip_filter_objects: Optional[Sequence['outputs.GetDragonflyDragonflyUserConfigIpFilterObjectResult']] = None,
|
|
22742
22766
|
ip_filter_strings: Optional[Sequence[str]] = None,
|
|
@@ -22752,6 +22776,7 @@ class GetDragonflyDragonflyUserConfigResult(dict):
|
|
|
22752
22776
|
static_ips: Optional[bool] = None):
|
|
22753
22777
|
"""
|
|
22754
22778
|
:param bool cache_mode: Evict entries when getting close to maxmemory limit. The default value is `false`.
|
|
22779
|
+
:param str dragonfly_persistence: When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
22755
22780
|
:param bool dragonfly_ssl: Require SSL to access Dragonfly. The default value is `true`.
|
|
22756
22781
|
:param Sequence['GetDragonflyDragonflyUserConfigIpFilterObjectArgs'] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'
|
|
22757
22782
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
@@ -22768,6 +22793,8 @@ class GetDragonflyDragonflyUserConfigResult(dict):
|
|
|
22768
22793
|
"""
|
|
22769
22794
|
if cache_mode is not None:
|
|
22770
22795
|
pulumi.set(__self__, "cache_mode", cache_mode)
|
|
22796
|
+
if dragonfly_persistence is not None:
|
|
22797
|
+
pulumi.set(__self__, "dragonfly_persistence", dragonfly_persistence)
|
|
22771
22798
|
if dragonfly_ssl is not None:
|
|
22772
22799
|
pulumi.set(__self__, "dragonfly_ssl", dragonfly_ssl)
|
|
22773
22800
|
if ip_filter_objects is not None:
|
|
@@ -22803,6 +22830,14 @@ class GetDragonflyDragonflyUserConfigResult(dict):
|
|
|
22803
22830
|
"""
|
|
22804
22831
|
return pulumi.get(self, "cache_mode")
|
|
22805
22832
|
|
|
22833
|
+
@property
|
|
22834
|
+
@pulumi.getter(name="dragonflyPersistence")
|
|
22835
|
+
def dragonfly_persistence(self) -> Optional[str]:
|
|
22836
|
+
"""
|
|
22837
|
+
When persistence is 'rdb', Dragonfly does RDB dumps each 10 minutes. Also RDB dumps are done according to backup schedule for backup purposes. When persistence is 'off', no RDB dumps and backups are done, so data can be lost at any moment if service is restarted for any reason, or if service is powered off. Also service can't be forked.
|
|
22838
|
+
"""
|
|
22839
|
+
return pulumi.get(self, "dragonfly_persistence")
|
|
22840
|
+
|
|
22806
22841
|
@property
|
|
22807
22842
|
@pulumi.getter(name="dragonflySsl")
|
|
22808
22843
|
def dragonfly_ssl(self) -> Optional[bool]:
|
|
@@ -23737,12 +23772,6 @@ class GetGrafanaComponentResult(dict):
|
|
|
23737
23772
|
return pulumi.get(self, "usage")
|
|
23738
23773
|
|
|
23739
23774
|
|
|
23740
|
-
@pulumi.output_type
|
|
23741
|
-
class GetGrafanaGrafanaResult(dict):
|
|
23742
|
-
def __init__(__self__):
|
|
23743
|
-
pass
|
|
23744
|
-
|
|
23745
|
-
|
|
23746
23775
|
@pulumi.output_type
|
|
23747
23776
|
class GetGrafanaGrafanaUserConfigResult(dict):
|
|
23748
23777
|
def __init__(__self__, *,
|
|
@@ -23923,9 +23952,6 @@ class GetGrafanaGrafanaUserConfigResult(dict):
|
|
|
23923
23952
|
"""
|
|
23924
23953
|
Additional Cloud Regions for Backup Replication.
|
|
23925
23954
|
"""
|
|
23926
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
23927
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
23928
|
-
|
|
23929
23955
|
return pulumi.get(self, "additional_backup_regions")
|
|
23930
23956
|
|
|
23931
23957
|
@property
|
|
@@ -25865,12 +25891,6 @@ class GetKafkaConnectComponentResult(dict):
|
|
|
25865
25891
|
return pulumi.get(self, "usage")
|
|
25866
25892
|
|
|
25867
25893
|
|
|
25868
|
-
@pulumi.output_type
|
|
25869
|
-
class GetKafkaConnectKafkaConnectResult(dict):
|
|
25870
|
-
def __init__(__self__):
|
|
25871
|
-
pass
|
|
25872
|
-
|
|
25873
|
-
|
|
25874
25894
|
@pulumi.output_type
|
|
25875
25895
|
class GetKafkaConnectKafkaConnectUserConfigResult(dict):
|
|
25876
25896
|
def __init__(__self__, *,
|
|
@@ -26058,7 +26078,7 @@ class GetKafkaConnectKafkaConnectUserConfigKafkaConnectResult(dict):
|
|
|
26058
26078
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
26059
26079
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
26060
26080
|
:param str consumer_isolation_level: Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired.
|
|
26061
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26081
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26062
26082
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
26063
26083
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
26064
26084
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -26140,7 +26160,7 @@ class GetKafkaConnectKafkaConnectUserConfigKafkaConnectResult(dict):
|
|
|
26140
26160
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
26141
26161
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
26142
26162
|
"""
|
|
26143
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26163
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
26144
26164
|
"""
|
|
26145
26165
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
26146
26166
|
|
|
@@ -26452,11 +26472,11 @@ class GetKafkaKafkaResult(dict):
|
|
|
26452
26472
|
rest_uri: str,
|
|
26453
26473
|
schema_registry_uri: str):
|
|
26454
26474
|
"""
|
|
26455
|
-
:param str access_cert: The Kafka client certificate
|
|
26456
|
-
:param str access_key: The Kafka client certificate key
|
|
26457
|
-
:param str connect_uri: The Kafka Connect URI
|
|
26458
|
-
:param str rest_uri: The Kafka REST URI
|
|
26459
|
-
:param str schema_registry_uri: The Schema Registry URI
|
|
26475
|
+
:param str access_cert: The Kafka client certificate.
|
|
26476
|
+
:param str access_key: The Kafka client certificate key.
|
|
26477
|
+
:param str connect_uri: The Kafka Connect URI.
|
|
26478
|
+
:param str rest_uri: The Kafka REST URI.
|
|
26479
|
+
:param str schema_registry_uri: The Schema Registry URI.
|
|
26460
26480
|
"""
|
|
26461
26481
|
pulumi.set(__self__, "access_cert", access_cert)
|
|
26462
26482
|
pulumi.set(__self__, "access_key", access_key)
|
|
@@ -26468,7 +26488,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26468
26488
|
@pulumi.getter(name="accessCert")
|
|
26469
26489
|
def access_cert(self) -> str:
|
|
26470
26490
|
"""
|
|
26471
|
-
The Kafka client certificate
|
|
26491
|
+
The Kafka client certificate.
|
|
26472
26492
|
"""
|
|
26473
26493
|
return pulumi.get(self, "access_cert")
|
|
26474
26494
|
|
|
@@ -26476,7 +26496,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26476
26496
|
@pulumi.getter(name="accessKey")
|
|
26477
26497
|
def access_key(self) -> str:
|
|
26478
26498
|
"""
|
|
26479
|
-
The Kafka client certificate key
|
|
26499
|
+
The Kafka client certificate key.
|
|
26480
26500
|
"""
|
|
26481
26501
|
return pulumi.get(self, "access_key")
|
|
26482
26502
|
|
|
@@ -26484,7 +26504,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26484
26504
|
@pulumi.getter(name="connectUri")
|
|
26485
26505
|
def connect_uri(self) -> str:
|
|
26486
26506
|
"""
|
|
26487
|
-
The Kafka Connect URI
|
|
26507
|
+
The Kafka Connect URI.
|
|
26488
26508
|
"""
|
|
26489
26509
|
return pulumi.get(self, "connect_uri")
|
|
26490
26510
|
|
|
@@ -26492,7 +26512,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26492
26512
|
@pulumi.getter(name="restUri")
|
|
26493
26513
|
def rest_uri(self) -> str:
|
|
26494
26514
|
"""
|
|
26495
|
-
The Kafka REST URI
|
|
26515
|
+
The Kafka REST URI.
|
|
26496
26516
|
"""
|
|
26497
26517
|
return pulumi.get(self, "rest_uri")
|
|
26498
26518
|
|
|
@@ -26500,7 +26520,7 @@ class GetKafkaKafkaResult(dict):
|
|
|
26500
26520
|
@pulumi.getter(name="schemaRegistryUri")
|
|
26501
26521
|
def schema_registry_uri(self) -> str:
|
|
26502
26522
|
"""
|
|
26503
|
-
The Schema Registry URI
|
|
26523
|
+
The Schema Registry URI.
|
|
26504
26524
|
"""
|
|
26505
26525
|
return pulumi.get(self, "schema_registry_uri")
|
|
26506
26526
|
|
|
@@ -26880,7 +26900,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
26880
26900
|
:param int log_index_size_max_bytes: The maximum size in bytes of the offset index.
|
|
26881
26901
|
:param int log_local_retention_bytes: The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value.
|
|
26882
26902
|
:param int log_local_retention_ms: The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value.
|
|
26883
|
-
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
26903
|
+
:param bool log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
26884
26904
|
:param int log_message_timestamp_difference_max_ms: The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message.
|
|
26885
26905
|
:param str log_message_timestamp_type: Define whether the timestamp in the message is message create time or log append time.
|
|
26886
26906
|
:param bool log_preallocate: Should pre allocate file when create new segment?
|
|
@@ -26902,7 +26922,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
26902
26922
|
:param int replica_fetch_response_max_bytes: Maximum bytes expected for the entire fetch response (defaults to 10485760). Records are fetched in batches, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. As such, this is not an absolute maximum.
|
|
26903
26923
|
:param str sasl_oauthbearer_expected_audience: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences.
|
|
26904
26924
|
:param str sasl_oauthbearer_expected_issuer: Optional setting for the broker to use to verify that the JWT was created by the expected issuer.
|
|
26905
|
-
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
26925
|
+
:param str sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
26906
26926
|
:param str sasl_oauthbearer_sub_claim_name: Name of the scope from which to extract the subject claim from the JWT. Defaults to sub.
|
|
26907
26927
|
:param int socket_request_max_bytes: The maximum number of bytes in a socket request (defaults to 104857600).
|
|
26908
26928
|
:param bool transaction_partition_verification_enable: Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition.
|
|
@@ -27150,7 +27170,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
27150
27170
|
@pulumi.getter(name="logMessageDownconversionEnable")
|
|
27151
27171
|
def log_message_downconversion_enable(self) -> Optional[bool]:
|
|
27152
27172
|
"""
|
|
27153
|
-
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
27173
|
+
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
27154
27174
|
"""
|
|
27155
27175
|
return pulumi.get(self, "log_message_downconversion_enable")
|
|
27156
27176
|
|
|
@@ -27326,7 +27346,7 @@ class GetKafkaKafkaUserConfigKafkaResult(dict):
|
|
|
27326
27346
|
@pulumi.getter(name="saslOauthbearerJwksEndpointUrl")
|
|
27327
27347
|
def sasl_oauthbearer_jwks_endpoint_url(self) -> Optional[str]:
|
|
27328
27348
|
"""
|
|
27329
|
-
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
27349
|
+
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
27330
27350
|
"""
|
|
27331
27351
|
return pulumi.get(self, "sasl_oauthbearer_jwks_endpoint_url")
|
|
27332
27352
|
|
|
@@ -27426,7 +27446,7 @@ class GetKafkaKafkaUserConfigKafkaConnectConfigResult(dict):
|
|
|
27426
27446
|
:param str consumer_auto_offset_reset: What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server. Default is earliest.
|
|
27427
27447
|
:param int consumer_fetch_max_bytes: Records are fetched in batches by the consumer, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that the consumer can make progress. As such, this is not a absolute maximum.
|
|
27428
27448
|
:param str consumer_isolation_level: Transaction read isolation level. read_uncommitted is the default, but read_committed can be used if consume-exactly-once behavior is desired.
|
|
27429
|
-
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27449
|
+
:param int consumer_max_partition_fetch_bytes: Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27430
27450
|
:param int consumer_max_poll_interval_ms: The maximum delay in milliseconds between invocations of poll() when using consumer group management (defaults to 300000).
|
|
27431
27451
|
:param int consumer_max_poll_records: The maximum number of records returned in a single call to poll() (defaults to 500).
|
|
27432
27452
|
:param int offset_flush_interval_ms: The interval at which to try committing offsets for tasks (defaults to 60000).
|
|
@@ -27508,7 +27528,7 @@ class GetKafkaKafkaUserConfigKafkaConnectConfigResult(dict):
|
|
|
27508
27528
|
@pulumi.getter(name="consumerMaxPartitionFetchBytes")
|
|
27509
27529
|
def consumer_max_partition_fetch_bytes(self) -> Optional[int]:
|
|
27510
27530
|
"""
|
|
27511
|
-
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27531
|
+
Records are fetched in batches by the consumer.If the first record batch in the first non-empty partition of the fetch is larger than this limit, the batch will still be returned to ensure that the consumer can make progress.
|
|
27512
27532
|
"""
|
|
27513
27533
|
return pulumi.get(self, "consumer_max_partition_fetch_bytes")
|
|
27514
27534
|
|
|
@@ -28123,12 +28143,6 @@ class GetKafkaMirrorMakerComponentResult(dict):
|
|
|
28123
28143
|
return pulumi.get(self, "usage")
|
|
28124
28144
|
|
|
28125
28145
|
|
|
28126
|
-
@pulumi.output_type
|
|
28127
|
-
class GetKafkaMirrorMakerKafkaMirrormakerResult(dict):
|
|
28128
|
-
def __init__(__self__):
|
|
28129
|
-
pass
|
|
28130
|
-
|
|
28131
|
-
|
|
28132
28146
|
@pulumi.output_type
|
|
28133
28147
|
class GetKafkaMirrorMakerKafkaMirrormakerUserConfigResult(dict):
|
|
28134
28148
|
def __init__(__self__, *,
|
|
@@ -29030,12 +29044,6 @@ class GetM3AggregatorComponentResult(dict):
|
|
|
29030
29044
|
return pulumi.get(self, "usage")
|
|
29031
29045
|
|
|
29032
29046
|
|
|
29033
|
-
@pulumi.output_type
|
|
29034
|
-
class GetM3AggregatorM3aggregatorResult(dict):
|
|
29035
|
-
def __init__(__self__):
|
|
29036
|
-
pass
|
|
29037
|
-
|
|
29038
|
-
|
|
29039
29047
|
@pulumi.output_type
|
|
29040
29048
|
class GetM3AggregatorM3aggregatorUserConfigResult(dict):
|
|
29041
29049
|
def __init__(__self__, *,
|
|
@@ -29343,12 +29351,6 @@ class GetM3DbComponentResult(dict):
|
|
|
29343
29351
|
return pulumi.get(self, "usage")
|
|
29344
29352
|
|
|
29345
29353
|
|
|
29346
|
-
@pulumi.output_type
|
|
29347
|
-
class GetM3DbM3dbResult(dict):
|
|
29348
|
-
def __init__(__self__):
|
|
29349
|
-
pass
|
|
29350
|
-
|
|
29351
|
-
|
|
29352
29354
|
@pulumi.output_type
|
|
29353
29355
|
class GetM3DbM3dbUserConfigResult(dict):
|
|
29354
29356
|
def __init__(__self__, *,
|
|
@@ -29433,9 +29435,6 @@ class GetM3DbM3dbUserConfigResult(dict):
|
|
|
29433
29435
|
"""
|
|
29434
29436
|
Additional Cloud Regions for Backup Replication.
|
|
29435
29437
|
"""
|
|
29436
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
29437
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
29438
|
-
|
|
29439
29438
|
return pulumi.get(self, "additional_backup_regions")
|
|
29440
29439
|
|
|
29441
29440
|
@property
|
|
@@ -30291,12 +30290,6 @@ class GetMySqlComponentResult(dict):
|
|
|
30291
30290
|
return pulumi.get(self, "usage")
|
|
30292
30291
|
|
|
30293
30292
|
|
|
30294
|
-
@pulumi.output_type
|
|
30295
|
-
class GetMySqlMysqlResult(dict):
|
|
30296
|
-
def __init__(__self__):
|
|
30297
|
-
pass
|
|
30298
|
-
|
|
30299
|
-
|
|
30300
30293
|
@pulumi.output_type
|
|
30301
30294
|
class GetMySqlMysqlUserConfigResult(dict):
|
|
30302
30295
|
def __init__(__self__, *,
|
|
@@ -31399,7 +31392,7 @@ class GetOpenSearchOpensearchUserConfigResult(dict):
|
|
|
31399
31392
|
:param Sequence[str] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
31400
31393
|
:param Sequence[str] ip_filters: Allow incoming connections from CIDR address block, e.g. '10.20.0.0/16'.
|
|
31401
31394
|
:param bool keep_index_refresh_interval: Aiven automation resets index.refresh_interval to default value for every index to be sure that indices are always visible to search. If it doesn't fit your case, you can disable this by setting up this flag to true.
|
|
31402
|
-
:param int max_index_count:
|
|
31395
|
+
:param int max_index_count: Use index_patterns instead. The default value is `0`.
|
|
31403
31396
|
:param 'GetOpenSearchOpensearchUserConfigOpenidArgs' openid: OpenSearch OpenID Connect Configuration
|
|
31404
31397
|
:param 'GetOpenSearchOpensearchUserConfigOpensearchArgs' opensearch: OpenSearch settings
|
|
31405
31398
|
:param 'GetOpenSearchOpensearchUserConfigOpensearchDashboardsArgs' opensearch_dashboards: OpenSearch Dashboards settings
|
|
@@ -31467,9 +31460,6 @@ class GetOpenSearchOpensearchUserConfigResult(dict):
|
|
|
31467
31460
|
"""
|
|
31468
31461
|
Additional Cloud Regions for Backup Replication.
|
|
31469
31462
|
"""
|
|
31470
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
31471
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
31472
|
-
|
|
31473
31463
|
return pulumi.get(self, "additional_backup_regions")
|
|
31474
31464
|
|
|
31475
31465
|
@property
|
|
@@ -31543,7 +31533,7 @@ class GetOpenSearchOpensearchUserConfigResult(dict):
|
|
|
31543
31533
|
@pulumi.getter(name="maxIndexCount")
|
|
31544
31534
|
def max_index_count(self) -> Optional[int]:
|
|
31545
31535
|
"""
|
|
31546
|
-
|
|
31536
|
+
Use index_patterns instead. The default value is `0`.
|
|
31547
31537
|
"""
|
|
31548
31538
|
return pulumi.get(self, "max_index_count")
|
|
31549
31539
|
|
|
@@ -33777,7 +33767,7 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
33777
33767
|
:param int pg_partman_bgw_dot_interval: Sets the time interval to run pg_partman's scheduled tasks.
|
|
33778
33768
|
:param str pg_partman_bgw_dot_role: Controls which role to use for pg_partman's scheduled background tasks.
|
|
33779
33769
|
:param bool pg_stat_monitor_dot_pgsm_enable_query_plan: Enables or disables query plan monitoring.
|
|
33780
|
-
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets
|
|
33770
|
+
:param int pg_stat_monitor_dot_pgsm_max_buckets: Sets the maximum number of buckets.
|
|
33781
33771
|
:param str pg_stat_statements_dot_track: Controls which statements are counted. Specify top to track top-level statements (those issued directly by clients), all to also track nested statements (such as statements invoked within functions), or none to disable statement statistics collection. The default value is top.
|
|
33782
33772
|
:param int temp_file_limit: PostgreSQL temporary file limit in KiB, -1 for unlimited.
|
|
33783
33773
|
:param str timezone: PostgreSQL service timezone.
|
|
@@ -34203,7 +34193,7 @@ class GetPgPgUserConfigPgResult(dict):
|
|
|
34203
34193
|
@pulumi.getter(name="pgStatMonitorDotPgsmMaxBuckets")
|
|
34204
34194
|
def pg_stat_monitor_dot_pgsm_max_buckets(self) -> Optional[int]:
|
|
34205
34195
|
"""
|
|
34206
|
-
Sets the maximum number of buckets
|
|
34196
|
+
Sets the maximum number of buckets.
|
|
34207
34197
|
"""
|
|
34208
34198
|
return pulumi.get(self, "pg_stat_monitor_dot_pgsm_max_buckets")
|
|
34209
34199
|
|
|
@@ -34901,8 +34891,8 @@ class GetProjectTagResult(dict):
|
|
|
34901
34891
|
key: str,
|
|
34902
34892
|
value: str):
|
|
34903
34893
|
"""
|
|
34904
|
-
:param str key: Project tag key
|
|
34905
|
-
:param str value: Project tag value
|
|
34894
|
+
:param str key: Project tag key.
|
|
34895
|
+
:param str value: Project tag value.
|
|
34906
34896
|
"""
|
|
34907
34897
|
pulumi.set(__self__, "key", key)
|
|
34908
34898
|
pulumi.set(__self__, "value", value)
|
|
@@ -34911,7 +34901,7 @@ class GetProjectTagResult(dict):
|
|
|
34911
34901
|
@pulumi.getter
|
|
34912
34902
|
def key(self) -> str:
|
|
34913
34903
|
"""
|
|
34914
|
-
Project tag key
|
|
34904
|
+
Project tag key.
|
|
34915
34905
|
"""
|
|
34916
34906
|
return pulumi.get(self, "key")
|
|
34917
34907
|
|
|
@@ -34919,7 +34909,7 @@ class GetProjectTagResult(dict):
|
|
|
34919
34909
|
@pulumi.getter
|
|
34920
34910
|
def value(self) -> str:
|
|
34921
34911
|
"""
|
|
34922
|
-
Project tag value
|
|
34912
|
+
Project tag value.
|
|
34923
34913
|
"""
|
|
34924
34914
|
return pulumi.get(self, "value")
|
|
34925
34915
|
|
|
@@ -35019,12 +35009,6 @@ class GetRedisComponentResult(dict):
|
|
|
35019
35009
|
return pulumi.get(self, "usage")
|
|
35020
35010
|
|
|
35021
35011
|
|
|
35022
|
-
@pulumi.output_type
|
|
35023
|
-
class GetRedisRediResult(dict):
|
|
35024
|
-
def __init__(__self__):
|
|
35025
|
-
pass
|
|
35026
|
-
|
|
35027
|
-
|
|
35028
35012
|
@pulumi.output_type
|
|
35029
35013
|
class GetRedisRedisUserConfigResult(dict):
|
|
35030
35014
|
def __init__(__self__, *,
|
|
@@ -35137,9 +35121,6 @@ class GetRedisRedisUserConfigResult(dict):
|
|
|
35137
35121
|
"""
|
|
35138
35122
|
Additional Cloud Regions for Backup Replication.
|
|
35139
35123
|
"""
|
|
35140
|
-
warnings.warn("""This property is deprecated.""", DeprecationWarning)
|
|
35141
|
-
pulumi.log.warn("""additional_backup_regions is deprecated: This property is deprecated.""")
|
|
35142
|
-
|
|
35143
35124
|
return pulumi.get(self, "additional_backup_regions")
|
|
35144
35125
|
|
|
35145
35126
|
@property
|
|
@@ -36072,14 +36053,18 @@ class GetServiceIntegrationDatadogUserConfigDatadogTagResult(dict):
|
|
|
36072
36053
|
@pulumi.output_type
|
|
36073
36054
|
class GetServiceIntegrationDatadogUserConfigOpensearchResult(dict):
|
|
36074
36055
|
def __init__(__self__, *,
|
|
36056
|
+
cluster_stats_enabled: Optional[bool] = None,
|
|
36075
36057
|
index_stats_enabled: Optional[bool] = None,
|
|
36076
36058
|
pending_task_stats_enabled: Optional[bool] = None,
|
|
36077
36059
|
pshard_stats_enabled: Optional[bool] = None):
|
|
36078
36060
|
"""
|
|
36061
|
+
:param bool cluster_stats_enabled: Enable Datadog Opensearch Cluster Monitoring.
|
|
36079
36062
|
:param bool index_stats_enabled: Enable Datadog Opensearch Index Monitoring.
|
|
36080
36063
|
:param bool pending_task_stats_enabled: Enable Datadog Opensearch Pending Task Monitoring.
|
|
36081
36064
|
:param bool pshard_stats_enabled: Enable Datadog Opensearch Primary Shard Monitoring.
|
|
36082
36065
|
"""
|
|
36066
|
+
if cluster_stats_enabled is not None:
|
|
36067
|
+
pulumi.set(__self__, "cluster_stats_enabled", cluster_stats_enabled)
|
|
36083
36068
|
if index_stats_enabled is not None:
|
|
36084
36069
|
pulumi.set(__self__, "index_stats_enabled", index_stats_enabled)
|
|
36085
36070
|
if pending_task_stats_enabled is not None:
|
|
@@ -36087,6 +36072,14 @@ class GetServiceIntegrationDatadogUserConfigOpensearchResult(dict):
|
|
|
36087
36072
|
if pshard_stats_enabled is not None:
|
|
36088
36073
|
pulumi.set(__self__, "pshard_stats_enabled", pshard_stats_enabled)
|
|
36089
36074
|
|
|
36075
|
+
@property
|
|
36076
|
+
@pulumi.getter(name="clusterStatsEnabled")
|
|
36077
|
+
def cluster_stats_enabled(self) -> Optional[bool]:
|
|
36078
|
+
"""
|
|
36079
|
+
Enable Datadog Opensearch Cluster Monitoring.
|
|
36080
|
+
"""
|
|
36081
|
+
return pulumi.get(self, "cluster_stats_enabled")
|
|
36082
|
+
|
|
36090
36083
|
@property
|
|
36091
36084
|
@pulumi.getter(name="indexStatsEnabled")
|
|
36092
36085
|
def index_stats_enabled(self) -> Optional[bool]:
|
|
@@ -36426,7 +36419,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudBigqueryResult(dict):
|
|
|
36426
36419
|
service_account_credentials: str):
|
|
36427
36420
|
"""
|
|
36428
36421
|
:param str project_id: GCP project id.
|
|
36429
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36422
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36430
36423
|
"""
|
|
36431
36424
|
pulumi.set(__self__, "project_id", project_id)
|
|
36432
36425
|
pulumi.set(__self__, "service_account_credentials", service_account_credentials)
|
|
@@ -36443,7 +36436,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudBigqueryResult(dict):
|
|
|
36443
36436
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
36444
36437
|
def service_account_credentials(self) -> str:
|
|
36445
36438
|
"""
|
|
36446
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36439
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36447
36440
|
"""
|
|
36448
36441
|
return pulumi.get(self, "service_account_credentials")
|
|
36449
36442
|
|
|
@@ -36457,7 +36450,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfigResult(di
|
|
|
36457
36450
|
"""
|
|
36458
36451
|
:param str log_id: Google Cloud Logging log id.
|
|
36459
36452
|
:param str project_id: GCP project id.
|
|
36460
|
-
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36453
|
+
:param str service_account_credentials: This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36461
36454
|
"""
|
|
36462
36455
|
pulumi.set(__self__, "log_id", log_id)
|
|
36463
36456
|
pulumi.set(__self__, "project_id", project_id)
|
|
@@ -36483,7 +36476,7 @@ class GetServiceIntegrationEndpointExternalGoogleCloudLoggingUserConfigResult(di
|
|
|
36483
36476
|
@pulumi.getter(name="serviceAccountCredentials")
|
|
36484
36477
|
def service_account_credentials(self) -> str:
|
|
36485
36478
|
"""
|
|
36486
|
-
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys
|
|
36479
|
+
This is a JSON object with the fields documented in https://cloud.google.com/iam/docs/creating-managing-service-account-keys.
|
|
36487
36480
|
"""
|
|
36488
36481
|
return pulumi.get(self, "service_account_credentials")
|
|
36489
36482
|
|
|
@@ -37524,7 +37517,7 @@ class GetServiceIntegrationMetricsUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37524
37517
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
37525
37518
|
"""
|
|
37526
37519
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE_SCHEMA.EVENT_WAITS.
|
|
37527
|
-
:param bool gather_file_events_stats:
|
|
37520
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37528
37521
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMARY_BY_INDEX_USAGE.
|
|
37529
37522
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
37530
37523
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION_SCHEMA.INNODB_METRICS.
|
|
@@ -37579,7 +37572,7 @@ class GetServiceIntegrationMetricsUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37579
37572
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
37580
37573
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
37581
37574
|
"""
|
|
37582
|
-
|
|
37575
|
+
Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37583
37576
|
"""
|
|
37584
37577
|
return pulumi.get(self, "gather_file_events_stats")
|
|
37585
37578
|
|
|
@@ -37737,7 +37730,7 @@ class GetServiceIntegrationPrometheusUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37737
37730
|
perf_events_statements_time_limit: Optional[int] = None):
|
|
37738
37731
|
"""
|
|
37739
37732
|
:param bool gather_event_waits: Gather metrics from PERFORMANCE_SCHEMA.EVENT_WAITS.
|
|
37740
|
-
:param bool gather_file_events_stats:
|
|
37733
|
+
:param bool gather_file_events_stats: Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37741
37734
|
:param bool gather_index_io_waits: Gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMARY_BY_INDEX_USAGE.
|
|
37742
37735
|
:param bool gather_info_schema_auto_inc: Gather auto_increment columns and max values from information schema.
|
|
37743
37736
|
:param bool gather_innodb_metrics: Gather metrics from INFORMATION_SCHEMA.INNODB_METRICS.
|
|
@@ -37792,7 +37785,7 @@ class GetServiceIntegrationPrometheusUserConfigSourceMysqlTelegrafResult(dict):
|
|
|
37792
37785
|
@pulumi.getter(name="gatherFileEventsStats")
|
|
37793
37786
|
def gather_file_events_stats(self) -> Optional[bool]:
|
|
37794
37787
|
"""
|
|
37795
|
-
|
|
37788
|
+
Gather metrics from PERFORMANCE_SCHEMA.FILE_SUMMARY_BY_EVENT_NAME.
|
|
37796
37789
|
"""
|
|
37797
37790
|
return pulumi.get(self, "gather_file_events_stats")
|
|
37798
37791
|
|