pulumi-aiven 6.21.0a1722057299__py3-none-any.whl → 6.22.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pulumi-aiven might be problematic. Click here for more details.
- pulumi_aiven/_inputs.py +980 -158
- pulumi_aiven/account_authentication.py +5 -5
- pulumi_aiven/account_team_project.py +7 -7
- pulumi_aiven/aws_vpc_peering_connection.py +7 -7
- pulumi_aiven/azure_privatelink_connection_approval.py +16 -16
- pulumi_aiven/azure_vpc_peering_connection.py +7 -7
- pulumi_aiven/cassandra.py +46 -46
- pulumi_aiven/clickhouse.py +34 -34
- pulumi_aiven/clickhouse_grant.py +34 -34
- pulumi_aiven/dragonfly.py +40 -40
- pulumi_aiven/flink.py +40 -40
- pulumi_aiven/flink_application_deployment.py +16 -16
- pulumi_aiven/flink_application_version.py +36 -36
- pulumi_aiven/gcp_vpc_peering_connection.py +7 -7
- pulumi_aiven/get_account_team_project.py +1 -1
- pulumi_aiven/get_aws_vpc_peering_connection.py +1 -1
- pulumi_aiven/get_azure_vpc_peering_connection.py +1 -1
- pulumi_aiven/get_cassanda.py +1 -1
- pulumi_aiven/get_cassandra.py +1 -1
- pulumi_aiven/get_clickhouse.py +1 -1
- pulumi_aiven/get_dragonfly.py +1 -1
- pulumi_aiven/get_flink.py +1 -1
- pulumi_aiven/get_gcp_vpc_peering_connection.py +1 -1
- pulumi_aiven/get_grafana.py +1 -1
- pulumi_aiven/get_kafka.py +1 -1
- pulumi_aiven/get_kafka_connect.py +1 -1
- pulumi_aiven/get_kafka_mirror_maker.py +1 -1
- pulumi_aiven/get_m3_aggregator.py +1 -1
- pulumi_aiven/get_m3_db.py +1 -1
- pulumi_aiven/get_my_sql.py +1 -1
- pulumi_aiven/get_mysql_user.py +1 -1
- pulumi_aiven/get_open_search.py +1 -1
- pulumi_aiven/get_pg.py +8 -8
- pulumi_aiven/get_pg_database.py +8 -8
- pulumi_aiven/get_pg_user.py +16 -16
- pulumi_aiven/get_project_user.py +1 -1
- pulumi_aiven/get_project_vpc.py +1 -1
- pulumi_aiven/get_redis.py +1 -1
- pulumi_aiven/get_service_integration.py +3 -3
- pulumi_aiven/get_thanos.py +1 -1
- pulumi_aiven/get_transit_gateway_vpc_attachment.py +1 -1
- pulumi_aiven/get_valkey.py +1 -1
- pulumi_aiven/grafana.py +46 -46
- pulumi_aiven/influx_db.py +48 -41
- pulumi_aiven/kafka.py +62 -62
- pulumi_aiven/kafka_connect.py +59 -59
- pulumi_aiven/kafka_connector.py +2 -2
- pulumi_aiven/kafka_mirror_maker.py +45 -45
- pulumi_aiven/kafka_topic.py +18 -18
- pulumi_aiven/m3_aggregator.py +40 -40
- pulumi_aiven/m3_db.py +48 -48
- pulumi_aiven/my_sql.py +54 -54
- pulumi_aiven/mysql_user.py +7 -7
- pulumi_aiven/open_search.py +56 -56
- pulumi_aiven/organization.py +3 -3
- pulumi_aiven/organization_group_project.py +3 -3
- pulumi_aiven/organization_user_group_member.py +3 -3
- pulumi_aiven/outputs.py +1708 -317
- pulumi_aiven/pg.py +43 -43
- pulumi_aiven/pg_database.py +12 -12
- pulumi_aiven/pg_user.py +47 -47
- pulumi_aiven/project.py +5 -5
- pulumi_aiven/project_user.py +7 -7
- pulumi_aiven/project_vpc.py +4 -4
- pulumi_aiven/pulumi-plugin.json +1 -1
- pulumi_aiven/redis.py +46 -46
- pulumi_aiven/service_integration.py +72 -72
- pulumi_aiven/service_integration_endpoint.py +65 -65
- pulumi_aiven/thanos.py +34 -34
- pulumi_aiven/transit_gateway_vpc_attachment.py +7 -7
- pulumi_aiven/valkey.py +40 -40
- {pulumi_aiven-6.21.0a1722057299.dist-info → pulumi_aiven-6.22.0.dist-info}/METADATA +1 -1
- {pulumi_aiven-6.21.0a1722057299.dist-info → pulumi_aiven-6.22.0.dist-info}/RECORD +75 -75
- {pulumi_aiven-6.21.0a1722057299.dist-info → pulumi_aiven-6.22.0.dist-info}/WHEEL +1 -1
- {pulumi_aiven-6.21.0a1722057299.dist-info → pulumi_aiven-6.22.0.dist-info}/top_level.txt +0 -0
pulumi_aiven/_inputs.py
CHANGED
|
@@ -100,6 +100,7 @@ __all__ = [
|
|
|
100
100
|
'KafkaConnectorTaskArgs',
|
|
101
101
|
'KafkaKafkaArgs',
|
|
102
102
|
'KafkaKafkaUserConfigArgs',
|
|
103
|
+
'KafkaKafkaUserConfigFollowerFetchingArgs',
|
|
103
104
|
'KafkaKafkaUserConfigIpFilterObjectArgs',
|
|
104
105
|
'KafkaKafkaUserConfigKafkaArgs',
|
|
105
106
|
'KafkaKafkaUserConfigKafkaAuthenticationMethodsArgs',
|
|
@@ -108,6 +109,7 @@ __all__ = [
|
|
|
108
109
|
'KafkaKafkaUserConfigKafkaConnectSecretProviderAwsArgs',
|
|
109
110
|
'KafkaKafkaUserConfigKafkaConnectSecretProviderVaultArgs',
|
|
110
111
|
'KafkaKafkaUserConfigKafkaRestConfigArgs',
|
|
112
|
+
'KafkaKafkaUserConfigKafkaSaslMechanismsArgs',
|
|
111
113
|
'KafkaKafkaUserConfigPrivateAccessArgs',
|
|
112
114
|
'KafkaKafkaUserConfigPrivatelinkAccessArgs',
|
|
113
115
|
'KafkaKafkaUserConfigPublicAccessArgs',
|
|
@@ -168,7 +170,10 @@ __all__ = [
|
|
|
168
170
|
'OpenSearchComponentArgs',
|
|
169
171
|
'OpenSearchOpensearchArgs',
|
|
170
172
|
'OpenSearchOpensearchUserConfigArgs',
|
|
173
|
+
'OpenSearchOpensearchUserConfigAzureMigrationArgs',
|
|
174
|
+
'OpenSearchOpensearchUserConfigGcsMigrationArgs',
|
|
171
175
|
'OpenSearchOpensearchUserConfigIndexPatternArgs',
|
|
176
|
+
'OpenSearchOpensearchUserConfigIndexRollupArgs',
|
|
172
177
|
'OpenSearchOpensearchUserConfigIndexTemplateArgs',
|
|
173
178
|
'OpenSearchOpensearchUserConfigIpFilterObjectArgs',
|
|
174
179
|
'OpenSearchOpensearchUserConfigOpenidArgs',
|
|
@@ -180,6 +185,7 @@ __all__ = [
|
|
|
180
185
|
'OpenSearchOpensearchUserConfigPrivateAccessArgs',
|
|
181
186
|
'OpenSearchOpensearchUserConfigPrivatelinkAccessArgs',
|
|
182
187
|
'OpenSearchOpensearchUserConfigPublicAccessArgs',
|
|
188
|
+
'OpenSearchOpensearchUserConfigS3MigrationArgs',
|
|
183
189
|
'OpenSearchOpensearchUserConfigSamlArgs',
|
|
184
190
|
'OpenSearchServiceIntegrationArgs',
|
|
185
191
|
'OpenSearchTagArgs',
|
|
@@ -2373,6 +2379,7 @@ class DragonflyDragonflyUserConfigMigrationArgs:
|
|
|
2373
2379
|
port: pulumi.Input[int],
|
|
2374
2380
|
dbname: Optional[pulumi.Input[str]] = None,
|
|
2375
2381
|
ignore_dbs: Optional[pulumi.Input[str]] = None,
|
|
2382
|
+
ignore_roles: Optional[pulumi.Input[str]] = None,
|
|
2376
2383
|
method: Optional[pulumi.Input[str]] = None,
|
|
2377
2384
|
password: Optional[pulumi.Input[str]] = None,
|
|
2378
2385
|
ssl: Optional[pulumi.Input[bool]] = None,
|
|
@@ -2382,6 +2389,7 @@ class DragonflyDragonflyUserConfigMigrationArgs:
|
|
|
2382
2389
|
:param pulumi.Input[int] port: Port number of the server where to migrate data from. Example: `1234`.
|
|
2383
2390
|
:param pulumi.Input[str] dbname: Database name for bootstrapping the initial connection. Example: `defaultdb`.
|
|
2384
2391
|
:param pulumi.Input[str] ignore_dbs: Comma-separated list of databases, which should be ignored during migration (supported by MySQL and PostgreSQL only at the moment). Example: `db1,db2`.
|
|
2392
|
+
:param pulumi.Input[str] ignore_roles: Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
2385
2393
|
:param pulumi.Input[str] method: Enum: `dump`, `replication`. The migration method to be used (currently supported only by Redis, Dragonfly, MySQL and PostgreSQL service types).
|
|
2386
2394
|
:param pulumi.Input[str] password: Password for authentication with the server where to migrate data from. Example: `jjKk45Nnd`.
|
|
2387
2395
|
:param pulumi.Input[bool] ssl: The server where to migrate data from is secured with SSL. Default: `true`.
|
|
@@ -2393,6 +2401,8 @@ class DragonflyDragonflyUserConfigMigrationArgs:
|
|
|
2393
2401
|
pulumi.set(__self__, "dbname", dbname)
|
|
2394
2402
|
if ignore_dbs is not None:
|
|
2395
2403
|
pulumi.set(__self__, "ignore_dbs", ignore_dbs)
|
|
2404
|
+
if ignore_roles is not None:
|
|
2405
|
+
pulumi.set(__self__, "ignore_roles", ignore_roles)
|
|
2396
2406
|
if method is not None:
|
|
2397
2407
|
pulumi.set(__self__, "method", method)
|
|
2398
2408
|
if password is not None:
|
|
@@ -2450,6 +2460,18 @@ class DragonflyDragonflyUserConfigMigrationArgs:
|
|
|
2450
2460
|
def ignore_dbs(self, value: Optional[pulumi.Input[str]]):
|
|
2451
2461
|
pulumi.set(self, "ignore_dbs", value)
|
|
2452
2462
|
|
|
2463
|
+
@property
|
|
2464
|
+
@pulumi.getter(name="ignoreRoles")
|
|
2465
|
+
def ignore_roles(self) -> Optional[pulumi.Input[str]]:
|
|
2466
|
+
"""
|
|
2467
|
+
Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
2468
|
+
"""
|
|
2469
|
+
return pulumi.get(self, "ignore_roles")
|
|
2470
|
+
|
|
2471
|
+
@ignore_roles.setter
|
|
2472
|
+
def ignore_roles(self, value: Optional[pulumi.Input[str]]):
|
|
2473
|
+
pulumi.set(self, "ignore_roles", value)
|
|
2474
|
+
|
|
2453
2475
|
@property
|
|
2454
2476
|
@pulumi.getter
|
|
2455
2477
|
def method(self) -> Optional[pulumi.Input[str]]:
|
|
@@ -3480,7 +3502,8 @@ class GrafanaGrafanaUserConfigArgs:
|
|
|
3480
3502
|
unified_alerting_enabled: Optional[pulumi.Input[bool]] = None,
|
|
3481
3503
|
user_auto_assign_org: Optional[pulumi.Input[bool]] = None,
|
|
3482
3504
|
user_auto_assign_org_role: Optional[pulumi.Input[str]] = None,
|
|
3483
|
-
viewers_can_edit: Optional[pulumi.Input[bool]] = None
|
|
3505
|
+
viewers_can_edit: Optional[pulumi.Input[bool]] = None,
|
|
3506
|
+
wal: Optional[pulumi.Input[bool]] = None):
|
|
3484
3507
|
"""
|
|
3485
3508
|
:param pulumi.Input[str] additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
3486
3509
|
:param pulumi.Input[bool] alerting_enabled: Enable or disable Grafana legacy alerting functionality. This should not be enabled with unified*alerting*enabled.
|
|
@@ -3524,6 +3547,7 @@ class GrafanaGrafanaUserConfigArgs:
|
|
|
3524
3547
|
:param pulumi.Input[bool] user_auto_assign_org: Auto-assign new users on signup to main organization. Defaults to false.
|
|
3525
3548
|
:param pulumi.Input[str] user_auto_assign_org_role: Enum: `Viewer`, `Admin`, `Editor`. Set role for new signups. Defaults to Viewer.
|
|
3526
3549
|
:param pulumi.Input[bool] viewers_can_edit: Users with view-only permission can edit but not save dashboards.
|
|
3550
|
+
:param pulumi.Input[bool] wal: Setting to enable/disable Write-Ahead Logging. The default value is false (disabled).
|
|
3527
3551
|
"""
|
|
3528
3552
|
if additional_backup_regions is not None:
|
|
3529
3553
|
pulumi.set(__self__, "additional_backup_regions", additional_backup_regions)
|
|
@@ -3612,6 +3636,8 @@ class GrafanaGrafanaUserConfigArgs:
|
|
|
3612
3636
|
pulumi.set(__self__, "user_auto_assign_org_role", user_auto_assign_org_role)
|
|
3613
3637
|
if viewers_can_edit is not None:
|
|
3614
3638
|
pulumi.set(__self__, "viewers_can_edit", viewers_can_edit)
|
|
3639
|
+
if wal is not None:
|
|
3640
|
+
pulumi.set(__self__, "wal", wal)
|
|
3615
3641
|
|
|
3616
3642
|
@property
|
|
3617
3643
|
@pulumi.getter(name="additionalBackupRegions")
|
|
@@ -4118,6 +4144,18 @@ class GrafanaGrafanaUserConfigArgs:
|
|
|
4118
4144
|
def viewers_can_edit(self, value: Optional[pulumi.Input[bool]]):
|
|
4119
4145
|
pulumi.set(self, "viewers_can_edit", value)
|
|
4120
4146
|
|
|
4147
|
+
@property
|
|
4148
|
+
@pulumi.getter
|
|
4149
|
+
def wal(self) -> Optional[pulumi.Input[bool]]:
|
|
4150
|
+
"""
|
|
4151
|
+
Setting to enable/disable Write-Ahead Logging. The default value is false (disabled).
|
|
4152
|
+
"""
|
|
4153
|
+
return pulumi.get(self, "wal")
|
|
4154
|
+
|
|
4155
|
+
@wal.setter
|
|
4156
|
+
def wal(self, value: Optional[pulumi.Input[bool]]):
|
|
4157
|
+
pulumi.set(self, "wal", value)
|
|
4158
|
+
|
|
4121
4159
|
|
|
4122
4160
|
@pulumi.input_type
|
|
4123
4161
|
class GrafanaGrafanaUserConfigAuthAzureadArgs:
|
|
@@ -7344,6 +7382,7 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7344
7382
|
additional_backup_regions: Optional[pulumi.Input[str]] = None,
|
|
7345
7383
|
aiven_kafka_topic_messages: Optional[pulumi.Input[bool]] = None,
|
|
7346
7384
|
custom_domain: Optional[pulumi.Input[str]] = None,
|
|
7385
|
+
follower_fetching: Optional[pulumi.Input['KafkaKafkaUserConfigFollowerFetchingArgs']] = None,
|
|
7347
7386
|
ip_filter_objects: Optional[pulumi.Input[Sequence[pulumi.Input['KafkaKafkaUserConfigIpFilterObjectArgs']]]] = None,
|
|
7348
7387
|
ip_filter_strings: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
|
|
7349
7388
|
ip_filters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
|
|
@@ -7355,6 +7394,7 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7355
7394
|
kafka_rest: Optional[pulumi.Input[bool]] = None,
|
|
7356
7395
|
kafka_rest_authorization: Optional[pulumi.Input[bool]] = None,
|
|
7357
7396
|
kafka_rest_config: Optional[pulumi.Input['KafkaKafkaUserConfigKafkaRestConfigArgs']] = None,
|
|
7397
|
+
kafka_sasl_mechanisms: Optional[pulumi.Input['KafkaKafkaUserConfigKafkaSaslMechanismsArgs']] = None,
|
|
7358
7398
|
kafka_version: Optional[pulumi.Input[str]] = None,
|
|
7359
7399
|
letsencrypt_sasl_privatelink: Optional[pulumi.Input[bool]] = None,
|
|
7360
7400
|
private_access: Optional[pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs']] = None,
|
|
@@ -7369,6 +7409,7 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7369
7409
|
:param pulumi.Input[str] additional_backup_regions: Additional Cloud Regions for Backup Replication.
|
|
7370
7410
|
:param pulumi.Input[bool] aiven_kafka_topic_messages: Allow access to read Kafka topic messages in the Aiven Console and REST API.
|
|
7371
7411
|
:param pulumi.Input[str] custom_domain: Serve the web frontend using a custom CNAME pointing to the Aiven DNS name. Example: `grafana.example.org`.
|
|
7412
|
+
:param pulumi.Input['KafkaKafkaUserConfigFollowerFetchingArgs'] follower_fetching: Enable follower fetching
|
|
7372
7413
|
:param pulumi.Input[Sequence[pulumi.Input['KafkaKafkaUserConfigIpFilterObjectArgs']]] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. `10.20.0.0/16`
|
|
7373
7414
|
:param pulumi.Input[Sequence[pulumi.Input[str]]] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. `10.20.0.0/16`.
|
|
7374
7415
|
:param pulumi.Input[Sequence[pulumi.Input[str]]] ip_filters: Allow incoming connections from CIDR address block, e.g. `10.20.0.0/16`.
|
|
@@ -7379,6 +7420,7 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7379
7420
|
:param pulumi.Input[bool] kafka_rest: Enable Kafka-REST service. Default: `false`.
|
|
7380
7421
|
:param pulumi.Input[bool] kafka_rest_authorization: Enable authorization in Kafka-REST service.
|
|
7381
7422
|
:param pulumi.Input['KafkaKafkaUserConfigKafkaRestConfigArgs'] kafka_rest_config: Kafka REST configuration
|
|
7423
|
+
:param pulumi.Input['KafkaKafkaUserConfigKafkaSaslMechanismsArgs'] kafka_sasl_mechanisms: Kafka SASL mechanisms
|
|
7382
7424
|
:param pulumi.Input[str] kafka_version: Enum: `3.1`, `3.2`, `3.3`, `3.4`, `3.5`, `3.6`, `3.7`, and newer. Kafka major version.
|
|
7383
7425
|
:param pulumi.Input[bool] letsencrypt_sasl_privatelink: Use Letsencrypt CA for Kafka SASL via Privatelink.
|
|
7384
7426
|
:param pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs'] private_access: Allow access to selected service ports from private networks
|
|
@@ -7399,6 +7441,8 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7399
7441
|
pulumi.set(__self__, "aiven_kafka_topic_messages", aiven_kafka_topic_messages)
|
|
7400
7442
|
if custom_domain is not None:
|
|
7401
7443
|
pulumi.set(__self__, "custom_domain", custom_domain)
|
|
7444
|
+
if follower_fetching is not None:
|
|
7445
|
+
pulumi.set(__self__, "follower_fetching", follower_fetching)
|
|
7402
7446
|
if ip_filter_objects is not None:
|
|
7403
7447
|
pulumi.set(__self__, "ip_filter_objects", ip_filter_objects)
|
|
7404
7448
|
if ip_filter_strings is not None:
|
|
@@ -7424,6 +7468,8 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7424
7468
|
pulumi.set(__self__, "kafka_rest_authorization", kafka_rest_authorization)
|
|
7425
7469
|
if kafka_rest_config is not None:
|
|
7426
7470
|
pulumi.set(__self__, "kafka_rest_config", kafka_rest_config)
|
|
7471
|
+
if kafka_sasl_mechanisms is not None:
|
|
7472
|
+
pulumi.set(__self__, "kafka_sasl_mechanisms", kafka_sasl_mechanisms)
|
|
7427
7473
|
if kafka_version is not None:
|
|
7428
7474
|
pulumi.set(__self__, "kafka_version", kafka_version)
|
|
7429
7475
|
if letsencrypt_sasl_privatelink is not None:
|
|
@@ -7482,6 +7528,18 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7482
7528
|
def custom_domain(self, value: Optional[pulumi.Input[str]]):
|
|
7483
7529
|
pulumi.set(self, "custom_domain", value)
|
|
7484
7530
|
|
|
7531
|
+
@property
|
|
7532
|
+
@pulumi.getter(name="followerFetching")
|
|
7533
|
+
def follower_fetching(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigFollowerFetchingArgs']]:
|
|
7534
|
+
"""
|
|
7535
|
+
Enable follower fetching
|
|
7536
|
+
"""
|
|
7537
|
+
return pulumi.get(self, "follower_fetching")
|
|
7538
|
+
|
|
7539
|
+
@follower_fetching.setter
|
|
7540
|
+
def follower_fetching(self, value: Optional[pulumi.Input['KafkaKafkaUserConfigFollowerFetchingArgs']]):
|
|
7541
|
+
pulumi.set(self, "follower_fetching", value)
|
|
7542
|
+
|
|
7485
7543
|
@property
|
|
7486
7544
|
@pulumi.getter(name="ipFilterObjects")
|
|
7487
7545
|
def ip_filter_objects(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KafkaKafkaUserConfigIpFilterObjectArgs']]]]:
|
|
@@ -7612,6 +7670,18 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7612
7670
|
def kafka_rest_config(self, value: Optional[pulumi.Input['KafkaKafkaUserConfigKafkaRestConfigArgs']]):
|
|
7613
7671
|
pulumi.set(self, "kafka_rest_config", value)
|
|
7614
7672
|
|
|
7673
|
+
@property
|
|
7674
|
+
@pulumi.getter(name="kafkaSaslMechanisms")
|
|
7675
|
+
def kafka_sasl_mechanisms(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigKafkaSaslMechanismsArgs']]:
|
|
7676
|
+
"""
|
|
7677
|
+
Kafka SASL mechanisms
|
|
7678
|
+
"""
|
|
7679
|
+
return pulumi.get(self, "kafka_sasl_mechanisms")
|
|
7680
|
+
|
|
7681
|
+
@kafka_sasl_mechanisms.setter
|
|
7682
|
+
def kafka_sasl_mechanisms(self, value: Optional[pulumi.Input['KafkaKafkaUserConfigKafkaSaslMechanismsArgs']]):
|
|
7683
|
+
pulumi.set(self, "kafka_sasl_mechanisms", value)
|
|
7684
|
+
|
|
7615
7685
|
@property
|
|
7616
7686
|
@pulumi.getter(name="kafkaVersion")
|
|
7617
7687
|
def kafka_version(self) -> Optional[pulumi.Input[str]]:
|
|
@@ -7733,6 +7803,29 @@ class KafkaKafkaUserConfigArgs:
|
|
|
7733
7803
|
pulumi.set(self, "tiered_storage", value)
|
|
7734
7804
|
|
|
7735
7805
|
|
|
7806
|
+
@pulumi.input_type
|
|
7807
|
+
class KafkaKafkaUserConfigFollowerFetchingArgs:
|
|
7808
|
+
def __init__(__self__, *,
|
|
7809
|
+
enabled: Optional[pulumi.Input[bool]] = None):
|
|
7810
|
+
"""
|
|
7811
|
+
:param pulumi.Input[bool] enabled: Whether to enable the follower fetching functionality.
|
|
7812
|
+
"""
|
|
7813
|
+
if enabled is not None:
|
|
7814
|
+
pulumi.set(__self__, "enabled", enabled)
|
|
7815
|
+
|
|
7816
|
+
@property
|
|
7817
|
+
@pulumi.getter
|
|
7818
|
+
def enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
7819
|
+
"""
|
|
7820
|
+
Whether to enable the follower fetching functionality.
|
|
7821
|
+
"""
|
|
7822
|
+
return pulumi.get(self, "enabled")
|
|
7823
|
+
|
|
7824
|
+
@enabled.setter
|
|
7825
|
+
def enabled(self, value: Optional[pulumi.Input[bool]]):
|
|
7826
|
+
pulumi.set(self, "enabled", value)
|
|
7827
|
+
|
|
7828
|
+
|
|
7736
7829
|
@pulumi.input_type
|
|
7737
7830
|
class KafkaKafkaUserConfigIpFilterObjectArgs:
|
|
7738
7831
|
def __init__(__self__, *,
|
|
@@ -7821,52 +7914,52 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
7821
7914
|
transaction_remove_expired_transaction_cleanup_interval_ms: Optional[pulumi.Input[int]] = None,
|
|
7822
7915
|
transaction_state_log_segment_bytes: Optional[pulumi.Input[int]] = None):
|
|
7823
7916
|
"""
|
|
7824
|
-
:param pulumi.Input[bool] auto_create_topics_enable: Enable auto
|
|
7825
|
-
:param pulumi.Input[str] compression_type: Enum: `gzip`, `snappy`, `lz4`, `zstd`, `uncompressed`, `producer`. Specify the final compression type for a given topic. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `uncompressed` which is equivalent to no compression; and `producer` which means retain the original compression codec set by the producer.
|
|
7826
|
-
:param pulumi.Input[int] connections_max_idle_ms: Idle connections timeout: the server socket processor threads close the connections that idle for longer than this. Example: `540000`.
|
|
7827
|
-
:param pulumi.Input[int] default_replication_factor: Replication factor for
|
|
7828
|
-
:param pulumi.Input[int] group_initial_rebalance_delay_ms: The amount of time, in milliseconds, the group coordinator will wait for more consumers to join a new group before performing the first rebalance. A longer delay means potentially fewer rebalances, but increases the time until processing begins. The default value for this is 3 seconds. During development and testing it might be desirable to set this to 0 in order to not delay test execution time. Example: `3000`.
|
|
7829
|
-
:param pulumi.Input[int] group_max_session_timeout_ms: The maximum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. Example: `1800000`.
|
|
7830
|
-
:param pulumi.Input[int] group_min_session_timeout_ms: The minimum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. Example: `6000`.
|
|
7831
|
-
:param pulumi.Input[int] log_cleaner_delete_retention_ms: How long are delete records retained? Example: `86400000`.
|
|
7832
|
-
:param pulumi.Input[int] log_cleaner_max_compaction_lag_ms: The maximum amount of time message will remain uncompacted. Only applicable for logs that are being compacted.
|
|
7833
|
-
:param pulumi.Input[float] log_cleaner_min_cleanable_ratio: Controls log compactor frequency. Larger value means more frequent compactions but also more space wasted for logs. Consider setting log.cleaner.max.compaction.lag.ms to enforce compactions sooner, instead of setting a very high value for this option. Example: `0.5`.
|
|
7834
|
-
:param pulumi.Input[int] log_cleaner_min_compaction_lag_ms: The minimum time a message will remain uncompacted in the log. Only applicable for logs that are being compacted.
|
|
7835
|
-
:param pulumi.Input[str] log_cleanup_policy: Enum: `delete`, `compact`, `compact,delete`. The default cleanup policy for segments beyond the retention window.
|
|
7836
|
-
:param pulumi.Input[int] log_flush_interval_messages: The number of messages accumulated on a log partition before messages are flushed to disk. Example: `9223372036854775807`.
|
|
7837
|
-
:param pulumi.Input[int] log_flush_interval_ms: The maximum time in ms that a message in any topic is kept in memory before flushed to disk. If not set, the value in log.flush.scheduler.interval.ms is used.
|
|
7838
|
-
:param pulumi.Input[int] log_index_interval_bytes: The interval with which Kafka adds an entry to the offset index. Example: `4096`.
|
|
7839
|
-
:param pulumi.Input[int] log_index_size_max_bytes: The maximum size in bytes of the offset index. Example: `10485760`.
|
|
7840
|
-
:param pulumi.Input[int] log_local_retention_bytes: The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value.
|
|
7841
|
-
:param pulumi.Input[int] log_local_retention_ms: The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value.
|
|
7842
|
-
:param pulumi.Input[bool] log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
7843
|
-
:param pulumi.Input[int] log_message_timestamp_difference_max_ms: The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message.
|
|
7844
|
-
:param pulumi.Input[str] log_message_timestamp_type: Enum: `CreateTime`, `LogAppendTime`. Define whether the timestamp in the message is message create time or log append time.
|
|
7845
|
-
:param pulumi.Input[bool] log_preallocate: Should pre allocate file when create new segment?
|
|
7846
|
-
:param pulumi.Input[int] log_retention_bytes: The maximum size of the log before deleting messages.
|
|
7847
|
-
:param pulumi.Input[int] log_retention_hours: The number of hours to keep a log file before deleting it.
|
|
7848
|
-
:param pulumi.Input[int] log_retention_ms: The number of milliseconds to keep a log file before deleting it (in milliseconds), If not set, the value in log.retention.minutes is used. If set to -1, no time limit is applied.
|
|
7849
|
-
:param pulumi.Input[int] log_roll_jitter_ms: The maximum jitter to subtract from logRollTimeMillis (in milliseconds). If not set, the value in log.roll.jitter.hours is used.
|
|
7850
|
-
:param pulumi.Input[int] log_roll_ms: The maximum time before a new log segment is rolled out (in milliseconds).
|
|
7851
|
-
:param pulumi.Input[int] log_segment_bytes: The maximum size of a single log file.
|
|
7852
|
-
:param pulumi.Input[int] log_segment_delete_delay_ms: The amount of time to wait before deleting a file from the filesystem. Example: `60000`.
|
|
7853
|
-
:param pulumi.Input[int] max_connections_per_ip: The maximum number of connections allowed from each ip address (
|
|
7854
|
-
:param pulumi.Input[int] max_incremental_fetch_session_cache_slots: The maximum number of incremental fetch sessions that the broker will maintain. Example: `1000`.
|
|
7855
|
-
:param pulumi.Input[int] message_max_bytes: The maximum size of message that the server can receive. Example: `1048588`.
|
|
7856
|
-
:param pulumi.Input[int] min_insync_replicas: When a producer sets acks to `all` (or `-1`), min.insync.replicas specifies the minimum number of replicas that must acknowledge a write for the write to be considered successful. Example: `1`.
|
|
7857
|
-
:param pulumi.Input[int] num_partitions: Number of partitions for
|
|
7858
|
-
:param pulumi.Input[int] offsets_retention_minutes: Log retention window in minutes for offsets topic. Example: `10080`.
|
|
7859
|
-
:param pulumi.Input[int] producer_purgatory_purge_interval_requests: The purge interval (in number of requests) of the producer request purgatory(
|
|
7860
|
-
:param pulumi.Input[int] replica_fetch_max_bytes: The number of bytes of messages to attempt to fetch for each partition
|
|
7861
|
-
:param pulumi.Input[int] replica_fetch_response_max_bytes: Maximum bytes expected for the entire fetch response
|
|
7862
|
-
:param pulumi.Input[str] sasl_oauthbearer_expected_audience: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences.
|
|
7863
|
-
:param pulumi.Input[str] sasl_oauthbearer_expected_issuer: Optional setting for the broker to use to verify that the JWT was created by the expected issuer.
|
|
7864
|
-
:param pulumi.Input[str] sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
7865
|
-
:param pulumi.Input[str] sasl_oauthbearer_sub_claim_name: Name of the scope from which to extract the subject claim from the JWT.
|
|
7866
|
-
:param pulumi.Input[int] socket_request_max_bytes: The maximum number of bytes in a socket request (
|
|
7867
|
-
:param pulumi.Input[bool] transaction_partition_verification_enable: Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition.
|
|
7868
|
-
:param pulumi.Input[int] transaction_remove_expired_transaction_cleanup_interval_ms: The interval at which to remove transactions that have expired due to transactional.id.expiration.ms passing (
|
|
7869
|
-
:param pulumi.Input[int] transaction_state_log_segment_bytes: The transaction topic segment bytes should be kept relatively small in order to facilitate faster log compaction and cache loads (
|
|
7917
|
+
:param pulumi.Input[bool] auto_create_topics_enable: Enable auto-creation of topics. (Default: true).
|
|
7918
|
+
:param pulumi.Input[str] compression_type: Enum: `gzip`, `snappy`, `lz4`, `zstd`, `uncompressed`, `producer`. Specify the final compression type for a given topic. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `uncompressed` which is equivalent to no compression; and `producer` which means retain the original compression codec set by the producer.(Default: producer).
|
|
7919
|
+
:param pulumi.Input[int] connections_max_idle_ms: Idle connections timeout: the server socket processor threads close the connections that idle for longer than this. (Default: 600000 ms (10 minutes)). Example: `540000`.
|
|
7920
|
+
:param pulumi.Input[int] default_replication_factor: Replication factor for auto-created topics (Default: 3).
|
|
7921
|
+
:param pulumi.Input[int] group_initial_rebalance_delay_ms: The amount of time, in milliseconds, the group coordinator will wait for more consumers to join a new group before performing the first rebalance. A longer delay means potentially fewer rebalances, but increases the time until processing begins. The default value for this is 3 seconds. During development and testing it might be desirable to set this to 0 in order to not delay test execution time. (Default: 3000 ms (3 seconds)). Example: `3000`.
|
|
7922
|
+
:param pulumi.Input[int] group_max_session_timeout_ms: The maximum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. Default: 1800000 ms (30 minutes). Example: `1800000`.
|
|
7923
|
+
:param pulumi.Input[int] group_min_session_timeout_ms: The minimum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. (Default: 6000 ms (6 seconds)). Example: `6000`.
|
|
7924
|
+
:param pulumi.Input[int] log_cleaner_delete_retention_ms: How long are delete records retained? (Default: 86400000 (1 day)). Example: `86400000`.
|
|
7925
|
+
:param pulumi.Input[int] log_cleaner_max_compaction_lag_ms: The maximum amount of time message will remain uncompacted. Only applicable for logs that are being compacted. (Default: 9223372036854775807 ms (Long.MAX_VALUE)).
|
|
7926
|
+
:param pulumi.Input[float] log_cleaner_min_cleanable_ratio: Controls log compactor frequency. Larger value means more frequent compactions but also more space wasted for logs. Consider setting log.cleaner.max.compaction.lag.ms to enforce compactions sooner, instead of setting a very high value for this option. (Default: 0.5). Example: `0.5`.
|
|
7927
|
+
:param pulumi.Input[int] log_cleaner_min_compaction_lag_ms: The minimum time a message will remain uncompacted in the log. Only applicable for logs that are being compacted. (Default: 0 ms).
|
|
7928
|
+
:param pulumi.Input[str] log_cleanup_policy: Enum: `delete`, `compact`, `compact,delete`. The default cleanup policy for segments beyond the retention window (Default: delete).
|
|
7929
|
+
:param pulumi.Input[int] log_flush_interval_messages: The number of messages accumulated on a log partition before messages are flushed to disk (Default: 9223372036854775807 (Long.MAX_VALUE)). Example: `9223372036854775807`.
|
|
7930
|
+
:param pulumi.Input[int] log_flush_interval_ms: The maximum time in ms that a message in any topic is kept in memory (page-cache) before flushed to disk. If not set, the value in log.flush.scheduler.interval.ms is used (Default: null).
|
|
7931
|
+
:param pulumi.Input[int] log_index_interval_bytes: The interval with which Kafka adds an entry to the offset index (Default: 4096 bytes (4 kibibytes)). Example: `4096`.
|
|
7932
|
+
:param pulumi.Input[int] log_index_size_max_bytes: The maximum size in bytes of the offset index (Default: 10485760 (10 mebibytes)). Example: `10485760`.
|
|
7933
|
+
:param pulumi.Input[int] log_local_retention_bytes: The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value. (Default: -2).
|
|
7934
|
+
:param pulumi.Input[int] log_local_retention_ms: The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value. (Default: -2).
|
|
7935
|
+
:param pulumi.Input[bool] log_message_downconversion_enable: This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests. (Default: true).
|
|
7936
|
+
:param pulumi.Input[int] log_message_timestamp_difference_max_ms: The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message (Default: 9223372036854775807 (Long.MAX_VALUE)).
|
|
7937
|
+
:param pulumi.Input[str] log_message_timestamp_type: Enum: `CreateTime`, `LogAppendTime`. Define whether the timestamp in the message is message create time or log append time. (Default: CreateTime).
|
|
7938
|
+
:param pulumi.Input[bool] log_preallocate: Should pre allocate file when create new segment? (Default: false).
|
|
7939
|
+
:param pulumi.Input[int] log_retention_bytes: The maximum size of the log before deleting messages (Default: -1).
|
|
7940
|
+
:param pulumi.Input[int] log_retention_hours: The number of hours to keep a log file before deleting it (Default: 168 hours (1 week)).
|
|
7941
|
+
:param pulumi.Input[int] log_retention_ms: The number of milliseconds to keep a log file before deleting it (in milliseconds), If not set, the value in log.retention.minutes is used. If set to -1, no time limit is applied. (Default: null, log.retention.hours applies).
|
|
7942
|
+
:param pulumi.Input[int] log_roll_jitter_ms: The maximum jitter to subtract from logRollTimeMillis (in milliseconds). If not set, the value in log.roll.jitter.hours is used (Default: null).
|
|
7943
|
+
:param pulumi.Input[int] log_roll_ms: The maximum time before a new log segment is rolled out (in milliseconds). (Default: null, log.roll.hours applies (Default: 168, 7 days)).
|
|
7944
|
+
:param pulumi.Input[int] log_segment_bytes: The maximum size of a single log file (Default: 1073741824 bytes (1 gibibyte)).
|
|
7945
|
+
:param pulumi.Input[int] log_segment_delete_delay_ms: The amount of time to wait before deleting a file from the filesystem (Default: 60000 ms (1 minute)). Example: `60000`.
|
|
7946
|
+
:param pulumi.Input[int] max_connections_per_ip: The maximum number of connections allowed from each ip address (Default: 2147483647).
|
|
7947
|
+
:param pulumi.Input[int] max_incremental_fetch_session_cache_slots: The maximum number of incremental fetch sessions that the broker will maintain. (Default: 1000). Example: `1000`.
|
|
7948
|
+
:param pulumi.Input[int] message_max_bytes: The maximum size of message that the server can receive. (Default: 1048588 bytes (1 mebibyte + 12 bytes)). Example: `1048588`.
|
|
7949
|
+
:param pulumi.Input[int] min_insync_replicas: When a producer sets acks to `all` (or `-1`), min.insync.replicas specifies the minimum number of replicas that must acknowledge a write for the write to be considered successful. (Default: 1). Example: `1`.
|
|
7950
|
+
:param pulumi.Input[int] num_partitions: Number of partitions for auto-created topics (Default: 1).
|
|
7951
|
+
:param pulumi.Input[int] offsets_retention_minutes: Log retention window in minutes for offsets topic (Default: 10080 minutes (7 days)). Example: `10080`.
|
|
7952
|
+
:param pulumi.Input[int] producer_purgatory_purge_interval_requests: The purge interval (in number of requests) of the producer request purgatory (Default: 1000).
|
|
7953
|
+
:param pulumi.Input[int] replica_fetch_max_bytes: The number of bytes of messages to attempt to fetch for each partition . This is not an absolute maximum, if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. (Default: 1048576 bytes (1 mebibytes)).
|
|
7954
|
+
:param pulumi.Input[int] replica_fetch_response_max_bytes: Maximum bytes expected for the entire fetch response. Records are fetched in batches, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. As such, this is not an absolute maximum. (Default: 10485760 bytes (10 mebibytes)).
|
|
7955
|
+
:param pulumi.Input[str] sasl_oauthbearer_expected_audience: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences. (Default: null).
|
|
7956
|
+
:param pulumi.Input[str] sasl_oauthbearer_expected_issuer: Optional setting for the broker to use to verify that the JWT was created by the expected issuer.(Default: null).
|
|
7957
|
+
:param pulumi.Input[str] sasl_oauthbearer_jwks_endpoint_url: OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC. (Default: null).
|
|
7958
|
+
:param pulumi.Input[str] sasl_oauthbearer_sub_claim_name: Name of the scope from which to extract the subject claim from the JWT.(Default: sub).
|
|
7959
|
+
:param pulumi.Input[int] socket_request_max_bytes: The maximum number of bytes in a socket request (Default: 104857600 bytes).
|
|
7960
|
+
:param pulumi.Input[bool] transaction_partition_verification_enable: Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition. (Default: true).
|
|
7961
|
+
:param pulumi.Input[int] transaction_remove_expired_transaction_cleanup_interval_ms: The interval at which to remove transactions that have expired due to transactional.id.expiration.ms passing (Default: 3600000 ms (1 hour)). Example: `3600000`.
|
|
7962
|
+
:param pulumi.Input[int] transaction_state_log_segment_bytes: The transaction topic segment bytes should be kept relatively small in order to facilitate faster log compaction and cache loads (Default: 104857600 bytes (100 mebibytes)). Example: `104857600`.
|
|
7870
7963
|
"""
|
|
7871
7964
|
if auto_create_topics_enable is not None:
|
|
7872
7965
|
pulumi.set(__self__, "auto_create_topics_enable", auto_create_topics_enable)
|
|
@@ -7965,7 +8058,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
7965
8058
|
@pulumi.getter(name="autoCreateTopicsEnable")
|
|
7966
8059
|
def auto_create_topics_enable(self) -> Optional[pulumi.Input[bool]]:
|
|
7967
8060
|
"""
|
|
7968
|
-
Enable auto
|
|
8061
|
+
Enable auto-creation of topics. (Default: true).
|
|
7969
8062
|
"""
|
|
7970
8063
|
return pulumi.get(self, "auto_create_topics_enable")
|
|
7971
8064
|
|
|
@@ -7977,7 +8070,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
7977
8070
|
@pulumi.getter(name="compressionType")
|
|
7978
8071
|
def compression_type(self) -> Optional[pulumi.Input[str]]:
|
|
7979
8072
|
"""
|
|
7980
|
-
Enum: `gzip`, `snappy`, `lz4`, `zstd`, `uncompressed`, `producer`. Specify the final compression type for a given topic. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `uncompressed` which is equivalent to no compression; and `producer` which means retain the original compression codec set by the producer.
|
|
8073
|
+
Enum: `gzip`, `snappy`, `lz4`, `zstd`, `uncompressed`, `producer`. Specify the final compression type for a given topic. This configuration accepts the standard compression codecs (`gzip`, `snappy`, `lz4`, `zstd`). It additionally accepts `uncompressed` which is equivalent to no compression; and `producer` which means retain the original compression codec set by the producer.(Default: producer).
|
|
7981
8074
|
"""
|
|
7982
8075
|
return pulumi.get(self, "compression_type")
|
|
7983
8076
|
|
|
@@ -7989,7 +8082,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
7989
8082
|
@pulumi.getter(name="connectionsMaxIdleMs")
|
|
7990
8083
|
def connections_max_idle_ms(self) -> Optional[pulumi.Input[int]]:
|
|
7991
8084
|
"""
|
|
7992
|
-
Idle connections timeout: the server socket processor threads close the connections that idle for longer than this. Example: `540000`.
|
|
8085
|
+
Idle connections timeout: the server socket processor threads close the connections that idle for longer than this. (Default: 600000 ms (10 minutes)). Example: `540000`.
|
|
7993
8086
|
"""
|
|
7994
8087
|
return pulumi.get(self, "connections_max_idle_ms")
|
|
7995
8088
|
|
|
@@ -8001,7 +8094,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8001
8094
|
@pulumi.getter(name="defaultReplicationFactor")
|
|
8002
8095
|
def default_replication_factor(self) -> Optional[pulumi.Input[int]]:
|
|
8003
8096
|
"""
|
|
8004
|
-
Replication factor for
|
|
8097
|
+
Replication factor for auto-created topics (Default: 3).
|
|
8005
8098
|
"""
|
|
8006
8099
|
return pulumi.get(self, "default_replication_factor")
|
|
8007
8100
|
|
|
@@ -8013,7 +8106,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8013
8106
|
@pulumi.getter(name="groupInitialRebalanceDelayMs")
|
|
8014
8107
|
def group_initial_rebalance_delay_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8015
8108
|
"""
|
|
8016
|
-
The amount of time, in milliseconds, the group coordinator will wait for more consumers to join a new group before performing the first rebalance. A longer delay means potentially fewer rebalances, but increases the time until processing begins. The default value for this is 3 seconds. During development and testing it might be desirable to set this to 0 in order to not delay test execution time. Example: `3000`.
|
|
8109
|
+
The amount of time, in milliseconds, the group coordinator will wait for more consumers to join a new group before performing the first rebalance. A longer delay means potentially fewer rebalances, but increases the time until processing begins. The default value for this is 3 seconds. During development and testing it might be desirable to set this to 0 in order to not delay test execution time. (Default: 3000 ms (3 seconds)). Example: `3000`.
|
|
8017
8110
|
"""
|
|
8018
8111
|
return pulumi.get(self, "group_initial_rebalance_delay_ms")
|
|
8019
8112
|
|
|
@@ -8025,7 +8118,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8025
8118
|
@pulumi.getter(name="groupMaxSessionTimeoutMs")
|
|
8026
8119
|
def group_max_session_timeout_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8027
8120
|
"""
|
|
8028
|
-
The maximum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. Example: `1800000`.
|
|
8121
|
+
The maximum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. Default: 1800000 ms (30 minutes). Example: `1800000`.
|
|
8029
8122
|
"""
|
|
8030
8123
|
return pulumi.get(self, "group_max_session_timeout_ms")
|
|
8031
8124
|
|
|
@@ -8037,7 +8130,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8037
8130
|
@pulumi.getter(name="groupMinSessionTimeoutMs")
|
|
8038
8131
|
def group_min_session_timeout_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8039
8132
|
"""
|
|
8040
|
-
The minimum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. Example: `6000`.
|
|
8133
|
+
The minimum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. (Default: 6000 ms (6 seconds)). Example: `6000`.
|
|
8041
8134
|
"""
|
|
8042
8135
|
return pulumi.get(self, "group_min_session_timeout_ms")
|
|
8043
8136
|
|
|
@@ -8049,7 +8142,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8049
8142
|
@pulumi.getter(name="logCleanerDeleteRetentionMs")
|
|
8050
8143
|
def log_cleaner_delete_retention_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8051
8144
|
"""
|
|
8052
|
-
How long are delete records retained? Example: `86400000`.
|
|
8145
|
+
How long are delete records retained? (Default: 86400000 (1 day)). Example: `86400000`.
|
|
8053
8146
|
"""
|
|
8054
8147
|
return pulumi.get(self, "log_cleaner_delete_retention_ms")
|
|
8055
8148
|
|
|
@@ -8061,7 +8154,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8061
8154
|
@pulumi.getter(name="logCleanerMaxCompactionLagMs")
|
|
8062
8155
|
def log_cleaner_max_compaction_lag_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8063
8156
|
"""
|
|
8064
|
-
The maximum amount of time message will remain uncompacted. Only applicable for logs that are being compacted.
|
|
8157
|
+
The maximum amount of time message will remain uncompacted. Only applicable for logs that are being compacted. (Default: 9223372036854775807 ms (Long.MAX_VALUE)).
|
|
8065
8158
|
"""
|
|
8066
8159
|
return pulumi.get(self, "log_cleaner_max_compaction_lag_ms")
|
|
8067
8160
|
|
|
@@ -8073,7 +8166,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8073
8166
|
@pulumi.getter(name="logCleanerMinCleanableRatio")
|
|
8074
8167
|
def log_cleaner_min_cleanable_ratio(self) -> Optional[pulumi.Input[float]]:
|
|
8075
8168
|
"""
|
|
8076
|
-
Controls log compactor frequency. Larger value means more frequent compactions but also more space wasted for logs. Consider setting log.cleaner.max.compaction.lag.ms to enforce compactions sooner, instead of setting a very high value for this option. Example: `0.5`.
|
|
8169
|
+
Controls log compactor frequency. Larger value means more frequent compactions but also more space wasted for logs. Consider setting log.cleaner.max.compaction.lag.ms to enforce compactions sooner, instead of setting a very high value for this option. (Default: 0.5). Example: `0.5`.
|
|
8077
8170
|
"""
|
|
8078
8171
|
return pulumi.get(self, "log_cleaner_min_cleanable_ratio")
|
|
8079
8172
|
|
|
@@ -8085,7 +8178,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8085
8178
|
@pulumi.getter(name="logCleanerMinCompactionLagMs")
|
|
8086
8179
|
def log_cleaner_min_compaction_lag_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8087
8180
|
"""
|
|
8088
|
-
The minimum time a message will remain uncompacted in the log. Only applicable for logs that are being compacted.
|
|
8181
|
+
The minimum time a message will remain uncompacted in the log. Only applicable for logs that are being compacted. (Default: 0 ms).
|
|
8089
8182
|
"""
|
|
8090
8183
|
return pulumi.get(self, "log_cleaner_min_compaction_lag_ms")
|
|
8091
8184
|
|
|
@@ -8097,7 +8190,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8097
8190
|
@pulumi.getter(name="logCleanupPolicy")
|
|
8098
8191
|
def log_cleanup_policy(self) -> Optional[pulumi.Input[str]]:
|
|
8099
8192
|
"""
|
|
8100
|
-
Enum: `delete`, `compact`, `compact,delete`. The default cleanup policy for segments beyond the retention window.
|
|
8193
|
+
Enum: `delete`, `compact`, `compact,delete`. The default cleanup policy for segments beyond the retention window (Default: delete).
|
|
8101
8194
|
"""
|
|
8102
8195
|
return pulumi.get(self, "log_cleanup_policy")
|
|
8103
8196
|
|
|
@@ -8109,7 +8202,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8109
8202
|
@pulumi.getter(name="logFlushIntervalMessages")
|
|
8110
8203
|
def log_flush_interval_messages(self) -> Optional[pulumi.Input[int]]:
|
|
8111
8204
|
"""
|
|
8112
|
-
The number of messages accumulated on a log partition before messages are flushed to disk. Example: `9223372036854775807`.
|
|
8205
|
+
The number of messages accumulated on a log partition before messages are flushed to disk (Default: 9223372036854775807 (Long.MAX_VALUE)). Example: `9223372036854775807`.
|
|
8113
8206
|
"""
|
|
8114
8207
|
return pulumi.get(self, "log_flush_interval_messages")
|
|
8115
8208
|
|
|
@@ -8121,7 +8214,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8121
8214
|
@pulumi.getter(name="logFlushIntervalMs")
|
|
8122
8215
|
def log_flush_interval_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8123
8216
|
"""
|
|
8124
|
-
The maximum time in ms that a message in any topic is kept in memory before flushed to disk. If not set, the value in log.flush.scheduler.interval.ms is used.
|
|
8217
|
+
The maximum time in ms that a message in any topic is kept in memory (page-cache) before flushed to disk. If not set, the value in log.flush.scheduler.interval.ms is used (Default: null).
|
|
8125
8218
|
"""
|
|
8126
8219
|
return pulumi.get(self, "log_flush_interval_ms")
|
|
8127
8220
|
|
|
@@ -8133,7 +8226,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8133
8226
|
@pulumi.getter(name="logIndexIntervalBytes")
|
|
8134
8227
|
def log_index_interval_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8135
8228
|
"""
|
|
8136
|
-
The interval with which Kafka adds an entry to the offset index. Example: `4096`.
|
|
8229
|
+
The interval with which Kafka adds an entry to the offset index (Default: 4096 bytes (4 kibibytes)). Example: `4096`.
|
|
8137
8230
|
"""
|
|
8138
8231
|
return pulumi.get(self, "log_index_interval_bytes")
|
|
8139
8232
|
|
|
@@ -8145,7 +8238,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8145
8238
|
@pulumi.getter(name="logIndexSizeMaxBytes")
|
|
8146
8239
|
def log_index_size_max_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8147
8240
|
"""
|
|
8148
|
-
The maximum size in bytes of the offset index. Example: `10485760`.
|
|
8241
|
+
The maximum size in bytes of the offset index (Default: 10485760 (10 mebibytes)). Example: `10485760`.
|
|
8149
8242
|
"""
|
|
8150
8243
|
return pulumi.get(self, "log_index_size_max_bytes")
|
|
8151
8244
|
|
|
@@ -8157,7 +8250,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8157
8250
|
@pulumi.getter(name="logLocalRetentionBytes")
|
|
8158
8251
|
def log_local_retention_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8159
8252
|
"""
|
|
8160
|
-
The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value.
|
|
8253
|
+
The maximum size of local log segments that can grow for a partition before it gets eligible for deletion. If set to -2, the value of log.retention.bytes is used. The effective value should always be less than or equal to log.retention.bytes value. (Default: -2).
|
|
8161
8254
|
"""
|
|
8162
8255
|
return pulumi.get(self, "log_local_retention_bytes")
|
|
8163
8256
|
|
|
@@ -8169,7 +8262,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8169
8262
|
@pulumi.getter(name="logLocalRetentionMs")
|
|
8170
8263
|
def log_local_retention_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8171
8264
|
"""
|
|
8172
|
-
The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value.
|
|
8265
|
+
The number of milliseconds to keep the local log segments before it gets eligible for deletion. If set to -2, the value of log.retention.ms is used. The effective value should always be less than or equal to log.retention.ms value. (Default: -2).
|
|
8173
8266
|
"""
|
|
8174
8267
|
return pulumi.get(self, "log_local_retention_ms")
|
|
8175
8268
|
|
|
@@ -8181,7 +8274,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8181
8274
|
@pulumi.getter(name="logMessageDownconversionEnable")
|
|
8182
8275
|
def log_message_downconversion_enable(self) -> Optional[pulumi.Input[bool]]:
|
|
8183
8276
|
"""
|
|
8184
|
-
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests.
|
|
8277
|
+
This configuration controls whether down-conversion of message formats is enabled to satisfy consume requests. (Default: true).
|
|
8185
8278
|
"""
|
|
8186
8279
|
return pulumi.get(self, "log_message_downconversion_enable")
|
|
8187
8280
|
|
|
@@ -8193,7 +8286,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8193
8286
|
@pulumi.getter(name="logMessageTimestampDifferenceMaxMs")
|
|
8194
8287
|
def log_message_timestamp_difference_max_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8195
8288
|
"""
|
|
8196
|
-
The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message.
|
|
8289
|
+
The maximum difference allowed between the timestamp when a broker receives a message and the timestamp specified in the message (Default: 9223372036854775807 (Long.MAX_VALUE)).
|
|
8197
8290
|
"""
|
|
8198
8291
|
return pulumi.get(self, "log_message_timestamp_difference_max_ms")
|
|
8199
8292
|
|
|
@@ -8205,7 +8298,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8205
8298
|
@pulumi.getter(name="logMessageTimestampType")
|
|
8206
8299
|
def log_message_timestamp_type(self) -> Optional[pulumi.Input[str]]:
|
|
8207
8300
|
"""
|
|
8208
|
-
Enum: `CreateTime`, `LogAppendTime`. Define whether the timestamp in the message is message create time or log append time.
|
|
8301
|
+
Enum: `CreateTime`, `LogAppendTime`. Define whether the timestamp in the message is message create time or log append time. (Default: CreateTime).
|
|
8209
8302
|
"""
|
|
8210
8303
|
return pulumi.get(self, "log_message_timestamp_type")
|
|
8211
8304
|
|
|
@@ -8217,7 +8310,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8217
8310
|
@pulumi.getter(name="logPreallocate")
|
|
8218
8311
|
def log_preallocate(self) -> Optional[pulumi.Input[bool]]:
|
|
8219
8312
|
"""
|
|
8220
|
-
Should pre allocate file when create new segment?
|
|
8313
|
+
Should pre allocate file when create new segment? (Default: false).
|
|
8221
8314
|
"""
|
|
8222
8315
|
return pulumi.get(self, "log_preallocate")
|
|
8223
8316
|
|
|
@@ -8229,7 +8322,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8229
8322
|
@pulumi.getter(name="logRetentionBytes")
|
|
8230
8323
|
def log_retention_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8231
8324
|
"""
|
|
8232
|
-
The maximum size of the log before deleting messages.
|
|
8325
|
+
The maximum size of the log before deleting messages (Default: -1).
|
|
8233
8326
|
"""
|
|
8234
8327
|
return pulumi.get(self, "log_retention_bytes")
|
|
8235
8328
|
|
|
@@ -8241,7 +8334,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8241
8334
|
@pulumi.getter(name="logRetentionHours")
|
|
8242
8335
|
def log_retention_hours(self) -> Optional[pulumi.Input[int]]:
|
|
8243
8336
|
"""
|
|
8244
|
-
The number of hours to keep a log file before deleting it.
|
|
8337
|
+
The number of hours to keep a log file before deleting it (Default: 168 hours (1 week)).
|
|
8245
8338
|
"""
|
|
8246
8339
|
return pulumi.get(self, "log_retention_hours")
|
|
8247
8340
|
|
|
@@ -8253,7 +8346,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8253
8346
|
@pulumi.getter(name="logRetentionMs")
|
|
8254
8347
|
def log_retention_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8255
8348
|
"""
|
|
8256
|
-
The number of milliseconds to keep a log file before deleting it (in milliseconds), If not set, the value in log.retention.minutes is used. If set to -1, no time limit is applied.
|
|
8349
|
+
The number of milliseconds to keep a log file before deleting it (in milliseconds), If not set, the value in log.retention.minutes is used. If set to -1, no time limit is applied. (Default: null, log.retention.hours applies).
|
|
8257
8350
|
"""
|
|
8258
8351
|
return pulumi.get(self, "log_retention_ms")
|
|
8259
8352
|
|
|
@@ -8265,7 +8358,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8265
8358
|
@pulumi.getter(name="logRollJitterMs")
|
|
8266
8359
|
def log_roll_jitter_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8267
8360
|
"""
|
|
8268
|
-
The maximum jitter to subtract from logRollTimeMillis (in milliseconds). If not set, the value in log.roll.jitter.hours is used.
|
|
8361
|
+
The maximum jitter to subtract from logRollTimeMillis (in milliseconds). If not set, the value in log.roll.jitter.hours is used (Default: null).
|
|
8269
8362
|
"""
|
|
8270
8363
|
return pulumi.get(self, "log_roll_jitter_ms")
|
|
8271
8364
|
|
|
@@ -8277,7 +8370,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8277
8370
|
@pulumi.getter(name="logRollMs")
|
|
8278
8371
|
def log_roll_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8279
8372
|
"""
|
|
8280
|
-
The maximum time before a new log segment is rolled out (in milliseconds).
|
|
8373
|
+
The maximum time before a new log segment is rolled out (in milliseconds). (Default: null, log.roll.hours applies (Default: 168, 7 days)).
|
|
8281
8374
|
"""
|
|
8282
8375
|
return pulumi.get(self, "log_roll_ms")
|
|
8283
8376
|
|
|
@@ -8289,7 +8382,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8289
8382
|
@pulumi.getter(name="logSegmentBytes")
|
|
8290
8383
|
def log_segment_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8291
8384
|
"""
|
|
8292
|
-
The maximum size of a single log file.
|
|
8385
|
+
The maximum size of a single log file (Default: 1073741824 bytes (1 gibibyte)).
|
|
8293
8386
|
"""
|
|
8294
8387
|
return pulumi.get(self, "log_segment_bytes")
|
|
8295
8388
|
|
|
@@ -8301,7 +8394,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8301
8394
|
@pulumi.getter(name="logSegmentDeleteDelayMs")
|
|
8302
8395
|
def log_segment_delete_delay_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8303
8396
|
"""
|
|
8304
|
-
The amount of time to wait before deleting a file from the filesystem. Example: `60000`.
|
|
8397
|
+
The amount of time to wait before deleting a file from the filesystem (Default: 60000 ms (1 minute)). Example: `60000`.
|
|
8305
8398
|
"""
|
|
8306
8399
|
return pulumi.get(self, "log_segment_delete_delay_ms")
|
|
8307
8400
|
|
|
@@ -8313,7 +8406,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8313
8406
|
@pulumi.getter(name="maxConnectionsPerIp")
|
|
8314
8407
|
def max_connections_per_ip(self) -> Optional[pulumi.Input[int]]:
|
|
8315
8408
|
"""
|
|
8316
|
-
The maximum number of connections allowed from each ip address (
|
|
8409
|
+
The maximum number of connections allowed from each ip address (Default: 2147483647).
|
|
8317
8410
|
"""
|
|
8318
8411
|
return pulumi.get(self, "max_connections_per_ip")
|
|
8319
8412
|
|
|
@@ -8325,7 +8418,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8325
8418
|
@pulumi.getter(name="maxIncrementalFetchSessionCacheSlots")
|
|
8326
8419
|
def max_incremental_fetch_session_cache_slots(self) -> Optional[pulumi.Input[int]]:
|
|
8327
8420
|
"""
|
|
8328
|
-
The maximum number of incremental fetch sessions that the broker will maintain. Example: `1000`.
|
|
8421
|
+
The maximum number of incremental fetch sessions that the broker will maintain. (Default: 1000). Example: `1000`.
|
|
8329
8422
|
"""
|
|
8330
8423
|
return pulumi.get(self, "max_incremental_fetch_session_cache_slots")
|
|
8331
8424
|
|
|
@@ -8337,7 +8430,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8337
8430
|
@pulumi.getter(name="messageMaxBytes")
|
|
8338
8431
|
def message_max_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8339
8432
|
"""
|
|
8340
|
-
The maximum size of message that the server can receive. Example: `1048588`.
|
|
8433
|
+
The maximum size of message that the server can receive. (Default: 1048588 bytes (1 mebibyte + 12 bytes)). Example: `1048588`.
|
|
8341
8434
|
"""
|
|
8342
8435
|
return pulumi.get(self, "message_max_bytes")
|
|
8343
8436
|
|
|
@@ -8349,7 +8442,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8349
8442
|
@pulumi.getter(name="minInsyncReplicas")
|
|
8350
8443
|
def min_insync_replicas(self) -> Optional[pulumi.Input[int]]:
|
|
8351
8444
|
"""
|
|
8352
|
-
When a producer sets acks to `all` (or `-1`), min.insync.replicas specifies the minimum number of replicas that must acknowledge a write for the write to be considered successful. Example: `1`.
|
|
8445
|
+
When a producer sets acks to `all` (or `-1`), min.insync.replicas specifies the minimum number of replicas that must acknowledge a write for the write to be considered successful. (Default: 1). Example: `1`.
|
|
8353
8446
|
"""
|
|
8354
8447
|
return pulumi.get(self, "min_insync_replicas")
|
|
8355
8448
|
|
|
@@ -8361,7 +8454,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8361
8454
|
@pulumi.getter(name="numPartitions")
|
|
8362
8455
|
def num_partitions(self) -> Optional[pulumi.Input[int]]:
|
|
8363
8456
|
"""
|
|
8364
|
-
Number of partitions for
|
|
8457
|
+
Number of partitions for auto-created topics (Default: 1).
|
|
8365
8458
|
"""
|
|
8366
8459
|
return pulumi.get(self, "num_partitions")
|
|
8367
8460
|
|
|
@@ -8373,7 +8466,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8373
8466
|
@pulumi.getter(name="offsetsRetentionMinutes")
|
|
8374
8467
|
def offsets_retention_minutes(self) -> Optional[pulumi.Input[int]]:
|
|
8375
8468
|
"""
|
|
8376
|
-
Log retention window in minutes for offsets topic. Example: `10080`.
|
|
8469
|
+
Log retention window in minutes for offsets topic (Default: 10080 minutes (7 days)). Example: `10080`.
|
|
8377
8470
|
"""
|
|
8378
8471
|
return pulumi.get(self, "offsets_retention_minutes")
|
|
8379
8472
|
|
|
@@ -8385,7 +8478,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8385
8478
|
@pulumi.getter(name="producerPurgatoryPurgeIntervalRequests")
|
|
8386
8479
|
def producer_purgatory_purge_interval_requests(self) -> Optional[pulumi.Input[int]]:
|
|
8387
8480
|
"""
|
|
8388
|
-
The purge interval (in number of requests) of the producer request purgatory(
|
|
8481
|
+
The purge interval (in number of requests) of the producer request purgatory (Default: 1000).
|
|
8389
8482
|
"""
|
|
8390
8483
|
return pulumi.get(self, "producer_purgatory_purge_interval_requests")
|
|
8391
8484
|
|
|
@@ -8397,7 +8490,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8397
8490
|
@pulumi.getter(name="replicaFetchMaxBytes")
|
|
8398
8491
|
def replica_fetch_max_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8399
8492
|
"""
|
|
8400
|
-
The number of bytes of messages to attempt to fetch for each partition
|
|
8493
|
+
The number of bytes of messages to attempt to fetch for each partition . This is not an absolute maximum, if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. (Default: 1048576 bytes (1 mebibytes)).
|
|
8401
8494
|
"""
|
|
8402
8495
|
return pulumi.get(self, "replica_fetch_max_bytes")
|
|
8403
8496
|
|
|
@@ -8409,7 +8502,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8409
8502
|
@pulumi.getter(name="replicaFetchResponseMaxBytes")
|
|
8410
8503
|
def replica_fetch_response_max_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8411
8504
|
"""
|
|
8412
|
-
Maximum bytes expected for the entire fetch response
|
|
8505
|
+
Maximum bytes expected for the entire fetch response. Records are fetched in batches, and if the first record batch in the first non-empty partition of the fetch is larger than this value, the record batch will still be returned to ensure that progress can be made. As such, this is not an absolute maximum. (Default: 10485760 bytes (10 mebibytes)).
|
|
8413
8506
|
"""
|
|
8414
8507
|
return pulumi.get(self, "replica_fetch_response_max_bytes")
|
|
8415
8508
|
|
|
@@ -8421,7 +8514,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8421
8514
|
@pulumi.getter(name="saslOauthbearerExpectedAudience")
|
|
8422
8515
|
def sasl_oauthbearer_expected_audience(self) -> Optional[pulumi.Input[str]]:
|
|
8423
8516
|
"""
|
|
8424
|
-
The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences.
|
|
8517
|
+
The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences. (Default: null).
|
|
8425
8518
|
"""
|
|
8426
8519
|
return pulumi.get(self, "sasl_oauthbearer_expected_audience")
|
|
8427
8520
|
|
|
@@ -8433,7 +8526,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8433
8526
|
@pulumi.getter(name="saslOauthbearerExpectedIssuer")
|
|
8434
8527
|
def sasl_oauthbearer_expected_issuer(self) -> Optional[pulumi.Input[str]]:
|
|
8435
8528
|
"""
|
|
8436
|
-
Optional setting for the broker to use to verify that the JWT was created by the expected issuer.
|
|
8529
|
+
Optional setting for the broker to use to verify that the JWT was created by the expected issuer.(Default: null).
|
|
8437
8530
|
"""
|
|
8438
8531
|
return pulumi.get(self, "sasl_oauthbearer_expected_issuer")
|
|
8439
8532
|
|
|
@@ -8445,7 +8538,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8445
8538
|
@pulumi.getter(name="saslOauthbearerJwksEndpointUrl")
|
|
8446
8539
|
def sasl_oauthbearer_jwks_endpoint_url(self) -> Optional[pulumi.Input[str]]:
|
|
8447
8540
|
"""
|
|
8448
|
-
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC.
|
|
8541
|
+
OIDC JWKS endpoint URL. By setting this the SASL SSL OAuth2/OIDC authentication is enabled. See also other options for SASL OAuth2/OIDC. (Default: null).
|
|
8449
8542
|
"""
|
|
8450
8543
|
return pulumi.get(self, "sasl_oauthbearer_jwks_endpoint_url")
|
|
8451
8544
|
|
|
@@ -8457,7 +8550,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8457
8550
|
@pulumi.getter(name="saslOauthbearerSubClaimName")
|
|
8458
8551
|
def sasl_oauthbearer_sub_claim_name(self) -> Optional[pulumi.Input[str]]:
|
|
8459
8552
|
"""
|
|
8460
|
-
Name of the scope from which to extract the subject claim from the JWT.
|
|
8553
|
+
Name of the scope from which to extract the subject claim from the JWT.(Default: sub).
|
|
8461
8554
|
"""
|
|
8462
8555
|
return pulumi.get(self, "sasl_oauthbearer_sub_claim_name")
|
|
8463
8556
|
|
|
@@ -8469,7 +8562,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8469
8562
|
@pulumi.getter(name="socketRequestMaxBytes")
|
|
8470
8563
|
def socket_request_max_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8471
8564
|
"""
|
|
8472
|
-
The maximum number of bytes in a socket request (
|
|
8565
|
+
The maximum number of bytes in a socket request (Default: 104857600 bytes).
|
|
8473
8566
|
"""
|
|
8474
8567
|
return pulumi.get(self, "socket_request_max_bytes")
|
|
8475
8568
|
|
|
@@ -8481,7 +8574,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8481
8574
|
@pulumi.getter(name="transactionPartitionVerificationEnable")
|
|
8482
8575
|
def transaction_partition_verification_enable(self) -> Optional[pulumi.Input[bool]]:
|
|
8483
8576
|
"""
|
|
8484
|
-
Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition.
|
|
8577
|
+
Enable verification that checks that the partition has been added to the transaction before writing transactional records to the partition. (Default: true).
|
|
8485
8578
|
"""
|
|
8486
8579
|
return pulumi.get(self, "transaction_partition_verification_enable")
|
|
8487
8580
|
|
|
@@ -8493,7 +8586,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8493
8586
|
@pulumi.getter(name="transactionRemoveExpiredTransactionCleanupIntervalMs")
|
|
8494
8587
|
def transaction_remove_expired_transaction_cleanup_interval_ms(self) -> Optional[pulumi.Input[int]]:
|
|
8495
8588
|
"""
|
|
8496
|
-
The interval at which to remove transactions that have expired due to transactional.id.expiration.ms passing (
|
|
8589
|
+
The interval at which to remove transactions that have expired due to transactional.id.expiration.ms passing (Default: 3600000 ms (1 hour)). Example: `3600000`.
|
|
8497
8590
|
"""
|
|
8498
8591
|
return pulumi.get(self, "transaction_remove_expired_transaction_cleanup_interval_ms")
|
|
8499
8592
|
|
|
@@ -8505,7 +8598,7 @@ class KafkaKafkaUserConfigKafkaArgs:
|
|
|
8505
8598
|
@pulumi.getter(name="transactionStateLogSegmentBytes")
|
|
8506
8599
|
def transaction_state_log_segment_bytes(self) -> Optional[pulumi.Input[int]]:
|
|
8507
8600
|
"""
|
|
8508
|
-
The transaction topic segment bytes should be kept relatively small in order to facilitate faster log compaction and cache loads (
|
|
8601
|
+
The transaction topic segment bytes should be kept relatively small in order to facilitate faster log compaction and cache loads (Default: 104857600 bytes (100 mebibytes)). Example: `104857600`.
|
|
8509
8602
|
"""
|
|
8510
8603
|
return pulumi.get(self, "transaction_state_log_segment_bytes")
|
|
8511
8604
|
|
|
@@ -9175,6 +9268,61 @@ class KafkaKafkaUserConfigKafkaRestConfigArgs:
|
|
|
9175
9268
|
pulumi.set(self, "simpleconsumer_pool_size_max", value)
|
|
9176
9269
|
|
|
9177
9270
|
|
|
9271
|
+
@pulumi.input_type
|
|
9272
|
+
class KafkaKafkaUserConfigKafkaSaslMechanismsArgs:
|
|
9273
|
+
def __init__(__self__, *,
|
|
9274
|
+
plain: Optional[pulumi.Input[bool]] = None,
|
|
9275
|
+
scram_sha256: Optional[pulumi.Input[bool]] = None,
|
|
9276
|
+
scram_sha512: Optional[pulumi.Input[bool]] = None):
|
|
9277
|
+
"""
|
|
9278
|
+
:param pulumi.Input[bool] plain: Enable PLAIN mechanism. Default: `true`.
|
|
9279
|
+
:param pulumi.Input[bool] scram_sha256: Enable SCRAM-SHA-256 mechanism. Default: `true`.
|
|
9280
|
+
:param pulumi.Input[bool] scram_sha512: Enable SCRAM-SHA-512 mechanism. Default: `true`.
|
|
9281
|
+
"""
|
|
9282
|
+
if plain is not None:
|
|
9283
|
+
pulumi.set(__self__, "plain", plain)
|
|
9284
|
+
if scram_sha256 is not None:
|
|
9285
|
+
pulumi.set(__self__, "scram_sha256", scram_sha256)
|
|
9286
|
+
if scram_sha512 is not None:
|
|
9287
|
+
pulumi.set(__self__, "scram_sha512", scram_sha512)
|
|
9288
|
+
|
|
9289
|
+
@property
|
|
9290
|
+
@pulumi.getter
|
|
9291
|
+
def plain(self) -> Optional[pulumi.Input[bool]]:
|
|
9292
|
+
"""
|
|
9293
|
+
Enable PLAIN mechanism. Default: `true`.
|
|
9294
|
+
"""
|
|
9295
|
+
return pulumi.get(self, "plain")
|
|
9296
|
+
|
|
9297
|
+
@plain.setter
|
|
9298
|
+
def plain(self, value: Optional[pulumi.Input[bool]]):
|
|
9299
|
+
pulumi.set(self, "plain", value)
|
|
9300
|
+
|
|
9301
|
+
@property
|
|
9302
|
+
@pulumi.getter(name="scramSha256")
|
|
9303
|
+
def scram_sha256(self) -> Optional[pulumi.Input[bool]]:
|
|
9304
|
+
"""
|
|
9305
|
+
Enable SCRAM-SHA-256 mechanism. Default: `true`.
|
|
9306
|
+
"""
|
|
9307
|
+
return pulumi.get(self, "scram_sha256")
|
|
9308
|
+
|
|
9309
|
+
@scram_sha256.setter
|
|
9310
|
+
def scram_sha256(self, value: Optional[pulumi.Input[bool]]):
|
|
9311
|
+
pulumi.set(self, "scram_sha256", value)
|
|
9312
|
+
|
|
9313
|
+
@property
|
|
9314
|
+
@pulumi.getter(name="scramSha512")
|
|
9315
|
+
def scram_sha512(self) -> Optional[pulumi.Input[bool]]:
|
|
9316
|
+
"""
|
|
9317
|
+
Enable SCRAM-SHA-512 mechanism. Default: `true`.
|
|
9318
|
+
"""
|
|
9319
|
+
return pulumi.get(self, "scram_sha512")
|
|
9320
|
+
|
|
9321
|
+
@scram_sha512.setter
|
|
9322
|
+
def scram_sha512(self, value: Optional[pulumi.Input[bool]]):
|
|
9323
|
+
pulumi.set(self, "scram_sha512", value)
|
|
9324
|
+
|
|
9325
|
+
|
|
9178
9326
|
@pulumi.input_type
|
|
9179
9327
|
class KafkaKafkaUserConfigPrivateAccessArgs:
|
|
9180
9328
|
def __init__(__self__, *,
|
|
@@ -9864,6 +10012,7 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigIpFilterObjectArgs:
|
|
|
9864
10012
|
@pulumi.input_type
|
|
9865
10013
|
class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormakerArgs:
|
|
9866
10014
|
def __init__(__self__, *,
|
|
10015
|
+
admin_timeout_ms: Optional[pulumi.Input[int]] = None,
|
|
9867
10016
|
emit_checkpoints_enabled: Optional[pulumi.Input[bool]] = None,
|
|
9868
10017
|
emit_checkpoints_interval_seconds: Optional[pulumi.Input[int]] = None,
|
|
9869
10018
|
groups: Optional[pulumi.Input[str]] = None,
|
|
@@ -9878,6 +10027,7 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormakerArgs:
|
|
|
9878
10027
|
sync_topic_configs_enabled: Optional[pulumi.Input[bool]] = None,
|
|
9879
10028
|
tasks_max_per_cpu: Optional[pulumi.Input[int]] = None):
|
|
9880
10029
|
"""
|
|
10030
|
+
:param pulumi.Input[int] admin_timeout_ms: Timeout for administrative tasks, e.g. detecting new topics, loading of consumer group and offsets. Defaults to 60000 milliseconds (1 minute).
|
|
9881
10031
|
:param pulumi.Input[bool] emit_checkpoints_enabled: Whether to emit consumer group offset checkpoints to target cluster periodically (default: true).
|
|
9882
10032
|
:param pulumi.Input[int] emit_checkpoints_interval_seconds: Frequency at which consumer group offset checkpoints are emitted (default: 60, every minute). Example: `60`.
|
|
9883
10033
|
:param pulumi.Input[str] groups: Consumer groups to replicate. Supports comma-separated group IDs and regexes. Example: `.*`.
|
|
@@ -9892,6 +10042,8 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormakerArgs:
|
|
|
9892
10042
|
:param pulumi.Input[bool] sync_topic_configs_enabled: Whether to periodically configure remote topics to match their corresponding upstream topics.
|
|
9893
10043
|
:param pulumi.Input[int] tasks_max_per_cpu: `tasks.max` is set to this multiplied by the number of CPUs in the service. Default: `1`.
|
|
9894
10044
|
"""
|
|
10045
|
+
if admin_timeout_ms is not None:
|
|
10046
|
+
pulumi.set(__self__, "admin_timeout_ms", admin_timeout_ms)
|
|
9895
10047
|
if emit_checkpoints_enabled is not None:
|
|
9896
10048
|
pulumi.set(__self__, "emit_checkpoints_enabled", emit_checkpoints_enabled)
|
|
9897
10049
|
if emit_checkpoints_interval_seconds is not None:
|
|
@@ -9919,6 +10071,18 @@ class KafkaMirrorMakerKafkaMirrormakerUserConfigKafkaMirrormakerArgs:
|
|
|
9919
10071
|
if tasks_max_per_cpu is not None:
|
|
9920
10072
|
pulumi.set(__self__, "tasks_max_per_cpu", tasks_max_per_cpu)
|
|
9921
10073
|
|
|
10074
|
+
@property
|
|
10075
|
+
@pulumi.getter(name="adminTimeoutMs")
|
|
10076
|
+
def admin_timeout_ms(self) -> Optional[pulumi.Input[int]]:
|
|
10077
|
+
"""
|
|
10078
|
+
Timeout for administrative tasks, e.g. detecting new topics, loading of consumer group and offsets. Defaults to 60000 milliseconds (1 minute).
|
|
10079
|
+
"""
|
|
10080
|
+
return pulumi.get(self, "admin_timeout_ms")
|
|
10081
|
+
|
|
10082
|
+
@admin_timeout_ms.setter
|
|
10083
|
+
def admin_timeout_ms(self, value: Optional[pulumi.Input[int]]):
|
|
10084
|
+
pulumi.set(self, "admin_timeout_ms", value)
|
|
10085
|
+
|
|
9922
10086
|
@property
|
|
9923
10087
|
@pulumi.getter(name="emitCheckpointsEnabled")
|
|
9924
10088
|
def emit_checkpoints_enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
@@ -13225,6 +13389,7 @@ class MySqlMysqlUserConfigMigrationArgs:
|
|
|
13225
13389
|
port: pulumi.Input[int],
|
|
13226
13390
|
dbname: Optional[pulumi.Input[str]] = None,
|
|
13227
13391
|
ignore_dbs: Optional[pulumi.Input[str]] = None,
|
|
13392
|
+
ignore_roles: Optional[pulumi.Input[str]] = None,
|
|
13228
13393
|
method: Optional[pulumi.Input[str]] = None,
|
|
13229
13394
|
password: Optional[pulumi.Input[str]] = None,
|
|
13230
13395
|
ssl: Optional[pulumi.Input[bool]] = None,
|
|
@@ -13234,6 +13399,7 @@ class MySqlMysqlUserConfigMigrationArgs:
|
|
|
13234
13399
|
:param pulumi.Input[int] port: Port number of the server where to migrate data from. Example: `1234`.
|
|
13235
13400
|
:param pulumi.Input[str] dbname: Database name for bootstrapping the initial connection. Example: `defaultdb`.
|
|
13236
13401
|
:param pulumi.Input[str] ignore_dbs: Comma-separated list of databases, which should be ignored during migration (supported by MySQL and PostgreSQL only at the moment). Example: `db1,db2`.
|
|
13402
|
+
:param pulumi.Input[str] ignore_roles: Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
13237
13403
|
:param pulumi.Input[str] method: Enum: `dump`, `replication`. The migration method to be used (currently supported only by Redis, Dragonfly, MySQL and PostgreSQL service types).
|
|
13238
13404
|
:param pulumi.Input[str] password: Password for authentication with the server where to migrate data from. Example: `jjKk45Nnd`.
|
|
13239
13405
|
:param pulumi.Input[bool] ssl: The server where to migrate data from is secured with SSL. Default: `true`.
|
|
@@ -13245,6 +13411,8 @@ class MySqlMysqlUserConfigMigrationArgs:
|
|
|
13245
13411
|
pulumi.set(__self__, "dbname", dbname)
|
|
13246
13412
|
if ignore_dbs is not None:
|
|
13247
13413
|
pulumi.set(__self__, "ignore_dbs", ignore_dbs)
|
|
13414
|
+
if ignore_roles is not None:
|
|
13415
|
+
pulumi.set(__self__, "ignore_roles", ignore_roles)
|
|
13248
13416
|
if method is not None:
|
|
13249
13417
|
pulumi.set(__self__, "method", method)
|
|
13250
13418
|
if password is not None:
|
|
@@ -13302,6 +13470,18 @@ class MySqlMysqlUserConfigMigrationArgs:
|
|
|
13302
13470
|
def ignore_dbs(self, value: Optional[pulumi.Input[str]]):
|
|
13303
13471
|
pulumi.set(self, "ignore_dbs", value)
|
|
13304
13472
|
|
|
13473
|
+
@property
|
|
13474
|
+
@pulumi.getter(name="ignoreRoles")
|
|
13475
|
+
def ignore_roles(self) -> Optional[pulumi.Input[str]]:
|
|
13476
|
+
"""
|
|
13477
|
+
Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
13478
|
+
"""
|
|
13479
|
+
return pulumi.get(self, "ignore_roles")
|
|
13480
|
+
|
|
13481
|
+
@ignore_roles.setter
|
|
13482
|
+
def ignore_roles(self, value: Optional[pulumi.Input[str]]):
|
|
13483
|
+
pulumi.set(self, "ignore_roles", value)
|
|
13484
|
+
|
|
13305
13485
|
@property
|
|
13306
13486
|
@pulumi.getter
|
|
13307
13487
|
def method(self) -> Optional[pulumi.Input[str]]:
|
|
@@ -14325,9 +14505,12 @@ class OpenSearchOpensearchArgs:
|
|
|
14325
14505
|
class OpenSearchOpensearchUserConfigArgs:
|
|
14326
14506
|
def __init__(__self__, *,
|
|
14327
14507
|
additional_backup_regions: Optional[pulumi.Input[str]] = None,
|
|
14508
|
+
azure_migration: Optional[pulumi.Input['OpenSearchOpensearchUserConfigAzureMigrationArgs']] = None,
|
|
14328
14509
|
custom_domain: Optional[pulumi.Input[str]] = None,
|
|
14329
14510
|
disable_replication_factor_adjustment: Optional[pulumi.Input[bool]] = None,
|
|
14511
|
+
gcs_migration: Optional[pulumi.Input['OpenSearchOpensearchUserConfigGcsMigrationArgs']] = None,
|
|
14330
14512
|
index_patterns: Optional[pulumi.Input[Sequence[pulumi.Input['OpenSearchOpensearchUserConfigIndexPatternArgs']]]] = None,
|
|
14513
|
+
index_rollup: Optional[pulumi.Input['OpenSearchOpensearchUserConfigIndexRollupArgs']] = None,
|
|
14331
14514
|
index_template: Optional[pulumi.Input['OpenSearchOpensearchUserConfigIndexTemplateArgs']] = None,
|
|
14332
14515
|
ip_filter_objects: Optional[pulumi.Input[Sequence[pulumi.Input['OpenSearchOpensearchUserConfigIpFilterObjectArgs']]]] = None,
|
|
14333
14516
|
ip_filter_strings: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
|
|
@@ -14343,6 +14526,7 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14343
14526
|
project_to_fork_from: Optional[pulumi.Input[str]] = None,
|
|
14344
14527
|
public_access: Optional[pulumi.Input['OpenSearchOpensearchUserConfigPublicAccessArgs']] = None,
|
|
14345
14528
|
recovery_basebackup_name: Optional[pulumi.Input[str]] = None,
|
|
14529
|
+
s3_migration: Optional[pulumi.Input['OpenSearchOpensearchUserConfigS3MigrationArgs']] = None,
|
|
14346
14530
|
saml: Optional[pulumi.Input['OpenSearchOpensearchUserConfigSamlArgs']] = None,
|
|
14347
14531
|
service_log: Optional[pulumi.Input[bool]] = None,
|
|
14348
14532
|
service_to_fork_from: Optional[pulumi.Input[str]] = None,
|
|
@@ -14352,6 +14536,7 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14352
14536
|
:param pulumi.Input[str] custom_domain: Serve the web frontend using a custom CNAME pointing to the Aiven DNS name. Example: `grafana.example.org`.
|
|
14353
14537
|
:param pulumi.Input[bool] disable_replication_factor_adjustment: Disable automatic replication factor adjustment for multi-node services. By default, Aiven ensures all indexes are replicated at least to two nodes. Note: Due to potential data loss in case of losing a service node, this setting can no longer be activated.
|
|
14354
14538
|
:param pulumi.Input[Sequence[pulumi.Input['OpenSearchOpensearchUserConfigIndexPatternArgs']]] index_patterns: Index patterns
|
|
14539
|
+
:param pulumi.Input['OpenSearchOpensearchUserConfigIndexRollupArgs'] index_rollup: Index rollup settings
|
|
14355
14540
|
:param pulumi.Input['OpenSearchOpensearchUserConfigIndexTemplateArgs'] index_template: Template settings for all new indexes
|
|
14356
14541
|
:param pulumi.Input[Sequence[pulumi.Input['OpenSearchOpensearchUserConfigIpFilterObjectArgs']]] ip_filter_objects: Allow incoming connections from CIDR address block, e.g. `10.20.0.0/16`
|
|
14357
14542
|
:param pulumi.Input[Sequence[pulumi.Input[str]]] ip_filter_strings: Allow incoming connections from CIDR address block, e.g. `10.20.0.0/16`.
|
|
@@ -14374,12 +14559,18 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14374
14559
|
"""
|
|
14375
14560
|
if additional_backup_regions is not None:
|
|
14376
14561
|
pulumi.set(__self__, "additional_backup_regions", additional_backup_regions)
|
|
14562
|
+
if azure_migration is not None:
|
|
14563
|
+
pulumi.set(__self__, "azure_migration", azure_migration)
|
|
14377
14564
|
if custom_domain is not None:
|
|
14378
14565
|
pulumi.set(__self__, "custom_domain", custom_domain)
|
|
14379
14566
|
if disable_replication_factor_adjustment is not None:
|
|
14380
14567
|
pulumi.set(__self__, "disable_replication_factor_adjustment", disable_replication_factor_adjustment)
|
|
14568
|
+
if gcs_migration is not None:
|
|
14569
|
+
pulumi.set(__self__, "gcs_migration", gcs_migration)
|
|
14381
14570
|
if index_patterns is not None:
|
|
14382
14571
|
pulumi.set(__self__, "index_patterns", index_patterns)
|
|
14572
|
+
if index_rollup is not None:
|
|
14573
|
+
pulumi.set(__self__, "index_rollup", index_rollup)
|
|
14383
14574
|
if index_template is not None:
|
|
14384
14575
|
pulumi.set(__self__, "index_template", index_template)
|
|
14385
14576
|
if ip_filter_objects is not None:
|
|
@@ -14413,6 +14604,8 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14413
14604
|
pulumi.set(__self__, "public_access", public_access)
|
|
14414
14605
|
if recovery_basebackup_name is not None:
|
|
14415
14606
|
pulumi.set(__self__, "recovery_basebackup_name", recovery_basebackup_name)
|
|
14607
|
+
if s3_migration is not None:
|
|
14608
|
+
pulumi.set(__self__, "s3_migration", s3_migration)
|
|
14416
14609
|
if saml is not None:
|
|
14417
14610
|
pulumi.set(__self__, "saml", saml)
|
|
14418
14611
|
if service_log is not None:
|
|
@@ -14434,6 +14627,15 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14434
14627
|
def additional_backup_regions(self, value: Optional[pulumi.Input[str]]):
|
|
14435
14628
|
pulumi.set(self, "additional_backup_regions", value)
|
|
14436
14629
|
|
|
14630
|
+
@property
|
|
14631
|
+
@pulumi.getter(name="azureMigration")
|
|
14632
|
+
def azure_migration(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigAzureMigrationArgs']]:
|
|
14633
|
+
return pulumi.get(self, "azure_migration")
|
|
14634
|
+
|
|
14635
|
+
@azure_migration.setter
|
|
14636
|
+
def azure_migration(self, value: Optional[pulumi.Input['OpenSearchOpensearchUserConfigAzureMigrationArgs']]):
|
|
14637
|
+
pulumi.set(self, "azure_migration", value)
|
|
14638
|
+
|
|
14437
14639
|
@property
|
|
14438
14640
|
@pulumi.getter(name="customDomain")
|
|
14439
14641
|
def custom_domain(self) -> Optional[pulumi.Input[str]]:
|
|
@@ -14458,6 +14660,15 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14458
14660
|
def disable_replication_factor_adjustment(self, value: Optional[pulumi.Input[bool]]):
|
|
14459
14661
|
pulumi.set(self, "disable_replication_factor_adjustment", value)
|
|
14460
14662
|
|
|
14663
|
+
@property
|
|
14664
|
+
@pulumi.getter(name="gcsMigration")
|
|
14665
|
+
def gcs_migration(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigGcsMigrationArgs']]:
|
|
14666
|
+
return pulumi.get(self, "gcs_migration")
|
|
14667
|
+
|
|
14668
|
+
@gcs_migration.setter
|
|
14669
|
+
def gcs_migration(self, value: Optional[pulumi.Input['OpenSearchOpensearchUserConfigGcsMigrationArgs']]):
|
|
14670
|
+
pulumi.set(self, "gcs_migration", value)
|
|
14671
|
+
|
|
14461
14672
|
@property
|
|
14462
14673
|
@pulumi.getter(name="indexPatterns")
|
|
14463
14674
|
def index_patterns(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['OpenSearchOpensearchUserConfigIndexPatternArgs']]]]:
|
|
@@ -14470,6 +14681,18 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14470
14681
|
def index_patterns(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['OpenSearchOpensearchUserConfigIndexPatternArgs']]]]):
|
|
14471
14682
|
pulumi.set(self, "index_patterns", value)
|
|
14472
14683
|
|
|
14684
|
+
@property
|
|
14685
|
+
@pulumi.getter(name="indexRollup")
|
|
14686
|
+
def index_rollup(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigIndexRollupArgs']]:
|
|
14687
|
+
"""
|
|
14688
|
+
Index rollup settings
|
|
14689
|
+
"""
|
|
14690
|
+
return pulumi.get(self, "index_rollup")
|
|
14691
|
+
|
|
14692
|
+
@index_rollup.setter
|
|
14693
|
+
def index_rollup(self, value: Optional[pulumi.Input['OpenSearchOpensearchUserConfigIndexRollupArgs']]):
|
|
14694
|
+
pulumi.set(self, "index_rollup", value)
|
|
14695
|
+
|
|
14473
14696
|
@property
|
|
14474
14697
|
@pulumi.getter(name="indexTemplate")
|
|
14475
14698
|
def index_template(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigIndexTemplateArgs']]:
|
|
@@ -14651,6 +14874,15 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14651
14874
|
def recovery_basebackup_name(self, value: Optional[pulumi.Input[str]]):
|
|
14652
14875
|
pulumi.set(self, "recovery_basebackup_name", value)
|
|
14653
14876
|
|
|
14877
|
+
@property
|
|
14878
|
+
@pulumi.getter(name="s3Migration")
|
|
14879
|
+
def s3_migration(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigS3MigrationArgs']]:
|
|
14880
|
+
return pulumi.get(self, "s3_migration")
|
|
14881
|
+
|
|
14882
|
+
@s3_migration.setter
|
|
14883
|
+
def s3_migration(self, value: Optional[pulumi.Input['OpenSearchOpensearchUserConfigS3MigrationArgs']]):
|
|
14884
|
+
pulumi.set(self, "s3_migration", value)
|
|
14885
|
+
|
|
14654
14886
|
@property
|
|
14655
14887
|
@pulumi.getter
|
|
14656
14888
|
def saml(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigSamlArgs']]:
|
|
@@ -14700,6 +14932,252 @@ class OpenSearchOpensearchUserConfigArgs:
|
|
|
14700
14932
|
pulumi.set(self, "static_ips", value)
|
|
14701
14933
|
|
|
14702
14934
|
|
|
14935
|
+
@pulumi.input_type
|
|
14936
|
+
class OpenSearchOpensearchUserConfigAzureMigrationArgs:
|
|
14937
|
+
def __init__(__self__, *,
|
|
14938
|
+
account: pulumi.Input[str],
|
|
14939
|
+
base_path: pulumi.Input[str],
|
|
14940
|
+
container: pulumi.Input[str],
|
|
14941
|
+
snapshot_name: pulumi.Input[str],
|
|
14942
|
+
chunk_size: Optional[pulumi.Input[str]] = None,
|
|
14943
|
+
compress: Optional[pulumi.Input[bool]] = None,
|
|
14944
|
+
endpoint_suffix: Optional[pulumi.Input[str]] = None,
|
|
14945
|
+
key: Optional[pulumi.Input[str]] = None,
|
|
14946
|
+
sas_token: Optional[pulumi.Input[str]] = None):
|
|
14947
|
+
"""
|
|
14948
|
+
:param pulumi.Input[str] account: Azure account name.
|
|
14949
|
+
:param pulumi.Input[str] base_path: The path to the repository data within its container. The value of this setting should not start or end with a /.
|
|
14950
|
+
:param pulumi.Input[str] container: Azure container name.
|
|
14951
|
+
:param pulumi.Input[str] snapshot_name: The snapshot name to restore from.
|
|
14952
|
+
:param pulumi.Input[str] chunk_size: Big files can be broken down into chunks during snapshotting if needed. Should be the same as for the 3rd party repository.
|
|
14953
|
+
:param pulumi.Input[bool] compress: When set to true metadata files are stored in compressed format.
|
|
14954
|
+
:param pulumi.Input[str] endpoint_suffix: Defines the DNS suffix for Azure Storage endpoints.
|
|
14955
|
+
:param pulumi.Input[str] key: Azure account secret key. One of key or sas_token should be specified.
|
|
14956
|
+
:param pulumi.Input[str] sas_token: A shared access signatures (SAS) token. One of key or sas_token should be specified.
|
|
14957
|
+
"""
|
|
14958
|
+
pulumi.set(__self__, "account", account)
|
|
14959
|
+
pulumi.set(__self__, "base_path", base_path)
|
|
14960
|
+
pulumi.set(__self__, "container", container)
|
|
14961
|
+
pulumi.set(__self__, "snapshot_name", snapshot_name)
|
|
14962
|
+
if chunk_size is not None:
|
|
14963
|
+
pulumi.set(__self__, "chunk_size", chunk_size)
|
|
14964
|
+
if compress is not None:
|
|
14965
|
+
pulumi.set(__self__, "compress", compress)
|
|
14966
|
+
if endpoint_suffix is not None:
|
|
14967
|
+
pulumi.set(__self__, "endpoint_suffix", endpoint_suffix)
|
|
14968
|
+
if key is not None:
|
|
14969
|
+
pulumi.set(__self__, "key", key)
|
|
14970
|
+
if sas_token is not None:
|
|
14971
|
+
pulumi.set(__self__, "sas_token", sas_token)
|
|
14972
|
+
|
|
14973
|
+
@property
|
|
14974
|
+
@pulumi.getter
|
|
14975
|
+
def account(self) -> pulumi.Input[str]:
|
|
14976
|
+
"""
|
|
14977
|
+
Azure account name.
|
|
14978
|
+
"""
|
|
14979
|
+
return pulumi.get(self, "account")
|
|
14980
|
+
|
|
14981
|
+
@account.setter
|
|
14982
|
+
def account(self, value: pulumi.Input[str]):
|
|
14983
|
+
pulumi.set(self, "account", value)
|
|
14984
|
+
|
|
14985
|
+
@property
|
|
14986
|
+
@pulumi.getter(name="basePath")
|
|
14987
|
+
def base_path(self) -> pulumi.Input[str]:
|
|
14988
|
+
"""
|
|
14989
|
+
The path to the repository data within its container. The value of this setting should not start or end with a /.
|
|
14990
|
+
"""
|
|
14991
|
+
return pulumi.get(self, "base_path")
|
|
14992
|
+
|
|
14993
|
+
@base_path.setter
|
|
14994
|
+
def base_path(self, value: pulumi.Input[str]):
|
|
14995
|
+
pulumi.set(self, "base_path", value)
|
|
14996
|
+
|
|
14997
|
+
@property
|
|
14998
|
+
@pulumi.getter
|
|
14999
|
+
def container(self) -> pulumi.Input[str]:
|
|
15000
|
+
"""
|
|
15001
|
+
Azure container name.
|
|
15002
|
+
"""
|
|
15003
|
+
return pulumi.get(self, "container")
|
|
15004
|
+
|
|
15005
|
+
@container.setter
|
|
15006
|
+
def container(self, value: pulumi.Input[str]):
|
|
15007
|
+
pulumi.set(self, "container", value)
|
|
15008
|
+
|
|
15009
|
+
@property
|
|
15010
|
+
@pulumi.getter(name="snapshotName")
|
|
15011
|
+
def snapshot_name(self) -> pulumi.Input[str]:
|
|
15012
|
+
"""
|
|
15013
|
+
The snapshot name to restore from.
|
|
15014
|
+
"""
|
|
15015
|
+
return pulumi.get(self, "snapshot_name")
|
|
15016
|
+
|
|
15017
|
+
@snapshot_name.setter
|
|
15018
|
+
def snapshot_name(self, value: pulumi.Input[str]):
|
|
15019
|
+
pulumi.set(self, "snapshot_name", value)
|
|
15020
|
+
|
|
15021
|
+
@property
|
|
15022
|
+
@pulumi.getter(name="chunkSize")
|
|
15023
|
+
def chunk_size(self) -> Optional[pulumi.Input[str]]:
|
|
15024
|
+
"""
|
|
15025
|
+
Big files can be broken down into chunks during snapshotting if needed. Should be the same as for the 3rd party repository.
|
|
15026
|
+
"""
|
|
15027
|
+
return pulumi.get(self, "chunk_size")
|
|
15028
|
+
|
|
15029
|
+
@chunk_size.setter
|
|
15030
|
+
def chunk_size(self, value: Optional[pulumi.Input[str]]):
|
|
15031
|
+
pulumi.set(self, "chunk_size", value)
|
|
15032
|
+
|
|
15033
|
+
@property
|
|
15034
|
+
@pulumi.getter
|
|
15035
|
+
def compress(self) -> Optional[pulumi.Input[bool]]:
|
|
15036
|
+
"""
|
|
15037
|
+
When set to true metadata files are stored in compressed format.
|
|
15038
|
+
"""
|
|
15039
|
+
return pulumi.get(self, "compress")
|
|
15040
|
+
|
|
15041
|
+
@compress.setter
|
|
15042
|
+
def compress(self, value: Optional[pulumi.Input[bool]]):
|
|
15043
|
+
pulumi.set(self, "compress", value)
|
|
15044
|
+
|
|
15045
|
+
@property
|
|
15046
|
+
@pulumi.getter(name="endpointSuffix")
|
|
15047
|
+
def endpoint_suffix(self) -> Optional[pulumi.Input[str]]:
|
|
15048
|
+
"""
|
|
15049
|
+
Defines the DNS suffix for Azure Storage endpoints.
|
|
15050
|
+
"""
|
|
15051
|
+
return pulumi.get(self, "endpoint_suffix")
|
|
15052
|
+
|
|
15053
|
+
@endpoint_suffix.setter
|
|
15054
|
+
def endpoint_suffix(self, value: Optional[pulumi.Input[str]]):
|
|
15055
|
+
pulumi.set(self, "endpoint_suffix", value)
|
|
15056
|
+
|
|
15057
|
+
@property
|
|
15058
|
+
@pulumi.getter
|
|
15059
|
+
def key(self) -> Optional[pulumi.Input[str]]:
|
|
15060
|
+
"""
|
|
15061
|
+
Azure account secret key. One of key or sas_token should be specified.
|
|
15062
|
+
"""
|
|
15063
|
+
return pulumi.get(self, "key")
|
|
15064
|
+
|
|
15065
|
+
@key.setter
|
|
15066
|
+
def key(self, value: Optional[pulumi.Input[str]]):
|
|
15067
|
+
pulumi.set(self, "key", value)
|
|
15068
|
+
|
|
15069
|
+
@property
|
|
15070
|
+
@pulumi.getter(name="sasToken")
|
|
15071
|
+
def sas_token(self) -> Optional[pulumi.Input[str]]:
|
|
15072
|
+
"""
|
|
15073
|
+
A shared access signatures (SAS) token. One of key or sas_token should be specified.
|
|
15074
|
+
"""
|
|
15075
|
+
return pulumi.get(self, "sas_token")
|
|
15076
|
+
|
|
15077
|
+
@sas_token.setter
|
|
15078
|
+
def sas_token(self, value: Optional[pulumi.Input[str]]):
|
|
15079
|
+
pulumi.set(self, "sas_token", value)
|
|
15080
|
+
|
|
15081
|
+
|
|
15082
|
+
@pulumi.input_type
|
|
15083
|
+
class OpenSearchOpensearchUserConfigGcsMigrationArgs:
|
|
15084
|
+
def __init__(__self__, *,
|
|
15085
|
+
base_path: pulumi.Input[str],
|
|
15086
|
+
bucket: pulumi.Input[str],
|
|
15087
|
+
credentials: pulumi.Input[str],
|
|
15088
|
+
snapshot_name: pulumi.Input[str],
|
|
15089
|
+
chunk_size: Optional[pulumi.Input[str]] = None,
|
|
15090
|
+
compress: Optional[pulumi.Input[bool]] = None):
|
|
15091
|
+
"""
|
|
15092
|
+
:param pulumi.Input[str] base_path: The path to the repository data within its container. The value of this setting should not start or end with a /.
|
|
15093
|
+
:param pulumi.Input[str] bucket: The path to the repository data within its container.
|
|
15094
|
+
:param pulumi.Input[str] credentials: Google Cloud Storage credentials file content.
|
|
15095
|
+
:param pulumi.Input[str] snapshot_name: The snapshot name to restore from.
|
|
15096
|
+
:param pulumi.Input[str] chunk_size: Big files can be broken down into chunks during snapshotting if needed. Should be the same as for the 3rd party repository.
|
|
15097
|
+
:param pulumi.Input[bool] compress: When set to true metadata files are stored in compressed format.
|
|
15098
|
+
"""
|
|
15099
|
+
pulumi.set(__self__, "base_path", base_path)
|
|
15100
|
+
pulumi.set(__self__, "bucket", bucket)
|
|
15101
|
+
pulumi.set(__self__, "credentials", credentials)
|
|
15102
|
+
pulumi.set(__self__, "snapshot_name", snapshot_name)
|
|
15103
|
+
if chunk_size is not None:
|
|
15104
|
+
pulumi.set(__self__, "chunk_size", chunk_size)
|
|
15105
|
+
if compress is not None:
|
|
15106
|
+
pulumi.set(__self__, "compress", compress)
|
|
15107
|
+
|
|
15108
|
+
@property
|
|
15109
|
+
@pulumi.getter(name="basePath")
|
|
15110
|
+
def base_path(self) -> pulumi.Input[str]:
|
|
15111
|
+
"""
|
|
15112
|
+
The path to the repository data within its container. The value of this setting should not start or end with a /.
|
|
15113
|
+
"""
|
|
15114
|
+
return pulumi.get(self, "base_path")
|
|
15115
|
+
|
|
15116
|
+
@base_path.setter
|
|
15117
|
+
def base_path(self, value: pulumi.Input[str]):
|
|
15118
|
+
pulumi.set(self, "base_path", value)
|
|
15119
|
+
|
|
15120
|
+
@property
|
|
15121
|
+
@pulumi.getter
|
|
15122
|
+
def bucket(self) -> pulumi.Input[str]:
|
|
15123
|
+
"""
|
|
15124
|
+
The path to the repository data within its container.
|
|
15125
|
+
"""
|
|
15126
|
+
return pulumi.get(self, "bucket")
|
|
15127
|
+
|
|
15128
|
+
@bucket.setter
|
|
15129
|
+
def bucket(self, value: pulumi.Input[str]):
|
|
15130
|
+
pulumi.set(self, "bucket", value)
|
|
15131
|
+
|
|
15132
|
+
@property
|
|
15133
|
+
@pulumi.getter
|
|
15134
|
+
def credentials(self) -> pulumi.Input[str]:
|
|
15135
|
+
"""
|
|
15136
|
+
Google Cloud Storage credentials file content.
|
|
15137
|
+
"""
|
|
15138
|
+
return pulumi.get(self, "credentials")
|
|
15139
|
+
|
|
15140
|
+
@credentials.setter
|
|
15141
|
+
def credentials(self, value: pulumi.Input[str]):
|
|
15142
|
+
pulumi.set(self, "credentials", value)
|
|
15143
|
+
|
|
15144
|
+
@property
|
|
15145
|
+
@pulumi.getter(name="snapshotName")
|
|
15146
|
+
def snapshot_name(self) -> pulumi.Input[str]:
|
|
15147
|
+
"""
|
|
15148
|
+
The snapshot name to restore from.
|
|
15149
|
+
"""
|
|
15150
|
+
return pulumi.get(self, "snapshot_name")
|
|
15151
|
+
|
|
15152
|
+
@snapshot_name.setter
|
|
15153
|
+
def snapshot_name(self, value: pulumi.Input[str]):
|
|
15154
|
+
pulumi.set(self, "snapshot_name", value)
|
|
15155
|
+
|
|
15156
|
+
@property
|
|
15157
|
+
@pulumi.getter(name="chunkSize")
|
|
15158
|
+
def chunk_size(self) -> Optional[pulumi.Input[str]]:
|
|
15159
|
+
"""
|
|
15160
|
+
Big files can be broken down into chunks during snapshotting if needed. Should be the same as for the 3rd party repository.
|
|
15161
|
+
"""
|
|
15162
|
+
return pulumi.get(self, "chunk_size")
|
|
15163
|
+
|
|
15164
|
+
@chunk_size.setter
|
|
15165
|
+
def chunk_size(self, value: Optional[pulumi.Input[str]]):
|
|
15166
|
+
pulumi.set(self, "chunk_size", value)
|
|
15167
|
+
|
|
15168
|
+
@property
|
|
15169
|
+
@pulumi.getter
|
|
15170
|
+
def compress(self) -> Optional[pulumi.Input[bool]]:
|
|
15171
|
+
"""
|
|
15172
|
+
When set to true metadata files are stored in compressed format.
|
|
15173
|
+
"""
|
|
15174
|
+
return pulumi.get(self, "compress")
|
|
15175
|
+
|
|
15176
|
+
@compress.setter
|
|
15177
|
+
def compress(self, value: Optional[pulumi.Input[bool]]):
|
|
15178
|
+
pulumi.set(self, "compress", value)
|
|
15179
|
+
|
|
15180
|
+
|
|
14703
15181
|
@pulumi.input_type
|
|
14704
15182
|
class OpenSearchOpensearchUserConfigIndexPatternArgs:
|
|
14705
15183
|
def __init__(__self__, *,
|
|
@@ -14753,6 +15231,93 @@ class OpenSearchOpensearchUserConfigIndexPatternArgs:
|
|
|
14753
15231
|
pulumi.set(self, "sorting_algorithm", value)
|
|
14754
15232
|
|
|
14755
15233
|
|
|
15234
|
+
@pulumi.input_type
|
|
15235
|
+
class OpenSearchOpensearchUserConfigIndexRollupArgs:
|
|
15236
|
+
def __init__(__self__, *,
|
|
15237
|
+
rollup_dashboards_enabled: Optional[pulumi.Input[bool]] = None,
|
|
15238
|
+
rollup_enabled: Optional[pulumi.Input[bool]] = None,
|
|
15239
|
+
rollup_search_backoff_count: Optional[pulumi.Input[int]] = None,
|
|
15240
|
+
rollup_search_backoff_millis: Optional[pulumi.Input[int]] = None,
|
|
15241
|
+
rollup_search_search_all_jobs: Optional[pulumi.Input[bool]] = None):
|
|
15242
|
+
"""
|
|
15243
|
+
:param pulumi.Input[bool] rollup_dashboards_enabled: Whether rollups are enabled in OpenSearch Dashboards. Defaults to true.
|
|
15244
|
+
:param pulumi.Input[bool] rollup_enabled: Whether the rollup plugin is enabled. Defaults to true.
|
|
15245
|
+
:param pulumi.Input[int] rollup_search_backoff_count: How many retries the plugin should attempt for failed rollup jobs. Defaults to 5.
|
|
15246
|
+
:param pulumi.Input[int] rollup_search_backoff_millis: The backoff time between retries for failed rollup jobs. Defaults to 1000ms.
|
|
15247
|
+
:param pulumi.Input[bool] rollup_search_search_all_jobs: Whether OpenSearch should return all jobs that match all specified search terms. If disabled, OpenSearch returns just one, as opposed to all, of the jobs that matches the search terms. Defaults to false.
|
|
15248
|
+
"""
|
|
15249
|
+
if rollup_dashboards_enabled is not None:
|
|
15250
|
+
pulumi.set(__self__, "rollup_dashboards_enabled", rollup_dashboards_enabled)
|
|
15251
|
+
if rollup_enabled is not None:
|
|
15252
|
+
pulumi.set(__self__, "rollup_enabled", rollup_enabled)
|
|
15253
|
+
if rollup_search_backoff_count is not None:
|
|
15254
|
+
pulumi.set(__self__, "rollup_search_backoff_count", rollup_search_backoff_count)
|
|
15255
|
+
if rollup_search_backoff_millis is not None:
|
|
15256
|
+
pulumi.set(__self__, "rollup_search_backoff_millis", rollup_search_backoff_millis)
|
|
15257
|
+
if rollup_search_search_all_jobs is not None:
|
|
15258
|
+
pulumi.set(__self__, "rollup_search_search_all_jobs", rollup_search_search_all_jobs)
|
|
15259
|
+
|
|
15260
|
+
@property
|
|
15261
|
+
@pulumi.getter(name="rollupDashboardsEnabled")
|
|
15262
|
+
def rollup_dashboards_enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
15263
|
+
"""
|
|
15264
|
+
Whether rollups are enabled in OpenSearch Dashboards. Defaults to true.
|
|
15265
|
+
"""
|
|
15266
|
+
return pulumi.get(self, "rollup_dashboards_enabled")
|
|
15267
|
+
|
|
15268
|
+
@rollup_dashboards_enabled.setter
|
|
15269
|
+
def rollup_dashboards_enabled(self, value: Optional[pulumi.Input[bool]]):
|
|
15270
|
+
pulumi.set(self, "rollup_dashboards_enabled", value)
|
|
15271
|
+
|
|
15272
|
+
@property
|
|
15273
|
+
@pulumi.getter(name="rollupEnabled")
|
|
15274
|
+
def rollup_enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
15275
|
+
"""
|
|
15276
|
+
Whether the rollup plugin is enabled. Defaults to true.
|
|
15277
|
+
"""
|
|
15278
|
+
return pulumi.get(self, "rollup_enabled")
|
|
15279
|
+
|
|
15280
|
+
@rollup_enabled.setter
|
|
15281
|
+
def rollup_enabled(self, value: Optional[pulumi.Input[bool]]):
|
|
15282
|
+
pulumi.set(self, "rollup_enabled", value)
|
|
15283
|
+
|
|
15284
|
+
@property
|
|
15285
|
+
@pulumi.getter(name="rollupSearchBackoffCount")
|
|
15286
|
+
def rollup_search_backoff_count(self) -> Optional[pulumi.Input[int]]:
|
|
15287
|
+
"""
|
|
15288
|
+
How many retries the plugin should attempt for failed rollup jobs. Defaults to 5.
|
|
15289
|
+
"""
|
|
15290
|
+
return pulumi.get(self, "rollup_search_backoff_count")
|
|
15291
|
+
|
|
15292
|
+
@rollup_search_backoff_count.setter
|
|
15293
|
+
def rollup_search_backoff_count(self, value: Optional[pulumi.Input[int]]):
|
|
15294
|
+
pulumi.set(self, "rollup_search_backoff_count", value)
|
|
15295
|
+
|
|
15296
|
+
@property
|
|
15297
|
+
@pulumi.getter(name="rollupSearchBackoffMillis")
|
|
15298
|
+
def rollup_search_backoff_millis(self) -> Optional[pulumi.Input[int]]:
|
|
15299
|
+
"""
|
|
15300
|
+
The backoff time between retries for failed rollup jobs. Defaults to 1000ms.
|
|
15301
|
+
"""
|
|
15302
|
+
return pulumi.get(self, "rollup_search_backoff_millis")
|
|
15303
|
+
|
|
15304
|
+
@rollup_search_backoff_millis.setter
|
|
15305
|
+
def rollup_search_backoff_millis(self, value: Optional[pulumi.Input[int]]):
|
|
15306
|
+
pulumi.set(self, "rollup_search_backoff_millis", value)
|
|
15307
|
+
|
|
15308
|
+
@property
|
|
15309
|
+
@pulumi.getter(name="rollupSearchSearchAllJobs")
|
|
15310
|
+
def rollup_search_search_all_jobs(self) -> Optional[pulumi.Input[bool]]:
|
|
15311
|
+
"""
|
|
15312
|
+
Whether OpenSearch should return all jobs that match all specified search terms. If disabled, OpenSearch returns just one, as opposed to all, of the jobs that matches the search terms. Defaults to false.
|
|
15313
|
+
"""
|
|
15314
|
+
return pulumi.get(self, "rollup_search_search_all_jobs")
|
|
15315
|
+
|
|
15316
|
+
@rollup_search_search_all_jobs.setter
|
|
15317
|
+
def rollup_search_search_all_jobs(self, value: Optional[pulumi.Input[bool]]):
|
|
15318
|
+
pulumi.set(self, "rollup_search_search_all_jobs", value)
|
|
15319
|
+
|
|
15320
|
+
|
|
14756
15321
|
@pulumi.input_type
|
|
14757
15322
|
class OpenSearchOpensearchUserConfigIndexTemplateArgs:
|
|
14758
15323
|
def __init__(__self__, *,
|
|
@@ -14862,8 +15427,8 @@ class OpenSearchOpensearchUserConfigOpenidArgs:
|
|
|
14862
15427
|
scope: Optional[pulumi.Input[str]] = None,
|
|
14863
15428
|
subject_key: Optional[pulumi.Input[str]] = None):
|
|
14864
15429
|
"""
|
|
14865
|
-
:param pulumi.Input[str] client_id: The ID of the OpenID Connect client configured in your IdP. Required.
|
|
14866
|
-
:param pulumi.Input[str] client_secret: The client secret of the OpenID Connect client configured in your IdP. Required.
|
|
15430
|
+
:param pulumi.Input[str] client_id: The ID of the OpenID Connect client configured in your IdP. Required.
|
|
15431
|
+
:param pulumi.Input[str] client_secret: The client secret of the OpenID Connect client configured in your IdP. Required.
|
|
14867
15432
|
:param pulumi.Input[str] connect_url: The URL of your IdP where the Security plugin can find the OpenID Connect metadata/configuration settings. Example: `https://test-account.okta.com/app/exk491jujcVc83LEX697/sso/saml/metadata`.
|
|
14868
15433
|
:param pulumi.Input[bool] enabled: Enables or disables OpenID Connect authentication for OpenSearch. When enabled, users can authenticate using OpenID Connect with an Identity Provider. Default: `true`.
|
|
14869
15434
|
:param pulumi.Input[str] header: HTTP header name of the JWT token. Optional. Default is Authorization. Default: `Authorization`.
|
|
@@ -14872,7 +15437,7 @@ class OpenSearchOpensearchUserConfigOpenidArgs:
|
|
|
14872
15437
|
:param pulumi.Input[int] refresh_rate_limit_count: The maximum number of unknown key IDs in the time frame. Default is 10. Optional. Default: `10`.
|
|
14873
15438
|
:param pulumi.Input[int] refresh_rate_limit_time_window_ms: The time frame to use when checking the maximum number of unknown key IDs, in milliseconds. Optional.Default is 10000 (10 seconds). Default: `10000`.
|
|
14874
15439
|
:param pulumi.Input[str] roles_key: The key in the JSON payload that stores the user’s roles. The value of this key must be a comma-separated list of roles. Required only if you want to use roles in the JWT. Example: `roles`.
|
|
14875
|
-
:param pulumi.Input[str] scope: The scope of the identity token issued by the IdP. Optional. Default is openid profile email address phone.
|
|
15440
|
+
:param pulumi.Input[str] scope: The scope of the identity token issued by the IdP. Optional. Default is openid profile email address phone.
|
|
14876
15441
|
:param pulumi.Input[str] subject_key: The key in the JSON payload that stores the user’s name. If not defined, the subject registered claim is used. Most IdP providers use the preferred_username claim. Optional. Example: `preferred_username`.
|
|
14877
15442
|
"""
|
|
14878
15443
|
pulumi.set(__self__, "client_id", client_id)
|
|
@@ -14900,7 +15465,7 @@ class OpenSearchOpensearchUserConfigOpenidArgs:
|
|
|
14900
15465
|
@pulumi.getter(name="clientId")
|
|
14901
15466
|
def client_id(self) -> pulumi.Input[str]:
|
|
14902
15467
|
"""
|
|
14903
|
-
The ID of the OpenID Connect client configured in your IdP. Required.
|
|
15468
|
+
The ID of the OpenID Connect client configured in your IdP. Required.
|
|
14904
15469
|
"""
|
|
14905
15470
|
return pulumi.get(self, "client_id")
|
|
14906
15471
|
|
|
@@ -14912,7 +15477,7 @@ class OpenSearchOpensearchUserConfigOpenidArgs:
|
|
|
14912
15477
|
@pulumi.getter(name="clientSecret")
|
|
14913
15478
|
def client_secret(self) -> pulumi.Input[str]:
|
|
14914
15479
|
"""
|
|
14915
|
-
The client secret of the OpenID Connect client configured in your IdP. Required.
|
|
15480
|
+
The client secret of the OpenID Connect client configured in your IdP. Required.
|
|
14916
15481
|
"""
|
|
14917
15482
|
return pulumi.get(self, "client_secret")
|
|
14918
15483
|
|
|
@@ -15020,7 +15585,7 @@ class OpenSearchOpensearchUserConfigOpenidArgs:
|
|
|
15020
15585
|
@pulumi.getter
|
|
15021
15586
|
def scope(self) -> Optional[pulumi.Input[str]]:
|
|
15022
15587
|
"""
|
|
15023
|
-
The scope of the identity token issued by the IdP. Optional. Default is openid profile email address phone.
|
|
15588
|
+
The scope of the identity token issued by the IdP. Optional. Default is openid profile email address phone.
|
|
15024
15589
|
"""
|
|
15025
15590
|
return pulumi.get(self, "scope")
|
|
15026
15591
|
|
|
@@ -15097,7 +15662,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15097
15662
|
:param pulumi.Input[str] email_sender_name: Sender name placeholder to be used in Opensearch Dashboards and Opensearch keystore. Example: `alert-sender`.
|
|
15098
15663
|
:param pulumi.Input[str] email_sender_password: Sender password for Opensearch alerts to authenticate with SMTP server. Example: `very-secure-mail-password`.
|
|
15099
15664
|
:param pulumi.Input[str] email_sender_username: Sender username for Opensearch alerts. Example: `jane@example.com`.
|
|
15100
|
-
:param pulumi.Input[bool] enable_security_audit: Enable/Disable security audit.
|
|
15665
|
+
:param pulumi.Input[bool] enable_security_audit: Enable/Disable security audit.
|
|
15101
15666
|
:param pulumi.Input[int] http_max_content_length: Maximum content length for HTTP requests to the OpenSearch HTTP API, in bytes.
|
|
15102
15667
|
:param pulumi.Input[int] http_max_header_size: The max size of allowed headers, in bytes. Example: `8192`.
|
|
15103
15668
|
:param pulumi.Input[int] http_max_initial_line_length: The max length of an HTTP URL, in bytes. Example: `4096`.
|
|
@@ -15109,14 +15674,14 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15109
15674
|
:param pulumi.Input[int] indices_query_bool_max_clause_count: Maximum number of clauses Lucene BooleanQuery can have. The default value (1024) is relatively high, and increasing it may cause performance issues. Investigate other approaches first before increasing this value.
|
|
15110
15675
|
:param pulumi.Input[int] indices_recovery_max_bytes_per_sec: Limits total inbound and outbound recovery traffic for each node. Applies to both peer recoveries as well as snapshot recoveries (i.e., restores from a snapshot). Defaults to 40mb.
|
|
15111
15676
|
:param pulumi.Input[int] indices_recovery_max_concurrent_file_chunks: Number of file chunks sent in parallel for each recovery. Defaults to 2.
|
|
15112
|
-
:param pulumi.Input[bool] ism_enabled: Specifies whether ISM is enabled or not.
|
|
15113
|
-
:param pulumi.Input[bool] ism_history_enabled: Specifies whether audit history is enabled or not. The logs from ISM are automatically indexed to a logs document.
|
|
15114
|
-
:param pulumi.Input[int] ism_history_max_age: The maximum age before rolling over the audit history index in hours.
|
|
15115
|
-
:param pulumi.Input[int] ism_history_max_docs: The maximum number of documents before rolling over the audit history index.
|
|
15116
|
-
:param pulumi.Input[int] ism_history_rollover_check_period: The time between rollover checks for the audit history index in hours.
|
|
15117
|
-
:param pulumi.Input[int] ism_history_rollover_retention_period: How long audit history indices are kept in days.
|
|
15118
|
-
:param pulumi.Input[bool] knn_memory_circuit_breaker_enabled: Enable or disable KNN memory circuit breaker. Defaults to true.
|
|
15119
|
-
:param pulumi.Input[int] knn_memory_circuit_breaker_limit: Maximum amount of memory that can be used for KNN index. Defaults to 50% of the JVM heap size.
|
|
15677
|
+
:param pulumi.Input[bool] ism_enabled: Specifies whether ISM is enabled or not.
|
|
15678
|
+
:param pulumi.Input[bool] ism_history_enabled: Specifies whether audit history is enabled or not. The logs from ISM are automatically indexed to a logs document.
|
|
15679
|
+
:param pulumi.Input[int] ism_history_max_age: The maximum age before rolling over the audit history index in hours. Example: `24`.
|
|
15680
|
+
:param pulumi.Input[int] ism_history_max_docs: The maximum number of documents before rolling over the audit history index. Example: `2500000`.
|
|
15681
|
+
:param pulumi.Input[int] ism_history_rollover_check_period: The time between rollover checks for the audit history index in hours. Example: `8`.
|
|
15682
|
+
:param pulumi.Input[int] ism_history_rollover_retention_period: How long audit history indices are kept in days. Example: `30`.
|
|
15683
|
+
:param pulumi.Input[bool] knn_memory_circuit_breaker_enabled: Enable or disable KNN memory circuit breaker. Defaults to true.
|
|
15684
|
+
:param pulumi.Input[int] knn_memory_circuit_breaker_limit: Maximum amount of memory that can be used for KNN index. Defaults to 50% of the JVM heap size.
|
|
15120
15685
|
:param pulumi.Input[bool] override_main_response_version: Compatibility mode sets OpenSearch to report its version as 7.10 so clients continue to work. Default is false.
|
|
15121
15686
|
:param pulumi.Input[bool] plugins_alerting_filter_by_backend_roles: Enable or disable filtering of alerting by backend roles. Requires Security plugin. Defaults to false.
|
|
15122
15687
|
:param pulumi.Input[Sequence[pulumi.Input[str]]] reindex_remote_whitelists: Whitelisted addresses for reindexing. Changing this value will cause all OpenSearch instances to restart.
|
|
@@ -15323,7 +15888,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15323
15888
|
@pulumi.getter(name="enableSecurityAudit")
|
|
15324
15889
|
def enable_security_audit(self) -> Optional[pulumi.Input[bool]]:
|
|
15325
15890
|
"""
|
|
15326
|
-
Enable/Disable security audit.
|
|
15891
|
+
Enable/Disable security audit.
|
|
15327
15892
|
"""
|
|
15328
15893
|
return pulumi.get(self, "enable_security_audit")
|
|
15329
15894
|
|
|
@@ -15467,7 +16032,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15467
16032
|
@pulumi.getter(name="ismEnabled")
|
|
15468
16033
|
def ism_enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
15469
16034
|
"""
|
|
15470
|
-
Specifies whether ISM is enabled or not.
|
|
16035
|
+
Specifies whether ISM is enabled or not.
|
|
15471
16036
|
"""
|
|
15472
16037
|
return pulumi.get(self, "ism_enabled")
|
|
15473
16038
|
|
|
@@ -15479,7 +16044,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15479
16044
|
@pulumi.getter(name="ismHistoryEnabled")
|
|
15480
16045
|
def ism_history_enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
15481
16046
|
"""
|
|
15482
|
-
Specifies whether audit history is enabled or not. The logs from ISM are automatically indexed to a logs document.
|
|
16047
|
+
Specifies whether audit history is enabled or not. The logs from ISM are automatically indexed to a logs document.
|
|
15483
16048
|
"""
|
|
15484
16049
|
return pulumi.get(self, "ism_history_enabled")
|
|
15485
16050
|
|
|
@@ -15491,7 +16056,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15491
16056
|
@pulumi.getter(name="ismHistoryMaxAge")
|
|
15492
16057
|
def ism_history_max_age(self) -> Optional[pulumi.Input[int]]:
|
|
15493
16058
|
"""
|
|
15494
|
-
The maximum age before rolling over the audit history index in hours.
|
|
16059
|
+
The maximum age before rolling over the audit history index in hours. Example: `24`.
|
|
15495
16060
|
"""
|
|
15496
16061
|
return pulumi.get(self, "ism_history_max_age")
|
|
15497
16062
|
|
|
@@ -15503,7 +16068,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15503
16068
|
@pulumi.getter(name="ismHistoryMaxDocs")
|
|
15504
16069
|
def ism_history_max_docs(self) -> Optional[pulumi.Input[int]]:
|
|
15505
16070
|
"""
|
|
15506
|
-
The maximum number of documents before rolling over the audit history index.
|
|
16071
|
+
The maximum number of documents before rolling over the audit history index. Example: `2500000`.
|
|
15507
16072
|
"""
|
|
15508
16073
|
return pulumi.get(self, "ism_history_max_docs")
|
|
15509
16074
|
|
|
@@ -15515,7 +16080,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15515
16080
|
@pulumi.getter(name="ismHistoryRolloverCheckPeriod")
|
|
15516
16081
|
def ism_history_rollover_check_period(self) -> Optional[pulumi.Input[int]]:
|
|
15517
16082
|
"""
|
|
15518
|
-
The time between rollover checks for the audit history index in hours.
|
|
16083
|
+
The time between rollover checks for the audit history index in hours. Example: `8`.
|
|
15519
16084
|
"""
|
|
15520
16085
|
return pulumi.get(self, "ism_history_rollover_check_period")
|
|
15521
16086
|
|
|
@@ -15527,7 +16092,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15527
16092
|
@pulumi.getter(name="ismHistoryRolloverRetentionPeriod")
|
|
15528
16093
|
def ism_history_rollover_retention_period(self) -> Optional[pulumi.Input[int]]:
|
|
15529
16094
|
"""
|
|
15530
|
-
How long audit history indices are kept in days.
|
|
16095
|
+
How long audit history indices are kept in days. Example: `30`.
|
|
15531
16096
|
"""
|
|
15532
16097
|
return pulumi.get(self, "ism_history_rollover_retention_period")
|
|
15533
16098
|
|
|
@@ -15539,7 +16104,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15539
16104
|
@pulumi.getter(name="knnMemoryCircuitBreakerEnabled")
|
|
15540
16105
|
def knn_memory_circuit_breaker_enabled(self) -> Optional[pulumi.Input[bool]]:
|
|
15541
16106
|
"""
|
|
15542
|
-
Enable or disable KNN memory circuit breaker. Defaults to true.
|
|
16107
|
+
Enable or disable KNN memory circuit breaker. Defaults to true.
|
|
15543
16108
|
"""
|
|
15544
16109
|
return pulumi.get(self, "knn_memory_circuit_breaker_enabled")
|
|
15545
16110
|
|
|
@@ -15551,7 +16116,7 @@ class OpenSearchOpensearchUserConfigOpensearchArgs:
|
|
|
15551
16116
|
@pulumi.getter(name="knnMemoryCircuitBreakerLimit")
|
|
15552
16117
|
def knn_memory_circuit_breaker_limit(self) -> Optional[pulumi.Input[int]]:
|
|
15553
16118
|
"""
|
|
15554
|
-
Maximum amount of memory that can be used for KNN index. Defaults to 50% of the JVM heap size.
|
|
16119
|
+
Maximum amount of memory that can be used for KNN index. Defaults to 50% of the JVM heap size.
|
|
15555
16120
|
"""
|
|
15556
16121
|
return pulumi.get(self, "knn_memory_circuit_breaker_limit")
|
|
15557
16122
|
|
|
@@ -16229,6 +16794,167 @@ class OpenSearchOpensearchUserConfigPublicAccessArgs:
|
|
|
16229
16794
|
pulumi.set(self, "prometheus", value)
|
|
16230
16795
|
|
|
16231
16796
|
|
|
16797
|
+
@pulumi.input_type
|
|
16798
|
+
class OpenSearchOpensearchUserConfigS3MigrationArgs:
|
|
16799
|
+
def __init__(__self__, *,
|
|
16800
|
+
access_key: pulumi.Input[str],
|
|
16801
|
+
base_path: pulumi.Input[str],
|
|
16802
|
+
bucket: pulumi.Input[str],
|
|
16803
|
+
region: pulumi.Input[str],
|
|
16804
|
+
secret_key: pulumi.Input[str],
|
|
16805
|
+
snapshot_name: pulumi.Input[str],
|
|
16806
|
+
chunk_size: Optional[pulumi.Input[str]] = None,
|
|
16807
|
+
compress: Optional[pulumi.Input[bool]] = None,
|
|
16808
|
+
endpoint: Optional[pulumi.Input[str]] = None,
|
|
16809
|
+
server_side_encryption: Optional[pulumi.Input[bool]] = None):
|
|
16810
|
+
"""
|
|
16811
|
+
:param pulumi.Input[str] access_key: AWS Access key.
|
|
16812
|
+
:param pulumi.Input[str] base_path: The path to the repository data within its container. The value of this setting should not start or end with a /.
|
|
16813
|
+
:param pulumi.Input[str] bucket: S3 bucket name.
|
|
16814
|
+
:param pulumi.Input[str] region: S3 region.
|
|
16815
|
+
:param pulumi.Input[str] secret_key: AWS secret key.
|
|
16816
|
+
:param pulumi.Input[str] snapshot_name: The snapshot name to restore from.
|
|
16817
|
+
:param pulumi.Input[str] chunk_size: Big files can be broken down into chunks during snapshotting if needed. Should be the same as for the 3rd party repository.
|
|
16818
|
+
:param pulumi.Input[bool] compress: When set to true metadata files are stored in compressed format.
|
|
16819
|
+
:param pulumi.Input[str] endpoint: The S3 service endpoint to connect to. If you are using an S3-compatible service then you should set this to the service’s endpoint.
|
|
16820
|
+
:param pulumi.Input[bool] server_side_encryption: When set to true files are encrypted on server side.
|
|
16821
|
+
"""
|
|
16822
|
+
pulumi.set(__self__, "access_key", access_key)
|
|
16823
|
+
pulumi.set(__self__, "base_path", base_path)
|
|
16824
|
+
pulumi.set(__self__, "bucket", bucket)
|
|
16825
|
+
pulumi.set(__self__, "region", region)
|
|
16826
|
+
pulumi.set(__self__, "secret_key", secret_key)
|
|
16827
|
+
pulumi.set(__self__, "snapshot_name", snapshot_name)
|
|
16828
|
+
if chunk_size is not None:
|
|
16829
|
+
pulumi.set(__self__, "chunk_size", chunk_size)
|
|
16830
|
+
if compress is not None:
|
|
16831
|
+
pulumi.set(__self__, "compress", compress)
|
|
16832
|
+
if endpoint is not None:
|
|
16833
|
+
pulumi.set(__self__, "endpoint", endpoint)
|
|
16834
|
+
if server_side_encryption is not None:
|
|
16835
|
+
pulumi.set(__self__, "server_side_encryption", server_side_encryption)
|
|
16836
|
+
|
|
16837
|
+
@property
|
|
16838
|
+
@pulumi.getter(name="accessKey")
|
|
16839
|
+
def access_key(self) -> pulumi.Input[str]:
|
|
16840
|
+
"""
|
|
16841
|
+
AWS Access key.
|
|
16842
|
+
"""
|
|
16843
|
+
return pulumi.get(self, "access_key")
|
|
16844
|
+
|
|
16845
|
+
@access_key.setter
|
|
16846
|
+
def access_key(self, value: pulumi.Input[str]):
|
|
16847
|
+
pulumi.set(self, "access_key", value)
|
|
16848
|
+
|
|
16849
|
+
@property
|
|
16850
|
+
@pulumi.getter(name="basePath")
|
|
16851
|
+
def base_path(self) -> pulumi.Input[str]:
|
|
16852
|
+
"""
|
|
16853
|
+
The path to the repository data within its container. The value of this setting should not start or end with a /.
|
|
16854
|
+
"""
|
|
16855
|
+
return pulumi.get(self, "base_path")
|
|
16856
|
+
|
|
16857
|
+
@base_path.setter
|
|
16858
|
+
def base_path(self, value: pulumi.Input[str]):
|
|
16859
|
+
pulumi.set(self, "base_path", value)
|
|
16860
|
+
|
|
16861
|
+
@property
|
|
16862
|
+
@pulumi.getter
|
|
16863
|
+
def bucket(self) -> pulumi.Input[str]:
|
|
16864
|
+
"""
|
|
16865
|
+
S3 bucket name.
|
|
16866
|
+
"""
|
|
16867
|
+
return pulumi.get(self, "bucket")
|
|
16868
|
+
|
|
16869
|
+
@bucket.setter
|
|
16870
|
+
def bucket(self, value: pulumi.Input[str]):
|
|
16871
|
+
pulumi.set(self, "bucket", value)
|
|
16872
|
+
|
|
16873
|
+
@property
|
|
16874
|
+
@pulumi.getter
|
|
16875
|
+
def region(self) -> pulumi.Input[str]:
|
|
16876
|
+
"""
|
|
16877
|
+
S3 region.
|
|
16878
|
+
"""
|
|
16879
|
+
return pulumi.get(self, "region")
|
|
16880
|
+
|
|
16881
|
+
@region.setter
|
|
16882
|
+
def region(self, value: pulumi.Input[str]):
|
|
16883
|
+
pulumi.set(self, "region", value)
|
|
16884
|
+
|
|
16885
|
+
@property
|
|
16886
|
+
@pulumi.getter(name="secretKey")
|
|
16887
|
+
def secret_key(self) -> pulumi.Input[str]:
|
|
16888
|
+
"""
|
|
16889
|
+
AWS secret key.
|
|
16890
|
+
"""
|
|
16891
|
+
return pulumi.get(self, "secret_key")
|
|
16892
|
+
|
|
16893
|
+
@secret_key.setter
|
|
16894
|
+
def secret_key(self, value: pulumi.Input[str]):
|
|
16895
|
+
pulumi.set(self, "secret_key", value)
|
|
16896
|
+
|
|
16897
|
+
@property
|
|
16898
|
+
@pulumi.getter(name="snapshotName")
|
|
16899
|
+
def snapshot_name(self) -> pulumi.Input[str]:
|
|
16900
|
+
"""
|
|
16901
|
+
The snapshot name to restore from.
|
|
16902
|
+
"""
|
|
16903
|
+
return pulumi.get(self, "snapshot_name")
|
|
16904
|
+
|
|
16905
|
+
@snapshot_name.setter
|
|
16906
|
+
def snapshot_name(self, value: pulumi.Input[str]):
|
|
16907
|
+
pulumi.set(self, "snapshot_name", value)
|
|
16908
|
+
|
|
16909
|
+
@property
|
|
16910
|
+
@pulumi.getter(name="chunkSize")
|
|
16911
|
+
def chunk_size(self) -> Optional[pulumi.Input[str]]:
|
|
16912
|
+
"""
|
|
16913
|
+
Big files can be broken down into chunks during snapshotting if needed. Should be the same as for the 3rd party repository.
|
|
16914
|
+
"""
|
|
16915
|
+
return pulumi.get(self, "chunk_size")
|
|
16916
|
+
|
|
16917
|
+
@chunk_size.setter
|
|
16918
|
+
def chunk_size(self, value: Optional[pulumi.Input[str]]):
|
|
16919
|
+
pulumi.set(self, "chunk_size", value)
|
|
16920
|
+
|
|
16921
|
+
@property
|
|
16922
|
+
@pulumi.getter
|
|
16923
|
+
def compress(self) -> Optional[pulumi.Input[bool]]:
|
|
16924
|
+
"""
|
|
16925
|
+
When set to true metadata files are stored in compressed format.
|
|
16926
|
+
"""
|
|
16927
|
+
return pulumi.get(self, "compress")
|
|
16928
|
+
|
|
16929
|
+
@compress.setter
|
|
16930
|
+
def compress(self, value: Optional[pulumi.Input[bool]]):
|
|
16931
|
+
pulumi.set(self, "compress", value)
|
|
16932
|
+
|
|
16933
|
+
@property
|
|
16934
|
+
@pulumi.getter
|
|
16935
|
+
def endpoint(self) -> Optional[pulumi.Input[str]]:
|
|
16936
|
+
"""
|
|
16937
|
+
The S3 service endpoint to connect to. If you are using an S3-compatible service then you should set this to the service’s endpoint.
|
|
16938
|
+
"""
|
|
16939
|
+
return pulumi.get(self, "endpoint")
|
|
16940
|
+
|
|
16941
|
+
@endpoint.setter
|
|
16942
|
+
def endpoint(self, value: Optional[pulumi.Input[str]]):
|
|
16943
|
+
pulumi.set(self, "endpoint", value)
|
|
16944
|
+
|
|
16945
|
+
@property
|
|
16946
|
+
@pulumi.getter(name="serverSideEncryption")
|
|
16947
|
+
def server_side_encryption(self) -> Optional[pulumi.Input[bool]]:
|
|
16948
|
+
"""
|
|
16949
|
+
When set to true files are encrypted on server side.
|
|
16950
|
+
"""
|
|
16951
|
+
return pulumi.get(self, "server_side_encryption")
|
|
16952
|
+
|
|
16953
|
+
@server_side_encryption.setter
|
|
16954
|
+
def server_side_encryption(self, value: Optional[pulumi.Input[bool]]):
|
|
16955
|
+
pulumi.set(self, "server_side_encryption", value)
|
|
16956
|
+
|
|
16957
|
+
|
|
16232
16958
|
@pulumi.input_type
|
|
16233
16959
|
class OpenSearchOpensearchUserConfigSamlArgs:
|
|
16234
16960
|
def __init__(__self__, *,
|
|
@@ -16812,20 +17538,20 @@ class PgPgArgs:
|
|
|
16812
17538
|
uris: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
|
|
16813
17539
|
user: Optional[pulumi.Input[str]] = None):
|
|
16814
17540
|
"""
|
|
16815
|
-
:param pulumi.Input[str] bouncer:
|
|
16816
|
-
:param pulumi.Input[str] dbname: Primary PostgreSQL database name
|
|
16817
|
-
:param pulumi.Input[str] host: PostgreSQL
|
|
16818
|
-
:param pulumi.Input[int] max_connections:
|
|
16819
|
-
:param pulumi.Input[Sequence[pulumi.Input['PgPgParamArgs']]] params: PostgreSQL connection parameters
|
|
16820
|
-
:param pulumi.Input[str] password: PostgreSQL admin user password
|
|
16821
|
-
:param pulumi.Input[int] port: PostgreSQL port
|
|
16822
|
-
:param pulumi.Input[str] replica_uri: PostgreSQL replica URI for services with a replica
|
|
16823
|
-
:param pulumi.Input[str] sslmode: PostgreSQL
|
|
16824
|
-
:param pulumi.Input[Sequence[pulumi.Input[str]]] standby_uris: PostgreSQL standby connection URIs
|
|
16825
|
-
:param pulumi.Input[Sequence[pulumi.Input[str]]] syncing_uris: PostgreSQL syncing connection URIs
|
|
16826
|
-
:param pulumi.Input[str] uri: PostgreSQL
|
|
16827
|
-
:param pulumi.Input[Sequence[pulumi.Input[str]]] uris: PostgreSQL
|
|
16828
|
-
:param pulumi.Input[str] user: PostgreSQL admin user name
|
|
17541
|
+
:param pulumi.Input[str] bouncer: PgBouncer connection details for [connection pooling](https://aiven.io/docs/products/postgresql/concepts/pg-connection-pooling).
|
|
17542
|
+
:param pulumi.Input[str] dbname: Primary PostgreSQL database name.
|
|
17543
|
+
:param pulumi.Input[str] host: PostgreSQL primary node host IP or name.
|
|
17544
|
+
:param pulumi.Input[int] max_connections: The [number of allowed connections](https://aiven.io/docs/products/postgresql/reference/pg-connection-limits). Varies based on the service plan.
|
|
17545
|
+
:param pulumi.Input[Sequence[pulumi.Input['PgPgParamArgs']]] params: PostgreSQL connection parameters.
|
|
17546
|
+
:param pulumi.Input[str] password: PostgreSQL admin user password.
|
|
17547
|
+
:param pulumi.Input[int] port: PostgreSQL port.
|
|
17548
|
+
:param pulumi.Input[str] replica_uri: PostgreSQL replica URI for services with a replica.
|
|
17549
|
+
:param pulumi.Input[str] sslmode: PostgreSQL SSL mode setting.
|
|
17550
|
+
:param pulumi.Input[Sequence[pulumi.Input[str]]] standby_uris: PostgreSQL standby connection URIs.
|
|
17551
|
+
:param pulumi.Input[Sequence[pulumi.Input[str]]] syncing_uris: PostgreSQL syncing connection URIs.
|
|
17552
|
+
:param pulumi.Input[str] uri: PostgreSQL primary connection URI.
|
|
17553
|
+
:param pulumi.Input[Sequence[pulumi.Input[str]]] uris: PostgreSQL primary connection URIs.
|
|
17554
|
+
:param pulumi.Input[str] user: PostgreSQL admin user name.
|
|
16829
17555
|
"""
|
|
16830
17556
|
if bouncer is not None:
|
|
16831
17557
|
pulumi.set(__self__, "bouncer", bouncer)
|
|
@@ -16860,7 +17586,7 @@ class PgPgArgs:
|
|
|
16860
17586
|
@pulumi.getter
|
|
16861
17587
|
def bouncer(self) -> Optional[pulumi.Input[str]]:
|
|
16862
17588
|
"""
|
|
16863
|
-
|
|
17589
|
+
PgBouncer connection details for [connection pooling](https://aiven.io/docs/products/postgresql/concepts/pg-connection-pooling).
|
|
16864
17590
|
"""
|
|
16865
17591
|
return pulumi.get(self, "bouncer")
|
|
16866
17592
|
|
|
@@ -16872,7 +17598,7 @@ class PgPgArgs:
|
|
|
16872
17598
|
@pulumi.getter
|
|
16873
17599
|
def dbname(self) -> Optional[pulumi.Input[str]]:
|
|
16874
17600
|
"""
|
|
16875
|
-
Primary PostgreSQL database name
|
|
17601
|
+
Primary PostgreSQL database name.
|
|
16876
17602
|
"""
|
|
16877
17603
|
return pulumi.get(self, "dbname")
|
|
16878
17604
|
|
|
@@ -16884,7 +17610,7 @@ class PgPgArgs:
|
|
|
16884
17610
|
@pulumi.getter
|
|
16885
17611
|
def host(self) -> Optional[pulumi.Input[str]]:
|
|
16886
17612
|
"""
|
|
16887
|
-
PostgreSQL
|
|
17613
|
+
PostgreSQL primary node host IP or name.
|
|
16888
17614
|
"""
|
|
16889
17615
|
return pulumi.get(self, "host")
|
|
16890
17616
|
|
|
@@ -16896,7 +17622,7 @@ class PgPgArgs:
|
|
|
16896
17622
|
@pulumi.getter(name="maxConnections")
|
|
16897
17623
|
def max_connections(self) -> Optional[pulumi.Input[int]]:
|
|
16898
17624
|
"""
|
|
16899
|
-
|
|
17625
|
+
The [number of allowed connections](https://aiven.io/docs/products/postgresql/reference/pg-connection-limits). Varies based on the service plan.
|
|
16900
17626
|
"""
|
|
16901
17627
|
return pulumi.get(self, "max_connections")
|
|
16902
17628
|
|
|
@@ -16908,7 +17634,7 @@ class PgPgArgs:
|
|
|
16908
17634
|
@pulumi.getter
|
|
16909
17635
|
def params(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['PgPgParamArgs']]]]:
|
|
16910
17636
|
"""
|
|
16911
|
-
PostgreSQL connection parameters
|
|
17637
|
+
PostgreSQL connection parameters.
|
|
16912
17638
|
"""
|
|
16913
17639
|
return pulumi.get(self, "params")
|
|
16914
17640
|
|
|
@@ -16920,7 +17646,7 @@ class PgPgArgs:
|
|
|
16920
17646
|
@pulumi.getter
|
|
16921
17647
|
def password(self) -> Optional[pulumi.Input[str]]:
|
|
16922
17648
|
"""
|
|
16923
|
-
PostgreSQL admin user password
|
|
17649
|
+
PostgreSQL admin user password.
|
|
16924
17650
|
"""
|
|
16925
17651
|
return pulumi.get(self, "password")
|
|
16926
17652
|
|
|
@@ -16932,7 +17658,7 @@ class PgPgArgs:
|
|
|
16932
17658
|
@pulumi.getter
|
|
16933
17659
|
def port(self) -> Optional[pulumi.Input[int]]:
|
|
16934
17660
|
"""
|
|
16935
|
-
PostgreSQL port
|
|
17661
|
+
PostgreSQL port.
|
|
16936
17662
|
"""
|
|
16937
17663
|
return pulumi.get(self, "port")
|
|
16938
17664
|
|
|
@@ -16944,7 +17670,7 @@ class PgPgArgs:
|
|
|
16944
17670
|
@pulumi.getter(name="replicaUri")
|
|
16945
17671
|
def replica_uri(self) -> Optional[pulumi.Input[str]]:
|
|
16946
17672
|
"""
|
|
16947
|
-
PostgreSQL replica URI for services with a replica
|
|
17673
|
+
PostgreSQL replica URI for services with a replica.
|
|
16948
17674
|
"""
|
|
16949
17675
|
return pulumi.get(self, "replica_uri")
|
|
16950
17676
|
|
|
@@ -16956,7 +17682,7 @@ class PgPgArgs:
|
|
|
16956
17682
|
@pulumi.getter
|
|
16957
17683
|
def sslmode(self) -> Optional[pulumi.Input[str]]:
|
|
16958
17684
|
"""
|
|
16959
|
-
PostgreSQL
|
|
17685
|
+
PostgreSQL SSL mode setting.
|
|
16960
17686
|
"""
|
|
16961
17687
|
return pulumi.get(self, "sslmode")
|
|
16962
17688
|
|
|
@@ -16968,7 +17694,7 @@ class PgPgArgs:
|
|
|
16968
17694
|
@pulumi.getter(name="standbyUris")
|
|
16969
17695
|
def standby_uris(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
|
|
16970
17696
|
"""
|
|
16971
|
-
PostgreSQL standby connection URIs
|
|
17697
|
+
PostgreSQL standby connection URIs.
|
|
16972
17698
|
"""
|
|
16973
17699
|
return pulumi.get(self, "standby_uris")
|
|
16974
17700
|
|
|
@@ -16980,7 +17706,7 @@ class PgPgArgs:
|
|
|
16980
17706
|
@pulumi.getter(name="syncingUris")
|
|
16981
17707
|
def syncing_uris(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
|
|
16982
17708
|
"""
|
|
16983
|
-
PostgreSQL syncing connection URIs
|
|
17709
|
+
PostgreSQL syncing connection URIs.
|
|
16984
17710
|
"""
|
|
16985
17711
|
return pulumi.get(self, "syncing_uris")
|
|
16986
17712
|
|
|
@@ -16992,7 +17718,7 @@ class PgPgArgs:
|
|
|
16992
17718
|
@pulumi.getter
|
|
16993
17719
|
def uri(self) -> Optional[pulumi.Input[str]]:
|
|
16994
17720
|
"""
|
|
16995
|
-
PostgreSQL
|
|
17721
|
+
PostgreSQL primary connection URI.
|
|
16996
17722
|
"""
|
|
16997
17723
|
return pulumi.get(self, "uri")
|
|
16998
17724
|
|
|
@@ -17004,7 +17730,7 @@ class PgPgArgs:
|
|
|
17004
17730
|
@pulumi.getter
|
|
17005
17731
|
def uris(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
|
|
17006
17732
|
"""
|
|
17007
|
-
PostgreSQL
|
|
17733
|
+
PostgreSQL primary connection URIs.
|
|
17008
17734
|
"""
|
|
17009
17735
|
return pulumi.get(self, "uris")
|
|
17010
17736
|
|
|
@@ -17016,7 +17742,7 @@ class PgPgArgs:
|
|
|
17016
17742
|
@pulumi.getter
|
|
17017
17743
|
def user(self) -> Optional[pulumi.Input[str]]:
|
|
17018
17744
|
"""
|
|
17019
|
-
PostgreSQL admin user name
|
|
17745
|
+
PostgreSQL admin user name.
|
|
17020
17746
|
"""
|
|
17021
17747
|
return pulumi.get(self, "user")
|
|
17022
17748
|
|
|
@@ -17035,12 +17761,12 @@ class PgPgParamArgs:
|
|
|
17035
17761
|
sslmode: Optional[pulumi.Input[str]] = None,
|
|
17036
17762
|
user: Optional[pulumi.Input[str]] = None):
|
|
17037
17763
|
"""
|
|
17038
|
-
:param pulumi.Input[str] database_name: Primary PostgreSQL database name
|
|
17039
|
-
:param pulumi.Input[str] host: PostgreSQL host IP or name
|
|
17040
|
-
:param pulumi.Input[str] password: PostgreSQL admin user password
|
|
17041
|
-
:param pulumi.Input[int] port: PostgreSQL port
|
|
17042
|
-
:param pulumi.Input[str] sslmode: PostgreSQL
|
|
17043
|
-
:param pulumi.Input[str] user: PostgreSQL admin user name
|
|
17764
|
+
:param pulumi.Input[str] database_name: Primary PostgreSQL database name.
|
|
17765
|
+
:param pulumi.Input[str] host: PostgreSQL host IP or name.
|
|
17766
|
+
:param pulumi.Input[str] password: PostgreSQL admin user password.
|
|
17767
|
+
:param pulumi.Input[int] port: PostgreSQL port.
|
|
17768
|
+
:param pulumi.Input[str] sslmode: PostgreSQL SSL mode setting.
|
|
17769
|
+
:param pulumi.Input[str] user: PostgreSQL admin user name.
|
|
17044
17770
|
"""
|
|
17045
17771
|
if database_name is not None:
|
|
17046
17772
|
pulumi.set(__self__, "database_name", database_name)
|
|
@@ -17059,7 +17785,7 @@ class PgPgParamArgs:
|
|
|
17059
17785
|
@pulumi.getter(name="databaseName")
|
|
17060
17786
|
def database_name(self) -> Optional[pulumi.Input[str]]:
|
|
17061
17787
|
"""
|
|
17062
|
-
Primary PostgreSQL database name
|
|
17788
|
+
Primary PostgreSQL database name.
|
|
17063
17789
|
"""
|
|
17064
17790
|
return pulumi.get(self, "database_name")
|
|
17065
17791
|
|
|
@@ -17071,7 +17797,7 @@ class PgPgParamArgs:
|
|
|
17071
17797
|
@pulumi.getter
|
|
17072
17798
|
def host(self) -> Optional[pulumi.Input[str]]:
|
|
17073
17799
|
"""
|
|
17074
|
-
PostgreSQL host IP or name
|
|
17800
|
+
PostgreSQL host IP or name.
|
|
17075
17801
|
"""
|
|
17076
17802
|
return pulumi.get(self, "host")
|
|
17077
17803
|
|
|
@@ -17083,7 +17809,7 @@ class PgPgParamArgs:
|
|
|
17083
17809
|
@pulumi.getter
|
|
17084
17810
|
def password(self) -> Optional[pulumi.Input[str]]:
|
|
17085
17811
|
"""
|
|
17086
|
-
PostgreSQL admin user password
|
|
17812
|
+
PostgreSQL admin user password.
|
|
17087
17813
|
"""
|
|
17088
17814
|
return pulumi.get(self, "password")
|
|
17089
17815
|
|
|
@@ -17095,7 +17821,7 @@ class PgPgParamArgs:
|
|
|
17095
17821
|
@pulumi.getter
|
|
17096
17822
|
def port(self) -> Optional[pulumi.Input[int]]:
|
|
17097
17823
|
"""
|
|
17098
|
-
PostgreSQL port
|
|
17824
|
+
PostgreSQL port.
|
|
17099
17825
|
"""
|
|
17100
17826
|
return pulumi.get(self, "port")
|
|
17101
17827
|
|
|
@@ -17107,7 +17833,7 @@ class PgPgParamArgs:
|
|
|
17107
17833
|
@pulumi.getter
|
|
17108
17834
|
def sslmode(self) -> Optional[pulumi.Input[str]]:
|
|
17109
17835
|
"""
|
|
17110
|
-
PostgreSQL
|
|
17836
|
+
PostgreSQL SSL mode setting.
|
|
17111
17837
|
"""
|
|
17112
17838
|
return pulumi.get(self, "sslmode")
|
|
17113
17839
|
|
|
@@ -17119,7 +17845,7 @@ class PgPgParamArgs:
|
|
|
17119
17845
|
@pulumi.getter
|
|
17120
17846
|
def user(self) -> Optional[pulumi.Input[str]]:
|
|
17121
17847
|
"""
|
|
17122
|
-
PostgreSQL admin user name
|
|
17848
|
+
PostgreSQL admin user name.
|
|
17123
17849
|
"""
|
|
17124
17850
|
return pulumi.get(self, "user")
|
|
17125
17851
|
|
|
@@ -17704,6 +18430,7 @@ class PgPgUserConfigMigrationArgs:
|
|
|
17704
18430
|
port: pulumi.Input[int],
|
|
17705
18431
|
dbname: Optional[pulumi.Input[str]] = None,
|
|
17706
18432
|
ignore_dbs: Optional[pulumi.Input[str]] = None,
|
|
18433
|
+
ignore_roles: Optional[pulumi.Input[str]] = None,
|
|
17707
18434
|
method: Optional[pulumi.Input[str]] = None,
|
|
17708
18435
|
password: Optional[pulumi.Input[str]] = None,
|
|
17709
18436
|
ssl: Optional[pulumi.Input[bool]] = None,
|
|
@@ -17713,6 +18440,7 @@ class PgPgUserConfigMigrationArgs:
|
|
|
17713
18440
|
:param pulumi.Input[int] port: Port number of the server where to migrate data from. Example: `1234`.
|
|
17714
18441
|
:param pulumi.Input[str] dbname: Database name for bootstrapping the initial connection. Example: `defaultdb`.
|
|
17715
18442
|
:param pulumi.Input[str] ignore_dbs: Comma-separated list of databases, which should be ignored during migration (supported by MySQL and PostgreSQL only at the moment). Example: `db1,db2`.
|
|
18443
|
+
:param pulumi.Input[str] ignore_roles: Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
17716
18444
|
:param pulumi.Input[str] method: Enum: `dump`, `replication`. The migration method to be used (currently supported only by Redis, Dragonfly, MySQL and PostgreSQL service types).
|
|
17717
18445
|
:param pulumi.Input[str] password: Password for authentication with the server where to migrate data from. Example: `jjKk45Nnd`.
|
|
17718
18446
|
:param pulumi.Input[bool] ssl: The server where to migrate data from is secured with SSL. Default: `true`.
|
|
@@ -17724,6 +18452,8 @@ class PgPgUserConfigMigrationArgs:
|
|
|
17724
18452
|
pulumi.set(__self__, "dbname", dbname)
|
|
17725
18453
|
if ignore_dbs is not None:
|
|
17726
18454
|
pulumi.set(__self__, "ignore_dbs", ignore_dbs)
|
|
18455
|
+
if ignore_roles is not None:
|
|
18456
|
+
pulumi.set(__self__, "ignore_roles", ignore_roles)
|
|
17727
18457
|
if method is not None:
|
|
17728
18458
|
pulumi.set(__self__, "method", method)
|
|
17729
18459
|
if password is not None:
|
|
@@ -17781,6 +18511,18 @@ class PgPgUserConfigMigrationArgs:
|
|
|
17781
18511
|
def ignore_dbs(self, value: Optional[pulumi.Input[str]]):
|
|
17782
18512
|
pulumi.set(self, "ignore_dbs", value)
|
|
17783
18513
|
|
|
18514
|
+
@property
|
|
18515
|
+
@pulumi.getter(name="ignoreRoles")
|
|
18516
|
+
def ignore_roles(self) -> Optional[pulumi.Input[str]]:
|
|
18517
|
+
"""
|
|
18518
|
+
Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
18519
|
+
"""
|
|
18520
|
+
return pulumi.get(self, "ignore_roles")
|
|
18521
|
+
|
|
18522
|
+
@ignore_roles.setter
|
|
18523
|
+
def ignore_roles(self, value: Optional[pulumi.Input[str]]):
|
|
18524
|
+
pulumi.set(self, "ignore_roles", value)
|
|
18525
|
+
|
|
17784
18526
|
@property
|
|
17785
18527
|
@pulumi.getter
|
|
17786
18528
|
def method(self) -> Optional[pulumi.Input[str]]:
|
|
@@ -19023,6 +19765,7 @@ class PgPgUserConfigPgbouncerArgs:
|
|
|
19023
19765
|
autodb_pool_mode: Optional[pulumi.Input[str]] = None,
|
|
19024
19766
|
autodb_pool_size: Optional[pulumi.Input[int]] = None,
|
|
19025
19767
|
ignore_startup_parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
|
|
19768
|
+
max_prepared_statements: Optional[pulumi.Input[int]] = None,
|
|
19026
19769
|
min_pool_size: Optional[pulumi.Input[int]] = None,
|
|
19027
19770
|
server_idle_timeout: Optional[pulumi.Input[int]] = None,
|
|
19028
19771
|
server_lifetime: Optional[pulumi.Input[int]] = None,
|
|
@@ -19033,6 +19776,7 @@ class PgPgUserConfigPgbouncerArgs:
|
|
|
19033
19776
|
:param pulumi.Input[str] autodb_pool_mode: Enum: `session`, `transaction`, `statement`. PGBouncer pool mode. Default: `transaction`.
|
|
19034
19777
|
:param pulumi.Input[int] autodb_pool_size: If non-zero then create automatically a pool of that size per user when a pool doesn't exist. Default: `0`.
|
|
19035
19778
|
:param pulumi.Input[Sequence[pulumi.Input[str]]] ignore_startup_parameters: List of parameters to ignore when given in startup packet.
|
|
19779
|
+
:param pulumi.Input[int] max_prepared_statements: PgBouncer tracks protocol-level named prepared statements related commands sent by the client in transaction and statement pooling modes when max*prepared*statements is set to a non-zero value. Setting it to 0 disables prepared statements. max*prepared*statements defaults to 100, and its maximum is 3000. Default: `100`.
|
|
19036
19780
|
:param pulumi.Input[int] min_pool_size: Add more server connections to pool if below this number. Improves behavior when usual load comes suddenly back after period of total inactivity. The value is effectively capped at the pool size. Default: `0`.
|
|
19037
19781
|
:param pulumi.Input[int] server_idle_timeout: If a server connection has been idle more than this many seconds it will be dropped. If 0 then timeout is disabled. (seconds). Default: `600`.
|
|
19038
19782
|
:param pulumi.Input[int] server_lifetime: The pooler will close an unused server connection that has been connected longer than this. (seconds). Default: `3600`.
|
|
@@ -19048,6 +19792,8 @@ class PgPgUserConfigPgbouncerArgs:
|
|
|
19048
19792
|
pulumi.set(__self__, "autodb_pool_size", autodb_pool_size)
|
|
19049
19793
|
if ignore_startup_parameters is not None:
|
|
19050
19794
|
pulumi.set(__self__, "ignore_startup_parameters", ignore_startup_parameters)
|
|
19795
|
+
if max_prepared_statements is not None:
|
|
19796
|
+
pulumi.set(__self__, "max_prepared_statements", max_prepared_statements)
|
|
19051
19797
|
if min_pool_size is not None:
|
|
19052
19798
|
pulumi.set(__self__, "min_pool_size", min_pool_size)
|
|
19053
19799
|
if server_idle_timeout is not None:
|
|
@@ -19117,6 +19863,18 @@ class PgPgUserConfigPgbouncerArgs:
|
|
|
19117
19863
|
def ignore_startup_parameters(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
|
|
19118
19864
|
pulumi.set(self, "ignore_startup_parameters", value)
|
|
19119
19865
|
|
|
19866
|
+
@property
|
|
19867
|
+
@pulumi.getter(name="maxPreparedStatements")
|
|
19868
|
+
def max_prepared_statements(self) -> Optional[pulumi.Input[int]]:
|
|
19869
|
+
"""
|
|
19870
|
+
PgBouncer tracks protocol-level named prepared statements related commands sent by the client in transaction and statement pooling modes when max*prepared*statements is set to a non-zero value. Setting it to 0 disables prepared statements. max*prepared*statements defaults to 100, and its maximum is 3000. Default: `100`.
|
|
19871
|
+
"""
|
|
19872
|
+
return pulumi.get(self, "max_prepared_statements")
|
|
19873
|
+
|
|
19874
|
+
@max_prepared_statements.setter
|
|
19875
|
+
def max_prepared_statements(self, value: Optional[pulumi.Input[int]]):
|
|
19876
|
+
pulumi.set(self, "max_prepared_statements", value)
|
|
19877
|
+
|
|
19120
19878
|
@property
|
|
19121
19879
|
@pulumi.getter(name="minPoolSize")
|
|
19122
19880
|
def min_pool_size(self) -> Optional[pulumi.Input[int]]:
|
|
@@ -20172,6 +20930,7 @@ class RedisRedisUserConfigMigrationArgs:
|
|
|
20172
20930
|
port: pulumi.Input[int],
|
|
20173
20931
|
dbname: Optional[pulumi.Input[str]] = None,
|
|
20174
20932
|
ignore_dbs: Optional[pulumi.Input[str]] = None,
|
|
20933
|
+
ignore_roles: Optional[pulumi.Input[str]] = None,
|
|
20175
20934
|
method: Optional[pulumi.Input[str]] = None,
|
|
20176
20935
|
password: Optional[pulumi.Input[str]] = None,
|
|
20177
20936
|
ssl: Optional[pulumi.Input[bool]] = None,
|
|
@@ -20181,6 +20940,7 @@ class RedisRedisUserConfigMigrationArgs:
|
|
|
20181
20940
|
:param pulumi.Input[int] port: Port number of the server where to migrate data from. Example: `1234`.
|
|
20182
20941
|
:param pulumi.Input[str] dbname: Database name for bootstrapping the initial connection. Example: `defaultdb`.
|
|
20183
20942
|
:param pulumi.Input[str] ignore_dbs: Comma-separated list of databases, which should be ignored during migration (supported by MySQL and PostgreSQL only at the moment). Example: `db1,db2`.
|
|
20943
|
+
:param pulumi.Input[str] ignore_roles: Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
20184
20944
|
:param pulumi.Input[str] method: Enum: `dump`, `replication`. The migration method to be used (currently supported only by Redis, Dragonfly, MySQL and PostgreSQL service types).
|
|
20185
20945
|
:param pulumi.Input[str] password: Password for authentication with the server where to migrate data from. Example: `jjKk45Nnd`.
|
|
20186
20946
|
:param pulumi.Input[bool] ssl: The server where to migrate data from is secured with SSL. Default: `true`.
|
|
@@ -20192,6 +20952,8 @@ class RedisRedisUserConfigMigrationArgs:
|
|
|
20192
20952
|
pulumi.set(__self__, "dbname", dbname)
|
|
20193
20953
|
if ignore_dbs is not None:
|
|
20194
20954
|
pulumi.set(__self__, "ignore_dbs", ignore_dbs)
|
|
20955
|
+
if ignore_roles is not None:
|
|
20956
|
+
pulumi.set(__self__, "ignore_roles", ignore_roles)
|
|
20195
20957
|
if method is not None:
|
|
20196
20958
|
pulumi.set(__self__, "method", method)
|
|
20197
20959
|
if password is not None:
|
|
@@ -20249,6 +21011,18 @@ class RedisRedisUserConfigMigrationArgs:
|
|
|
20249
21011
|
def ignore_dbs(self, value: Optional[pulumi.Input[str]]):
|
|
20250
21012
|
pulumi.set(self, "ignore_dbs", value)
|
|
20251
21013
|
|
|
21014
|
+
@property
|
|
21015
|
+
@pulumi.getter(name="ignoreRoles")
|
|
21016
|
+
def ignore_roles(self) -> Optional[pulumi.Input[str]]:
|
|
21017
|
+
"""
|
|
21018
|
+
Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
21019
|
+
"""
|
|
21020
|
+
return pulumi.get(self, "ignore_roles")
|
|
21021
|
+
|
|
21022
|
+
@ignore_roles.setter
|
|
21023
|
+
def ignore_roles(self, value: Optional[pulumi.Input[str]]):
|
|
21024
|
+
pulumi.set(self, "ignore_roles", value)
|
|
21025
|
+
|
|
20252
21026
|
@property
|
|
20253
21027
|
@pulumi.getter
|
|
20254
21028
|
def method(self) -> Optional[pulumi.Input[str]]:
|
|
@@ -20549,7 +21323,9 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableArgs:
|
|
|
20549
21323
|
max_rows_per_message: Optional[pulumi.Input[int]] = None,
|
|
20550
21324
|
num_consumers: Optional[pulumi.Input[int]] = None,
|
|
20551
21325
|
poll_max_batch_size: Optional[pulumi.Input[int]] = None,
|
|
20552
|
-
|
|
21326
|
+
poll_max_timeout_ms: Optional[pulumi.Input[int]] = None,
|
|
21327
|
+
skip_broken_messages: Optional[pulumi.Input[int]] = None,
|
|
21328
|
+
thread_per_consumer: Optional[pulumi.Input[bool]] = None):
|
|
20553
21329
|
"""
|
|
20554
21330
|
:param pulumi.Input[Sequence[pulumi.Input['ServiceIntegrationClickhouseKafkaUserConfigTableColumnArgs']]] columns: Table columns
|
|
20555
21331
|
:param pulumi.Input[str] data_format: Enum: `Avro`, `CSV`, `JSONAsString`, `JSONCompactEachRow`, `JSONCompactStringsEachRow`, `JSONEachRow`, `JSONStringsEachRow`, `MsgPack`, `TSKV`, `TSV`, `TabSeparated`, `RawBLOB`, `AvroConfluent`, `Parquet`. Message data format. Default: `JSONEachRow`.
|
|
@@ -20563,7 +21339,9 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableArgs:
|
|
|
20563
21339
|
:param pulumi.Input[int] max_rows_per_message: The maximum number of rows produced in one kafka message for row-based formats. Default: `1`.
|
|
20564
21340
|
:param pulumi.Input[int] num_consumers: The number of consumers per table per replica. Default: `1`.
|
|
20565
21341
|
:param pulumi.Input[int] poll_max_batch_size: Maximum amount of messages to be polled in a single Kafka poll. Default: `0`.
|
|
21342
|
+
:param pulumi.Input[int] poll_max_timeout_ms: Timeout in milliseconds for a single poll from Kafka. Takes the value of the stream*flush*interval_ms server setting by default (500ms). Default: `0`.
|
|
20566
21343
|
:param pulumi.Input[int] skip_broken_messages: Skip at least this number of broken messages from Kafka topic per block. Default: `0`.
|
|
21344
|
+
:param pulumi.Input[bool] thread_per_consumer: Provide an independent thread for each consumer. All consumers run in the same thread by default. Default: `false`.
|
|
20567
21345
|
"""
|
|
20568
21346
|
pulumi.set(__self__, "columns", columns)
|
|
20569
21347
|
pulumi.set(__self__, "data_format", data_format)
|
|
@@ -20584,8 +21362,12 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableArgs:
|
|
|
20584
21362
|
pulumi.set(__self__, "num_consumers", num_consumers)
|
|
20585
21363
|
if poll_max_batch_size is not None:
|
|
20586
21364
|
pulumi.set(__self__, "poll_max_batch_size", poll_max_batch_size)
|
|
21365
|
+
if poll_max_timeout_ms is not None:
|
|
21366
|
+
pulumi.set(__self__, "poll_max_timeout_ms", poll_max_timeout_ms)
|
|
20587
21367
|
if skip_broken_messages is not None:
|
|
20588
21368
|
pulumi.set(__self__, "skip_broken_messages", skip_broken_messages)
|
|
21369
|
+
if thread_per_consumer is not None:
|
|
21370
|
+
pulumi.set(__self__, "thread_per_consumer", thread_per_consumer)
|
|
20589
21371
|
|
|
20590
21372
|
@property
|
|
20591
21373
|
@pulumi.getter
|
|
@@ -20731,6 +21513,18 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableArgs:
|
|
|
20731
21513
|
def poll_max_batch_size(self, value: Optional[pulumi.Input[int]]):
|
|
20732
21514
|
pulumi.set(self, "poll_max_batch_size", value)
|
|
20733
21515
|
|
|
21516
|
+
@property
|
|
21517
|
+
@pulumi.getter(name="pollMaxTimeoutMs")
|
|
21518
|
+
def poll_max_timeout_ms(self) -> Optional[pulumi.Input[int]]:
|
|
21519
|
+
"""
|
|
21520
|
+
Timeout in milliseconds for a single poll from Kafka. Takes the value of the stream*flush*interval_ms server setting by default (500ms). Default: `0`.
|
|
21521
|
+
"""
|
|
21522
|
+
return pulumi.get(self, "poll_max_timeout_ms")
|
|
21523
|
+
|
|
21524
|
+
@poll_max_timeout_ms.setter
|
|
21525
|
+
def poll_max_timeout_ms(self, value: Optional[pulumi.Input[int]]):
|
|
21526
|
+
pulumi.set(self, "poll_max_timeout_ms", value)
|
|
21527
|
+
|
|
20734
21528
|
@property
|
|
20735
21529
|
@pulumi.getter(name="skipBrokenMessages")
|
|
20736
21530
|
def skip_broken_messages(self) -> Optional[pulumi.Input[int]]:
|
|
@@ -20743,6 +21537,18 @@ class ServiceIntegrationClickhouseKafkaUserConfigTableArgs:
|
|
|
20743
21537
|
def skip_broken_messages(self, value: Optional[pulumi.Input[int]]):
|
|
20744
21538
|
pulumi.set(self, "skip_broken_messages", value)
|
|
20745
21539
|
|
|
21540
|
+
@property
|
|
21541
|
+
@pulumi.getter(name="threadPerConsumer")
|
|
21542
|
+
def thread_per_consumer(self) -> Optional[pulumi.Input[bool]]:
|
|
21543
|
+
"""
|
|
21544
|
+
Provide an independent thread for each consumer. All consumers run in the same thread by default. Default: `false`.
|
|
21545
|
+
"""
|
|
21546
|
+
return pulumi.get(self, "thread_per_consumer")
|
|
21547
|
+
|
|
21548
|
+
@thread_per_consumer.setter
|
|
21549
|
+
def thread_per_consumer(self, value: Optional[pulumi.Input[bool]]):
|
|
21550
|
+
pulumi.set(self, "thread_per_consumer", value)
|
|
21551
|
+
|
|
20746
21552
|
|
|
20747
21553
|
@pulumi.input_type
|
|
20748
21554
|
class ServiceIntegrationClickhouseKafkaUserConfigTableColumnArgs:
|
|
@@ -25083,6 +25889,7 @@ class ValkeyValkeyUserConfigMigrationArgs:
|
|
|
25083
25889
|
port: pulumi.Input[int],
|
|
25084
25890
|
dbname: Optional[pulumi.Input[str]] = None,
|
|
25085
25891
|
ignore_dbs: Optional[pulumi.Input[str]] = None,
|
|
25892
|
+
ignore_roles: Optional[pulumi.Input[str]] = None,
|
|
25086
25893
|
method: Optional[pulumi.Input[str]] = None,
|
|
25087
25894
|
password: Optional[pulumi.Input[str]] = None,
|
|
25088
25895
|
ssl: Optional[pulumi.Input[bool]] = None,
|
|
@@ -25092,6 +25899,7 @@ class ValkeyValkeyUserConfigMigrationArgs:
|
|
|
25092
25899
|
:param pulumi.Input[int] port: Port number of the server where to migrate data from. Example: `1234`.
|
|
25093
25900
|
:param pulumi.Input[str] dbname: Database name for bootstrapping the initial connection. Example: `defaultdb`.
|
|
25094
25901
|
:param pulumi.Input[str] ignore_dbs: Comma-separated list of databases, which should be ignored during migration (supported by MySQL and PostgreSQL only at the moment). Example: `db1,db2`.
|
|
25902
|
+
:param pulumi.Input[str] ignore_roles: Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
25095
25903
|
:param pulumi.Input[str] method: Enum: `dump`, `replication`. The migration method to be used (currently supported only by Redis, Dragonfly, MySQL and PostgreSQL service types).
|
|
25096
25904
|
:param pulumi.Input[str] password: Password for authentication with the server where to migrate data from. Example: `jjKk45Nnd`.
|
|
25097
25905
|
:param pulumi.Input[bool] ssl: The server where to migrate data from is secured with SSL. Default: `true`.
|
|
@@ -25103,6 +25911,8 @@ class ValkeyValkeyUserConfigMigrationArgs:
|
|
|
25103
25911
|
pulumi.set(__self__, "dbname", dbname)
|
|
25104
25912
|
if ignore_dbs is not None:
|
|
25105
25913
|
pulumi.set(__self__, "ignore_dbs", ignore_dbs)
|
|
25914
|
+
if ignore_roles is not None:
|
|
25915
|
+
pulumi.set(__self__, "ignore_roles", ignore_roles)
|
|
25106
25916
|
if method is not None:
|
|
25107
25917
|
pulumi.set(__self__, "method", method)
|
|
25108
25918
|
if password is not None:
|
|
@@ -25160,6 +25970,18 @@ class ValkeyValkeyUserConfigMigrationArgs:
|
|
|
25160
25970
|
def ignore_dbs(self, value: Optional[pulumi.Input[str]]):
|
|
25161
25971
|
pulumi.set(self, "ignore_dbs", value)
|
|
25162
25972
|
|
|
25973
|
+
@property
|
|
25974
|
+
@pulumi.getter(name="ignoreRoles")
|
|
25975
|
+
def ignore_roles(self) -> Optional[pulumi.Input[str]]:
|
|
25976
|
+
"""
|
|
25977
|
+
Comma-separated list of database roles, which should be ignored during migration (supported by PostgreSQL only at the moment). Example: `role1,role2`.
|
|
25978
|
+
"""
|
|
25979
|
+
return pulumi.get(self, "ignore_roles")
|
|
25980
|
+
|
|
25981
|
+
@ignore_roles.setter
|
|
25982
|
+
def ignore_roles(self, value: Optional[pulumi.Input[str]]):
|
|
25983
|
+
pulumi.set(self, "ignore_roles", value)
|
|
25984
|
+
|
|
25163
25985
|
@property
|
|
25164
25986
|
@pulumi.getter
|
|
25165
25987
|
def method(self) -> Optional[pulumi.Input[str]]:
|