pulumi-gcp 8.1.0a1726293903__py3-none-any.whl → 8.2.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- pulumi_gcp/__init__.py +62 -0
- pulumi_gcp/alloydb/_inputs.py +94 -0
- pulumi_gcp/alloydb/cluster.py +94 -1
- pulumi_gcp/alloydb/outputs.py +79 -0
- pulumi_gcp/assuredworkloads/workload.py +135 -16
- pulumi_gcp/backupdisasterrecovery/__init__.py +1 -0
- pulumi_gcp/backupdisasterrecovery/backup_vault.py +1203 -0
- pulumi_gcp/bigquery/__init__.py +1 -0
- pulumi_gcp/bigquery/_inputs.py +33 -0
- pulumi_gcp/bigquery/data_transfer_config.py +134 -6
- pulumi_gcp/bigquery/get_tables.py +143 -0
- pulumi_gcp/bigquery/outputs.py +66 -0
- pulumi_gcp/bigqueryanalyticshub/_inputs.py +169 -7
- pulumi_gcp/bigqueryanalyticshub/data_exchange.py +96 -3
- pulumi_gcp/bigqueryanalyticshub/listing.py +114 -0
- pulumi_gcp/bigqueryanalyticshub/outputs.py +138 -5
- pulumi_gcp/bigtable/_inputs.py +21 -1
- pulumi_gcp/bigtable/outputs.py +13 -1
- pulumi_gcp/bigtable/table.py +34 -0
- pulumi_gcp/certificateauthority/authority.py +14 -7
- pulumi_gcp/certificatemanager/__init__.py +1 -0
- pulumi_gcp/certificatemanager/certificate.py +28 -0
- pulumi_gcp/certificatemanager/get_certificates.py +150 -0
- pulumi_gcp/certificatemanager/outputs.py +322 -0
- pulumi_gcp/cloudbuild/_inputs.py +6 -6
- pulumi_gcp/cloudbuild/outputs.py +4 -4
- pulumi_gcp/cloudrun/_inputs.py +6 -12
- pulumi_gcp/cloudrun/outputs.py +8 -16
- pulumi_gcp/cloudrunv2/_inputs.py +72 -15
- pulumi_gcp/cloudrunv2/outputs.py +82 -16
- pulumi_gcp/cloudrunv2/service.py +50 -4
- pulumi_gcp/cloudtasks/_inputs.py +630 -0
- pulumi_gcp/cloudtasks/outputs.py +479 -0
- pulumi_gcp/cloudtasks/queue.py +238 -0
- pulumi_gcp/compute/__init__.py +1 -0
- pulumi_gcp/compute/_inputs.py +2300 -257
- pulumi_gcp/compute/get_instance.py +4 -1
- pulumi_gcp/compute/health_check.py +114 -0
- pulumi_gcp/compute/instance.py +86 -4
- pulumi_gcp/compute/instance_template.py +66 -0
- pulumi_gcp/compute/interconnect.py +28 -21
- pulumi_gcp/compute/network_firewall_policy_with_rules.py +826 -0
- pulumi_gcp/compute/node_template.py +93 -0
- pulumi_gcp/compute/outputs.py +1491 -6
- pulumi_gcp/compute/region_target_http_proxy.py +159 -0
- pulumi_gcp/compute/region_target_https_proxy.py +175 -0
- pulumi_gcp/compute/service_attachment.py +75 -0
- pulumi_gcp/compute/target_http_proxy.py +49 -28
- pulumi_gcp/compute/target_https_proxy.py +77 -28
- pulumi_gcp/config/__init__.pyi +4 -0
- pulumi_gcp/config/vars.py +8 -0
- pulumi_gcp/container/_inputs.py +236 -3
- pulumi_gcp/container/attached_cluster.py +61 -8
- pulumi_gcp/container/outputs.py +276 -2
- pulumi_gcp/databasemigrationservice/_inputs.py +176 -148
- pulumi_gcp/databasemigrationservice/connection_profile.py +206 -0
- pulumi_gcp/databasemigrationservice/outputs.py +109 -87
- pulumi_gcp/dataloss/_inputs.py +353 -1
- pulumi_gcp/dataloss/outputs.py +274 -3
- pulumi_gcp/dataproc/_inputs.py +27 -27
- pulumi_gcp/dataproc/metastore_federation.py +8 -8
- pulumi_gcp/dataproc/metastore_service.py +2 -0
- pulumi_gcp/dataproc/outputs.py +18 -18
- pulumi_gcp/datastream/_inputs.py +69 -1
- pulumi_gcp/datastream/outputs.py +44 -2
- pulumi_gcp/datastream/stream.py +201 -7
- pulumi_gcp/developerconnect/__init__.py +11 -0
- pulumi_gcp/developerconnect/_inputs.py +301 -0
- pulumi_gcp/developerconnect/connection.py +1034 -0
- pulumi_gcp/developerconnect/git_repository_link.py +873 -0
- pulumi_gcp/developerconnect/outputs.py +247 -0
- pulumi_gcp/discoveryengine/_inputs.py +188 -0
- pulumi_gcp/discoveryengine/data_store.py +14 -14
- pulumi_gcp/discoveryengine/outputs.py +153 -1
- pulumi_gcp/firebase/database_instance.py +7 -7
- pulumi_gcp/gkehub/_inputs.py +25 -1
- pulumi_gcp/gkehub/feature_membership.py +12 -6
- pulumi_gcp/gkehub/outputs.py +17 -1
- pulumi_gcp/gkeonprem/_inputs.py +3 -3
- pulumi_gcp/gkeonprem/outputs.py +2 -2
- pulumi_gcp/iam/_inputs.py +196 -0
- pulumi_gcp/iam/get_workload_identity_pool_provider.py +13 -3
- pulumi_gcp/iam/outputs.py +295 -0
- pulumi_gcp/iam/workload_identity_pool_provider.py +164 -3
- pulumi_gcp/kms/__init__.py +2 -0
- pulumi_gcp/kms/autokey_config.py +10 -2
- pulumi_gcp/kms/get_crypto_key_latest_version.py +222 -0
- pulumi_gcp/kms/get_crypto_key_versions.py +175 -0
- pulumi_gcp/kms/outputs.py +164 -0
- pulumi_gcp/memorystore/__init__.py +10 -0
- pulumi_gcp/memorystore/_inputs.py +731 -0
- pulumi_gcp/memorystore/instance.py +1663 -0
- pulumi_gcp/memorystore/outputs.py +598 -0
- pulumi_gcp/netapp/active_directory.py +6 -6
- pulumi_gcp/netapp/backup.py +6 -6
- pulumi_gcp/netapp/backup_policy.py +6 -6
- pulumi_gcp/netapp/backup_vault.py +6 -6
- pulumi_gcp/netapp/storage_pool.py +4 -4
- pulumi_gcp/netapp/volume.py +108 -0
- pulumi_gcp/networkconnectivity/_inputs.py +21 -1
- pulumi_gcp/networkconnectivity/outputs.py +15 -1
- pulumi_gcp/networkconnectivity/spoke.py +8 -0
- pulumi_gcp/networksecurity/client_tls_policy.py +24 -22
- pulumi_gcp/networksecurity/server_tls_policy.py +20 -32
- pulumi_gcp/organizations/folder.py +52 -33
- pulumi_gcp/organizations/get_project.py +13 -3
- pulumi_gcp/organizations/project.py +88 -3
- pulumi_gcp/parallelstore/instance.py +121 -121
- pulumi_gcp/projects/iam_member_remove.py +26 -0
- pulumi_gcp/projects/usage_export_bucket.py +38 -0
- pulumi_gcp/provider.py +40 -0
- pulumi_gcp/pubsub/_inputs.py +40 -0
- pulumi_gcp/pubsub/outputs.py +51 -1
- pulumi_gcp/pubsub/subscription.py +6 -0
- pulumi_gcp/pulumi-plugin.json +1 -1
- pulumi_gcp/redis/_inputs.py +419 -0
- pulumi_gcp/redis/cluster.py +123 -0
- pulumi_gcp/redis/outputs.py +315 -0
- pulumi_gcp/securitycenter/__init__.py +2 -0
- pulumi_gcp/securitycenter/v2_folder_scc_big_query_export.py +857 -0
- pulumi_gcp/securitycenter/v2_organization_scc_big_query_exports.py +4 -4
- pulumi_gcp/securitycenter/v2_project_mute_config.py +2 -2
- pulumi_gcp/securitycenter/v2_project_scc_big_query_export.py +796 -0
- pulumi_gcp/vpcaccess/connector.py +21 -28
- {pulumi_gcp-8.1.0a1726293903.dist-info → pulumi_gcp-8.2.0.dist-info}/METADATA +1 -1
- {pulumi_gcp-8.1.0a1726293903.dist-info → pulumi_gcp-8.2.0.dist-info}/RECORD +128 -111
- {pulumi_gcp-8.1.0a1726293903.dist-info → pulumi_gcp-8.2.0.dist-info}/WHEEL +1 -1
- {pulumi_gcp-8.1.0a1726293903.dist-info → pulumi_gcp-8.2.0.dist-info}/top_level.txt +0 -0
@@ -373,7 +373,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
373
373
|
import pulumi_gcp as gcp
|
374
374
|
|
375
375
|
default_metastore_service = gcp.dataproc.MetastoreService("default",
|
376
|
-
service_id="",
|
376
|
+
service_id="metastore-service",
|
377
377
|
location="us-central1",
|
378
378
|
tier="DEVELOPER",
|
379
379
|
hive_metastore_config={
|
@@ -382,7 +382,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
382
382
|
})
|
383
383
|
default = gcp.dataproc.MetastoreFederation("default",
|
384
384
|
location="us-central1",
|
385
|
-
federation_id="",
|
385
|
+
federation_id="metastore-fed",
|
386
386
|
version="3.1.2",
|
387
387
|
backend_metastores=[{
|
388
388
|
"rank": "1",
|
@@ -397,7 +397,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
397
397
|
import pulumi_gcp as gcp
|
398
398
|
|
399
399
|
default_metastore_service = gcp.dataproc.MetastoreService("default",
|
400
|
-
service_id="",
|
400
|
+
service_id="metastore-service",
|
401
401
|
location="us-central1",
|
402
402
|
tier="DEVELOPER",
|
403
403
|
hive_metastore_config={
|
@@ -407,7 +407,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
407
407
|
project = gcp.organizations.get_project()
|
408
408
|
default = gcp.dataproc.MetastoreFederation("default",
|
409
409
|
location="us-central1",
|
410
|
-
federation_id="",
|
410
|
+
federation_id="metastore-fed",
|
411
411
|
version="3.1.2",
|
412
412
|
backend_metastores=[
|
413
413
|
{
|
@@ -478,7 +478,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
478
478
|
import pulumi_gcp as gcp
|
479
479
|
|
480
480
|
default_metastore_service = gcp.dataproc.MetastoreService("default",
|
481
|
-
service_id="",
|
481
|
+
service_id="metastore-service",
|
482
482
|
location="us-central1",
|
483
483
|
tier="DEVELOPER",
|
484
484
|
hive_metastore_config={
|
@@ -487,7 +487,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
487
487
|
})
|
488
488
|
default = gcp.dataproc.MetastoreFederation("default",
|
489
489
|
location="us-central1",
|
490
|
-
federation_id="",
|
490
|
+
federation_id="metastore-fed",
|
491
491
|
version="3.1.2",
|
492
492
|
backend_metastores=[{
|
493
493
|
"rank": "1",
|
@@ -502,7 +502,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
502
502
|
import pulumi_gcp as gcp
|
503
503
|
|
504
504
|
default_metastore_service = gcp.dataproc.MetastoreService("default",
|
505
|
-
service_id="",
|
505
|
+
service_id="metastore-service",
|
506
506
|
location="us-central1",
|
507
507
|
tier="DEVELOPER",
|
508
508
|
hive_metastore_config={
|
@@ -512,7 +512,7 @@ class MetastoreFederation(pulumi.CustomResource):
|
|
512
512
|
project = gcp.organizations.get_project()
|
513
513
|
default = gcp.dataproc.MetastoreFederation("default",
|
514
514
|
location="us-central1",
|
515
|
-
federation_id="",
|
515
|
+
federation_id="metastore-fed",
|
516
516
|
version="3.1.2",
|
517
517
|
backend_metastores=[
|
518
518
|
{
|
@@ -908,6 +908,7 @@ class MetastoreService(pulumi.CustomResource):
|
|
908
908
|
default = gcp.dataproc.MetastoreService("default",
|
909
909
|
service_id="metastore-srv",
|
910
910
|
location="us-central1",
|
911
|
+
tier="DEVELOPER",
|
911
912
|
hive_metastore_config={
|
912
913
|
"version": "3.1.2",
|
913
914
|
},
|
@@ -1247,6 +1248,7 @@ class MetastoreService(pulumi.CustomResource):
|
|
1247
1248
|
default = gcp.dataproc.MetastoreService("default",
|
1248
1249
|
service_id="metastore-srv",
|
1249
1250
|
location="us-central1",
|
1251
|
+
tier="DEVELOPER",
|
1250
1252
|
hive_metastore_config={
|
1251
1253
|
"version": "3.1.2",
|
1252
1254
|
},
|
pulumi_gcp/dataproc/outputs.py
CHANGED
@@ -8232,7 +8232,7 @@ class WorkflowTemplatePlacementManagedClusterConfig(dict):
|
|
8232
8232
|
:param 'WorkflowTemplatePlacementManagedClusterConfigSecondaryWorkerConfigArgs' secondary_worker_config: The Compute Engine config settings for additional worker instances in a cluster.
|
8233
8233
|
:param 'WorkflowTemplatePlacementManagedClusterConfigSecurityConfigArgs' security_config: Security settings for the cluster.
|
8234
8234
|
:param 'WorkflowTemplatePlacementManagedClusterConfigSoftwareConfigArgs' software_config: The config settings for software inside the cluster.
|
8235
|
-
:param str staging_bucket: A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see (https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
8235
|
+
:param str staging_bucket: A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see [Dataproc staging and temp buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
8236
8236
|
:param str temp_bucket: A Cloud Storage bucket used to store ephemeral cluster and jobs data, such as Spark and MapReduce history files. If you do not specify a temp bucket, Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's temp bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket. The default bucket has a TTL of 90 days, but you can use any TTL (or none) if you specify a bucket.
|
8237
8237
|
:param 'WorkflowTemplatePlacementManagedClusterConfigWorkerConfigArgs' worker_config: The Compute Engine config settings for additional worker instances in a cluster.
|
8238
8238
|
|
@@ -8369,7 +8369,7 @@ class WorkflowTemplatePlacementManagedClusterConfig(dict):
|
|
8369
8369
|
@pulumi.getter(name="stagingBucket")
|
8370
8370
|
def staging_bucket(self) -> Optional[str]:
|
8371
8371
|
"""
|
8372
|
-
A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see (https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
8372
|
+
A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see [Dataproc staging and temp buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
8373
8373
|
"""
|
8374
8374
|
return pulumi.get(self, "staging_bucket")
|
8375
8375
|
|
@@ -8543,7 +8543,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8543
8543
|
zone: Optional[str] = None):
|
8544
8544
|
"""
|
8545
8545
|
:param bool internal_ip_only: If true, all instances in the cluster will only have internal IP addresses. By default, clusters are not restricted to internal IP addresses, and will have ephemeral external IP addresses assigned to each instance. This `internal_ip_only` restriction can only be enabled for subnetwork enabled networks, and all off-cluster dependencies must be configured to be accessible without external IP addresses.
|
8546
|
-
:param Mapping[str, str] metadata: The Compute Engine metadata entries to add to all instances (see (https://cloud.google.com/compute/docs/
|
8546
|
+
:param Mapping[str, str] metadata: The Compute Engine metadata entries to add to all instances (see [About VM metadata](https://cloud.google.com/compute/docs/metadata/overview)).
|
8547
8547
|
:param str network: The Compute Engine network to be used for machine communications. Cannot be specified with subnetwork_uri. If neither `network_uri` nor `subnetwork_uri` is specified, the "default" network of the project is used, if it exists. Cannot be a "Custom Subnet Network" (see /regions/global/default` * `default`
|
8548
8548
|
:param 'WorkflowTemplatePlacementManagedClusterConfigGceClusterConfigNodeGroupAffinityArgs' node_group_affinity: Node Group Affinity for sole-tenant clusters.
|
8549
8549
|
:param str private_ipv6_google_access: The type of IPv6 access for a cluster. Possible values: PRIVATE_IPV6_GOOGLE_ACCESS_UNSPECIFIED, INHERIT_FROM_SUBNETWORK, OUTBOUND, BIDIRECTIONAL
|
@@ -8552,7 +8552,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8552
8552
|
:param Sequence[str] service_account_scopes: The URIs of service account scopes to be included in Compute Engine instances. The following base set of scopes is always included: * https://www.googleapis.com/auth/cloud.useraccounts.readonly * https://www.googleapis.com/auth/devstorage.read_write * https://www.googleapis.com/auth/logging.write If no scopes are specified, the following defaults are also provided: * https://www.googleapis.com/auth/bigquery * https://www.googleapis.com/auth/bigtable.admin.table * https://www.googleapis.com/auth/bigtable.data * https://www.googleapis.com/auth/devstorage.full_control
|
8553
8553
|
:param 'WorkflowTemplatePlacementManagedClusterConfigGceClusterConfigShieldedInstanceConfigArgs' shielded_instance_config: Shielded Instance Config for clusters using [Compute Engine Shielded VMs](https://cloud.google.com/security/shielded-cloud/shielded-vm). Structure defined below.
|
8554
8554
|
:param str subnetwork: The Compute Engine subnetwork to be used for machine communications. Cannot be specified with network_uri. A full URL, partial URI, or short name are valid. Examples: * `https://www.googleapis.com/compute/v1/projects//regions/us-east1/subnetworks/sub0` * `sub0`
|
8555
|
-
:param Sequence[str] tags: The Compute Engine tags to add to all instances (see (https://cloud.google.com/compute/docs/
|
8555
|
+
:param Sequence[str] tags: The Compute Engine tags to add to all instances (see [Manage tags for resources](https://cloud.google.com/compute/docs/tag-resources)).
|
8556
8556
|
:param str zone: The zone where the Compute Engine cluster will be located. On a create request, it is required in the "global" region. If omitted in a non-global Dataproc region, the service will pick a zone in the corresponding Compute Engine region. On a get request, zone will always be present. A full URL, partial URI, or short name are valid. Examples: * `https://www.googleapis.com/compute/v1/projects/` * `us-central1-f`
|
8557
8557
|
"""
|
8558
8558
|
if internal_ip_only is not None:
|
@@ -8592,7 +8592,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8592
8592
|
@pulumi.getter
|
8593
8593
|
def metadata(self) -> Optional[Mapping[str, str]]:
|
8594
8594
|
"""
|
8595
|
-
The Compute Engine metadata entries to add to all instances (see (https://cloud.google.com/compute/docs/
|
8595
|
+
The Compute Engine metadata entries to add to all instances (see [About VM metadata](https://cloud.google.com/compute/docs/metadata/overview)).
|
8596
8596
|
"""
|
8597
8597
|
return pulumi.get(self, "metadata")
|
8598
8598
|
|
@@ -8664,7 +8664,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8664
8664
|
@pulumi.getter
|
8665
8665
|
def tags(self) -> Optional[Sequence[str]]:
|
8666
8666
|
"""
|
8667
|
-
The Compute Engine tags to add to all instances (see (https://cloud.google.com/compute/docs/
|
8667
|
+
The Compute Engine tags to add to all instances (see [Manage tags for resources](https://cloud.google.com/compute/docs/tag-resources)).
|
8668
8668
|
"""
|
8669
8669
|
return pulumi.get(self, "tags")
|
8670
8670
|
|
@@ -8948,7 +8948,7 @@ class WorkflowTemplatePlacementManagedClusterConfigInitializationAction(dict):
|
|
8948
8948
|
execution_timeout: Optional[str] = None):
|
8949
8949
|
"""
|
8950
8950
|
:param str executable_file: Required. Cloud Storage URI of executable file.
|
8951
|
-
:param str execution_timeout: Amount of time executable has to complete. Default is 10 minutes (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)). Cluster creation fails with an explanatory error message (the name of the executable that caused the error and the exceeded timeout period) if the executable is not completed at end of the timeout period.
|
8951
|
+
:param str execution_timeout: Amount of time executable has to complete. Default is 10 minutes (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)). Cluster creation fails with an explanatory error message (the name of the executable that caused the error and the exceeded timeout period) if the executable is not completed at end of the timeout period.
|
8952
8952
|
"""
|
8953
8953
|
if executable_file is not None:
|
8954
8954
|
pulumi.set(__self__, "executable_file", executable_file)
|
@@ -8967,7 +8967,7 @@ class WorkflowTemplatePlacementManagedClusterConfigInitializationAction(dict):
|
|
8967
8967
|
@pulumi.getter(name="executionTimeout")
|
8968
8968
|
def execution_timeout(self) -> Optional[str]:
|
8969
8969
|
"""
|
8970
|
-
Amount of time executable has to complete. Default is 10 minutes (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)). Cluster creation fails with an explanatory error message (the name of the executable that caused the error and the exceeded timeout period) if the executable is not completed at end of the timeout period.
|
8970
|
+
Amount of time executable has to complete. Default is 10 minutes (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)). Cluster creation fails with an explanatory error message (the name of the executable that caused the error and the exceeded timeout period) if the executable is not completed at end of the timeout period.
|
8971
8971
|
"""
|
8972
8972
|
return pulumi.get(self, "execution_timeout")
|
8973
8973
|
|
@@ -9003,10 +9003,10 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9003
9003
|
idle_delete_ttl: Optional[str] = None,
|
9004
9004
|
idle_start_time: Optional[str] = None):
|
9005
9005
|
"""
|
9006
|
-
:param str auto_delete_time: The time when cluster will be auto-deleted (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9007
|
-
:param str auto_delete_ttl: The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9008
|
-
:param str idle_delete_ttl: The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json).
|
9009
|
-
:param str idle_start_time: Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9006
|
+
:param str auto_delete_time: The time when cluster will be auto-deleted (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9007
|
+
:param str auto_delete_ttl: The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9008
|
+
:param str idle_delete_ttl: The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json).
|
9009
|
+
:param str idle_start_time: Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9010
9010
|
"""
|
9011
9011
|
if auto_delete_time is not None:
|
9012
9012
|
pulumi.set(__self__, "auto_delete_time", auto_delete_time)
|
@@ -9021,7 +9021,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9021
9021
|
@pulumi.getter(name="autoDeleteTime")
|
9022
9022
|
def auto_delete_time(self) -> Optional[str]:
|
9023
9023
|
"""
|
9024
|
-
The time when cluster will be auto-deleted (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9024
|
+
The time when cluster will be auto-deleted (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9025
9025
|
"""
|
9026
9026
|
return pulumi.get(self, "auto_delete_time")
|
9027
9027
|
|
@@ -9029,7 +9029,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9029
9029
|
@pulumi.getter(name="autoDeleteTtl")
|
9030
9030
|
def auto_delete_ttl(self) -> Optional[str]:
|
9031
9031
|
"""
|
9032
|
-
The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9032
|
+
The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9033
9033
|
"""
|
9034
9034
|
return pulumi.get(self, "auto_delete_ttl")
|
9035
9035
|
|
@@ -9037,7 +9037,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9037
9037
|
@pulumi.getter(name="idleDeleteTtl")
|
9038
9038
|
def idle_delete_ttl(self) -> Optional[str]:
|
9039
9039
|
"""
|
9040
|
-
The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json).
|
9040
|
+
The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json).
|
9041
9041
|
"""
|
9042
9042
|
return pulumi.get(self, "idle_delete_ttl")
|
9043
9043
|
|
@@ -9045,7 +9045,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9045
9045
|
@pulumi.getter(name="idleStartTime")
|
9046
9046
|
def idle_start_time(self) -> Optional[str]:
|
9047
9047
|
"""
|
9048
|
-
Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9048
|
+
Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of [JSON Mapping - Language Guide (proto 3)](https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9049
9049
|
"""
|
9050
9050
|
return pulumi.get(self, "idle_start_time")
|
9051
9051
|
|
@@ -9100,7 +9100,7 @@ class WorkflowTemplatePlacementManagedClusterConfigMasterConfig(dict):
|
|
9100
9100
|
:param bool is_preemptible: Output only. Specifies that this instance group contains preemptible instances.
|
9101
9101
|
:param str machine_type: The Compute Engine machine type used for cluster instances. A full URL, partial URI, or short name are valid. Examples: * `https://www.googleapis.com/compute/v1/projects/(https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) feature, you must use the short name of the machine type resource, for example, `n1-standard-2`.
|
9102
9102
|
:param Sequence['WorkflowTemplatePlacementManagedClusterConfigMasterConfigManagedGroupConfigArgs'] managed_group_configs: Output only. The config for Compute Engine Instance Group Manager that manages this group. This is only used for preemptible instance groups.
|
9103
|
-
:param str min_cpu_platform: Specifies the minimum cpu platform for the Instance Group. See (https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
|
9103
|
+
:param str min_cpu_platform: Specifies the minimum cpu platform for the Instance Group. See [Minimum CPU platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
|
9104
9104
|
:param int num_instances: The number of VM instances in the instance group. For master instance groups, must be set to 1.
|
9105
9105
|
:param str preemptibility: Specifies the preemptibility of the instance group. The default value for master and worker groups is `NON_PREEMPTIBLE`. This default cannot be changed. The default value for secondary instances is `PREEMPTIBLE`. Possible values: PREEMPTIBILITY_UNSPECIFIED, NON_PREEMPTIBLE, PREEMPTIBLE
|
9106
9106
|
"""
|
@@ -9185,7 +9185,7 @@ class WorkflowTemplatePlacementManagedClusterConfigMasterConfig(dict):
|
|
9185
9185
|
@pulumi.getter(name="minCpuPlatform")
|
9186
9186
|
def min_cpu_platform(self) -> Optional[str]:
|
9187
9187
|
"""
|
9188
|
-
Specifies the minimum cpu platform for the Instance Group. See (https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
|
9188
|
+
Specifies the minimum cpu platform for the Instance Group. See [Minimum CPU platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
|
9189
9189
|
"""
|
9190
9190
|
return pulumi.get(self, "min_cpu_platform")
|
9191
9191
|
|
pulumi_gcp/datastream/_inputs.py
CHANGED
@@ -155,6 +155,8 @@ __all__ = [
|
|
155
155
|
'StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPostgresqlTablePostgresqlColumnArgsDict',
|
156
156
|
'StreamSourceConfigSqlServerSourceConfigArgs',
|
157
157
|
'StreamSourceConfigSqlServerSourceConfigArgsDict',
|
158
|
+
'StreamSourceConfigSqlServerSourceConfigChangeTablesArgs',
|
159
|
+
'StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict',
|
158
160
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs',
|
159
161
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgsDict',
|
160
162
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsSchemaArgs',
|
@@ -171,6 +173,8 @@ __all__ = [
|
|
171
173
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableArgsDict',
|
172
174
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumnArgs',
|
173
175
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumnArgsDict',
|
176
|
+
'StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs',
|
177
|
+
'StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict',
|
174
178
|
]
|
175
179
|
|
176
180
|
MYPY = False
|
@@ -5645,6 +5649,10 @@ class StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPost
|
|
5645
5649
|
|
5646
5650
|
if not MYPY:
|
5647
5651
|
class StreamSourceConfigSqlServerSourceConfigArgsDict(TypedDict):
|
5652
|
+
change_tables: NotRequired[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict']]
|
5653
|
+
"""
|
5654
|
+
CDC reader reads from change tables.
|
5655
|
+
"""
|
5648
5656
|
exclude_objects: NotRequired[pulumi.Input['StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgsDict']]
|
5649
5657
|
"""
|
5650
5658
|
SQL Server objects to exclude from the stream.
|
@@ -5663,24 +5671,34 @@ if not MYPY:
|
|
5663
5671
|
"""
|
5664
5672
|
Max concurrent CDC tasks.
|
5665
5673
|
"""
|
5674
|
+
transaction_logs: NotRequired[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict']]
|
5675
|
+
"""
|
5676
|
+
CDC reader reads from transaction logs.
|
5677
|
+
"""
|
5666
5678
|
elif False:
|
5667
5679
|
StreamSourceConfigSqlServerSourceConfigArgsDict: TypeAlias = Mapping[str, Any]
|
5668
5680
|
|
5669
5681
|
@pulumi.input_type
|
5670
5682
|
class StreamSourceConfigSqlServerSourceConfigArgs:
|
5671
5683
|
def __init__(__self__, *,
|
5684
|
+
change_tables: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs']] = None,
|
5672
5685
|
exclude_objects: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs']] = None,
|
5673
5686
|
include_objects: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigIncludeObjectsArgs']] = None,
|
5674
5687
|
max_concurrent_backfill_tasks: Optional[pulumi.Input[int]] = None,
|
5675
|
-
max_concurrent_cdc_tasks: Optional[pulumi.Input[int]] = None
|
5688
|
+
max_concurrent_cdc_tasks: Optional[pulumi.Input[int]] = None,
|
5689
|
+
transaction_logs: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs']] = None):
|
5676
5690
|
"""
|
5691
|
+
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs'] change_tables: CDC reader reads from change tables.
|
5677
5692
|
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs'] exclude_objects: SQL Server objects to exclude from the stream.
|
5678
5693
|
Structure is documented below.
|
5679
5694
|
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigIncludeObjectsArgs'] include_objects: SQL Server objects to retrieve from the source.
|
5680
5695
|
Structure is documented below.
|
5681
5696
|
:param pulumi.Input[int] max_concurrent_backfill_tasks: Max concurrent backfill tasks.
|
5682
5697
|
:param pulumi.Input[int] max_concurrent_cdc_tasks: Max concurrent CDC tasks.
|
5698
|
+
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs'] transaction_logs: CDC reader reads from transaction logs.
|
5683
5699
|
"""
|
5700
|
+
if change_tables is not None:
|
5701
|
+
pulumi.set(__self__, "change_tables", change_tables)
|
5684
5702
|
if exclude_objects is not None:
|
5685
5703
|
pulumi.set(__self__, "exclude_objects", exclude_objects)
|
5686
5704
|
if include_objects is not None:
|
@@ -5689,6 +5707,20 @@ class StreamSourceConfigSqlServerSourceConfigArgs:
|
|
5689
5707
|
pulumi.set(__self__, "max_concurrent_backfill_tasks", max_concurrent_backfill_tasks)
|
5690
5708
|
if max_concurrent_cdc_tasks is not None:
|
5691
5709
|
pulumi.set(__self__, "max_concurrent_cdc_tasks", max_concurrent_cdc_tasks)
|
5710
|
+
if transaction_logs is not None:
|
5711
|
+
pulumi.set(__self__, "transaction_logs", transaction_logs)
|
5712
|
+
|
5713
|
+
@property
|
5714
|
+
@pulumi.getter(name="changeTables")
|
5715
|
+
def change_tables(self) -> Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs']]:
|
5716
|
+
"""
|
5717
|
+
CDC reader reads from change tables.
|
5718
|
+
"""
|
5719
|
+
return pulumi.get(self, "change_tables")
|
5720
|
+
|
5721
|
+
@change_tables.setter
|
5722
|
+
def change_tables(self, value: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs']]):
|
5723
|
+
pulumi.set(self, "change_tables", value)
|
5692
5724
|
|
5693
5725
|
@property
|
5694
5726
|
@pulumi.getter(name="excludeObjects")
|
@@ -5740,6 +5772,30 @@ class StreamSourceConfigSqlServerSourceConfigArgs:
|
|
5740
5772
|
def max_concurrent_cdc_tasks(self, value: Optional[pulumi.Input[int]]):
|
5741
5773
|
pulumi.set(self, "max_concurrent_cdc_tasks", value)
|
5742
5774
|
|
5775
|
+
@property
|
5776
|
+
@pulumi.getter(name="transactionLogs")
|
5777
|
+
def transaction_logs(self) -> Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs']]:
|
5778
|
+
"""
|
5779
|
+
CDC reader reads from transaction logs.
|
5780
|
+
"""
|
5781
|
+
return pulumi.get(self, "transaction_logs")
|
5782
|
+
|
5783
|
+
@transaction_logs.setter
|
5784
|
+
def transaction_logs(self, value: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs']]):
|
5785
|
+
pulumi.set(self, "transaction_logs", value)
|
5786
|
+
|
5787
|
+
|
5788
|
+
if not MYPY:
|
5789
|
+
class StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict(TypedDict):
|
5790
|
+
pass
|
5791
|
+
elif False:
|
5792
|
+
StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict: TypeAlias = Mapping[str, Any]
|
5793
|
+
|
5794
|
+
@pulumi.input_type
|
5795
|
+
class StreamSourceConfigSqlServerSourceConfigChangeTablesArgs:
|
5796
|
+
def __init__(__self__):
|
5797
|
+
pass
|
5798
|
+
|
5743
5799
|
|
5744
5800
|
if not MYPY:
|
5745
5801
|
class StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgsDict(TypedDict):
|
@@ -6411,3 +6467,15 @@ class StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumnArgs
|
|
6411
6467
|
pulumi.set(self, "scale", value)
|
6412
6468
|
|
6413
6469
|
|
6470
|
+
if not MYPY:
|
6471
|
+
class StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict(TypedDict):
|
6472
|
+
pass
|
6473
|
+
elif False:
|
6474
|
+
StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict: TypeAlias = Mapping[str, Any]
|
6475
|
+
|
6476
|
+
@pulumi.input_type
|
6477
|
+
class StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs:
|
6478
|
+
def __init__(__self__):
|
6479
|
+
pass
|
6480
|
+
|
6481
|
+
|
pulumi_gcp/datastream/outputs.py
CHANGED
@@ -86,6 +86,7 @@ __all__ = [
|
|
86
86
|
'StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPostgresqlTable',
|
87
87
|
'StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPostgresqlTablePostgresqlColumn',
|
88
88
|
'StreamSourceConfigSqlServerSourceConfig',
|
89
|
+
'StreamSourceConfigSqlServerSourceConfigChangeTables',
|
89
90
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjects',
|
90
91
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsSchema',
|
91
92
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsSchemaTable',
|
@@ -94,6 +95,7 @@ __all__ = [
|
|
94
95
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchema',
|
95
96
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTable',
|
96
97
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumn',
|
98
|
+
'StreamSourceConfigSqlServerSourceConfigTransactionLogs',
|
97
99
|
]
|
98
100
|
|
99
101
|
@pulumi.output_type
|
@@ -4408,7 +4410,9 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4408
4410
|
@staticmethod
|
4409
4411
|
def __key_warning(key: str):
|
4410
4412
|
suggest = None
|
4411
|
-
if key == "
|
4413
|
+
if key == "changeTables":
|
4414
|
+
suggest = "change_tables"
|
4415
|
+
elif key == "excludeObjects":
|
4412
4416
|
suggest = "exclude_objects"
|
4413
4417
|
elif key == "includeObjects":
|
4414
4418
|
suggest = "include_objects"
|
@@ -4416,6 +4420,8 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4416
4420
|
suggest = "max_concurrent_backfill_tasks"
|
4417
4421
|
elif key == "maxConcurrentCdcTasks":
|
4418
4422
|
suggest = "max_concurrent_cdc_tasks"
|
4423
|
+
elif key == "transactionLogs":
|
4424
|
+
suggest = "transaction_logs"
|
4419
4425
|
|
4420
4426
|
if suggest:
|
4421
4427
|
pulumi.log.warn(f"Key '{key}' not found in StreamSourceConfigSqlServerSourceConfig. Access the value via the '{suggest}' property getter instead.")
|
@@ -4429,18 +4435,24 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4429
4435
|
return super().get(key, default)
|
4430
4436
|
|
4431
4437
|
def __init__(__self__, *,
|
4438
|
+
change_tables: Optional['outputs.StreamSourceConfigSqlServerSourceConfigChangeTables'] = None,
|
4432
4439
|
exclude_objects: Optional['outputs.StreamSourceConfigSqlServerSourceConfigExcludeObjects'] = None,
|
4433
4440
|
include_objects: Optional['outputs.StreamSourceConfigSqlServerSourceConfigIncludeObjects'] = None,
|
4434
4441
|
max_concurrent_backfill_tasks: Optional[int] = None,
|
4435
|
-
max_concurrent_cdc_tasks: Optional[int] = None
|
4442
|
+
max_concurrent_cdc_tasks: Optional[int] = None,
|
4443
|
+
transaction_logs: Optional['outputs.StreamSourceConfigSqlServerSourceConfigTransactionLogs'] = None):
|
4436
4444
|
"""
|
4445
|
+
:param 'StreamSourceConfigSqlServerSourceConfigChangeTablesArgs' change_tables: CDC reader reads from change tables.
|
4437
4446
|
:param 'StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs' exclude_objects: SQL Server objects to exclude from the stream.
|
4438
4447
|
Structure is documented below.
|
4439
4448
|
:param 'StreamSourceConfigSqlServerSourceConfigIncludeObjectsArgs' include_objects: SQL Server objects to retrieve from the source.
|
4440
4449
|
Structure is documented below.
|
4441
4450
|
:param int max_concurrent_backfill_tasks: Max concurrent backfill tasks.
|
4442
4451
|
:param int max_concurrent_cdc_tasks: Max concurrent CDC tasks.
|
4452
|
+
:param 'StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs' transaction_logs: CDC reader reads from transaction logs.
|
4443
4453
|
"""
|
4454
|
+
if change_tables is not None:
|
4455
|
+
pulumi.set(__self__, "change_tables", change_tables)
|
4444
4456
|
if exclude_objects is not None:
|
4445
4457
|
pulumi.set(__self__, "exclude_objects", exclude_objects)
|
4446
4458
|
if include_objects is not None:
|
@@ -4449,6 +4461,16 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4449
4461
|
pulumi.set(__self__, "max_concurrent_backfill_tasks", max_concurrent_backfill_tasks)
|
4450
4462
|
if max_concurrent_cdc_tasks is not None:
|
4451
4463
|
pulumi.set(__self__, "max_concurrent_cdc_tasks", max_concurrent_cdc_tasks)
|
4464
|
+
if transaction_logs is not None:
|
4465
|
+
pulumi.set(__self__, "transaction_logs", transaction_logs)
|
4466
|
+
|
4467
|
+
@property
|
4468
|
+
@pulumi.getter(name="changeTables")
|
4469
|
+
def change_tables(self) -> Optional['outputs.StreamSourceConfigSqlServerSourceConfigChangeTables']:
|
4470
|
+
"""
|
4471
|
+
CDC reader reads from change tables.
|
4472
|
+
"""
|
4473
|
+
return pulumi.get(self, "change_tables")
|
4452
4474
|
|
4453
4475
|
@property
|
4454
4476
|
@pulumi.getter(name="excludeObjects")
|
@@ -4484,6 +4506,20 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4484
4506
|
"""
|
4485
4507
|
return pulumi.get(self, "max_concurrent_cdc_tasks")
|
4486
4508
|
|
4509
|
+
@property
|
4510
|
+
@pulumi.getter(name="transactionLogs")
|
4511
|
+
def transaction_logs(self) -> Optional['outputs.StreamSourceConfigSqlServerSourceConfigTransactionLogs']:
|
4512
|
+
"""
|
4513
|
+
CDC reader reads from transaction logs.
|
4514
|
+
"""
|
4515
|
+
return pulumi.get(self, "transaction_logs")
|
4516
|
+
|
4517
|
+
|
4518
|
+
@pulumi.output_type
|
4519
|
+
class StreamSourceConfigSqlServerSourceConfigChangeTables(dict):
|
4520
|
+
def __init__(__self__):
|
4521
|
+
pass
|
4522
|
+
|
4487
4523
|
|
4488
4524
|
@pulumi.output_type
|
4489
4525
|
class StreamSourceConfigSqlServerSourceConfigExcludeObjects(dict):
|
@@ -4929,3 +4965,9 @@ class StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumn(dic
|
|
4929
4965
|
return pulumi.get(self, "scale")
|
4930
4966
|
|
4931
4967
|
|
4968
|
+
@pulumi.output_type
|
4969
|
+
class StreamSourceConfigSqlServerSourceConfigTransactionLogs(dict):
|
4970
|
+
def __init__(__self__):
|
4971
|
+
pass
|
4972
|
+
|
4973
|
+
|