pulumi-gcp 8.1.0__py3-none-any.whl → 8.1.0a1726293903__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- pulumi_gcp/__init__.py +0 -24
- pulumi_gcp/alloydb/_inputs.py +0 -94
- pulumi_gcp/alloydb/cluster.py +1 -94
- pulumi_gcp/alloydb/outputs.py +0 -79
- pulumi_gcp/assuredworkloads/workload.py +16 -135
- pulumi_gcp/backupdisasterrecovery/__init__.py +0 -1
- pulumi_gcp/bigquery/_inputs.py +0 -33
- pulumi_gcp/bigquery/data_transfer_config.py +6 -134
- pulumi_gcp/bigquery/outputs.py +0 -36
- pulumi_gcp/bigqueryanalyticshub/_inputs.py +7 -169
- pulumi_gcp/bigqueryanalyticshub/data_exchange.py +3 -96
- pulumi_gcp/bigqueryanalyticshub/listing.py +0 -114
- pulumi_gcp/bigqueryanalyticshub/outputs.py +5 -138
- pulumi_gcp/bigtable/_inputs.py +1 -21
- pulumi_gcp/bigtable/outputs.py +1 -13
- pulumi_gcp/bigtable/table.py +0 -34
- pulumi_gcp/certificateauthority/authority.py +7 -14
- pulumi_gcp/certificatemanager/__init__.py +0 -1
- pulumi_gcp/certificatemanager/certificate.py +0 -28
- pulumi_gcp/certificatemanager/outputs.py +0 -322
- pulumi_gcp/cloudbuild/_inputs.py +6 -6
- pulumi_gcp/cloudbuild/outputs.py +4 -4
- pulumi_gcp/cloudrun/_inputs.py +12 -6
- pulumi_gcp/cloudrun/outputs.py +16 -8
- pulumi_gcp/cloudrunv2/_inputs.py +15 -72
- pulumi_gcp/cloudrunv2/outputs.py +16 -82
- pulumi_gcp/cloudrunv2/service.py +4 -50
- pulumi_gcp/cloudtasks/_inputs.py +0 -630
- pulumi_gcp/cloudtasks/outputs.py +0 -479
- pulumi_gcp/cloudtasks/queue.py +0 -238
- pulumi_gcp/compute/_inputs.py +9 -129
- pulumi_gcp/compute/get_instance.py +1 -1
- pulumi_gcp/compute/health_check.py +0 -114
- pulumi_gcp/compute/instance.py +4 -86
- pulumi_gcp/compute/instance_template.py +0 -66
- pulumi_gcp/compute/interconnect.py +21 -28
- pulumi_gcp/compute/node_template.py +0 -93
- pulumi_gcp/compute/outputs.py +6 -108
- pulumi_gcp/compute/target_https_proxy.py +0 -28
- pulumi_gcp/container/_inputs.py +3 -140
- pulumi_gcp/container/attached_cluster.py +7 -7
- pulumi_gcp/container/outputs.py +2 -174
- pulumi_gcp/databasemigrationservice/_inputs.py +148 -176
- pulumi_gcp/databasemigrationservice/connection_profile.py +0 -206
- pulumi_gcp/databasemigrationservice/outputs.py +87 -109
- pulumi_gcp/dataloss/_inputs.py +1 -353
- pulumi_gcp/dataloss/outputs.py +3 -274
- pulumi_gcp/dataproc/_inputs.py +27 -27
- pulumi_gcp/dataproc/outputs.py +18 -18
- pulumi_gcp/datastream/_inputs.py +1 -69
- pulumi_gcp/datastream/outputs.py +2 -44
- pulumi_gcp/datastream/stream.py +7 -194
- pulumi_gcp/discoveryengine/_inputs.py +0 -188
- pulumi_gcp/discoveryengine/data_store.py +14 -14
- pulumi_gcp/discoveryengine/outputs.py +1 -153
- pulumi_gcp/firebase/database_instance.py +7 -7
- pulumi_gcp/gkehub/_inputs.py +1 -25
- pulumi_gcp/gkehub/feature_membership.py +6 -12
- pulumi_gcp/gkehub/outputs.py +1 -17
- pulumi_gcp/iam/_inputs.py +0 -196
- pulumi_gcp/iam/get_workload_identity_pool_provider.py +3 -13
- pulumi_gcp/iam/outputs.py +0 -295
- pulumi_gcp/iam/workload_identity_pool_provider.py +3 -164
- pulumi_gcp/kms/__init__.py +0 -2
- pulumi_gcp/kms/autokey_config.py +2 -10
- pulumi_gcp/kms/outputs.py +0 -164
- pulumi_gcp/netapp/active_directory.py +6 -6
- pulumi_gcp/netapp/backup.py +6 -6
- pulumi_gcp/netapp/backup_policy.py +6 -6
- pulumi_gcp/netapp/backup_vault.py +6 -6
- pulumi_gcp/netapp/storage_pool.py +4 -4
- pulumi_gcp/netapp/volume.py +0 -7
- pulumi_gcp/networkconnectivity/_inputs.py +1 -21
- pulumi_gcp/networkconnectivity/outputs.py +1 -15
- pulumi_gcp/networkconnectivity/spoke.py +0 -8
- pulumi_gcp/networksecurity/client_tls_policy.py +22 -24
- pulumi_gcp/networksecurity/server_tls_policy.py +32 -20
- pulumi_gcp/organizations/get_project.py +3 -13
- pulumi_gcp/organizations/project.py +3 -88
- pulumi_gcp/parallelstore/instance.py +121 -121
- pulumi_gcp/projects/iam_member_remove.py +0 -26
- pulumi_gcp/projects/usage_export_bucket.py +0 -38
- pulumi_gcp/pubsub/_inputs.py +0 -40
- pulumi_gcp/pubsub/outputs.py +1 -51
- pulumi_gcp/pubsub/subscription.py +0 -6
- pulumi_gcp/pulumi-plugin.json +1 -1
- pulumi_gcp/redis/_inputs.py +0 -419
- pulumi_gcp/redis/cluster.py +0 -123
- pulumi_gcp/redis/outputs.py +0 -315
- pulumi_gcp/securitycenter/__init__.py +0 -2
- pulumi_gcp/securitycenter/v2_organization_scc_big_query_exports.py +4 -4
- {pulumi_gcp-8.1.0.dist-info → pulumi_gcp-8.1.0a1726293903.dist-info}/METADATA +1 -1
- {pulumi_gcp-8.1.0.dist-info → pulumi_gcp-8.1.0a1726293903.dist-info}/RECORD +95 -101
- {pulumi_gcp-8.1.0.dist-info → pulumi_gcp-8.1.0a1726293903.dist-info}/WHEEL +1 -1
- pulumi_gcp/backupdisasterrecovery/backup_vault.py +0 -1203
- pulumi_gcp/certificatemanager/get_certificates.py +0 -150
- pulumi_gcp/kms/get_crypto_key_latest_version.py +0 -222
- pulumi_gcp/kms/get_crypto_key_versions.py +0 -175
- pulumi_gcp/securitycenter/v2_folder_scc_big_query_export.py +0 -857
- pulumi_gcp/securitycenter/v2_project_scc_big_query_export.py +0 -796
- {pulumi_gcp-8.1.0.dist-info → pulumi_gcp-8.1.0a1726293903.dist-info}/top_level.txt +0 -0
pulumi_gcp/dataproc/outputs.py
CHANGED
@@ -8232,7 +8232,7 @@ class WorkflowTemplatePlacementManagedClusterConfig(dict):
|
|
8232
8232
|
:param 'WorkflowTemplatePlacementManagedClusterConfigSecondaryWorkerConfigArgs' secondary_worker_config: The Compute Engine config settings for additional worker instances in a cluster.
|
8233
8233
|
:param 'WorkflowTemplatePlacementManagedClusterConfigSecurityConfigArgs' security_config: Security settings for the cluster.
|
8234
8234
|
:param 'WorkflowTemplatePlacementManagedClusterConfigSoftwareConfigArgs' software_config: The config settings for software inside the cluster.
|
8235
|
-
:param str staging_bucket: A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see
|
8235
|
+
:param str staging_bucket: A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see (https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
8236
8236
|
:param str temp_bucket: A Cloud Storage bucket used to store ephemeral cluster and jobs data, such as Spark and MapReduce history files. If you do not specify a temp bucket, Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's temp bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket. The default bucket has a TTL of 90 days, but you can use any TTL (or none) if you specify a bucket.
|
8237
8237
|
:param 'WorkflowTemplatePlacementManagedClusterConfigWorkerConfigArgs' worker_config: The Compute Engine config settings for additional worker instances in a cluster.
|
8238
8238
|
|
@@ -8369,7 +8369,7 @@ class WorkflowTemplatePlacementManagedClusterConfig(dict):
|
|
8369
8369
|
@pulumi.getter(name="stagingBucket")
|
8370
8370
|
def staging_bucket(self) -> Optional[str]:
|
8371
8371
|
"""
|
8372
|
-
A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see
|
8372
|
+
A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see (https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
8373
8373
|
"""
|
8374
8374
|
return pulumi.get(self, "staging_bucket")
|
8375
8375
|
|
@@ -8543,7 +8543,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8543
8543
|
zone: Optional[str] = None):
|
8544
8544
|
"""
|
8545
8545
|
:param bool internal_ip_only: If true, all instances in the cluster will only have internal IP addresses. By default, clusters are not restricted to internal IP addresses, and will have ephemeral external IP addresses assigned to each instance. This `internal_ip_only` restriction can only be enabled for subnetwork enabled networks, and all off-cluster dependencies must be configured to be accessible without external IP addresses.
|
8546
|
-
:param Mapping[str, str] metadata: The Compute Engine metadata entries to add to all instances (see
|
8546
|
+
:param Mapping[str, str] metadata: The Compute Engine metadata entries to add to all instances (see (https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)).
|
8547
8547
|
:param str network: The Compute Engine network to be used for machine communications. Cannot be specified with subnetwork_uri. If neither `network_uri` nor `subnetwork_uri` is specified, the "default" network of the project is used, if it exists. Cannot be a "Custom Subnet Network" (see /regions/global/default` * `default`
|
8548
8548
|
:param 'WorkflowTemplatePlacementManagedClusterConfigGceClusterConfigNodeGroupAffinityArgs' node_group_affinity: Node Group Affinity for sole-tenant clusters.
|
8549
8549
|
:param str private_ipv6_google_access: The type of IPv6 access for a cluster. Possible values: PRIVATE_IPV6_GOOGLE_ACCESS_UNSPECIFIED, INHERIT_FROM_SUBNETWORK, OUTBOUND, BIDIRECTIONAL
|
@@ -8552,7 +8552,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8552
8552
|
:param Sequence[str] service_account_scopes: The URIs of service account scopes to be included in Compute Engine instances. The following base set of scopes is always included: * https://www.googleapis.com/auth/cloud.useraccounts.readonly * https://www.googleapis.com/auth/devstorage.read_write * https://www.googleapis.com/auth/logging.write If no scopes are specified, the following defaults are also provided: * https://www.googleapis.com/auth/bigquery * https://www.googleapis.com/auth/bigtable.admin.table * https://www.googleapis.com/auth/bigtable.data * https://www.googleapis.com/auth/devstorage.full_control
|
8553
8553
|
:param 'WorkflowTemplatePlacementManagedClusterConfigGceClusterConfigShieldedInstanceConfigArgs' shielded_instance_config: Shielded Instance Config for clusters using [Compute Engine Shielded VMs](https://cloud.google.com/security/shielded-cloud/shielded-vm). Structure defined below.
|
8554
8554
|
:param str subnetwork: The Compute Engine subnetwork to be used for machine communications. Cannot be specified with network_uri. A full URL, partial URI, or short name are valid. Examples: * `https://www.googleapis.com/compute/v1/projects//regions/us-east1/subnetworks/sub0` * `sub0`
|
8555
|
-
:param Sequence[str] tags: The Compute Engine tags to add to all instances (see
|
8555
|
+
:param Sequence[str] tags: The Compute Engine tags to add to all instances (see (https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
|
8556
8556
|
:param str zone: The zone where the Compute Engine cluster will be located. On a create request, it is required in the "global" region. If omitted in a non-global Dataproc region, the service will pick a zone in the corresponding Compute Engine region. On a get request, zone will always be present. A full URL, partial URI, or short name are valid. Examples: * `https://www.googleapis.com/compute/v1/projects/` * `us-central1-f`
|
8557
8557
|
"""
|
8558
8558
|
if internal_ip_only is not None:
|
@@ -8592,7 +8592,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8592
8592
|
@pulumi.getter
|
8593
8593
|
def metadata(self) -> Optional[Mapping[str, str]]:
|
8594
8594
|
"""
|
8595
|
-
The Compute Engine metadata entries to add to all instances (see
|
8595
|
+
The Compute Engine metadata entries to add to all instances (see (https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)).
|
8596
8596
|
"""
|
8597
8597
|
return pulumi.get(self, "metadata")
|
8598
8598
|
|
@@ -8664,7 +8664,7 @@ class WorkflowTemplatePlacementManagedClusterConfigGceClusterConfig(dict):
|
|
8664
8664
|
@pulumi.getter
|
8665
8665
|
def tags(self) -> Optional[Sequence[str]]:
|
8666
8666
|
"""
|
8667
|
-
The Compute Engine tags to add to all instances (see
|
8667
|
+
The Compute Engine tags to add to all instances (see (https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
|
8668
8668
|
"""
|
8669
8669
|
return pulumi.get(self, "tags")
|
8670
8670
|
|
@@ -8948,7 +8948,7 @@ class WorkflowTemplatePlacementManagedClusterConfigInitializationAction(dict):
|
|
8948
8948
|
execution_timeout: Optional[str] = None):
|
8949
8949
|
"""
|
8950
8950
|
:param str executable_file: Required. Cloud Storage URI of executable file.
|
8951
|
-
:param str execution_timeout: Amount of time executable has to complete. Default is 10 minutes (see JSON representation of
|
8951
|
+
:param str execution_timeout: Amount of time executable has to complete. Default is 10 minutes (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)). Cluster creation fails with an explanatory error message (the name of the executable that caused the error and the exceeded timeout period) if the executable is not completed at end of the timeout period.
|
8952
8952
|
"""
|
8953
8953
|
if executable_file is not None:
|
8954
8954
|
pulumi.set(__self__, "executable_file", executable_file)
|
@@ -8967,7 +8967,7 @@ class WorkflowTemplatePlacementManagedClusterConfigInitializationAction(dict):
|
|
8967
8967
|
@pulumi.getter(name="executionTimeout")
|
8968
8968
|
def execution_timeout(self) -> Optional[str]:
|
8969
8969
|
"""
|
8970
|
-
Amount of time executable has to complete. Default is 10 minutes (see JSON representation of
|
8970
|
+
Amount of time executable has to complete. Default is 10 minutes (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)). Cluster creation fails with an explanatory error message (the name of the executable that caused the error and the exceeded timeout period) if the executable is not completed at end of the timeout period.
|
8971
8971
|
"""
|
8972
8972
|
return pulumi.get(self, "execution_timeout")
|
8973
8973
|
|
@@ -9003,10 +9003,10 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9003
9003
|
idle_delete_ttl: Optional[str] = None,
|
9004
9004
|
idle_start_time: Optional[str] = None):
|
9005
9005
|
"""
|
9006
|
-
:param str auto_delete_time: The time when cluster will be auto-deleted (see JSON representation of
|
9007
|
-
:param str auto_delete_ttl: The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of
|
9008
|
-
:param str idle_delete_ttl: The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of
|
9009
|
-
:param str idle_start_time: Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of
|
9006
|
+
:param str auto_delete_time: The time when cluster will be auto-deleted (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9007
|
+
:param str auto_delete_ttl: The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9008
|
+
:param str idle_delete_ttl: The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json).
|
9009
|
+
:param str idle_start_time: Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9010
9010
|
"""
|
9011
9011
|
if auto_delete_time is not None:
|
9012
9012
|
pulumi.set(__self__, "auto_delete_time", auto_delete_time)
|
@@ -9021,7 +9021,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9021
9021
|
@pulumi.getter(name="autoDeleteTime")
|
9022
9022
|
def auto_delete_time(self) -> Optional[str]:
|
9023
9023
|
"""
|
9024
|
-
The time when cluster will be auto-deleted (see JSON representation of
|
9024
|
+
The time when cluster will be auto-deleted (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9025
9025
|
"""
|
9026
9026
|
return pulumi.get(self, "auto_delete_time")
|
9027
9027
|
|
@@ -9029,7 +9029,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9029
9029
|
@pulumi.getter(name="autoDeleteTtl")
|
9030
9030
|
def auto_delete_ttl(self) -> Optional[str]:
|
9031
9031
|
"""
|
9032
|
-
The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of
|
9032
|
+
The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9033
9033
|
"""
|
9034
9034
|
return pulumi.get(self, "auto_delete_ttl")
|
9035
9035
|
|
@@ -9037,7 +9037,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9037
9037
|
@pulumi.getter(name="idleDeleteTtl")
|
9038
9038
|
def idle_delete_ttl(self) -> Optional[str]:
|
9039
9039
|
"""
|
9040
|
-
The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of
|
9040
|
+
The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json).
|
9041
9041
|
"""
|
9042
9042
|
return pulumi.get(self, "idle_delete_ttl")
|
9043
9043
|
|
@@ -9045,7 +9045,7 @@ class WorkflowTemplatePlacementManagedClusterConfigLifecycleConfig(dict):
|
|
9045
9045
|
@pulumi.getter(name="idleStartTime")
|
9046
9046
|
def idle_start_time(self) -> Optional[str]:
|
9047
9047
|
"""
|
9048
|
-
Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of
|
9048
|
+
Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of (https://developers.google.com/protocol-buffers/docs/proto3#json)).
|
9049
9049
|
"""
|
9050
9050
|
return pulumi.get(self, "idle_start_time")
|
9051
9051
|
|
@@ -9100,7 +9100,7 @@ class WorkflowTemplatePlacementManagedClusterConfigMasterConfig(dict):
|
|
9100
9100
|
:param bool is_preemptible: Output only. Specifies that this instance group contains preemptible instances.
|
9101
9101
|
:param str machine_type: The Compute Engine machine type used for cluster instances. A full URL, partial URI, or short name are valid. Examples: * `https://www.googleapis.com/compute/v1/projects/(https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) feature, you must use the short name of the machine type resource, for example, `n1-standard-2`.
|
9102
9102
|
:param Sequence['WorkflowTemplatePlacementManagedClusterConfigMasterConfigManagedGroupConfigArgs'] managed_group_configs: Output only. The config for Compute Engine Instance Group Manager that manages this group. This is only used for preemptible instance groups.
|
9103
|
-
:param str min_cpu_platform: Specifies the minimum cpu platform for the Instance Group. See
|
9103
|
+
:param str min_cpu_platform: Specifies the minimum cpu platform for the Instance Group. See (https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
|
9104
9104
|
:param int num_instances: The number of VM instances in the instance group. For master instance groups, must be set to 1.
|
9105
9105
|
:param str preemptibility: Specifies the preemptibility of the instance group. The default value for master and worker groups is `NON_PREEMPTIBLE`. This default cannot be changed. The default value for secondary instances is `PREEMPTIBLE`. Possible values: PREEMPTIBILITY_UNSPECIFIED, NON_PREEMPTIBLE, PREEMPTIBLE
|
9106
9106
|
"""
|
@@ -9185,7 +9185,7 @@ class WorkflowTemplatePlacementManagedClusterConfigMasterConfig(dict):
|
|
9185
9185
|
@pulumi.getter(name="minCpuPlatform")
|
9186
9186
|
def min_cpu_platform(self) -> Optional[str]:
|
9187
9187
|
"""
|
9188
|
-
Specifies the minimum cpu platform for the Instance Group. See
|
9188
|
+
Specifies the minimum cpu platform for the Instance Group. See (https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
|
9189
9189
|
"""
|
9190
9190
|
return pulumi.get(self, "min_cpu_platform")
|
9191
9191
|
|
pulumi_gcp/datastream/_inputs.py
CHANGED
@@ -155,8 +155,6 @@ __all__ = [
|
|
155
155
|
'StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPostgresqlTablePostgresqlColumnArgsDict',
|
156
156
|
'StreamSourceConfigSqlServerSourceConfigArgs',
|
157
157
|
'StreamSourceConfigSqlServerSourceConfigArgsDict',
|
158
|
-
'StreamSourceConfigSqlServerSourceConfigChangeTablesArgs',
|
159
|
-
'StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict',
|
160
158
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs',
|
161
159
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgsDict',
|
162
160
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsSchemaArgs',
|
@@ -173,8 +171,6 @@ __all__ = [
|
|
173
171
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableArgsDict',
|
174
172
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumnArgs',
|
175
173
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumnArgsDict',
|
176
|
-
'StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs',
|
177
|
-
'StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict',
|
178
174
|
]
|
179
175
|
|
180
176
|
MYPY = False
|
@@ -5649,10 +5645,6 @@ class StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPost
|
|
5649
5645
|
|
5650
5646
|
if not MYPY:
|
5651
5647
|
class StreamSourceConfigSqlServerSourceConfigArgsDict(TypedDict):
|
5652
|
-
change_tables: NotRequired[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict']]
|
5653
|
-
"""
|
5654
|
-
CDC reader reads from change tables.
|
5655
|
-
"""
|
5656
5648
|
exclude_objects: NotRequired[pulumi.Input['StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgsDict']]
|
5657
5649
|
"""
|
5658
5650
|
SQL Server objects to exclude from the stream.
|
@@ -5671,34 +5663,24 @@ if not MYPY:
|
|
5671
5663
|
"""
|
5672
5664
|
Max concurrent CDC tasks.
|
5673
5665
|
"""
|
5674
|
-
transaction_logs: NotRequired[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict']]
|
5675
|
-
"""
|
5676
|
-
CDC reader reads from transaction logs.
|
5677
|
-
"""
|
5678
5666
|
elif False:
|
5679
5667
|
StreamSourceConfigSqlServerSourceConfigArgsDict: TypeAlias = Mapping[str, Any]
|
5680
5668
|
|
5681
5669
|
@pulumi.input_type
|
5682
5670
|
class StreamSourceConfigSqlServerSourceConfigArgs:
|
5683
5671
|
def __init__(__self__, *,
|
5684
|
-
change_tables: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs']] = None,
|
5685
5672
|
exclude_objects: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs']] = None,
|
5686
5673
|
include_objects: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigIncludeObjectsArgs']] = None,
|
5687
5674
|
max_concurrent_backfill_tasks: Optional[pulumi.Input[int]] = None,
|
5688
|
-
max_concurrent_cdc_tasks: Optional[pulumi.Input[int]] = None
|
5689
|
-
transaction_logs: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs']] = None):
|
5675
|
+
max_concurrent_cdc_tasks: Optional[pulumi.Input[int]] = None):
|
5690
5676
|
"""
|
5691
|
-
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs'] change_tables: CDC reader reads from change tables.
|
5692
5677
|
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs'] exclude_objects: SQL Server objects to exclude from the stream.
|
5693
5678
|
Structure is documented below.
|
5694
5679
|
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigIncludeObjectsArgs'] include_objects: SQL Server objects to retrieve from the source.
|
5695
5680
|
Structure is documented below.
|
5696
5681
|
:param pulumi.Input[int] max_concurrent_backfill_tasks: Max concurrent backfill tasks.
|
5697
5682
|
:param pulumi.Input[int] max_concurrent_cdc_tasks: Max concurrent CDC tasks.
|
5698
|
-
:param pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs'] transaction_logs: CDC reader reads from transaction logs.
|
5699
5683
|
"""
|
5700
|
-
if change_tables is not None:
|
5701
|
-
pulumi.set(__self__, "change_tables", change_tables)
|
5702
5684
|
if exclude_objects is not None:
|
5703
5685
|
pulumi.set(__self__, "exclude_objects", exclude_objects)
|
5704
5686
|
if include_objects is not None:
|
@@ -5707,20 +5689,6 @@ class StreamSourceConfigSqlServerSourceConfigArgs:
|
|
5707
5689
|
pulumi.set(__self__, "max_concurrent_backfill_tasks", max_concurrent_backfill_tasks)
|
5708
5690
|
if max_concurrent_cdc_tasks is not None:
|
5709
5691
|
pulumi.set(__self__, "max_concurrent_cdc_tasks", max_concurrent_cdc_tasks)
|
5710
|
-
if transaction_logs is not None:
|
5711
|
-
pulumi.set(__self__, "transaction_logs", transaction_logs)
|
5712
|
-
|
5713
|
-
@property
|
5714
|
-
@pulumi.getter(name="changeTables")
|
5715
|
-
def change_tables(self) -> Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs']]:
|
5716
|
-
"""
|
5717
|
-
CDC reader reads from change tables.
|
5718
|
-
"""
|
5719
|
-
return pulumi.get(self, "change_tables")
|
5720
|
-
|
5721
|
-
@change_tables.setter
|
5722
|
-
def change_tables(self, value: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigChangeTablesArgs']]):
|
5723
|
-
pulumi.set(self, "change_tables", value)
|
5724
5692
|
|
5725
5693
|
@property
|
5726
5694
|
@pulumi.getter(name="excludeObjects")
|
@@ -5772,30 +5740,6 @@ class StreamSourceConfigSqlServerSourceConfigArgs:
|
|
5772
5740
|
def max_concurrent_cdc_tasks(self, value: Optional[pulumi.Input[int]]):
|
5773
5741
|
pulumi.set(self, "max_concurrent_cdc_tasks", value)
|
5774
5742
|
|
5775
|
-
@property
|
5776
|
-
@pulumi.getter(name="transactionLogs")
|
5777
|
-
def transaction_logs(self) -> Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs']]:
|
5778
|
-
"""
|
5779
|
-
CDC reader reads from transaction logs.
|
5780
|
-
"""
|
5781
|
-
return pulumi.get(self, "transaction_logs")
|
5782
|
-
|
5783
|
-
@transaction_logs.setter
|
5784
|
-
def transaction_logs(self, value: Optional[pulumi.Input['StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs']]):
|
5785
|
-
pulumi.set(self, "transaction_logs", value)
|
5786
|
-
|
5787
|
-
|
5788
|
-
if not MYPY:
|
5789
|
-
class StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict(TypedDict):
|
5790
|
-
pass
|
5791
|
-
elif False:
|
5792
|
-
StreamSourceConfigSqlServerSourceConfigChangeTablesArgsDict: TypeAlias = Mapping[str, Any]
|
5793
|
-
|
5794
|
-
@pulumi.input_type
|
5795
|
-
class StreamSourceConfigSqlServerSourceConfigChangeTablesArgs:
|
5796
|
-
def __init__(__self__):
|
5797
|
-
pass
|
5798
|
-
|
5799
5743
|
|
5800
5744
|
if not MYPY:
|
5801
5745
|
class StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgsDict(TypedDict):
|
@@ -6467,15 +6411,3 @@ class StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumnArgs
|
|
6467
6411
|
pulumi.set(self, "scale", value)
|
6468
6412
|
|
6469
6413
|
|
6470
|
-
if not MYPY:
|
6471
|
-
class StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict(TypedDict):
|
6472
|
-
pass
|
6473
|
-
elif False:
|
6474
|
-
StreamSourceConfigSqlServerSourceConfigTransactionLogsArgsDict: TypeAlias = Mapping[str, Any]
|
6475
|
-
|
6476
|
-
@pulumi.input_type
|
6477
|
-
class StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs:
|
6478
|
-
def __init__(__self__):
|
6479
|
-
pass
|
6480
|
-
|
6481
|
-
|
pulumi_gcp/datastream/outputs.py
CHANGED
@@ -86,7 +86,6 @@ __all__ = [
|
|
86
86
|
'StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPostgresqlTable',
|
87
87
|
'StreamSourceConfigPostgresqlSourceConfigIncludeObjectsPostgresqlSchemaPostgresqlTablePostgresqlColumn',
|
88
88
|
'StreamSourceConfigSqlServerSourceConfig',
|
89
|
-
'StreamSourceConfigSqlServerSourceConfigChangeTables',
|
90
89
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjects',
|
91
90
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsSchema',
|
92
91
|
'StreamSourceConfigSqlServerSourceConfigExcludeObjectsSchemaTable',
|
@@ -95,7 +94,6 @@ __all__ = [
|
|
95
94
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchema',
|
96
95
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTable',
|
97
96
|
'StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumn',
|
98
|
-
'StreamSourceConfigSqlServerSourceConfigTransactionLogs',
|
99
97
|
]
|
100
98
|
|
101
99
|
@pulumi.output_type
|
@@ -4410,9 +4408,7 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4410
4408
|
@staticmethod
|
4411
4409
|
def __key_warning(key: str):
|
4412
4410
|
suggest = None
|
4413
|
-
if key == "
|
4414
|
-
suggest = "change_tables"
|
4415
|
-
elif key == "excludeObjects":
|
4411
|
+
if key == "excludeObjects":
|
4416
4412
|
suggest = "exclude_objects"
|
4417
4413
|
elif key == "includeObjects":
|
4418
4414
|
suggest = "include_objects"
|
@@ -4420,8 +4416,6 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4420
4416
|
suggest = "max_concurrent_backfill_tasks"
|
4421
4417
|
elif key == "maxConcurrentCdcTasks":
|
4422
4418
|
suggest = "max_concurrent_cdc_tasks"
|
4423
|
-
elif key == "transactionLogs":
|
4424
|
-
suggest = "transaction_logs"
|
4425
4419
|
|
4426
4420
|
if suggest:
|
4427
4421
|
pulumi.log.warn(f"Key '{key}' not found in StreamSourceConfigSqlServerSourceConfig. Access the value via the '{suggest}' property getter instead.")
|
@@ -4435,24 +4429,18 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4435
4429
|
return super().get(key, default)
|
4436
4430
|
|
4437
4431
|
def __init__(__self__, *,
|
4438
|
-
change_tables: Optional['outputs.StreamSourceConfigSqlServerSourceConfigChangeTables'] = None,
|
4439
4432
|
exclude_objects: Optional['outputs.StreamSourceConfigSqlServerSourceConfigExcludeObjects'] = None,
|
4440
4433
|
include_objects: Optional['outputs.StreamSourceConfigSqlServerSourceConfigIncludeObjects'] = None,
|
4441
4434
|
max_concurrent_backfill_tasks: Optional[int] = None,
|
4442
|
-
max_concurrent_cdc_tasks: Optional[int] = None
|
4443
|
-
transaction_logs: Optional['outputs.StreamSourceConfigSqlServerSourceConfigTransactionLogs'] = None):
|
4435
|
+
max_concurrent_cdc_tasks: Optional[int] = None):
|
4444
4436
|
"""
|
4445
|
-
:param 'StreamSourceConfigSqlServerSourceConfigChangeTablesArgs' change_tables: CDC reader reads from change tables.
|
4446
4437
|
:param 'StreamSourceConfigSqlServerSourceConfigExcludeObjectsArgs' exclude_objects: SQL Server objects to exclude from the stream.
|
4447
4438
|
Structure is documented below.
|
4448
4439
|
:param 'StreamSourceConfigSqlServerSourceConfigIncludeObjectsArgs' include_objects: SQL Server objects to retrieve from the source.
|
4449
4440
|
Structure is documented below.
|
4450
4441
|
:param int max_concurrent_backfill_tasks: Max concurrent backfill tasks.
|
4451
4442
|
:param int max_concurrent_cdc_tasks: Max concurrent CDC tasks.
|
4452
|
-
:param 'StreamSourceConfigSqlServerSourceConfigTransactionLogsArgs' transaction_logs: CDC reader reads from transaction logs.
|
4453
4443
|
"""
|
4454
|
-
if change_tables is not None:
|
4455
|
-
pulumi.set(__self__, "change_tables", change_tables)
|
4456
4444
|
if exclude_objects is not None:
|
4457
4445
|
pulumi.set(__self__, "exclude_objects", exclude_objects)
|
4458
4446
|
if include_objects is not None:
|
@@ -4461,16 +4449,6 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4461
4449
|
pulumi.set(__self__, "max_concurrent_backfill_tasks", max_concurrent_backfill_tasks)
|
4462
4450
|
if max_concurrent_cdc_tasks is not None:
|
4463
4451
|
pulumi.set(__self__, "max_concurrent_cdc_tasks", max_concurrent_cdc_tasks)
|
4464
|
-
if transaction_logs is not None:
|
4465
|
-
pulumi.set(__self__, "transaction_logs", transaction_logs)
|
4466
|
-
|
4467
|
-
@property
|
4468
|
-
@pulumi.getter(name="changeTables")
|
4469
|
-
def change_tables(self) -> Optional['outputs.StreamSourceConfigSqlServerSourceConfigChangeTables']:
|
4470
|
-
"""
|
4471
|
-
CDC reader reads from change tables.
|
4472
|
-
"""
|
4473
|
-
return pulumi.get(self, "change_tables")
|
4474
4452
|
|
4475
4453
|
@property
|
4476
4454
|
@pulumi.getter(name="excludeObjects")
|
@@ -4506,20 +4484,6 @@ class StreamSourceConfigSqlServerSourceConfig(dict):
|
|
4506
4484
|
"""
|
4507
4485
|
return pulumi.get(self, "max_concurrent_cdc_tasks")
|
4508
4486
|
|
4509
|
-
@property
|
4510
|
-
@pulumi.getter(name="transactionLogs")
|
4511
|
-
def transaction_logs(self) -> Optional['outputs.StreamSourceConfigSqlServerSourceConfigTransactionLogs']:
|
4512
|
-
"""
|
4513
|
-
CDC reader reads from transaction logs.
|
4514
|
-
"""
|
4515
|
-
return pulumi.get(self, "transaction_logs")
|
4516
|
-
|
4517
|
-
|
4518
|
-
@pulumi.output_type
|
4519
|
-
class StreamSourceConfigSqlServerSourceConfigChangeTables(dict):
|
4520
|
-
def __init__(__self__):
|
4521
|
-
pass
|
4522
|
-
|
4523
4487
|
|
4524
4488
|
@pulumi.output_type
|
4525
4489
|
class StreamSourceConfigSqlServerSourceConfigExcludeObjects(dict):
|
@@ -4965,9 +4929,3 @@ class StreamSourceConfigSqlServerSourceConfigIncludeObjectsSchemaTableColumn(dic
|
|
4965
4929
|
return pulumi.get(self, "scale")
|
4966
4930
|
|
4967
4931
|
|
4968
|
-
@pulumi.output_type
|
4969
|
-
class StreamSourceConfigSqlServerSourceConfigTransactionLogs(dict):
|
4970
|
-
def __init__(__self__):
|
4971
|
-
pass
|
4972
|
-
|
4973
|
-
|