pulumi-gcp 8.2.0a1726671837__py3-none-any.whl → 8.3.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- pulumi_gcp/__init__.py +158 -0
- pulumi_gcp/alloydb/_inputs.py +20 -0
- pulumi_gcp/alloydb/instance.py +36 -0
- pulumi_gcp/alloydb/outputs.py +14 -0
- pulumi_gcp/bigquery/__init__.py +1 -0
- pulumi_gcp/bigquery/get_tables.py +143 -0
- pulumi_gcp/bigquery/outputs.py +30 -0
- pulumi_gcp/compute/__init__.py +2 -0
- pulumi_gcp/compute/_inputs.py +12481 -9158
- pulumi_gcp/compute/attached_disk.py +103 -0
- pulumi_gcp/compute/backend_service.py +29 -22
- pulumi_gcp/compute/firewall_policy_with_rules.py +769 -0
- pulumi_gcp/compute/get_instance.py +3 -0
- pulumi_gcp/compute/network_firewall_policy_with_rules.py +826 -0
- pulumi_gcp/compute/outputs.py +3695 -1293
- pulumi_gcp/compute/region_backend_service.py +29 -22
- pulumi_gcp/compute/region_target_http_proxy.py +159 -0
- pulumi_gcp/compute/region_target_https_proxy.py +175 -0
- pulumi_gcp/compute/router_nat.py +27 -66
- pulumi_gcp/compute/service_attachment.py +75 -0
- pulumi_gcp/compute/target_http_proxy.py +49 -28
- pulumi_gcp/compute/target_https_proxy.py +49 -28
- pulumi_gcp/config/__init__.pyi +6 -0
- pulumi_gcp/config/vars.py +12 -0
- pulumi_gcp/container/_inputs.py +168 -99
- pulumi_gcp/container/attached_cluster.py +54 -1
- pulumi_gcp/container/outputs.py +155 -70
- pulumi_gcp/databasemigrationservice/__init__.py +1 -0
- pulumi_gcp/databasemigrationservice/_inputs.py +362 -0
- pulumi_gcp/databasemigrationservice/migration_job.py +1739 -0
- pulumi_gcp/databasemigrationservice/outputs.py +268 -0
- pulumi_gcp/dataproc/get_metastore_service.py +11 -1
- pulumi_gcp/dataproc/metastore_federation.py +8 -8
- pulumi_gcp/dataproc/metastore_service.py +95 -0
- pulumi_gcp/datastream/stream.py +21 -14
- pulumi_gcp/developerconnect/__init__.py +11 -0
- pulumi_gcp/developerconnect/_inputs.py +304 -0
- pulumi_gcp/developerconnect/connection.py +1037 -0
- pulumi_gcp/developerconnect/git_repository_link.py +873 -0
- pulumi_gcp/developerconnect/outputs.py +249 -0
- pulumi_gcp/discoveryengine/__init__.py +1 -0
- pulumi_gcp/discoveryengine/_inputs.py +131 -0
- pulumi_gcp/discoveryengine/outputs.py +131 -0
- pulumi_gcp/discoveryengine/target_site.py +870 -0
- pulumi_gcp/dns/managed_zone.py +1 -1
- pulumi_gcp/filestore/get_instance.py +21 -1
- pulumi_gcp/filestore/instance.py +94 -0
- pulumi_gcp/gkehub/feature_membership.py +140 -62
- pulumi_gcp/gkeonprem/_inputs.py +3 -3
- pulumi_gcp/gkeonprem/outputs.py +2 -2
- pulumi_gcp/healthcare/__init__.py +1 -0
- pulumi_gcp/healthcare/_inputs.py +39 -0
- pulumi_gcp/healthcare/outputs.py +40 -0
- pulumi_gcp/healthcare/workspace.py +465 -0
- pulumi_gcp/looker/instance.py +81 -0
- pulumi_gcp/memorystore/__init__.py +10 -0
- pulumi_gcp/memorystore/_inputs.py +731 -0
- pulumi_gcp/memorystore/instance.py +1663 -0
- pulumi_gcp/memorystore/outputs.py +598 -0
- pulumi_gcp/netapp/_inputs.py +63 -0
- pulumi_gcp/netapp/outputs.py +57 -0
- pulumi_gcp/netapp/storage_pool.py +54 -0
- pulumi_gcp/netapp/volume.py +183 -0
- pulumi_gcp/organizations/folder.py +52 -33
- pulumi_gcp/provider.py +60 -0
- pulumi_gcp/pubsub/subscription.py +43 -7
- pulumi_gcp/pulumi-plugin.json +1 -1
- pulumi_gcp/secretmanager/__init__.py +8 -0
- pulumi_gcp/secretmanager/_inputs.py +308 -0
- pulumi_gcp/secretmanager/get_regional_secret.py +279 -0
- pulumi_gcp/secretmanager/get_regional_secret_iam_policy.py +173 -0
- pulumi_gcp/secretmanager/get_regional_secret_version.py +241 -0
- pulumi_gcp/secretmanager/outputs.py +336 -0
- pulumi_gcp/secretmanager/regional_secret.py +1433 -0
- pulumi_gcp/secretmanager/regional_secret_iam_binding.py +1082 -0
- pulumi_gcp/secretmanager/regional_secret_iam_member.py +1082 -0
- pulumi_gcp/secretmanager/regional_secret_iam_policy.py +901 -0
- pulumi_gcp/secretmanager/regional_secret_version.py +753 -0
- pulumi_gcp/securitycenter/__init__.py +4 -0
- pulumi_gcp/securitycenter/folder_scc_big_query_export.py +795 -0
- pulumi_gcp/securitycenter/organization_scc_big_query_export.py +738 -0
- pulumi_gcp/securitycenter/project_scc_big_query_export.py +749 -0
- pulumi_gcp/securitycenter/v2_organization_scc_big_query_export.py +862 -0
- pulumi_gcp/securitycenter/v2_organization_scc_big_query_exports.py +6 -2
- pulumi_gcp/securitycenter/v2_project_mute_config.py +2 -2
- pulumi_gcp/siteverification/__init__.py +3 -0
- pulumi_gcp/siteverification/_inputs.py +85 -0
- pulumi_gcp/siteverification/outputs.py +57 -0
- pulumi_gcp/siteverification/web_resource.py +398 -0
- pulumi_gcp/spanner/__init__.py +1 -0
- pulumi_gcp/spanner/_inputs.py +129 -0
- pulumi_gcp/spanner/backup_schedule.py +748 -0
- pulumi_gcp/spanner/get_instance.py +11 -1
- pulumi_gcp/spanner/instance.py +56 -0
- pulumi_gcp/spanner/outputs.py +93 -0
- pulumi_gcp/vpcaccess/connector.py +21 -28
- {pulumi_gcp-8.2.0a1726671837.dist-info → pulumi_gcp-8.3.0.dist-info}/METADATA +1 -1
- {pulumi_gcp-8.2.0a1726671837.dist-info → pulumi_gcp-8.3.0.dist-info}/RECORD +100 -69
- {pulumi_gcp-8.2.0a1726671837.dist-info → pulumi_gcp-8.3.0.dist-info}/WHEEL +0 -0
- {pulumi_gcp-8.2.0a1726671837.dist-info → pulumi_gcp-8.3.0.dist-info}/top_level.txt +0 -0
pulumi_gcp/container/outputs.py
CHANGED
@@ -27,6 +27,7 @@ __all__ = [
|
|
27
27
|
'AttachedClusterOidcConfig',
|
28
28
|
'AttachedClusterProxyConfig',
|
29
29
|
'AttachedClusterProxyConfigKubernetesSecret',
|
30
|
+
'AttachedClusterSecurityPostureConfig',
|
30
31
|
'AttachedClusterWorkloadIdentityConfig',
|
31
32
|
'AwsClusterAuthorization',
|
32
33
|
'AwsClusterAuthorizationAdminGroup',
|
@@ -839,6 +840,43 @@ class AttachedClusterProxyConfigKubernetesSecret(dict):
|
|
839
840
|
return pulumi.get(self, "namespace")
|
840
841
|
|
841
842
|
|
843
|
+
@pulumi.output_type
|
844
|
+
class AttachedClusterSecurityPostureConfig(dict):
|
845
|
+
@staticmethod
|
846
|
+
def __key_warning(key: str):
|
847
|
+
suggest = None
|
848
|
+
if key == "vulnerabilityMode":
|
849
|
+
suggest = "vulnerability_mode"
|
850
|
+
|
851
|
+
if suggest:
|
852
|
+
pulumi.log.warn(f"Key '{key}' not found in AttachedClusterSecurityPostureConfig. Access the value via the '{suggest}' property getter instead.")
|
853
|
+
|
854
|
+
def __getitem__(self, key: str) -> Any:
|
855
|
+
AttachedClusterSecurityPostureConfig.__key_warning(key)
|
856
|
+
return super().__getitem__(key)
|
857
|
+
|
858
|
+
def get(self, key: str, default = None) -> Any:
|
859
|
+
AttachedClusterSecurityPostureConfig.__key_warning(key)
|
860
|
+
return super().get(key, default)
|
861
|
+
|
862
|
+
def __init__(__self__, *,
|
863
|
+
vulnerability_mode: str):
|
864
|
+
"""
|
865
|
+
:param str vulnerability_mode: Sets the mode of the Kubernetes security posture API's workload vulnerability scanning.
|
866
|
+
Possible values are: `VULNERABILITY_DISABLED`, `VULNERABILITY_ENTERPRISE`.
|
867
|
+
"""
|
868
|
+
pulumi.set(__self__, "vulnerability_mode", vulnerability_mode)
|
869
|
+
|
870
|
+
@property
|
871
|
+
@pulumi.getter(name="vulnerabilityMode")
|
872
|
+
def vulnerability_mode(self) -> str:
|
873
|
+
"""
|
874
|
+
Sets the mode of the Kubernetes security posture API's workload vulnerability scanning.
|
875
|
+
Possible values are: `VULNERABILITY_DISABLED`, `VULNERABILITY_ENTERPRISE`.
|
876
|
+
"""
|
877
|
+
return pulumi.get(self, "vulnerability_mode")
|
878
|
+
|
879
|
+
|
842
880
|
@pulumi.output_type
|
843
881
|
class AttachedClusterWorkloadIdentityConfig(dict):
|
844
882
|
@staticmethod
|
@@ -6631,6 +6669,8 @@ class ClusterNodeConfig(dict):
|
|
6631
6669
|
suggest = "shielded_instance_config"
|
6632
6670
|
elif key == "soleTenantConfig":
|
6633
6671
|
suggest = "sole_tenant_config"
|
6672
|
+
elif key == "storagePools":
|
6673
|
+
suggest = "storage_pools"
|
6634
6674
|
elif key == "workloadMetadataConfig":
|
6635
6675
|
suggest = "workload_metadata_config"
|
6636
6676
|
|
@@ -6683,6 +6723,7 @@ class ClusterNodeConfig(dict):
|
|
6683
6723
|
shielded_instance_config: Optional['outputs.ClusterNodeConfigShieldedInstanceConfig'] = None,
|
6684
6724
|
sole_tenant_config: Optional['outputs.ClusterNodeConfigSoleTenantConfig'] = None,
|
6685
6725
|
spot: Optional[bool] = None,
|
6726
|
+
storage_pools: Optional[Sequence[str]] = None,
|
6686
6727
|
tags: Optional[Sequence[str]] = None,
|
6687
6728
|
taints: Optional[Sequence['outputs.ClusterNodeConfigTaint']] = None,
|
6688
6729
|
workload_metadata_config: Optional['outputs.ClusterNodeConfigWorkloadMetadataConfig'] = None):
|
@@ -6773,6 +6814,7 @@ class ClusterNodeConfig(dict):
|
|
6773
6814
|
:param bool spot: A boolean that represents whether the underlying node VMs are spot.
|
6774
6815
|
See the [official documentation](https://cloud.google.com/kubernetes-engine/docs/concepts/spot-vms)
|
6775
6816
|
for more information. Defaults to false.
|
6817
|
+
:param Sequence[str] storage_pools: The list of Storage Pools where boot disks are provisioned.
|
6776
6818
|
:param Sequence[str] tags: The list of instance tags applied to all nodes. Tags are used to identify
|
6777
6819
|
valid sources or targets for network firewalls.
|
6778
6820
|
:param Sequence['ClusterNodeConfigTaintArgs'] taints: A list of [Kubernetes taints](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
|
@@ -6860,6 +6902,8 @@ class ClusterNodeConfig(dict):
|
|
6860
6902
|
pulumi.set(__self__, "sole_tenant_config", sole_tenant_config)
|
6861
6903
|
if spot is not None:
|
6862
6904
|
pulumi.set(__self__, "spot", spot)
|
6905
|
+
if storage_pools is not None:
|
6906
|
+
pulumi.set(__self__, "storage_pools", storage_pools)
|
6863
6907
|
if tags is not None:
|
6864
6908
|
pulumi.set(__self__, "tags", tags)
|
6865
6909
|
if taints is not None:
|
@@ -7212,6 +7256,14 @@ class ClusterNodeConfig(dict):
|
|
7212
7256
|
"""
|
7213
7257
|
return pulumi.get(self, "spot")
|
7214
7258
|
|
7259
|
+
@property
|
7260
|
+
@pulumi.getter(name="storagePools")
|
7261
|
+
def storage_pools(self) -> Optional[Sequence[str]]:
|
7262
|
+
"""
|
7263
|
+
The list of Storage Pools where boot disks are provisioned.
|
7264
|
+
"""
|
7265
|
+
return pulumi.get(self, "storage_pools")
|
7266
|
+
|
7215
7267
|
@property
|
7216
7268
|
@pulumi.getter
|
7217
7269
|
def tags(self) -> Optional[Sequence[str]]:
|
@@ -7871,12 +7923,12 @@ class ClusterNodeConfigKubeletConfig(dict):
|
|
7871
7923
|
@staticmethod
|
7872
7924
|
def __key_warning(key: str):
|
7873
7925
|
suggest = None
|
7874
|
-
if key == "
|
7875
|
-
suggest = "cpu_manager_policy"
|
7876
|
-
elif key == "cpuCfsQuota":
|
7926
|
+
if key == "cpuCfsQuota":
|
7877
7927
|
suggest = "cpu_cfs_quota"
|
7878
7928
|
elif key == "cpuCfsQuotaPeriod":
|
7879
7929
|
suggest = "cpu_cfs_quota_period"
|
7930
|
+
elif key == "cpuManagerPolicy":
|
7931
|
+
suggest = "cpu_manager_policy"
|
7880
7932
|
elif key == "insecureKubeletReadonlyPortEnabled":
|
7881
7933
|
suggest = "insecure_kubelet_readonly_port_enabled"
|
7882
7934
|
elif key == "podPidsLimit":
|
@@ -7894,49 +7946,35 @@ class ClusterNodeConfigKubeletConfig(dict):
|
|
7894
7946
|
return super().get(key, default)
|
7895
7947
|
|
7896
7948
|
def __init__(__self__, *,
|
7897
|
-
cpu_manager_policy: str,
|
7898
7949
|
cpu_cfs_quota: Optional[bool] = None,
|
7899
7950
|
cpu_cfs_quota_period: Optional[str] = None,
|
7951
|
+
cpu_manager_policy: Optional[str] = None,
|
7900
7952
|
insecure_kubelet_readonly_port_enabled: Optional[str] = None,
|
7901
7953
|
pod_pids_limit: Optional[int] = None):
|
7902
7954
|
"""
|
7903
|
-
:param str cpu_manager_policy: The CPU management policy on the node. See
|
7904
|
-
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
7905
|
-
One of `"none"` or `"static"`. Defaults to `none` when `kubelet_config` is unset.
|
7906
7955
|
:param bool cpu_cfs_quota: If true, enables CPU CFS quota enforcement for
|
7907
7956
|
containers that specify CPU limits.
|
7908
7957
|
:param str cpu_cfs_quota_period: The CPU CFS quota period value. Specified
|
7909
7958
|
as a sequence of decimal numbers, each with optional fraction and a unit suffix,
|
7910
7959
|
such as `"300ms"`. Valid time units are "ns", "us" (or "µs"), "ms", "s", "m",
|
7911
7960
|
"h". The value must be a positive duration.
|
7912
|
-
|
7913
|
-
|
7914
|
-
|
7915
|
-
not specifying the `kubelet_config` block should be the equivalent of specifying
|
7916
|
-
`none`.
|
7961
|
+
:param str cpu_manager_policy: The CPU management policy on the node. See
|
7962
|
+
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
7963
|
+
One of `"none"` or `"static"`. If unset (or set to the empty string `""`), the API will treat the field as if set to "none".
|
7917
7964
|
:param str insecure_kubelet_readonly_port_enabled: Controls whether the kubelet read-only port is enabled. It is strongly recommended to set this to `FALSE`. Possible values: `TRUE`, `FALSE`.
|
7918
7965
|
:param int pod_pids_limit: Controls the maximum number of processes allowed to run in a pod. The value must be greater than or equal to 1024 and less than 4194304.
|
7919
7966
|
"""
|
7920
|
-
pulumi.set(__self__, "cpu_manager_policy", cpu_manager_policy)
|
7921
7967
|
if cpu_cfs_quota is not None:
|
7922
7968
|
pulumi.set(__self__, "cpu_cfs_quota", cpu_cfs_quota)
|
7923
7969
|
if cpu_cfs_quota_period is not None:
|
7924
7970
|
pulumi.set(__self__, "cpu_cfs_quota_period", cpu_cfs_quota_period)
|
7971
|
+
if cpu_manager_policy is not None:
|
7972
|
+
pulumi.set(__self__, "cpu_manager_policy", cpu_manager_policy)
|
7925
7973
|
if insecure_kubelet_readonly_port_enabled is not None:
|
7926
7974
|
pulumi.set(__self__, "insecure_kubelet_readonly_port_enabled", insecure_kubelet_readonly_port_enabled)
|
7927
7975
|
if pod_pids_limit is not None:
|
7928
7976
|
pulumi.set(__self__, "pod_pids_limit", pod_pids_limit)
|
7929
7977
|
|
7930
|
-
@property
|
7931
|
-
@pulumi.getter(name="cpuManagerPolicy")
|
7932
|
-
def cpu_manager_policy(self) -> str:
|
7933
|
-
"""
|
7934
|
-
The CPU management policy on the node. See
|
7935
|
-
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
7936
|
-
One of `"none"` or `"static"`. Defaults to `none` when `kubelet_config` is unset.
|
7937
|
-
"""
|
7938
|
-
return pulumi.get(self, "cpu_manager_policy")
|
7939
|
-
|
7940
7978
|
@property
|
7941
7979
|
@pulumi.getter(name="cpuCfsQuota")
|
7942
7980
|
def cpu_cfs_quota(self) -> Optional[bool]:
|
@@ -7954,14 +7992,19 @@ class ClusterNodeConfigKubeletConfig(dict):
|
|
7954
7992
|
as a sequence of decimal numbers, each with optional fraction and a unit suffix,
|
7955
7993
|
such as `"300ms"`. Valid time units are "ns", "us" (or "µs"), "ms", "s", "m",
|
7956
7994
|
"h". The value must be a positive duration.
|
7957
|
-
|
7958
|
-
> Note: At the time of writing (2020/08/18) the GKE API rejects the `none`
|
7959
|
-
value and accepts an invalid `default` value instead. While this remains true,
|
7960
|
-
not specifying the `kubelet_config` block should be the equivalent of specifying
|
7961
|
-
`none`.
|
7962
7995
|
"""
|
7963
7996
|
return pulumi.get(self, "cpu_cfs_quota_period")
|
7964
7997
|
|
7998
|
+
@property
|
7999
|
+
@pulumi.getter(name="cpuManagerPolicy")
|
8000
|
+
def cpu_manager_policy(self) -> Optional[str]:
|
8001
|
+
"""
|
8002
|
+
The CPU management policy on the node. See
|
8003
|
+
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
8004
|
+
One of `"none"` or `"static"`. If unset (or set to the empty string `""`), the API will treat the field as if set to "none".
|
8005
|
+
"""
|
8006
|
+
return pulumi.get(self, "cpu_manager_policy")
|
8007
|
+
|
7965
8008
|
@property
|
7966
8009
|
@pulumi.getter(name="insecureKubeletReadonlyPortEnabled")
|
7967
8010
|
def insecure_kubelet_readonly_port_enabled(self) -> Optional[str]:
|
@@ -9640,6 +9683,8 @@ class ClusterNodePoolNodeConfig(dict):
|
|
9640
9683
|
suggest = "shielded_instance_config"
|
9641
9684
|
elif key == "soleTenantConfig":
|
9642
9685
|
suggest = "sole_tenant_config"
|
9686
|
+
elif key == "storagePools":
|
9687
|
+
suggest = "storage_pools"
|
9643
9688
|
elif key == "workloadMetadataConfig":
|
9644
9689
|
suggest = "workload_metadata_config"
|
9645
9690
|
|
@@ -9692,6 +9737,7 @@ class ClusterNodePoolNodeConfig(dict):
|
|
9692
9737
|
shielded_instance_config: Optional['outputs.ClusterNodePoolNodeConfigShieldedInstanceConfig'] = None,
|
9693
9738
|
sole_tenant_config: Optional['outputs.ClusterNodePoolNodeConfigSoleTenantConfig'] = None,
|
9694
9739
|
spot: Optional[bool] = None,
|
9740
|
+
storage_pools: Optional[Sequence[str]] = None,
|
9695
9741
|
tags: Optional[Sequence[str]] = None,
|
9696
9742
|
taints: Optional[Sequence['outputs.ClusterNodePoolNodeConfigTaint']] = None,
|
9697
9743
|
workload_metadata_config: Optional['outputs.ClusterNodePoolNodeConfigWorkloadMetadataConfig'] = None):
|
@@ -9782,6 +9828,7 @@ class ClusterNodePoolNodeConfig(dict):
|
|
9782
9828
|
:param bool spot: A boolean that represents whether the underlying node VMs are spot.
|
9783
9829
|
See the [official documentation](https://cloud.google.com/kubernetes-engine/docs/concepts/spot-vms)
|
9784
9830
|
for more information. Defaults to false.
|
9831
|
+
:param Sequence[str] storage_pools: The list of Storage Pools where boot disks are provisioned.
|
9785
9832
|
:param Sequence[str] tags: The list of instance tags applied to all nodes. Tags are used to identify
|
9786
9833
|
valid sources or targets for network firewalls.
|
9787
9834
|
:param Sequence['ClusterNodePoolNodeConfigTaintArgs'] taints: A list of [Kubernetes taints](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
|
@@ -9869,6 +9916,8 @@ class ClusterNodePoolNodeConfig(dict):
|
|
9869
9916
|
pulumi.set(__self__, "sole_tenant_config", sole_tenant_config)
|
9870
9917
|
if spot is not None:
|
9871
9918
|
pulumi.set(__self__, "spot", spot)
|
9919
|
+
if storage_pools is not None:
|
9920
|
+
pulumi.set(__self__, "storage_pools", storage_pools)
|
9872
9921
|
if tags is not None:
|
9873
9922
|
pulumi.set(__self__, "tags", tags)
|
9874
9923
|
if taints is not None:
|
@@ -10221,6 +10270,14 @@ class ClusterNodePoolNodeConfig(dict):
|
|
10221
10270
|
"""
|
10222
10271
|
return pulumi.get(self, "spot")
|
10223
10272
|
|
10273
|
+
@property
|
10274
|
+
@pulumi.getter(name="storagePools")
|
10275
|
+
def storage_pools(self) -> Optional[Sequence[str]]:
|
10276
|
+
"""
|
10277
|
+
The list of Storage Pools where boot disks are provisioned.
|
10278
|
+
"""
|
10279
|
+
return pulumi.get(self, "storage_pools")
|
10280
|
+
|
10224
10281
|
@property
|
10225
10282
|
@pulumi.getter
|
10226
10283
|
def tags(self) -> Optional[Sequence[str]]:
|
@@ -10880,12 +10937,12 @@ class ClusterNodePoolNodeConfigKubeletConfig(dict):
|
|
10880
10937
|
@staticmethod
|
10881
10938
|
def __key_warning(key: str):
|
10882
10939
|
suggest = None
|
10883
|
-
if key == "
|
10884
|
-
suggest = "cpu_manager_policy"
|
10885
|
-
elif key == "cpuCfsQuota":
|
10940
|
+
if key == "cpuCfsQuota":
|
10886
10941
|
suggest = "cpu_cfs_quota"
|
10887
10942
|
elif key == "cpuCfsQuotaPeriod":
|
10888
10943
|
suggest = "cpu_cfs_quota_period"
|
10944
|
+
elif key == "cpuManagerPolicy":
|
10945
|
+
suggest = "cpu_manager_policy"
|
10889
10946
|
elif key == "insecureKubeletReadonlyPortEnabled":
|
10890
10947
|
suggest = "insecure_kubelet_readonly_port_enabled"
|
10891
10948
|
elif key == "podPidsLimit":
|
@@ -10903,49 +10960,35 @@ class ClusterNodePoolNodeConfigKubeletConfig(dict):
|
|
10903
10960
|
return super().get(key, default)
|
10904
10961
|
|
10905
10962
|
def __init__(__self__, *,
|
10906
|
-
cpu_manager_policy: str,
|
10907
10963
|
cpu_cfs_quota: Optional[bool] = None,
|
10908
10964
|
cpu_cfs_quota_period: Optional[str] = None,
|
10965
|
+
cpu_manager_policy: Optional[str] = None,
|
10909
10966
|
insecure_kubelet_readonly_port_enabled: Optional[str] = None,
|
10910
10967
|
pod_pids_limit: Optional[int] = None):
|
10911
10968
|
"""
|
10912
|
-
:param str cpu_manager_policy: The CPU management policy on the node. See
|
10913
|
-
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
10914
|
-
One of `"none"` or `"static"`. Defaults to `none` when `kubelet_config` is unset.
|
10915
10969
|
:param bool cpu_cfs_quota: If true, enables CPU CFS quota enforcement for
|
10916
10970
|
containers that specify CPU limits.
|
10917
10971
|
:param str cpu_cfs_quota_period: The CPU CFS quota period value. Specified
|
10918
10972
|
as a sequence of decimal numbers, each with optional fraction and a unit suffix,
|
10919
10973
|
such as `"300ms"`. Valid time units are "ns", "us" (or "µs"), "ms", "s", "m",
|
10920
10974
|
"h". The value must be a positive duration.
|
10921
|
-
|
10922
|
-
|
10923
|
-
|
10924
|
-
not specifying the `kubelet_config` block should be the equivalent of specifying
|
10925
|
-
`none`.
|
10975
|
+
:param str cpu_manager_policy: The CPU management policy on the node. See
|
10976
|
+
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
10977
|
+
One of `"none"` or `"static"`. If unset (or set to the empty string `""`), the API will treat the field as if set to "none".
|
10926
10978
|
:param str insecure_kubelet_readonly_port_enabled: Controls whether the kubelet read-only port is enabled. It is strongly recommended to set this to `FALSE`. Possible values: `TRUE`, `FALSE`.
|
10927
10979
|
:param int pod_pids_limit: Controls the maximum number of processes allowed to run in a pod. The value must be greater than or equal to 1024 and less than 4194304.
|
10928
10980
|
"""
|
10929
|
-
pulumi.set(__self__, "cpu_manager_policy", cpu_manager_policy)
|
10930
10981
|
if cpu_cfs_quota is not None:
|
10931
10982
|
pulumi.set(__self__, "cpu_cfs_quota", cpu_cfs_quota)
|
10932
10983
|
if cpu_cfs_quota_period is not None:
|
10933
10984
|
pulumi.set(__self__, "cpu_cfs_quota_period", cpu_cfs_quota_period)
|
10985
|
+
if cpu_manager_policy is not None:
|
10986
|
+
pulumi.set(__self__, "cpu_manager_policy", cpu_manager_policy)
|
10934
10987
|
if insecure_kubelet_readonly_port_enabled is not None:
|
10935
10988
|
pulumi.set(__self__, "insecure_kubelet_readonly_port_enabled", insecure_kubelet_readonly_port_enabled)
|
10936
10989
|
if pod_pids_limit is not None:
|
10937
10990
|
pulumi.set(__self__, "pod_pids_limit", pod_pids_limit)
|
10938
10991
|
|
10939
|
-
@property
|
10940
|
-
@pulumi.getter(name="cpuManagerPolicy")
|
10941
|
-
def cpu_manager_policy(self) -> str:
|
10942
|
-
"""
|
10943
|
-
The CPU management policy on the node. See
|
10944
|
-
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
10945
|
-
One of `"none"` or `"static"`. Defaults to `none` when `kubelet_config` is unset.
|
10946
|
-
"""
|
10947
|
-
return pulumi.get(self, "cpu_manager_policy")
|
10948
|
-
|
10949
10992
|
@property
|
10950
10993
|
@pulumi.getter(name="cpuCfsQuota")
|
10951
10994
|
def cpu_cfs_quota(self) -> Optional[bool]:
|
@@ -10963,14 +11006,19 @@ class ClusterNodePoolNodeConfigKubeletConfig(dict):
|
|
10963
11006
|
as a sequence of decimal numbers, each with optional fraction and a unit suffix,
|
10964
11007
|
such as `"300ms"`. Valid time units are "ns", "us" (or "µs"), "ms", "s", "m",
|
10965
11008
|
"h". The value must be a positive duration.
|
10966
|
-
|
10967
|
-
> Note: At the time of writing (2020/08/18) the GKE API rejects the `none`
|
10968
|
-
value and accepts an invalid `default` value instead. While this remains true,
|
10969
|
-
not specifying the `kubelet_config` block should be the equivalent of specifying
|
10970
|
-
`none`.
|
10971
11009
|
"""
|
10972
11010
|
return pulumi.get(self, "cpu_cfs_quota_period")
|
10973
11011
|
|
11012
|
+
@property
|
11013
|
+
@pulumi.getter(name="cpuManagerPolicy")
|
11014
|
+
def cpu_manager_policy(self) -> Optional[str]:
|
11015
|
+
"""
|
11016
|
+
The CPU management policy on the node. See
|
11017
|
+
[K8S CPU Management Policies](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/).
|
11018
|
+
One of `"none"` or `"static"`. If unset (or set to the empty string `""`), the API will treat the field as if set to "none".
|
11019
|
+
"""
|
11020
|
+
return pulumi.get(self, "cpu_manager_policy")
|
11021
|
+
|
10974
11022
|
@property
|
10975
11023
|
@pulumi.getter(name="insecureKubeletReadonlyPortEnabled")
|
10976
11024
|
def insecure_kubelet_readonly_port_enabled(self) -> Optional[str]:
|
@@ -12996,6 +13044,8 @@ class NodePoolNodeConfig(dict):
|
|
12996
13044
|
suggest = "shielded_instance_config"
|
12997
13045
|
elif key == "soleTenantConfig":
|
12998
13046
|
suggest = "sole_tenant_config"
|
13047
|
+
elif key == "storagePools":
|
13048
|
+
suggest = "storage_pools"
|
12999
13049
|
elif key == "workloadMetadataConfig":
|
13000
13050
|
suggest = "workload_metadata_config"
|
13001
13051
|
|
@@ -13048,6 +13098,7 @@ class NodePoolNodeConfig(dict):
|
|
13048
13098
|
shielded_instance_config: Optional['outputs.NodePoolNodeConfigShieldedInstanceConfig'] = None,
|
13049
13099
|
sole_tenant_config: Optional['outputs.NodePoolNodeConfigSoleTenantConfig'] = None,
|
13050
13100
|
spot: Optional[bool] = None,
|
13101
|
+
storage_pools: Optional[Sequence[str]] = None,
|
13051
13102
|
tags: Optional[Sequence[str]] = None,
|
13052
13103
|
taints: Optional[Sequence['outputs.NodePoolNodeConfigTaint']] = None,
|
13053
13104
|
workload_metadata_config: Optional['outputs.NodePoolNodeConfigWorkloadMetadataConfig'] = None):
|
@@ -13092,6 +13143,7 @@ class NodePoolNodeConfig(dict):
|
|
13092
13143
|
:param 'NodePoolNodeConfigShieldedInstanceConfigArgs' shielded_instance_config: Shielded Instance options.
|
13093
13144
|
:param 'NodePoolNodeConfigSoleTenantConfigArgs' sole_tenant_config: Node affinity options for sole tenant node pools.
|
13094
13145
|
:param bool spot: Whether the nodes are created as spot VM instances.
|
13146
|
+
:param Sequence[str] storage_pools: The list of Storage Pools where boot disks are provisioned.
|
13095
13147
|
:param Sequence[str] tags: The list of instance tags applied to all nodes.
|
13096
13148
|
:param Sequence['NodePoolNodeConfigTaintArgs'] taints: List of Kubernetes taints to be applied to each node.
|
13097
13149
|
:param 'NodePoolNodeConfigWorkloadMetadataConfigArgs' workload_metadata_config: The workload metadata configuration for this node.
|
@@ -13170,6 +13222,8 @@ class NodePoolNodeConfig(dict):
|
|
13170
13222
|
pulumi.set(__self__, "sole_tenant_config", sole_tenant_config)
|
13171
13223
|
if spot is not None:
|
13172
13224
|
pulumi.set(__self__, "spot", spot)
|
13225
|
+
if storage_pools is not None:
|
13226
|
+
pulumi.set(__self__, "storage_pools", storage_pools)
|
13173
13227
|
if tags is not None:
|
13174
13228
|
pulumi.set(__self__, "tags", tags)
|
13175
13229
|
if taints is not None:
|
@@ -13476,6 +13530,14 @@ class NodePoolNodeConfig(dict):
|
|
13476
13530
|
"""
|
13477
13531
|
return pulumi.get(self, "spot")
|
13478
13532
|
|
13533
|
+
@property
|
13534
|
+
@pulumi.getter(name="storagePools")
|
13535
|
+
def storage_pools(self) -> Optional[Sequence[str]]:
|
13536
|
+
"""
|
13537
|
+
The list of Storage Pools where boot disks are provisioned.
|
13538
|
+
"""
|
13539
|
+
return pulumi.get(self, "storage_pools")
|
13540
|
+
|
13479
13541
|
@property
|
13480
13542
|
@pulumi.getter
|
13481
13543
|
def tags(self) -> Optional[Sequence[str]]:
|
@@ -14108,12 +14170,12 @@ class NodePoolNodeConfigKubeletConfig(dict):
|
|
14108
14170
|
@staticmethod
|
14109
14171
|
def __key_warning(key: str):
|
14110
14172
|
suggest = None
|
14111
|
-
if key == "
|
14112
|
-
suggest = "cpu_manager_policy"
|
14113
|
-
elif key == "cpuCfsQuota":
|
14173
|
+
if key == "cpuCfsQuota":
|
14114
14174
|
suggest = "cpu_cfs_quota"
|
14115
14175
|
elif key == "cpuCfsQuotaPeriod":
|
14116
14176
|
suggest = "cpu_cfs_quota_period"
|
14177
|
+
elif key == "cpuManagerPolicy":
|
14178
|
+
suggest = "cpu_manager_policy"
|
14117
14179
|
elif key == "insecureKubeletReadonlyPortEnabled":
|
14118
14180
|
suggest = "insecure_kubelet_readonly_port_enabled"
|
14119
14181
|
elif key == "podPidsLimit":
|
@@ -14131,36 +14193,29 @@ class NodePoolNodeConfigKubeletConfig(dict):
|
|
14131
14193
|
return super().get(key, default)
|
14132
14194
|
|
14133
14195
|
def __init__(__self__, *,
|
14134
|
-
cpu_manager_policy: str,
|
14135
14196
|
cpu_cfs_quota: Optional[bool] = None,
|
14136
14197
|
cpu_cfs_quota_period: Optional[str] = None,
|
14198
|
+
cpu_manager_policy: Optional[str] = None,
|
14137
14199
|
insecure_kubelet_readonly_port_enabled: Optional[str] = None,
|
14138
14200
|
pod_pids_limit: Optional[int] = None):
|
14139
14201
|
"""
|
14140
|
-
:param str cpu_manager_policy: Control the CPU management policy on the node.
|
14141
14202
|
:param bool cpu_cfs_quota: Enable CPU CFS quota enforcement for containers that specify CPU limits.
|
14142
14203
|
:param str cpu_cfs_quota_period: Set the CPU CFS quota period value 'cpu.cfs_period_us'.
|
14204
|
+
:param str cpu_manager_policy: Control the CPU management policy on the node.
|
14143
14205
|
:param str insecure_kubelet_readonly_port_enabled: Controls whether the kubelet read-only port is enabled. It is strongly recommended to set this to `FALSE`. Possible values: `TRUE`, `FALSE`.
|
14144
14206
|
:param int pod_pids_limit: Controls the maximum number of processes allowed to run in a pod.
|
14145
14207
|
"""
|
14146
|
-
pulumi.set(__self__, "cpu_manager_policy", cpu_manager_policy)
|
14147
14208
|
if cpu_cfs_quota is not None:
|
14148
14209
|
pulumi.set(__self__, "cpu_cfs_quota", cpu_cfs_quota)
|
14149
14210
|
if cpu_cfs_quota_period is not None:
|
14150
14211
|
pulumi.set(__self__, "cpu_cfs_quota_period", cpu_cfs_quota_period)
|
14212
|
+
if cpu_manager_policy is not None:
|
14213
|
+
pulumi.set(__self__, "cpu_manager_policy", cpu_manager_policy)
|
14151
14214
|
if insecure_kubelet_readonly_port_enabled is not None:
|
14152
14215
|
pulumi.set(__self__, "insecure_kubelet_readonly_port_enabled", insecure_kubelet_readonly_port_enabled)
|
14153
14216
|
if pod_pids_limit is not None:
|
14154
14217
|
pulumi.set(__self__, "pod_pids_limit", pod_pids_limit)
|
14155
14218
|
|
14156
|
-
@property
|
14157
|
-
@pulumi.getter(name="cpuManagerPolicy")
|
14158
|
-
def cpu_manager_policy(self) -> str:
|
14159
|
-
"""
|
14160
|
-
Control the CPU management policy on the node.
|
14161
|
-
"""
|
14162
|
-
return pulumi.get(self, "cpu_manager_policy")
|
14163
|
-
|
14164
14219
|
@property
|
14165
14220
|
@pulumi.getter(name="cpuCfsQuota")
|
14166
14221
|
def cpu_cfs_quota(self) -> Optional[bool]:
|
@@ -14177,6 +14232,14 @@ class NodePoolNodeConfigKubeletConfig(dict):
|
|
14177
14232
|
"""
|
14178
14233
|
return pulumi.get(self, "cpu_cfs_quota_period")
|
14179
14234
|
|
14235
|
+
@property
|
14236
|
+
@pulumi.getter(name="cpuManagerPolicy")
|
14237
|
+
def cpu_manager_policy(self) -> Optional[str]:
|
14238
|
+
"""
|
14239
|
+
Control the CPU management policy on the node.
|
14240
|
+
"""
|
14241
|
+
return pulumi.get(self, "cpu_manager_policy")
|
14242
|
+
|
14180
14243
|
@property
|
14181
14244
|
@pulumi.getter(name="insecureKubeletReadonlyPortEnabled")
|
14182
14245
|
def insecure_kubelet_readonly_port_enabled(self) -> Optional[str]:
|
@@ -16643,6 +16706,7 @@ class GetClusterNodeConfigResult(dict):
|
|
16643
16706
|
shielded_instance_configs: Sequence['outputs.GetClusterNodeConfigShieldedInstanceConfigResult'],
|
16644
16707
|
sole_tenant_configs: Sequence['outputs.GetClusterNodeConfigSoleTenantConfigResult'],
|
16645
16708
|
spot: bool,
|
16709
|
+
storage_pools: Sequence[str],
|
16646
16710
|
tags: Sequence[str],
|
16647
16711
|
taints: Sequence['outputs.GetClusterNodeConfigTaintResult'],
|
16648
16712
|
workload_metadata_configs: Sequence['outputs.GetClusterNodeConfigWorkloadMetadataConfigResult']):
|
@@ -16684,6 +16748,7 @@ class GetClusterNodeConfigResult(dict):
|
|
16684
16748
|
:param Sequence['GetClusterNodeConfigShieldedInstanceConfigArgs'] shielded_instance_configs: Shielded Instance options.
|
16685
16749
|
:param Sequence['GetClusterNodeConfigSoleTenantConfigArgs'] sole_tenant_configs: Node affinity options for sole tenant node pools.
|
16686
16750
|
:param bool spot: Whether the nodes are created as spot VM instances.
|
16751
|
+
:param Sequence[str] storage_pools: The list of Storage Pools where boot disks are provisioned.
|
16687
16752
|
:param Sequence[str] tags: The list of instance tags applied to all nodes.
|
16688
16753
|
:param Sequence['GetClusterNodeConfigTaintArgs'] taints: List of Kubernetes taints to be applied to each node.
|
16689
16754
|
:param Sequence['GetClusterNodeConfigWorkloadMetadataConfigArgs'] workload_metadata_configs: The workload metadata configuration for this node.
|
@@ -16725,6 +16790,7 @@ class GetClusterNodeConfigResult(dict):
|
|
16725
16790
|
pulumi.set(__self__, "shielded_instance_configs", shielded_instance_configs)
|
16726
16791
|
pulumi.set(__self__, "sole_tenant_configs", sole_tenant_configs)
|
16727
16792
|
pulumi.set(__self__, "spot", spot)
|
16793
|
+
pulumi.set(__self__, "storage_pools", storage_pools)
|
16728
16794
|
pulumi.set(__self__, "tags", tags)
|
16729
16795
|
pulumi.set(__self__, "taints", taints)
|
16730
16796
|
pulumi.set(__self__, "workload_metadata_configs", workload_metadata_configs)
|
@@ -17025,6 +17091,14 @@ class GetClusterNodeConfigResult(dict):
|
|
17025
17091
|
"""
|
17026
17092
|
return pulumi.get(self, "spot")
|
17027
17093
|
|
17094
|
+
@property
|
17095
|
+
@pulumi.getter(name="storagePools")
|
17096
|
+
def storage_pools(self) -> Sequence[str]:
|
17097
|
+
"""
|
17098
|
+
The list of Storage Pools where boot disks are provisioned.
|
17099
|
+
"""
|
17100
|
+
return pulumi.get(self, "storage_pools")
|
17101
|
+
|
17028
17102
|
@property
|
17029
17103
|
@pulumi.getter
|
17030
17104
|
def tags(self) -> Sequence[str]:
|
@@ -18550,6 +18624,7 @@ class GetClusterNodePoolNodeConfigResult(dict):
|
|
18550
18624
|
shielded_instance_configs: Sequence['outputs.GetClusterNodePoolNodeConfigShieldedInstanceConfigResult'],
|
18551
18625
|
sole_tenant_configs: Sequence['outputs.GetClusterNodePoolNodeConfigSoleTenantConfigResult'],
|
18552
18626
|
spot: bool,
|
18627
|
+
storage_pools: Sequence[str],
|
18553
18628
|
tags: Sequence[str],
|
18554
18629
|
taints: Sequence['outputs.GetClusterNodePoolNodeConfigTaintResult'],
|
18555
18630
|
workload_metadata_configs: Sequence['outputs.GetClusterNodePoolNodeConfigWorkloadMetadataConfigResult']):
|
@@ -18591,6 +18666,7 @@ class GetClusterNodePoolNodeConfigResult(dict):
|
|
18591
18666
|
:param Sequence['GetClusterNodePoolNodeConfigShieldedInstanceConfigArgs'] shielded_instance_configs: Shielded Instance options.
|
18592
18667
|
:param Sequence['GetClusterNodePoolNodeConfigSoleTenantConfigArgs'] sole_tenant_configs: Node affinity options for sole tenant node pools.
|
18593
18668
|
:param bool spot: Whether the nodes are created as spot VM instances.
|
18669
|
+
:param Sequence[str] storage_pools: The list of Storage Pools where boot disks are provisioned.
|
18594
18670
|
:param Sequence[str] tags: The list of instance tags applied to all nodes.
|
18595
18671
|
:param Sequence['GetClusterNodePoolNodeConfigTaintArgs'] taints: List of Kubernetes taints to be applied to each node.
|
18596
18672
|
:param Sequence['GetClusterNodePoolNodeConfigWorkloadMetadataConfigArgs'] workload_metadata_configs: The workload metadata configuration for this node.
|
@@ -18632,6 +18708,7 @@ class GetClusterNodePoolNodeConfigResult(dict):
|
|
18632
18708
|
pulumi.set(__self__, "shielded_instance_configs", shielded_instance_configs)
|
18633
18709
|
pulumi.set(__self__, "sole_tenant_configs", sole_tenant_configs)
|
18634
18710
|
pulumi.set(__self__, "spot", spot)
|
18711
|
+
pulumi.set(__self__, "storage_pools", storage_pools)
|
18635
18712
|
pulumi.set(__self__, "tags", tags)
|
18636
18713
|
pulumi.set(__self__, "taints", taints)
|
18637
18714
|
pulumi.set(__self__, "workload_metadata_configs", workload_metadata_configs)
|
@@ -18932,6 +19009,14 @@ class GetClusterNodePoolNodeConfigResult(dict):
|
|
18932
19009
|
"""
|
18933
19010
|
return pulumi.get(self, "spot")
|
18934
19011
|
|
19012
|
+
@property
|
19013
|
+
@pulumi.getter(name="storagePools")
|
19014
|
+
def storage_pools(self) -> Sequence[str]:
|
19015
|
+
"""
|
19016
|
+
The list of Storage Pools where boot disks are provisioned.
|
19017
|
+
"""
|
19018
|
+
return pulumi.get(self, "storage_pools")
|
19019
|
+
|
18935
19020
|
@property
|
18936
19021
|
@pulumi.getter
|
18937
19022
|
def tags(self) -> Sequence[str]:
|