pulumi-gcp 7.16.0a1711520590__py3-none-any.whl → 7.17.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- pulumi_gcp/__init__.py +35 -0
- pulumi_gcp/accesscontextmanager/__init__.py +1 -0
- pulumi_gcp/accesscontextmanager/service_perimeter_dry_run_resource.py +345 -0
- pulumi_gcp/accesscontextmanager/service_perimeter_egress_policy.py +80 -0
- pulumi_gcp/accesscontextmanager/service_perimeter_ingress_policy.py +86 -0
- pulumi_gcp/accesscontextmanager/service_perimeter_resource.py +4 -2
- pulumi_gcp/apphub/__init__.py +1 -0
- pulumi_gcp/apphub/get_application.py +220 -0
- pulumi_gcp/apphub/outputs.py +214 -0
- pulumi_gcp/applicationintegration/__init__.py +10 -0
- pulumi_gcp/applicationintegration/_inputs.py +119 -0
- pulumi_gcp/applicationintegration/client.py +566 -0
- pulumi_gcp/applicationintegration/outputs.py +122 -0
- pulumi_gcp/bigquery/_inputs.py +16 -0
- pulumi_gcp/bigquery/outputs.py +14 -0
- pulumi_gcp/bigquery/routine.py +98 -0
- pulumi_gcp/bigtable/_inputs.py +4 -4
- pulumi_gcp/bigtable/gc_policy.py +8 -0
- pulumi_gcp/bigtable/outputs.py +4 -4
- pulumi_gcp/billing/_inputs.py +4 -4
- pulumi_gcp/billing/outputs.py +4 -4
- pulumi_gcp/billing/project_info.py +4 -4
- pulumi_gcp/cloudfunctionsv2/_inputs.py +2 -2
- pulumi_gcp/cloudfunctionsv2/outputs.py +4 -4
- pulumi_gcp/cloudquota/__init__.py +3 -0
- pulumi_gcp/cloudquota/_inputs.py +131 -0
- pulumi_gcp/cloudquota/get_s_quota_infos.py +136 -0
- pulumi_gcp/cloudquota/outputs.py +388 -0
- pulumi_gcp/cloudquota/s_quota_preference.py +777 -0
- pulumi_gcp/cloudrunv2/_inputs.py +59 -2
- pulumi_gcp/cloudrunv2/outputs.py +107 -4
- pulumi_gcp/cloudrunv2/service.py +13 -13
- pulumi_gcp/composer/_inputs.py +30 -4
- pulumi_gcp/composer/outputs.py +45 -10
- pulumi_gcp/compute/_inputs.py +145 -51
- pulumi_gcp/compute/autoscaler.py +14 -14
- pulumi_gcp/compute/interconnect_attachment.py +64 -0
- pulumi_gcp/compute/network_endpoint.py +8 -0
- pulumi_gcp/compute/network_endpoint_list.py +8 -0
- pulumi_gcp/compute/outputs.py +202 -69
- pulumi_gcp/compute/region_autoscaler.py +14 -14
- pulumi_gcp/compute/region_backend_service.py +28 -0
- pulumi_gcp/compute/region_url_map.py +152 -0
- pulumi_gcp/compute/target_instance.py +4 -4
- pulumi_gcp/config/__init__.pyi +2 -0
- pulumi_gcp/config/vars.py +4 -0
- pulumi_gcp/container/_inputs.py +148 -16
- pulumi_gcp/container/outputs.py +148 -16
- pulumi_gcp/databasemigrationservice/connection_profile.py +6 -6
- pulumi_gcp/dataflow/flex_template_job.py +84 -112
- pulumi_gcp/dataform/repository.py +4 -74
- pulumi_gcp/dataloss/_inputs.py +6 -6
- pulumi_gcp/dataloss/outputs.py +6 -6
- pulumi_gcp/dataplex/task.py +16 -16
- pulumi_gcp/dataproc/_inputs.py +85 -10
- pulumi_gcp/dataproc/get_metastore_service.py +11 -1
- pulumi_gcp/dataproc/metastore_service.py +120 -0
- pulumi_gcp/dataproc/outputs.py +142 -10
- pulumi_gcp/firebase/android_app.py +41 -40
- pulumi_gcp/firebase/app_check_service_config.py +2 -2
- pulumi_gcp/firestore/backup_schedule.py +14 -14
- pulumi_gcp/firestore/field.py +4 -4
- pulumi_gcp/firestore/index.py +34 -48
- pulumi_gcp/gkehub/membership_binding.py +6 -6
- pulumi_gcp/gkehub/membership_rbac_role_binding.py +4 -4
- pulumi_gcp/gkehub/namespace.py +4 -4
- pulumi_gcp/gkehub/scope_rbac_role_binding.py +4 -4
- pulumi_gcp/iam/_inputs.py +76 -0
- pulumi_gcp/iam/outputs.py +76 -0
- pulumi_gcp/iam/workforce_pool_provider.py +35 -0
- pulumi_gcp/iam/workload_identity_pool_provider.py +140 -0
- pulumi_gcp/iap/tunnel_dest_group.py +2 -2
- pulumi_gcp/kms/_inputs.py +46 -0
- pulumi_gcp/kms/crypto_key.py +54 -0
- pulumi_gcp/kms/crypto_key_version.py +54 -0
- pulumi_gcp/kms/get_kms_crypto_key.py +11 -1
- pulumi_gcp/kms/outputs.py +54 -0
- pulumi_gcp/logging/_inputs.py +8 -8
- pulumi_gcp/logging/metric.py +7 -7
- pulumi_gcp/logging/outputs.py +8 -8
- pulumi_gcp/monitoring/_inputs.py +4 -2
- pulumi_gcp/monitoring/outputs.py +4 -2
- pulumi_gcp/monitoring/slo.py +4 -4
- pulumi_gcp/networksecurity/firewall_endpoint.py +48 -0
- pulumi_gcp/networkservices/_inputs.py +6 -6
- pulumi_gcp/networkservices/outputs.py +6 -6
- pulumi_gcp/orgpolicy/policy.py +2 -2
- pulumi_gcp/provider.py +20 -0
- pulumi_gcp/pubsub/_inputs.py +108 -0
- pulumi_gcp/pubsub/get_topic.py +11 -1
- pulumi_gcp/pubsub/outputs.py +213 -0
- pulumi_gcp/pubsub/subscription.py +4 -4
- pulumi_gcp/pubsub/topic.py +92 -0
- pulumi_gcp/serviceusage/consumer_quota_override.py +7 -7
- pulumi_gcp/sql/_inputs.py +20 -2
- pulumi_gcp/sql/database_instance.py +2 -2
- pulumi_gcp/sql/outputs.py +20 -2
- pulumi_gcp/storage/_inputs.py +42 -2
- pulumi_gcp/storage/bucket.py +54 -0
- pulumi_gcp/storage/get_bucket.py +11 -1
- pulumi_gcp/storage/outputs.py +83 -2
- pulumi_gcp/vertex/__init__.py +1 -0
- pulumi_gcp/vertex/_inputs.py +175 -8
- pulumi_gcp/vertex/ai_deployment_resource_pool.py +477 -0
- pulumi_gcp/vertex/outputs.py +202 -10
- pulumi_gcp/vpcaccess/connector.py +77 -28
- pulumi_gcp/workstations/_inputs.py +113 -0
- pulumi_gcp/workstations/outputs.py +109 -1
- pulumi_gcp/workstations/workstation_config.py +106 -0
- {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/METADATA +1 -1
- {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/RECORD +113 -103
- {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/WHEEL +0 -0
- {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/top_level.txt +0 -0
pulumi_gcp/vertex/_inputs.py
CHANGED
@@ -11,6 +11,9 @@ from .. import _utilities
|
|
11
11
|
|
12
12
|
__all__ = [
|
13
13
|
'AiDatasetEncryptionSpecArgs',
|
14
|
+
'AiDeploymentResourcePoolDedicatedResourcesArgs',
|
15
|
+
'AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs',
|
16
|
+
'AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs',
|
14
17
|
'AiEndpointDeployedModelArgs',
|
15
18
|
'AiEndpointDeployedModelAutomaticResourceArgs',
|
16
19
|
'AiEndpointDeployedModelDedicatedResourceArgs',
|
@@ -85,6 +88,172 @@ class AiDatasetEncryptionSpecArgs:
|
|
85
88
|
pulumi.set(self, "kms_key_name", value)
|
86
89
|
|
87
90
|
|
91
|
+
@pulumi.input_type
|
92
|
+
class AiDeploymentResourcePoolDedicatedResourcesArgs:
|
93
|
+
def __init__(__self__, *,
|
94
|
+
machine_spec: pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs'],
|
95
|
+
min_replica_count: pulumi.Input[int],
|
96
|
+
autoscaling_metric_specs: Optional[pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]]] = None,
|
97
|
+
max_replica_count: Optional[pulumi.Input[int]] = None):
|
98
|
+
"""
|
99
|
+
:param pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs'] machine_spec: The specification of a single machine used by the prediction
|
100
|
+
Structure is documented below.
|
101
|
+
:param pulumi.Input[int] min_replica_count: The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
|
102
|
+
:param pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]] autoscaling_metric_specs: A list of the metric specifications that overrides a resource utilization metric.
|
103
|
+
Structure is documented below.
|
104
|
+
:param pulumi.Input[int] max_replica_count: The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
|
105
|
+
"""
|
106
|
+
pulumi.set(__self__, "machine_spec", machine_spec)
|
107
|
+
pulumi.set(__self__, "min_replica_count", min_replica_count)
|
108
|
+
if autoscaling_metric_specs is not None:
|
109
|
+
pulumi.set(__self__, "autoscaling_metric_specs", autoscaling_metric_specs)
|
110
|
+
if max_replica_count is not None:
|
111
|
+
pulumi.set(__self__, "max_replica_count", max_replica_count)
|
112
|
+
|
113
|
+
@property
|
114
|
+
@pulumi.getter(name="machineSpec")
|
115
|
+
def machine_spec(self) -> pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs']:
|
116
|
+
"""
|
117
|
+
The specification of a single machine used by the prediction
|
118
|
+
Structure is documented below.
|
119
|
+
"""
|
120
|
+
return pulumi.get(self, "machine_spec")
|
121
|
+
|
122
|
+
@machine_spec.setter
|
123
|
+
def machine_spec(self, value: pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs']):
|
124
|
+
pulumi.set(self, "machine_spec", value)
|
125
|
+
|
126
|
+
@property
|
127
|
+
@pulumi.getter(name="minReplicaCount")
|
128
|
+
def min_replica_count(self) -> pulumi.Input[int]:
|
129
|
+
"""
|
130
|
+
The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
|
131
|
+
"""
|
132
|
+
return pulumi.get(self, "min_replica_count")
|
133
|
+
|
134
|
+
@min_replica_count.setter
|
135
|
+
def min_replica_count(self, value: pulumi.Input[int]):
|
136
|
+
pulumi.set(self, "min_replica_count", value)
|
137
|
+
|
138
|
+
@property
|
139
|
+
@pulumi.getter(name="autoscalingMetricSpecs")
|
140
|
+
def autoscaling_metric_specs(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]]]:
|
141
|
+
"""
|
142
|
+
A list of the metric specifications that overrides a resource utilization metric.
|
143
|
+
Structure is documented below.
|
144
|
+
"""
|
145
|
+
return pulumi.get(self, "autoscaling_metric_specs")
|
146
|
+
|
147
|
+
@autoscaling_metric_specs.setter
|
148
|
+
def autoscaling_metric_specs(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]]]):
|
149
|
+
pulumi.set(self, "autoscaling_metric_specs", value)
|
150
|
+
|
151
|
+
@property
|
152
|
+
@pulumi.getter(name="maxReplicaCount")
|
153
|
+
def max_replica_count(self) -> Optional[pulumi.Input[int]]:
|
154
|
+
"""
|
155
|
+
The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
|
156
|
+
"""
|
157
|
+
return pulumi.get(self, "max_replica_count")
|
158
|
+
|
159
|
+
@max_replica_count.setter
|
160
|
+
def max_replica_count(self, value: Optional[pulumi.Input[int]]):
|
161
|
+
pulumi.set(self, "max_replica_count", value)
|
162
|
+
|
163
|
+
|
164
|
+
@pulumi.input_type
|
165
|
+
class AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs:
|
166
|
+
def __init__(__self__, *,
|
167
|
+
metric_name: pulumi.Input[str],
|
168
|
+
target: Optional[pulumi.Input[int]] = None):
|
169
|
+
"""
|
170
|
+
:param pulumi.Input[str] metric_name: The resource metric name. Supported metrics: For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
|
171
|
+
:param pulumi.Input[int] target: The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
|
172
|
+
"""
|
173
|
+
pulumi.set(__self__, "metric_name", metric_name)
|
174
|
+
if target is not None:
|
175
|
+
pulumi.set(__self__, "target", target)
|
176
|
+
|
177
|
+
@property
|
178
|
+
@pulumi.getter(name="metricName")
|
179
|
+
def metric_name(self) -> pulumi.Input[str]:
|
180
|
+
"""
|
181
|
+
The resource metric name. Supported metrics: For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
|
182
|
+
"""
|
183
|
+
return pulumi.get(self, "metric_name")
|
184
|
+
|
185
|
+
@metric_name.setter
|
186
|
+
def metric_name(self, value: pulumi.Input[str]):
|
187
|
+
pulumi.set(self, "metric_name", value)
|
188
|
+
|
189
|
+
@property
|
190
|
+
@pulumi.getter
|
191
|
+
def target(self) -> Optional[pulumi.Input[int]]:
|
192
|
+
"""
|
193
|
+
The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
|
194
|
+
"""
|
195
|
+
return pulumi.get(self, "target")
|
196
|
+
|
197
|
+
@target.setter
|
198
|
+
def target(self, value: Optional[pulumi.Input[int]]):
|
199
|
+
pulumi.set(self, "target", value)
|
200
|
+
|
201
|
+
|
202
|
+
@pulumi.input_type
|
203
|
+
class AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs:
|
204
|
+
def __init__(__self__, *,
|
205
|
+
accelerator_count: Optional[pulumi.Input[int]] = None,
|
206
|
+
accelerator_type: Optional[pulumi.Input[str]] = None,
|
207
|
+
machine_type: Optional[pulumi.Input[str]] = None):
|
208
|
+
"""
|
209
|
+
:param pulumi.Input[int] accelerator_count: The number of accelerators to attach to the machine.
|
210
|
+
:param pulumi.Input[str] accelerator_type: The type of accelerator(s) that may be attached to the machine as per accelerator_count. See possible values [here](https://cloud.google.com/vertex-ai/docs/reference/rest/v1/MachineSpec#AcceleratorType).
|
211
|
+
:param pulumi.Input[str] machine_type: The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types).
|
212
|
+
"""
|
213
|
+
if accelerator_count is not None:
|
214
|
+
pulumi.set(__self__, "accelerator_count", accelerator_count)
|
215
|
+
if accelerator_type is not None:
|
216
|
+
pulumi.set(__self__, "accelerator_type", accelerator_type)
|
217
|
+
if machine_type is not None:
|
218
|
+
pulumi.set(__self__, "machine_type", machine_type)
|
219
|
+
|
220
|
+
@property
|
221
|
+
@pulumi.getter(name="acceleratorCount")
|
222
|
+
def accelerator_count(self) -> Optional[pulumi.Input[int]]:
|
223
|
+
"""
|
224
|
+
The number of accelerators to attach to the machine.
|
225
|
+
"""
|
226
|
+
return pulumi.get(self, "accelerator_count")
|
227
|
+
|
228
|
+
@accelerator_count.setter
|
229
|
+
def accelerator_count(self, value: Optional[pulumi.Input[int]]):
|
230
|
+
pulumi.set(self, "accelerator_count", value)
|
231
|
+
|
232
|
+
@property
|
233
|
+
@pulumi.getter(name="acceleratorType")
|
234
|
+
def accelerator_type(self) -> Optional[pulumi.Input[str]]:
|
235
|
+
"""
|
236
|
+
The type of accelerator(s) that may be attached to the machine as per accelerator_count. See possible values [here](https://cloud.google.com/vertex-ai/docs/reference/rest/v1/MachineSpec#AcceleratorType).
|
237
|
+
"""
|
238
|
+
return pulumi.get(self, "accelerator_type")
|
239
|
+
|
240
|
+
@accelerator_type.setter
|
241
|
+
def accelerator_type(self, value: Optional[pulumi.Input[str]]):
|
242
|
+
pulumi.set(self, "accelerator_type", value)
|
243
|
+
|
244
|
+
@property
|
245
|
+
@pulumi.getter(name="machineType")
|
246
|
+
def machine_type(self) -> Optional[pulumi.Input[str]]:
|
247
|
+
"""
|
248
|
+
The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types).
|
249
|
+
"""
|
250
|
+
return pulumi.get(self, "machine_type")
|
251
|
+
|
252
|
+
@machine_type.setter
|
253
|
+
def machine_type(self, value: Optional[pulumi.Input[str]]):
|
254
|
+
pulumi.set(self, "machine_type", value)
|
255
|
+
|
256
|
+
|
88
257
|
@pulumi.input_type
|
89
258
|
class AiEndpointDeployedModelArgs:
|
90
259
|
def __init__(__self__, *,
|
@@ -447,7 +616,7 @@ class AiEndpointDeployedModelDedicatedResourceAutoscalingMetricSpecArgs:
|
|
447
616
|
:param pulumi.Input[str] metric_name: (Output)
|
448
617
|
The resource metric name. Supported metrics: * For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
|
449
618
|
:param pulumi.Input[int] target: (Output)
|
450
|
-
The target resource utilization in percentage (1%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60
|
619
|
+
The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
|
451
620
|
"""
|
452
621
|
if metric_name is not None:
|
453
622
|
pulumi.set(__self__, "metric_name", metric_name)
|
@@ -472,7 +641,7 @@ class AiEndpointDeployedModelDedicatedResourceAutoscalingMetricSpecArgs:
|
|
472
641
|
def target(self) -> Optional[pulumi.Input[int]]:
|
473
642
|
"""
|
474
643
|
(Output)
|
475
|
-
The target resource utilization in percentage (1%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60
|
644
|
+
The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
|
476
645
|
"""
|
477
646
|
return pulumi.get(self, "target")
|
478
647
|
|
@@ -816,8 +985,7 @@ class AiFeatureOnlineStoreBigtableAutoScalingArgs:
|
|
816
985
|
"""
|
817
986
|
:param pulumi.Input[int] max_node_count: The maximum number of nodes to scale up to. Must be greater than or equal to minNodeCount, and less than or equal to 10 times of 'minNodeCount'.
|
818
987
|
:param pulumi.Input[int] min_node_count: The minimum number of nodes to scale down to. Must be greater than or equal to 1.
|
819
|
-
:param pulumi.Input[int] cpu_utilization_target: A percentage of the cluster's CPU capacity. Can be from 10
|
820
|
-
(MISSING)
|
988
|
+
:param pulumi.Input[int] cpu_utilization_target: A percentage of the cluster's CPU capacity. Can be from 10% to 80%. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%.
|
821
989
|
"""
|
822
990
|
pulumi.set(__self__, "max_node_count", max_node_count)
|
823
991
|
pulumi.set(__self__, "min_node_count", min_node_count)
|
@@ -852,8 +1020,7 @@ class AiFeatureOnlineStoreBigtableAutoScalingArgs:
|
|
852
1020
|
@pulumi.getter(name="cpuUtilizationTarget")
|
853
1021
|
def cpu_utilization_target(self) -> Optional[pulumi.Input[int]]:
|
854
1022
|
"""
|
855
|
-
A percentage of the cluster's CPU capacity. Can be from 10
|
856
|
-
(MISSING)
|
1023
|
+
A percentage of the cluster's CPU capacity. Can be from 10% to 80%. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%.
|
857
1024
|
"""
|
858
1025
|
return pulumi.get(self, "cpu_utilization_target")
|
859
1026
|
|
@@ -2162,7 +2329,7 @@ class AiIndexMetadataConfigAlgorithmConfigTreeAhConfigArgs:
|
|
2162
2329
|
"""
|
2163
2330
|
:param pulumi.Input[int] leaf_node_embedding_count: Number of embeddings on each leaf node. The default value is 1000 if not set.
|
2164
2331
|
:param pulumi.Input[int] leaf_nodes_to_search_percent: The default percentage of leaf nodes that any query may be searched. Must be in
|
2165
|
-
range 1-100, inclusive. The default value is 10 (means 10
|
2332
|
+
range 1-100, inclusive. The default value is 10 (means 10%) if not set.
|
2166
2333
|
"""
|
2167
2334
|
if leaf_node_embedding_count is not None:
|
2168
2335
|
pulumi.set(__self__, "leaf_node_embedding_count", leaf_node_embedding_count)
|
@@ -2186,7 +2353,7 @@ class AiIndexMetadataConfigAlgorithmConfigTreeAhConfigArgs:
|
|
2186
2353
|
def leaf_nodes_to_search_percent(self) -> Optional[pulumi.Input[int]]:
|
2187
2354
|
"""
|
2188
2355
|
The default percentage of leaf nodes that any query may be searched. Must be in
|
2189
|
-
range 1-100, inclusive. The default value is 10 (means 10
|
2356
|
+
range 1-100, inclusive. The default value is 10 (means 10%) if not set.
|
2190
2357
|
"""
|
2191
2358
|
return pulumi.get(self, "leaf_nodes_to_search_percent")
|
2192
2359
|
|