pulumi-gcp 7.16.0a1711520590__py3-none-any.whl → 7.17.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (113) hide show
  1. pulumi_gcp/__init__.py +35 -0
  2. pulumi_gcp/accesscontextmanager/__init__.py +1 -0
  3. pulumi_gcp/accesscontextmanager/service_perimeter_dry_run_resource.py +345 -0
  4. pulumi_gcp/accesscontextmanager/service_perimeter_egress_policy.py +80 -0
  5. pulumi_gcp/accesscontextmanager/service_perimeter_ingress_policy.py +86 -0
  6. pulumi_gcp/accesscontextmanager/service_perimeter_resource.py +4 -2
  7. pulumi_gcp/apphub/__init__.py +1 -0
  8. pulumi_gcp/apphub/get_application.py +220 -0
  9. pulumi_gcp/apphub/outputs.py +214 -0
  10. pulumi_gcp/applicationintegration/__init__.py +10 -0
  11. pulumi_gcp/applicationintegration/_inputs.py +119 -0
  12. pulumi_gcp/applicationintegration/client.py +566 -0
  13. pulumi_gcp/applicationintegration/outputs.py +122 -0
  14. pulumi_gcp/bigquery/_inputs.py +16 -0
  15. pulumi_gcp/bigquery/outputs.py +14 -0
  16. pulumi_gcp/bigquery/routine.py +98 -0
  17. pulumi_gcp/bigtable/_inputs.py +4 -4
  18. pulumi_gcp/bigtable/gc_policy.py +8 -0
  19. pulumi_gcp/bigtable/outputs.py +4 -4
  20. pulumi_gcp/billing/_inputs.py +4 -4
  21. pulumi_gcp/billing/outputs.py +4 -4
  22. pulumi_gcp/billing/project_info.py +4 -4
  23. pulumi_gcp/cloudfunctionsv2/_inputs.py +2 -2
  24. pulumi_gcp/cloudfunctionsv2/outputs.py +4 -4
  25. pulumi_gcp/cloudquota/__init__.py +3 -0
  26. pulumi_gcp/cloudquota/_inputs.py +131 -0
  27. pulumi_gcp/cloudquota/get_s_quota_infos.py +136 -0
  28. pulumi_gcp/cloudquota/outputs.py +388 -0
  29. pulumi_gcp/cloudquota/s_quota_preference.py +777 -0
  30. pulumi_gcp/cloudrunv2/_inputs.py +59 -2
  31. pulumi_gcp/cloudrunv2/outputs.py +107 -4
  32. pulumi_gcp/cloudrunv2/service.py +13 -13
  33. pulumi_gcp/composer/_inputs.py +30 -4
  34. pulumi_gcp/composer/outputs.py +45 -10
  35. pulumi_gcp/compute/_inputs.py +145 -51
  36. pulumi_gcp/compute/autoscaler.py +14 -14
  37. pulumi_gcp/compute/interconnect_attachment.py +64 -0
  38. pulumi_gcp/compute/network_endpoint.py +8 -0
  39. pulumi_gcp/compute/network_endpoint_list.py +8 -0
  40. pulumi_gcp/compute/outputs.py +202 -69
  41. pulumi_gcp/compute/region_autoscaler.py +14 -14
  42. pulumi_gcp/compute/region_backend_service.py +28 -0
  43. pulumi_gcp/compute/region_url_map.py +152 -0
  44. pulumi_gcp/compute/target_instance.py +4 -4
  45. pulumi_gcp/config/__init__.pyi +2 -0
  46. pulumi_gcp/config/vars.py +4 -0
  47. pulumi_gcp/container/_inputs.py +148 -16
  48. pulumi_gcp/container/outputs.py +148 -16
  49. pulumi_gcp/databasemigrationservice/connection_profile.py +6 -6
  50. pulumi_gcp/dataflow/flex_template_job.py +84 -112
  51. pulumi_gcp/dataform/repository.py +4 -74
  52. pulumi_gcp/dataloss/_inputs.py +6 -6
  53. pulumi_gcp/dataloss/outputs.py +6 -6
  54. pulumi_gcp/dataplex/task.py +16 -16
  55. pulumi_gcp/dataproc/_inputs.py +85 -10
  56. pulumi_gcp/dataproc/get_metastore_service.py +11 -1
  57. pulumi_gcp/dataproc/metastore_service.py +120 -0
  58. pulumi_gcp/dataproc/outputs.py +142 -10
  59. pulumi_gcp/firebase/android_app.py +41 -40
  60. pulumi_gcp/firebase/app_check_service_config.py +2 -2
  61. pulumi_gcp/firestore/backup_schedule.py +14 -14
  62. pulumi_gcp/firestore/field.py +4 -4
  63. pulumi_gcp/firestore/index.py +34 -48
  64. pulumi_gcp/gkehub/membership_binding.py +6 -6
  65. pulumi_gcp/gkehub/membership_rbac_role_binding.py +4 -4
  66. pulumi_gcp/gkehub/namespace.py +4 -4
  67. pulumi_gcp/gkehub/scope_rbac_role_binding.py +4 -4
  68. pulumi_gcp/iam/_inputs.py +76 -0
  69. pulumi_gcp/iam/outputs.py +76 -0
  70. pulumi_gcp/iam/workforce_pool_provider.py +35 -0
  71. pulumi_gcp/iam/workload_identity_pool_provider.py +140 -0
  72. pulumi_gcp/iap/tunnel_dest_group.py +2 -2
  73. pulumi_gcp/kms/_inputs.py +46 -0
  74. pulumi_gcp/kms/crypto_key.py +54 -0
  75. pulumi_gcp/kms/crypto_key_version.py +54 -0
  76. pulumi_gcp/kms/get_kms_crypto_key.py +11 -1
  77. pulumi_gcp/kms/outputs.py +54 -0
  78. pulumi_gcp/logging/_inputs.py +8 -8
  79. pulumi_gcp/logging/metric.py +7 -7
  80. pulumi_gcp/logging/outputs.py +8 -8
  81. pulumi_gcp/monitoring/_inputs.py +4 -2
  82. pulumi_gcp/monitoring/outputs.py +4 -2
  83. pulumi_gcp/monitoring/slo.py +4 -4
  84. pulumi_gcp/networksecurity/firewall_endpoint.py +48 -0
  85. pulumi_gcp/networkservices/_inputs.py +6 -6
  86. pulumi_gcp/networkservices/outputs.py +6 -6
  87. pulumi_gcp/orgpolicy/policy.py +2 -2
  88. pulumi_gcp/provider.py +20 -0
  89. pulumi_gcp/pubsub/_inputs.py +108 -0
  90. pulumi_gcp/pubsub/get_topic.py +11 -1
  91. pulumi_gcp/pubsub/outputs.py +213 -0
  92. pulumi_gcp/pubsub/subscription.py +4 -4
  93. pulumi_gcp/pubsub/topic.py +92 -0
  94. pulumi_gcp/serviceusage/consumer_quota_override.py +7 -7
  95. pulumi_gcp/sql/_inputs.py +20 -2
  96. pulumi_gcp/sql/database_instance.py +2 -2
  97. pulumi_gcp/sql/outputs.py +20 -2
  98. pulumi_gcp/storage/_inputs.py +42 -2
  99. pulumi_gcp/storage/bucket.py +54 -0
  100. pulumi_gcp/storage/get_bucket.py +11 -1
  101. pulumi_gcp/storage/outputs.py +83 -2
  102. pulumi_gcp/vertex/__init__.py +1 -0
  103. pulumi_gcp/vertex/_inputs.py +175 -8
  104. pulumi_gcp/vertex/ai_deployment_resource_pool.py +477 -0
  105. pulumi_gcp/vertex/outputs.py +202 -10
  106. pulumi_gcp/vpcaccess/connector.py +77 -28
  107. pulumi_gcp/workstations/_inputs.py +113 -0
  108. pulumi_gcp/workstations/outputs.py +109 -1
  109. pulumi_gcp/workstations/workstation_config.py +106 -0
  110. {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/METADATA +1 -1
  111. {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/RECORD +113 -103
  112. {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/WHEEL +0 -0
  113. {pulumi_gcp-7.16.0a1711520590.dist-info → pulumi_gcp-7.17.0.dist-info}/top_level.txt +0 -0
@@ -11,6 +11,9 @@ from .. import _utilities
11
11
 
12
12
  __all__ = [
13
13
  'AiDatasetEncryptionSpecArgs',
14
+ 'AiDeploymentResourcePoolDedicatedResourcesArgs',
15
+ 'AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs',
16
+ 'AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs',
14
17
  'AiEndpointDeployedModelArgs',
15
18
  'AiEndpointDeployedModelAutomaticResourceArgs',
16
19
  'AiEndpointDeployedModelDedicatedResourceArgs',
@@ -85,6 +88,172 @@ class AiDatasetEncryptionSpecArgs:
85
88
  pulumi.set(self, "kms_key_name", value)
86
89
 
87
90
 
91
+ @pulumi.input_type
92
+ class AiDeploymentResourcePoolDedicatedResourcesArgs:
93
+ def __init__(__self__, *,
94
+ machine_spec: pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs'],
95
+ min_replica_count: pulumi.Input[int],
96
+ autoscaling_metric_specs: Optional[pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]]] = None,
97
+ max_replica_count: Optional[pulumi.Input[int]] = None):
98
+ """
99
+ :param pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs'] machine_spec: The specification of a single machine used by the prediction
100
+ Structure is documented below.
101
+ :param pulumi.Input[int] min_replica_count: The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
102
+ :param pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]] autoscaling_metric_specs: A list of the metric specifications that overrides a resource utilization metric.
103
+ Structure is documented below.
104
+ :param pulumi.Input[int] max_replica_count: The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
105
+ """
106
+ pulumi.set(__self__, "machine_spec", machine_spec)
107
+ pulumi.set(__self__, "min_replica_count", min_replica_count)
108
+ if autoscaling_metric_specs is not None:
109
+ pulumi.set(__self__, "autoscaling_metric_specs", autoscaling_metric_specs)
110
+ if max_replica_count is not None:
111
+ pulumi.set(__self__, "max_replica_count", max_replica_count)
112
+
113
+ @property
114
+ @pulumi.getter(name="machineSpec")
115
+ def machine_spec(self) -> pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs']:
116
+ """
117
+ The specification of a single machine used by the prediction
118
+ Structure is documented below.
119
+ """
120
+ return pulumi.get(self, "machine_spec")
121
+
122
+ @machine_spec.setter
123
+ def machine_spec(self, value: pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs']):
124
+ pulumi.set(self, "machine_spec", value)
125
+
126
+ @property
127
+ @pulumi.getter(name="minReplicaCount")
128
+ def min_replica_count(self) -> pulumi.Input[int]:
129
+ """
130
+ The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
131
+ """
132
+ return pulumi.get(self, "min_replica_count")
133
+
134
+ @min_replica_count.setter
135
+ def min_replica_count(self, value: pulumi.Input[int]):
136
+ pulumi.set(self, "min_replica_count", value)
137
+
138
+ @property
139
+ @pulumi.getter(name="autoscalingMetricSpecs")
140
+ def autoscaling_metric_specs(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]]]:
141
+ """
142
+ A list of the metric specifications that overrides a resource utilization metric.
143
+ Structure is documented below.
144
+ """
145
+ return pulumi.get(self, "autoscaling_metric_specs")
146
+
147
+ @autoscaling_metric_specs.setter
148
+ def autoscaling_metric_specs(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs']]]]):
149
+ pulumi.set(self, "autoscaling_metric_specs", value)
150
+
151
+ @property
152
+ @pulumi.getter(name="maxReplicaCount")
153
+ def max_replica_count(self) -> Optional[pulumi.Input[int]]:
154
+ """
155
+ The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
156
+ """
157
+ return pulumi.get(self, "max_replica_count")
158
+
159
+ @max_replica_count.setter
160
+ def max_replica_count(self, value: Optional[pulumi.Input[int]]):
161
+ pulumi.set(self, "max_replica_count", value)
162
+
163
+
164
+ @pulumi.input_type
165
+ class AiDeploymentResourcePoolDedicatedResourcesAutoscalingMetricSpecArgs:
166
+ def __init__(__self__, *,
167
+ metric_name: pulumi.Input[str],
168
+ target: Optional[pulumi.Input[int]] = None):
169
+ """
170
+ :param pulumi.Input[str] metric_name: The resource metric name. Supported metrics: For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
171
+ :param pulumi.Input[int] target: The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
172
+ """
173
+ pulumi.set(__self__, "metric_name", metric_name)
174
+ if target is not None:
175
+ pulumi.set(__self__, "target", target)
176
+
177
+ @property
178
+ @pulumi.getter(name="metricName")
179
+ def metric_name(self) -> pulumi.Input[str]:
180
+ """
181
+ The resource metric name. Supported metrics: For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
182
+ """
183
+ return pulumi.get(self, "metric_name")
184
+
185
+ @metric_name.setter
186
+ def metric_name(self, value: pulumi.Input[str]):
187
+ pulumi.set(self, "metric_name", value)
188
+
189
+ @property
190
+ @pulumi.getter
191
+ def target(self) -> Optional[pulumi.Input[int]]:
192
+ """
193
+ The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
194
+ """
195
+ return pulumi.get(self, "target")
196
+
197
+ @target.setter
198
+ def target(self, value: Optional[pulumi.Input[int]]):
199
+ pulumi.set(self, "target", value)
200
+
201
+
202
+ @pulumi.input_type
203
+ class AiDeploymentResourcePoolDedicatedResourcesMachineSpecArgs:
204
+ def __init__(__self__, *,
205
+ accelerator_count: Optional[pulumi.Input[int]] = None,
206
+ accelerator_type: Optional[pulumi.Input[str]] = None,
207
+ machine_type: Optional[pulumi.Input[str]] = None):
208
+ """
209
+ :param pulumi.Input[int] accelerator_count: The number of accelerators to attach to the machine.
210
+ :param pulumi.Input[str] accelerator_type: The type of accelerator(s) that may be attached to the machine as per accelerator_count. See possible values [here](https://cloud.google.com/vertex-ai/docs/reference/rest/v1/MachineSpec#AcceleratorType).
211
+ :param pulumi.Input[str] machine_type: The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types).
212
+ """
213
+ if accelerator_count is not None:
214
+ pulumi.set(__self__, "accelerator_count", accelerator_count)
215
+ if accelerator_type is not None:
216
+ pulumi.set(__self__, "accelerator_type", accelerator_type)
217
+ if machine_type is not None:
218
+ pulumi.set(__self__, "machine_type", machine_type)
219
+
220
+ @property
221
+ @pulumi.getter(name="acceleratorCount")
222
+ def accelerator_count(self) -> Optional[pulumi.Input[int]]:
223
+ """
224
+ The number of accelerators to attach to the machine.
225
+ """
226
+ return pulumi.get(self, "accelerator_count")
227
+
228
+ @accelerator_count.setter
229
+ def accelerator_count(self, value: Optional[pulumi.Input[int]]):
230
+ pulumi.set(self, "accelerator_count", value)
231
+
232
+ @property
233
+ @pulumi.getter(name="acceleratorType")
234
+ def accelerator_type(self) -> Optional[pulumi.Input[str]]:
235
+ """
236
+ The type of accelerator(s) that may be attached to the machine as per accelerator_count. See possible values [here](https://cloud.google.com/vertex-ai/docs/reference/rest/v1/MachineSpec#AcceleratorType).
237
+ """
238
+ return pulumi.get(self, "accelerator_type")
239
+
240
+ @accelerator_type.setter
241
+ def accelerator_type(self, value: Optional[pulumi.Input[str]]):
242
+ pulumi.set(self, "accelerator_type", value)
243
+
244
+ @property
245
+ @pulumi.getter(name="machineType")
246
+ def machine_type(self) -> Optional[pulumi.Input[str]]:
247
+ """
248
+ The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types).
249
+ """
250
+ return pulumi.get(self, "machine_type")
251
+
252
+ @machine_type.setter
253
+ def machine_type(self, value: Optional[pulumi.Input[str]]):
254
+ pulumi.set(self, "machine_type", value)
255
+
256
+
88
257
  @pulumi.input_type
89
258
  class AiEndpointDeployedModelArgs:
90
259
  def __init__(__self__, *,
@@ -447,7 +616,7 @@ class AiEndpointDeployedModelDedicatedResourceAutoscalingMetricSpecArgs:
447
616
  :param pulumi.Input[str] metric_name: (Output)
448
617
  The resource metric name. Supported metrics: * For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
449
618
  :param pulumi.Input[int] target: (Output)
450
- The target resource utilization in percentage (1%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%!)(MISSING) if not provided.
619
+ The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
451
620
  """
452
621
  if metric_name is not None:
453
622
  pulumi.set(__self__, "metric_name", metric_name)
@@ -472,7 +641,7 @@ class AiEndpointDeployedModelDedicatedResourceAutoscalingMetricSpecArgs:
472
641
  def target(self) -> Optional[pulumi.Input[int]]:
473
642
  """
474
643
  (Output)
475
- The target resource utilization in percentage (1%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%!)(MISSING) if not provided.
644
+ The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
476
645
  """
477
646
  return pulumi.get(self, "target")
478
647
 
@@ -816,8 +985,7 @@ class AiFeatureOnlineStoreBigtableAutoScalingArgs:
816
985
  """
817
986
  :param pulumi.Input[int] max_node_count: The maximum number of nodes to scale up to. Must be greater than or equal to minNodeCount, and less than or equal to 10 times of 'minNodeCount'.
818
987
  :param pulumi.Input[int] min_node_count: The minimum number of nodes to scale down to. Must be greater than or equal to 1.
819
- :param pulumi.Input[int] cpu_utilization_target: A percentage of the cluster's CPU capacity. Can be from 10%!t(MISSING)o 80%! (MISSING)When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%!
820
- (MISSING)
988
+ :param pulumi.Input[int] cpu_utilization_target: A percentage of the cluster's CPU capacity. Can be from 10% to 80%. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%.
821
989
  """
822
990
  pulumi.set(__self__, "max_node_count", max_node_count)
823
991
  pulumi.set(__self__, "min_node_count", min_node_count)
@@ -852,8 +1020,7 @@ class AiFeatureOnlineStoreBigtableAutoScalingArgs:
852
1020
  @pulumi.getter(name="cpuUtilizationTarget")
853
1021
  def cpu_utilization_target(self) -> Optional[pulumi.Input[int]]:
854
1022
  """
855
- A percentage of the cluster's CPU capacity. Can be from 10%!t(MISSING)o 80%! (MISSING)When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%!
856
- (MISSING)
1023
+ A percentage of the cluster's CPU capacity. Can be from 10% to 80%. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%.
857
1024
  """
858
1025
  return pulumi.get(self, "cpu_utilization_target")
859
1026
 
@@ -2162,7 +2329,7 @@ class AiIndexMetadataConfigAlgorithmConfigTreeAhConfigArgs:
2162
2329
  """
2163
2330
  :param pulumi.Input[int] leaf_node_embedding_count: Number of embeddings on each leaf node. The default value is 1000 if not set.
2164
2331
  :param pulumi.Input[int] leaf_nodes_to_search_percent: The default percentage of leaf nodes that any query may be searched. Must be in
2165
- range 1-100, inclusive. The default value is 10 (means 10%!)(MISSING) if not set.
2332
+ range 1-100, inclusive. The default value is 10 (means 10%) if not set.
2166
2333
  """
2167
2334
  if leaf_node_embedding_count is not None:
2168
2335
  pulumi.set(__self__, "leaf_node_embedding_count", leaf_node_embedding_count)
@@ -2186,7 +2353,7 @@ class AiIndexMetadataConfigAlgorithmConfigTreeAhConfigArgs:
2186
2353
  def leaf_nodes_to_search_percent(self) -> Optional[pulumi.Input[int]]:
2187
2354
  """
2188
2355
  The default percentage of leaf nodes that any query may be searched. Must be in
2189
- range 1-100, inclusive. The default value is 10 (means 10%!)(MISSING) if not set.
2356
+ range 1-100, inclusive. The default value is 10 (means 10%) if not set.
2190
2357
  """
2191
2358
  return pulumi.get(self, "leaf_nodes_to_search_percent")
2192
2359