google-cloud-ai_platform-v1 0.16.0 → 0.18.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/google/cloud/ai_platform/v1/dataset_service/client.rb +10 -7
- data/lib/google/cloud/ai_platform/v1/dataset_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/endpoint_service/client.rb +28 -21
- data/lib/google/cloud/ai_platform/v1/endpoint_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/client.rb +10 -9
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/featurestore_service/client.rb +53 -50
- data/lib/google/cloud/ai_platform/v1/featurestore_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/client.rb +18 -14
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/index_service/client.rb +8 -4
- data/lib/google/cloud/ai_platform/v1/index_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/job_service/client.rb +53 -36
- data/lib/google/cloud/ai_platform/v1/job_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/metadata_service/client.rb +33 -36
- data/lib/google/cloud/ai_platform/v1/metadata_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/migration_service/client.rb +3 -4
- data/lib/google/cloud/ai_platform/v1/migration_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/model_service/client.rb +44 -22
- data/lib/google/cloud/ai_platform/v1/model_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/pipeline_service/client.rb +28 -17
- data/lib/google/cloud/ai_platform/v1/pipeline_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/prediction_service/client.rb +28 -19
- data/lib/google/cloud/ai_platform/v1/prediction_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/client.rb +5 -3
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/client.rb +165 -71
- data/lib/google/cloud/ai_platform/v1/tensorboard_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/version.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/vizier_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1.rb +2 -2
- data/lib/google/cloud/aiplatform/v1/batch_prediction_job_pb.rb +8 -0
- data/lib/google/cloud/aiplatform/v1/custom_job_pb.rb +1 -0
- data/lib/google/cloud/aiplatform/v1/dataset_pb.rb +2 -0
- data/lib/google/cloud/aiplatform/v1/index_service_services_pb.rb +2 -1
- data/lib/google/cloud/aiplatform/v1/job_service_services_pb.rb +24 -14
- data/lib/google/cloud/aiplatform/v1/model_service_pb.rb +1 -0
- data/lib/google/cloud/aiplatform/v1/model_service_services_pb.rb +10 -5
- data/lib/google/cloud/aiplatform/v1/pipeline_service_services_pb.rb +16 -9
- data/lib/google/cloud/aiplatform/v1/prediction_service_services_pb.rb +10 -6
- data/lib/google/cloud/aiplatform/v1/study_pb.rb +1 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_pb.rb +17 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_services_pb.rb +8 -8
- data/proto_docs/google/cloud/aiplatform/v1/annotation.rb +9 -6
- data/proto_docs/google/cloud/aiplatform/v1/annotation_spec.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/batch_prediction_job.rb +183 -50
- data/proto_docs/google/cloud/aiplatform/v1/completion_stats.rb +8 -7
- data/proto_docs/google/cloud/aiplatform/v1/context.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/custom_job.rb +34 -17
- data/proto_docs/google/cloud/aiplatform/v1/data_item.rb +6 -5
- data/proto_docs/google/cloud/aiplatform/v1/data_labeling_job.rb +16 -16
- data/proto_docs/google/cloud/aiplatform/v1/dataset.rb +33 -17
- data/proto_docs/google/cloud/aiplatform/v1/dataset_service.rb +59 -32
- data/proto_docs/google/cloud/aiplatform/v1/encryption_spec.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/endpoint.rb +38 -23
- data/proto_docs/google/cloud/aiplatform/v1/endpoint_service.rb +57 -35
- data/proto_docs/google/cloud/aiplatform/v1/entity_type.rb +8 -6
- data/proto_docs/google/cloud/aiplatform/v1/explanation.rb +111 -75
- data/proto_docs/google/cloud/aiplatform/v1/explanation_metadata.rb +58 -38
- data/proto_docs/google/cloud/aiplatform/v1/feature.rb +9 -6
- data/proto_docs/google/cloud/aiplatform/v1/feature_monitoring_stats.rb +2 -1
- data/proto_docs/google/cloud/aiplatform/v1/featurestore.rb +12 -11
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_monitoring.rb +6 -2
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_online_service.rb +27 -19
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_service.rb +138 -99
- data/proto_docs/google/cloud/aiplatform/v1/hyperparameter_tuning_job.rb +9 -7
- data/proto_docs/google/cloud/aiplatform/v1/index.rb +18 -17
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint.rb +36 -24
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint_service.rb +49 -29
- data/proto_docs/google/cloud/aiplatform/v1/index_service.rb +35 -17
- data/proto_docs/google/cloud/aiplatform/v1/io.rb +2 -1
- data/proto_docs/google/cloud/aiplatform/v1/job_service.rb +94 -57
- data/proto_docs/google/cloud/aiplatform/v1/machine_resources.rb +53 -43
- data/proto_docs/google/cloud/aiplatform/v1/manual_batch_tuning_parameters.rb +5 -5
- data/proto_docs/google/cloud/aiplatform/v1/metadata_schema.rb +3 -3
- data/proto_docs/google/cloud/aiplatform/v1/metadata_service.rb +141 -91
- data/proto_docs/google/cloud/aiplatform/v1/migratable_resource.rb +3 -3
- data/proto_docs/google/cloud/aiplatform/v1/migration_service.rb +18 -15
- data/proto_docs/google/cloud/aiplatform/v1/model.rb +179 -132
- data/proto_docs/google/cloud/aiplatform/v1/model_deployment_monitoring_job.rb +21 -17
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation.rb +7 -6
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation_slice.rb +11 -7
- data/proto_docs/google/cloud/aiplatform/v1/model_monitoring.rb +4 -2
- data/proto_docs/google/cloud/aiplatform/v1/model_service.rb +110 -54
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_job.rb +56 -42
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_service.rb +42 -22
- data/proto_docs/google/cloud/aiplatform/v1/prediction_service.rb +42 -28
- data/proto_docs/google/cloud/aiplatform/v1/saved_query.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/specialist_pool_service.rb +19 -11
- data/proto_docs/google/cloud/aiplatform/v1/study.rb +42 -25
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_data.rb +8 -7
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_experiment.rb +2 -1
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_service.rb +203 -109
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_time_series.rb +8 -6
- data/proto_docs/google/cloud/aiplatform/v1/training_pipeline.rb +84 -59
- data/proto_docs/google/cloud/aiplatform/v1/unmanaged_container_model.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/vizier_service.rb +40 -20
- data/proto_docs/google/rpc/status.rb +4 -2
- metadata +9 -9
@@ -21,13 +21,13 @@ module Google
|
|
21
21
|
module Cloud
|
22
22
|
module AIPlatform
|
23
23
|
module V1
|
24
|
-
# Request message for
|
24
|
+
# Request message for
|
25
|
+
# {::Google::Cloud::AIPlatform::V1::MigrationService::Client#search_migratable_resources MigrationService.SearchMigratableResources}.
|
25
26
|
# @!attribute [rw] parent
|
26
27
|
# @return [::String]
|
27
|
-
# Required. The location that the migratable resources should be searched
|
28
|
-
# It's the Vertex AI location that the resources can be migrated to,
|
29
|
-
# the resources' original location.
|
30
|
-
# Format:
|
28
|
+
# Required. The location that the migratable resources should be searched
|
29
|
+
# from. It's the Vertex AI location that the resources can be migrated to,
|
30
|
+
# not the resources' original location. Format:
|
31
31
|
# `projects/{project}/locations/{location}`
|
32
32
|
# @!attribute [rw] page_size
|
33
33
|
# @return [::Integer]
|
@@ -55,7 +55,8 @@ module Google
|
|
55
55
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
56
56
|
end
|
57
57
|
|
58
|
-
# Response message for
|
58
|
+
# Response message for
|
59
|
+
# {::Google::Cloud::AIPlatform::V1::MigrationService::Client#search_migratable_resources MigrationService.SearchMigratableResources}.
|
59
60
|
# @!attribute [rw] migratable_resources
|
60
61
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::MigratableResource>]
|
61
62
|
# All migratable resources that can be migrated to the
|
@@ -70,7 +71,8 @@ module Google
|
|
70
71
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
71
72
|
end
|
72
73
|
|
73
|
-
# Request message for
|
74
|
+
# Request message for
|
75
|
+
# {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
|
74
76
|
# @!attribute [rw] parent
|
75
77
|
# @return [::String]
|
76
78
|
# Required. The location of the migrated resource will live in.
|
@@ -109,9 +111,8 @@ module Google
|
|
109
111
|
# Config for migrating version in ml.googleapis.com to Vertex AI's Model.
|
110
112
|
# @!attribute [rw] endpoint
|
111
113
|
# @return [::String]
|
112
|
-
# Required. The ml.googleapis.com endpoint that this model version should
|
113
|
-
# from.
|
114
|
-
# Example values:
|
114
|
+
# Required. The ml.googleapis.com endpoint that this model version should
|
115
|
+
# be migrated from. Example values:
|
115
116
|
#
|
116
117
|
# * ml.googleapis.com
|
117
118
|
#
|
@@ -177,9 +178,9 @@ module Google
|
|
177
178
|
# System will pick a display name if unspecified.
|
178
179
|
# @!attribute [rw] migrate_data_labeling_annotated_dataset_configs
|
179
180
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::MigrateResourceRequest::MigrateDataLabelingDatasetConfig::MigrateDataLabelingAnnotatedDatasetConfig>]
|
180
|
-
# Optional. Configs for migrating AnnotatedDataset in
|
181
|
-
# Vertex AI's SavedQuery. The specified
|
182
|
-
# to the datalabeling Dataset.
|
181
|
+
# Optional. Configs for migrating AnnotatedDataset in
|
182
|
+
# datalabeling.googleapis.com to Vertex AI's SavedQuery. The specified
|
183
|
+
# AnnotatedDatasets have to belong to the datalabeling Dataset.
|
183
184
|
class MigrateDataLabelingDatasetConfig
|
184
185
|
include ::Google::Protobuf::MessageExts
|
185
186
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -198,7 +199,8 @@ module Google
|
|
198
199
|
end
|
199
200
|
end
|
200
201
|
|
201
|
-
# Response message for
|
202
|
+
# Response message for
|
203
|
+
# {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
|
202
204
|
# @!attribute [rw] migrate_resource_responses
|
203
205
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::MigrateResourceResponse>]
|
204
206
|
# Successfully migrated resources.
|
@@ -223,7 +225,8 @@ module Google
|
|
223
225
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
224
226
|
end
|
225
227
|
|
226
|
-
# Runtime operation information for
|
228
|
+
# Runtime operation information for
|
229
|
+
# {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
|
227
230
|
# @!attribute [rw] generic_metadata
|
228
231
|
# @return [::Google::Cloud::AIPlatform::V1::GenericOperationMetadata]
|
229
232
|
# The common part of the operation metadata.
|
@@ -63,13 +63,15 @@ module Google
|
|
63
63
|
# @return [::Google::Cloud::AIPlatform::V1::PredictSchemata]
|
64
64
|
# The schemata that describe formats of the Model's predictions and
|
65
65
|
# explanations as given and returned via
|
66
|
-
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
|
66
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
|
67
|
+
# and
|
68
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
|
67
69
|
# @!attribute [rw] metadata_schema_uri
|
68
70
|
# @return [::String]
|
69
|
-
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
70
|
-
# information about the Model, that is specific to it. Unset if
|
71
|
-
# does not have any additional information.
|
72
|
-
#
|
71
|
+
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
72
|
+
# additional information about the Model, that is specific to it. Unset if
|
73
|
+
# the Model does not have any additional information. The schema is defined
|
74
|
+
# as an OpenAPI 3.0.2 [Schema
|
73
75
|
# Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
|
74
76
|
# AutoML Models always have this field populated by Vertex AI, if no
|
75
77
|
# additional metadata is needed, this field is set to an empty string.
|
@@ -78,48 +80,55 @@ module Google
|
|
78
80
|
# point to a location where the user only has a read access.
|
79
81
|
# @!attribute [rw] metadata
|
80
82
|
# @return [::Google::Protobuf::Value]
|
81
|
-
# Immutable. An additional information about the Model; the schema of the
|
82
|
-
# be found in
|
83
|
+
# Immutable. An additional information about the Model; the schema of the
|
84
|
+
# metadata can be found in
|
85
|
+
# {::Google::Cloud::AIPlatform::V1::Model#metadata_schema_uri metadata_schema}.
|
83
86
|
# Unset if the Model does not have any additional information.
|
84
87
|
# @!attribute [r] supported_export_formats
|
85
88
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::Model::ExportFormat>]
|
86
|
-
# Output only. The formats in which this Model may be exported. If empty,
|
87
|
-
# not available for export.
|
89
|
+
# Output only. The formats in which this Model may be exported. If empty,
|
90
|
+
# this Model is not available for export.
|
88
91
|
# @!attribute [r] training_pipeline
|
89
92
|
# @return [::String]
|
90
|
-
# Output only. The resource name of the TrainingPipeline that uploaded this
|
91
|
-
# any.
|
93
|
+
# Output only. The resource name of the TrainingPipeline that uploaded this
|
94
|
+
# Model, if any.
|
92
95
|
# @!attribute [rw] container_spec
|
93
96
|
# @return [::Google::Cloud::AIPlatform::V1::ModelContainerSpec]
|
94
|
-
# Input only. The specification of the container that is to be used when
|
95
|
-
# this Model. The specification is ingested upon
|
96
|
-
# {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel},
|
97
|
-
# and stored internally by Vertex AI.
|
97
|
+
# Input only. The specification of the container that is to be used when
|
98
|
+
# deploying this Model. The specification is ingested upon
|
99
|
+
# {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel},
|
100
|
+
# and all binaries it contains are copied and stored internally by Vertex AI.
|
98
101
|
# Not present for AutoML Models.
|
99
102
|
# @!attribute [rw] artifact_uri
|
100
103
|
# @return [::String]
|
101
|
-
# Immutable. The path to the directory containing the Model artifact and any
|
102
|
-
# supporting files.
|
103
|
-
# Not present for AutoML Models.
|
104
|
+
# Immutable. The path to the directory containing the Model artifact and any
|
105
|
+
# of its supporting files. Not present for AutoML Models.
|
104
106
|
# @!attribute [r] supported_deployment_resources_types
|
105
107
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::Model::DeploymentResourcesType>]
|
106
|
-
# Output only. When this Model is deployed, its prediction resources are
|
107
|
-
# `prediction_resources` field of the
|
108
|
-
#
|
109
|
-
#
|
108
|
+
# Output only. When this Model is deployed, its prediction resources are
|
109
|
+
# described by the `prediction_resources` field of the
|
110
|
+
# {::Google::Cloud::AIPlatform::V1::Endpoint#deployed_models Endpoint.deployed_models}
|
111
|
+
# object. Because not all Models support all resource configuration types,
|
112
|
+
# the configuration types this Model supports are listed here. If no
|
110
113
|
# configuration types are listed, the Model cannot be deployed to an
|
111
114
|
# {::Google::Cloud::AIPlatform::V1::Endpoint Endpoint} and does not support
|
112
|
-
# online predictions
|
113
|
-
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#
|
114
|
-
#
|
115
|
-
# {::Google::Cloud::AIPlatform::V1::
|
115
|
+
# online predictions
|
116
|
+
# ({::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
|
117
|
+
# or
|
118
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}).
|
119
|
+
# Such a Model can serve predictions by using a
|
120
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}, if it
|
121
|
+
# has at least one entry each in
|
122
|
+
# {::Google::Cloud::AIPlatform::V1::Model#supported_input_storage_formats supported_input_storage_formats}
|
123
|
+
# and
|
116
124
|
# {::Google::Cloud::AIPlatform::V1::Model#supported_output_storage_formats supported_output_storage_formats}.
|
117
125
|
# @!attribute [r] supported_input_storage_formats
|
118
126
|
# @return [::Array<::String>]
|
119
127
|
# Output only. The formats this Model supports in
|
120
|
-
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#input_config BatchPredictionJob.input_config}.
|
121
|
-
#
|
122
|
-
#
|
128
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#input_config BatchPredictionJob.input_config}.
|
129
|
+
# If
|
130
|
+
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri PredictSchemata.instance_schema_uri}
|
131
|
+
# exists, the instances should be given as per that schema.
|
123
132
|
#
|
124
133
|
# The possible formats are:
|
125
134
|
#
|
@@ -130,11 +139,13 @@ module Google
|
|
130
139
|
# * `csv`
|
131
140
|
# The CSV format, where each instance is a single comma-separated line.
|
132
141
|
# The first line in the file is the header, containing comma-separated field
|
133
|
-
# names. Uses
|
142
|
+
# names. Uses
|
143
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig#gcs_source GcsSource}.
|
134
144
|
#
|
135
145
|
# * `tf-record`
|
136
146
|
# The TFRecord format, where each instance is a single record in tfrecord
|
137
|
-
# syntax. Uses
|
147
|
+
# syntax. Uses
|
148
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig#gcs_source GcsSource}.
|
138
149
|
#
|
139
150
|
# * `tf-record-gzip`
|
140
151
|
# Similar to `tf-record`, but the file is gzipped. Uses
|
@@ -147,23 +158,30 @@ module Google
|
|
147
158
|
# * `file-list`
|
148
159
|
# Each line of the file is the location of an instance to process, uses
|
149
160
|
# `gcs_source` field of the
|
150
|
-
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig InputConfig}
|
161
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig InputConfig}
|
162
|
+
# object.
|
151
163
|
#
|
152
164
|
#
|
153
165
|
# If this Model doesn't support any of these formats it means it cannot be
|
154
|
-
# used with a
|
155
|
-
# {::Google::Cloud::AIPlatform::V1::
|
156
|
-
#
|
166
|
+
# used with a
|
167
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
|
168
|
+
# However, if it has
|
169
|
+
# {::Google::Cloud::AIPlatform::V1::Model#supported_deployment_resources_types supported_deployment_resources_types},
|
170
|
+
# it could serve online predictions by using
|
171
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
|
172
|
+
# or
|
157
173
|
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
|
158
174
|
# @!attribute [r] supported_output_storage_formats
|
159
175
|
# @return [::Array<::String>]
|
160
176
|
# Output only. The formats this Model supports in
|
161
|
-
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#output_config BatchPredictionJob.output_config}.
|
162
|
-
#
|
163
|
-
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#
|
164
|
-
#
|
165
|
-
#
|
166
|
-
#
|
177
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#output_config BatchPredictionJob.output_config}.
|
178
|
+
# If both
|
179
|
+
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri PredictSchemata.instance_schema_uri}
|
180
|
+
# and
|
181
|
+
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#prediction_schema_uri PredictSchemata.prediction_schema_uri}
|
182
|
+
# exist, the predictions are returned together with their instances. In other
|
183
|
+
# words, the prediction has the original instance data first, followed by the
|
184
|
+
# actual prediction content (as per the schema).
|
167
185
|
#
|
168
186
|
# The possible formats are:
|
169
187
|
#
|
@@ -184,9 +202,13 @@ module Google
|
|
184
202
|
#
|
185
203
|
#
|
186
204
|
# If this Model doesn't support any of these formats it means it cannot be
|
187
|
-
# used with a
|
188
|
-
# {::Google::Cloud::AIPlatform::V1::
|
189
|
-
#
|
205
|
+
# used with a
|
206
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
|
207
|
+
# However, if it has
|
208
|
+
# {::Google::Cloud::AIPlatform::V1::Model#supported_deployment_resources_types supported_deployment_resources_types},
|
209
|
+
# it could serve online predictions by using
|
210
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
|
211
|
+
# or
|
190
212
|
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
|
191
213
|
# @!attribute [r] create_time
|
192
214
|
# @return [::Google::Protobuf::Timestamp]
|
@@ -196,32 +218,36 @@ module Google
|
|
196
218
|
# Output only. Timestamp when this Model was most recently updated.
|
197
219
|
# @!attribute [r] deployed_models
|
198
220
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::DeployedModelRef>]
|
199
|
-
# Output only. The pointers to DeployedModels created from this Model. Note
|
200
|
-
# Model could have been deployed to Endpoints in different Locations.
|
221
|
+
# Output only. The pointers to DeployedModels created from this Model. Note
|
222
|
+
# that Model could have been deployed to Endpoints in different Locations.
|
201
223
|
# @!attribute [rw] explanation_spec
|
202
224
|
# @return [::Google::Cloud::AIPlatform::V1::ExplanationSpec]
|
203
225
|
# The default explanation specification for this Model.
|
204
226
|
#
|
205
227
|
# The Model can be used for [requesting
|
206
228
|
# explanation][PredictionService.Explain] after being
|
207
|
-
# {::Google::Cloud::AIPlatform::V1::EndpointService::Client#deploy_model deployed} if it is
|
208
|
-
# The Model can be used for [batch
|
229
|
+
# {::Google::Cloud::AIPlatform::V1::EndpointService::Client#deploy_model deployed} if it is
|
230
|
+
# populated. The Model can be used for [batch
|
209
231
|
# explanation][BatchPredictionJob.generate_explanation] if it is populated.
|
210
232
|
#
|
211
233
|
# All fields of the explanation_spec can be overridden by
|
212
|
-
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
213
|
-
#
|
214
|
-
# {::Google::Cloud::AIPlatform::V1::
|
215
|
-
#
|
234
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
235
|
+
# of
|
236
|
+
# {::Google::Cloud::AIPlatform::V1::DeployModelRequest#deployed_model DeployModelRequest.deployed_model},
|
237
|
+
# or
|
238
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#explanation_spec explanation_spec}
|
239
|
+
# of {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
|
216
240
|
#
|
217
241
|
# If the default explanation specification is not set for this Model, this
|
218
242
|
# Model can still be used for [requesting
|
219
243
|
# explanation][PredictionService.Explain] by setting
|
220
|
-
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
221
|
-
#
|
222
|
-
#
|
223
|
-
#
|
224
|
-
#
|
244
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
245
|
+
# of
|
246
|
+
# {::Google::Cloud::AIPlatform::V1::DeployModelRequest#deployed_model DeployModelRequest.deployed_model}
|
247
|
+
# and for [batch explanation][BatchPredictionJob.generate_explanation] by
|
248
|
+
# setting
|
249
|
+
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#explanation_spec explanation_spec}
|
250
|
+
# of {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
|
225
251
|
# @!attribute [rw] etag
|
226
252
|
# @return [::String]
|
227
253
|
# Used to perform consistent read-modify-write updates. If not set, a blind
|
@@ -241,12 +267,13 @@ module Google
|
|
241
267
|
# Model and all sub-resources of this Model will be secured by this key.
|
242
268
|
# @!attribute [r] model_source_info
|
243
269
|
# @return [::Google::Cloud::AIPlatform::V1::ModelSourceInfo]
|
244
|
-
# Output only. Source of a model. It can either be automl training pipeline,
|
245
|
-
# training pipeline, BigQuery ML, or existing Vertex AI Model.
|
270
|
+
# Output only. Source of a model. It can either be automl training pipeline,
|
271
|
+
# custom training pipeline, BigQuery ML, or existing Vertex AI Model.
|
246
272
|
# @!attribute [r] metadata_artifact
|
247
273
|
# @return [::String]
|
248
|
-
# Output only. The resource name of the Artifact that was created in
|
249
|
-
# creating the Model. The Artifact resource name pattern
|
274
|
+
# Output only. The resource name of the Artifact that was created in
|
275
|
+
# MetadataStore when creating the Model. The Artifact resource name pattern
|
276
|
+
# is
|
250
277
|
# `projects/{project}/locations/{location}/metadataStores/{metadata_store}/artifacts/{artifact}`.
|
251
278
|
class Model
|
252
279
|
include ::Google::Protobuf::MessageExts
|
@@ -291,12 +318,15 @@ module Google
|
|
291
318
|
|
292
319
|
# Model artifact and any of its supported files. Will be exported to the
|
293
320
|
# location specified by the `artifactDestination` field of the
|
294
|
-
# {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config}
|
321
|
+
# {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config}
|
322
|
+
# object.
|
295
323
|
ARTIFACT = 1
|
296
324
|
|
297
325
|
# The container image that is to be used when deploying this Model. Will
|
298
326
|
# be exported to the location specified by the `imageDestination` field
|
299
|
-
# of the
|
327
|
+
# of the
|
328
|
+
# {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config}
|
329
|
+
# object.
|
300
330
|
IMAGE = 2
|
301
331
|
end
|
302
332
|
end
|
@@ -315,28 +345,33 @@ module Google
|
|
315
345
|
# Should not be used.
|
316
346
|
DEPLOYMENT_RESOURCES_TYPE_UNSPECIFIED = 0
|
317
347
|
|
318
|
-
# Resources that are dedicated to the
|
319
|
-
#
|
348
|
+
# Resources that are dedicated to the
|
349
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModel}, and that need
|
350
|
+
# a higher degree of manual configuration.
|
320
351
|
DEDICATED_RESOURCES = 1
|
321
352
|
|
322
353
|
# Resources that to large degree are decided by Vertex AI, and require
|
323
354
|
# only a modest additional configuration.
|
324
355
|
AUTOMATIC_RESOURCES = 2
|
325
356
|
|
326
|
-
# Resources that can be shared by multiple
|
327
|
-
#
|
357
|
+
# Resources that can be shared by multiple
|
358
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModels}. A
|
359
|
+
# pre-configured [DeploymentResourcePool][] is required.
|
328
360
|
SHARED_RESOURCES = 3
|
329
361
|
end
|
330
362
|
end
|
331
363
|
|
332
364
|
# Contains the schemata used in Model's predictions and explanations via
|
333
|
-
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict},
|
334
|
-
# {::Google::Cloud::AIPlatform::V1::
|
365
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict},
|
366
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}
|
367
|
+
# and {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
|
335
368
|
# @!attribute [rw] instance_schema_uri
|
336
369
|
# @return [::String]
|
337
|
-
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
338
|
-
# of a single instance, which are used in
|
339
|
-
# {::Google::Cloud::AIPlatform::V1::
|
370
|
+
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
371
|
+
# the format of a single instance, which are used in
|
372
|
+
# {::Google::Cloud::AIPlatform::V1::PredictRequest#instances PredictRequest.instances},
|
373
|
+
# {::Google::Cloud::AIPlatform::V1::ExplainRequest#instances ExplainRequest.instances}
|
374
|
+
# and
|
340
375
|
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#input_config BatchPredictionJob.input_config}.
|
341
376
|
# The schema is defined as an OpenAPI 3.0.2 [Schema
|
342
377
|
# Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
|
@@ -346,9 +381,11 @@ module Google
|
|
346
381
|
# point to a location where the user only has a read access.
|
347
382
|
# @!attribute [rw] parameters_schema_uri
|
348
383
|
# @return [::String]
|
349
|
-
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
350
|
-
# parameters of prediction and explanation via
|
351
|
-
# {::Google::Cloud::AIPlatform::V1::PredictRequest#parameters PredictRequest.parameters},
|
384
|
+
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
385
|
+
# the parameters of prediction and explanation via
|
386
|
+
# {::Google::Cloud::AIPlatform::V1::PredictRequest#parameters PredictRequest.parameters},
|
387
|
+
# {::Google::Cloud::AIPlatform::V1::ExplainRequest#parameters ExplainRequest.parameters}
|
388
|
+
# and
|
352
389
|
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#model_parameters BatchPredictionJob.model_parameters}.
|
353
390
|
# The schema is defined as an OpenAPI 3.0.2 [Schema
|
354
391
|
# Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
|
@@ -359,9 +396,12 @@ module Google
|
|
359
396
|
# point to a location where the user only has a read access.
|
360
397
|
# @!attribute [rw] prediction_schema_uri
|
361
398
|
# @return [::String]
|
362
|
-
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
363
|
-
# of a single prediction produced by this Model, which are
|
364
|
-
#
|
399
|
+
# Immutable. Points to a YAML file stored on Google Cloud Storage describing
|
400
|
+
# the format of a single prediction produced by this Model, which are
|
401
|
+
# returned via
|
402
|
+
# {::Google::Cloud::AIPlatform::V1::PredictResponse#predictions PredictResponse.predictions},
|
403
|
+
# {::Google::Cloud::AIPlatform::V1::ExplainResponse#explanations ExplainResponse.explanations},
|
404
|
+
# and
|
365
405
|
# {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#output_config BatchPredictionJob.output_config}.
|
366
406
|
# The schema is defined as an OpenAPI 3.0.2 [Schema
|
367
407
|
# Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
|
@@ -379,14 +419,16 @@ module Google
|
|
379
419
|
# specification](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
|
380
420
|
# @!attribute [rw] image_uri
|
381
421
|
# @return [::String]
|
382
|
-
# Required. Immutable. URI of the Docker image to be used as the custom
|
383
|
-
# predictions. This URI must identify an image in
|
384
|
-
# Container Registry. Learn more about the [container
|
422
|
+
# Required. Immutable. URI of the Docker image to be used as the custom
|
423
|
+
# container for serving predictions. This URI must identify an image in
|
424
|
+
# Artifact Registry or Container Registry. Learn more about the [container
|
425
|
+
# publishing
|
385
426
|
# requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing),
|
386
427
|
# including permissions requirements for the Vertex AI Service Agent.
|
387
428
|
#
|
388
|
-
# The container image is ingested upon
|
389
|
-
#
|
429
|
+
# The container image is ingested upon
|
430
|
+
# {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel},
|
431
|
+
# stored internally, and this original path is afterwards not used.
|
390
432
|
#
|
391
433
|
# To learn about the requirements for the Docker image itself, see
|
392
434
|
# [Custom container
|
@@ -397,14 +439,15 @@ module Google
|
|
397
439
|
# in this field.
|
398
440
|
# @!attribute [rw] command
|
399
441
|
# @return [::Array<::String>]
|
400
|
-
# Immutable. Specifies the command that runs when the container starts. This
|
401
|
-
# the container's
|
442
|
+
# Immutable. Specifies the command that runs when the container starts. This
|
443
|
+
# overrides the container's
|
402
444
|
# [ENTRYPOINT](https://docs.docker.com/engine/reference/builder/#entrypoint).
|
403
445
|
# Specify this field as an array of executable and arguments, similar to a
|
404
446
|
# Docker `ENTRYPOINT`'s "exec" form, not its "shell" form.
|
405
447
|
#
|
406
448
|
# If you do not specify this field, then the container's `ENTRYPOINT` runs,
|
407
|
-
# in conjunction with the
|
449
|
+
# in conjunction with the
|
450
|
+
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#args args} field or the
|
408
451
|
# container's [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd),
|
409
452
|
# if either exists. If this field is not specified and the container does not
|
410
453
|
# have an `ENTRYPOINT`, then refer to the Docker documentation about [how
|
@@ -420,32 +463,30 @@ module Google
|
|
420
463
|
#
|
421
464
|
# In this field, you can reference [environment variables set by Vertex
|
422
465
|
# AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables)
|
423
|
-
# and environment variables set in the
|
424
|
-
#
|
425
|
-
#
|
426
|
-
# following
|
427
|
-
# <code>$(<var>VARIABLE_NAME</var>)</code>
|
428
|
-
#
|
429
|
-
#
|
430
|
-
#
|
431
|
-
#
|
432
|
-
#
|
433
|
-
# This field corresponds to the `command` field of the Kubernetes Containers
|
434
|
-
# [v1 core
|
466
|
+
# and environment variables set in the
|
467
|
+
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#env env} field. You cannot
|
468
|
+
# reference environment variables set in the Docker image. In order for
|
469
|
+
# environment variables to be expanded, reference them by using the following
|
470
|
+
# syntax: <code>$(<var>VARIABLE_NAME</var>)</code> Note that this differs
|
471
|
+
# from Bash variable expansion, which does not use parentheses. If a variable
|
472
|
+
# cannot be resolved, the reference in the input string is used unchanged. To
|
473
|
+
# avoid variable expansion, you can escape this syntax with `$$`; for
|
474
|
+
# example: <code>$$(<var>VARIABLE_NAME</var>)</code> This field corresponds
|
475
|
+
# to the `command` field of the Kubernetes Containers [v1 core
|
435
476
|
# API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
|
436
477
|
# @!attribute [rw] args
|
437
478
|
# @return [::Array<::String>]
|
438
|
-
# Immutable. Specifies arguments for the command that runs when the container
|
439
|
-
# This overrides the container's
|
479
|
+
# Immutable. Specifies arguments for the command that runs when the container
|
480
|
+
# starts. This overrides the container's
|
440
481
|
# [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd). Specify
|
441
482
|
# this field as an array of executable and arguments, similar to a Docker
|
442
483
|
# `CMD`'s "default parameters" form.
|
443
484
|
#
|
444
485
|
# If you don't specify this field but do specify the
|
445
|
-
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} field,
|
446
|
-
# `command` field runs without any additional
|
447
|
-
# [Kubernetes documentation about how the
|
448
|
-
# `
|
486
|
+
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} field,
|
487
|
+
# then the command from the `command` field runs without any additional
|
488
|
+
# arguments. See the [Kubernetes documentation about how the `command` and
|
489
|
+
# `args` fields interact with a container's `ENTRYPOINT` and
|
449
490
|
# `CMD`](https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#notes).
|
450
491
|
#
|
451
492
|
# If you don't specify this field and don't specify the `command` field,
|
@@ -458,29 +499,29 @@ module Google
|
|
458
499
|
# In this field, you can reference [environment variables
|
459
500
|
# set by Vertex
|
460
501
|
# AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables)
|
461
|
-
# and environment variables set in the
|
462
|
-
#
|
463
|
-
#
|
464
|
-
# following
|
465
|
-
# <code>$(<var>VARIABLE_NAME</var>)</code>
|
466
|
-
#
|
467
|
-
#
|
468
|
-
#
|
469
|
-
#
|
470
|
-
#
|
471
|
-
# This field corresponds to the `args` field of the Kubernetes Containers
|
472
|
-
# [v1 core
|
502
|
+
# and environment variables set in the
|
503
|
+
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#env env} field. You cannot
|
504
|
+
# reference environment variables set in the Docker image. In order for
|
505
|
+
# environment variables to be expanded, reference them by using the following
|
506
|
+
# syntax: <code>$(<var>VARIABLE_NAME</var>)</code> Note that this differs
|
507
|
+
# from Bash variable expansion, which does not use parentheses. If a variable
|
508
|
+
# cannot be resolved, the reference in the input string is used unchanged. To
|
509
|
+
# avoid variable expansion, you can escape this syntax with `$$`; for
|
510
|
+
# example: <code>$$(<var>VARIABLE_NAME</var>)</code> This field corresponds
|
511
|
+
# to the `args` field of the Kubernetes Containers [v1 core
|
473
512
|
# API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
|
474
513
|
# @!attribute [rw] env
|
475
514
|
# @return [::Array<::Google::Cloud::AIPlatform::V1::EnvVar>]
|
476
|
-
# Immutable. List of environment variables to set in the container. After the
|
477
|
-
# starts running, code running in the container can read these
|
478
|
-
# variables.
|
515
|
+
# Immutable. List of environment variables to set in the container. After the
|
516
|
+
# container starts running, code running in the container can read these
|
517
|
+
# environment variables.
|
479
518
|
#
|
480
|
-
# Additionally, the
|
481
|
-
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#
|
482
|
-
#
|
483
|
-
#
|
519
|
+
# Additionally, the
|
520
|
+
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} and
|
521
|
+
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#args args} fields can
|
522
|
+
# reference these variables. Later entries in this list can also reference
|
523
|
+
# earlier entries. For example, the following example sets the variable
|
524
|
+
# `VAR_2` to have the value `foo bar`:
|
484
525
|
#
|
485
526
|
# ```json
|
486
527
|
# [
|
@@ -526,11 +567,11 @@ module Google
|
|
526
567
|
# API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
|
527
568
|
# @!attribute [rw] predict_route
|
528
569
|
# @return [::String]
|
529
|
-
# Immutable. HTTP path on the container to send prediction requests to.
|
530
|
-
# forwards requests sent using
|
531
|
-
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict projects.locations.endpoints.predict}
|
532
|
-
# path on the container's IP address and port. Vertex AI then returns
|
533
|
-
# container's response in the API response.
|
570
|
+
# Immutable. HTTP path on the container to send prediction requests to.
|
571
|
+
# Vertex AI forwards requests sent using
|
572
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict projects.locations.endpoints.predict}
|
573
|
+
# to this path on the container's IP address and port. Vertex AI then returns
|
574
|
+
# the container's response in the API response.
|
534
575
|
#
|
535
576
|
# For example, if you set this field to `/foo`, then when Vertex AI
|
536
577
|
# receives a prediction request, it forwards the request body in a POST
|
@@ -539,7 +580,8 @@ module Google
|
|
539
580
|
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#ports ports} field.
|
540
581
|
#
|
541
582
|
# If you don't specify this field, it defaults to the following value when
|
542
|
-
# you
|
583
|
+
# you [deploy this Model to an
|
584
|
+
# Endpoint][google.cloud.aiplatform.v1.EndpointService.DeployModel]:
|
543
585
|
# <code>/v1/endpoints/<var>ENDPOINT</var>/deployedModels/<var>DEPLOYED_MODEL</var>:predict</code>
|
544
586
|
# The placeholders in this value are replaced as follows:
|
545
587
|
#
|
@@ -549,7 +591,9 @@ module Google
|
|
549
591
|
# as the [`AIP_ENDPOINT_ID` environment
|
550
592
|
# variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
|
551
593
|
#
|
552
|
-
# * <var>DEPLOYED_MODEL</var>:
|
594
|
+
# * <var>DEPLOYED_MODEL</var>:
|
595
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel#id DeployedModel.id} of the
|
596
|
+
# `DeployedModel`.
|
553
597
|
# (Vertex AI makes this value available to your container code
|
554
598
|
# as the [`AIP_DEPLOYED_MODEL_ID` environment
|
555
599
|
# variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
|
@@ -567,7 +611,8 @@ module Google
|
|
567
611
|
# {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#ports ports} field.
|
568
612
|
#
|
569
613
|
# If you don't specify this field, it defaults to the following value when
|
570
|
-
# you
|
614
|
+
# you [deploy this Model to an
|
615
|
+
# Endpoint][google.cloud.aiplatform.v1.EndpointService.DeployModel]:
|
571
616
|
# <code>/v1/endpoints/<var>ENDPOINT</var>/deployedModels/<var>DEPLOYED_MODEL</var>:predict</code>
|
572
617
|
# The placeholders in this value are replaced as follows:
|
573
618
|
#
|
@@ -577,7 +622,9 @@ module Google
|
|
577
622
|
# as the [`AIP_ENDPOINT_ID` environment
|
578
623
|
# variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
|
579
624
|
#
|
580
|
-
# * <var>DEPLOYED_MODEL</var>:
|
625
|
+
# * <var>DEPLOYED_MODEL</var>:
|
626
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel#id DeployedModel.id} of the
|
627
|
+
# `DeployedModel`.
|
581
628
|
# (Vertex AI makes this value available to your container code as the
|
582
629
|
# [`AIP_DEPLOYED_MODEL_ID` environment
|
583
630
|
# variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
|