google-cloud-ai_platform-v1 0.15.0 → 0.17.0

Sign up to get free protection for your applications and to get access to all the features.
Files changed (109) hide show
  1. checksums.yaml +4 -4
  2. data/README.md +1 -1
  3. data/lib/google/cloud/ai_platform/v1/dataset_service/client.rb +164 -4
  4. data/lib/google/cloud/ai_platform/v1/dataset_service/paths.rb +21 -0
  5. data/lib/google/cloud/ai_platform/v1/dataset_service.rb +1 -1
  6. data/lib/google/cloud/ai_platform/v1/endpoint_service/client.rb +28 -21
  7. data/lib/google/cloud/ai_platform/v1/endpoint_service.rb +1 -1
  8. data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/client.rb +109 -5
  9. data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service.rb +1 -1
  10. data/lib/google/cloud/ai_platform/v1/featurestore_service/client.rb +53 -50
  11. data/lib/google/cloud/ai_platform/v1/featurestore_service.rb +1 -1
  12. data/lib/google/cloud/ai_platform/v1/index_endpoint_service/client.rb +18 -14
  13. data/lib/google/cloud/ai_platform/v1/index_endpoint_service.rb +1 -1
  14. data/lib/google/cloud/ai_platform/v1/index_service/client.rb +8 -4
  15. data/lib/google/cloud/ai_platform/v1/index_service.rb +1 -1
  16. data/lib/google/cloud/ai_platform/v1/job_service/client.rb +53 -36
  17. data/lib/google/cloud/ai_platform/v1/job_service.rb +1 -1
  18. data/lib/google/cloud/ai_platform/v1/metadata_service/client.rb +33 -36
  19. data/lib/google/cloud/ai_platform/v1/metadata_service.rb +1 -1
  20. data/lib/google/cloud/ai_platform/v1/migration_service/client.rb +3 -4
  21. data/lib/google/cloud/ai_platform/v1/migration_service.rb +1 -1
  22. data/lib/google/cloud/ai_platform/v1/model_service/client.rb +53 -23
  23. data/lib/google/cloud/ai_platform/v1/model_service.rb +1 -1
  24. data/lib/google/cloud/ai_platform/v1/pipeline_service/client.rb +28 -17
  25. data/lib/google/cloud/ai_platform/v1/pipeline_service.rb +1 -1
  26. data/lib/google/cloud/ai_platform/v1/prediction_service/client.rb +28 -19
  27. data/lib/google/cloud/ai_platform/v1/prediction_service.rb +1 -1
  28. data/lib/google/cloud/ai_platform/v1/specialist_pool_service/client.rb +5 -3
  29. data/lib/google/cloud/ai_platform/v1/specialist_pool_service.rb +1 -1
  30. data/lib/google/cloud/ai_platform/v1/tensorboard_service/client.rb +165 -71
  31. data/lib/google/cloud/ai_platform/v1/tensorboard_service.rb +1 -1
  32. data/lib/google/cloud/ai_platform/v1/version.rb +1 -1
  33. data/lib/google/cloud/ai_platform/v1/vizier_service.rb +1 -1
  34. data/lib/google/cloud/ai_platform/v1.rb +2 -2
  35. data/lib/google/cloud/aiplatform/v1/batch_prediction_job_pb.rb +8 -0
  36. data/lib/google/cloud/aiplatform/v1/dataset_pb.rb +3 -0
  37. data/lib/google/cloud/aiplatform/v1/dataset_service_pb.rb +34 -0
  38. data/lib/google/cloud/aiplatform/v1/dataset_service_services_pb.rb +2 -0
  39. data/lib/google/cloud/aiplatform/v1/featurestore_online_service_pb.rb +13 -0
  40. data/lib/google/cloud/aiplatform/v1/featurestore_online_service_services_pb.rb +6 -0
  41. data/lib/google/cloud/aiplatform/v1/featurestore_service_pb.rb +1 -0
  42. data/lib/google/cloud/aiplatform/v1/index_service_services_pb.rb +2 -1
  43. data/lib/google/cloud/aiplatform/v1/job_service_services_pb.rb +24 -14
  44. data/lib/google/cloud/aiplatform/v1/model_service_pb.rb +2 -0
  45. data/lib/google/cloud/aiplatform/v1/model_service_services_pb.rb +10 -5
  46. data/lib/google/cloud/aiplatform/v1/pipeline_service_services_pb.rb +16 -9
  47. data/lib/google/cloud/aiplatform/v1/prediction_service_services_pb.rb +10 -6
  48. data/lib/google/cloud/aiplatform/v1/study_pb.rb +1 -0
  49. data/lib/google/cloud/aiplatform/v1/tensorboard_service_pb.rb +17 -0
  50. data/lib/google/cloud/aiplatform/v1/tensorboard_service_services_pb.rb +8 -8
  51. data/proto_docs/google/cloud/aiplatform/v1/annotation.rb +9 -6
  52. data/proto_docs/google/cloud/aiplatform/v1/annotation_spec.rb +3 -3
  53. data/proto_docs/google/cloud/aiplatform/v1/batch_prediction_job.rb +185 -52
  54. data/proto_docs/google/cloud/aiplatform/v1/completion_stats.rb +8 -7
  55. data/proto_docs/google/cloud/aiplatform/v1/context.rb +2 -2
  56. data/proto_docs/google/cloud/aiplatform/v1/custom_job.rb +23 -18
  57. data/proto_docs/google/cloud/aiplatform/v1/data_item.rb +6 -5
  58. data/proto_docs/google/cloud/aiplatform/v1/data_labeling_job.rb +17 -17
  59. data/proto_docs/google/cloud/aiplatform/v1/dataset.rb +38 -17
  60. data/proto_docs/google/cloud/aiplatform/v1/dataset_service.rb +180 -24
  61. data/proto_docs/google/cloud/aiplatform/v1/encryption_spec.rb +2 -2
  62. data/proto_docs/google/cloud/aiplatform/v1/endpoint.rb +39 -24
  63. data/proto_docs/google/cloud/aiplatform/v1/endpoint_service.rb +57 -35
  64. data/proto_docs/google/cloud/aiplatform/v1/entity_type.rb +8 -6
  65. data/proto_docs/google/cloud/aiplatform/v1/explanation.rb +111 -75
  66. data/proto_docs/google/cloud/aiplatform/v1/explanation_metadata.rb +58 -38
  67. data/proto_docs/google/cloud/aiplatform/v1/feature.rb +9 -6
  68. data/proto_docs/google/cloud/aiplatform/v1/feature_monitoring_stats.rb +2 -1
  69. data/proto_docs/google/cloud/aiplatform/v1/featurestore.rb +12 -11
  70. data/proto_docs/google/cloud/aiplatform/v1/featurestore_monitoring.rb +6 -2
  71. data/proto_docs/google/cloud/aiplatform/v1/featurestore_online_service.rb +65 -10
  72. data/proto_docs/google/cloud/aiplatform/v1/featurestore_service.rb +141 -99
  73. data/proto_docs/google/cloud/aiplatform/v1/hyperparameter_tuning_job.rb +10 -8
  74. data/proto_docs/google/cloud/aiplatform/v1/index.rb +19 -18
  75. data/proto_docs/google/cloud/aiplatform/v1/index_endpoint.rb +36 -24
  76. data/proto_docs/google/cloud/aiplatform/v1/index_endpoint_service.rb +49 -29
  77. data/proto_docs/google/cloud/aiplatform/v1/index_service.rb +35 -17
  78. data/proto_docs/google/cloud/aiplatform/v1/io.rb +2 -1
  79. data/proto_docs/google/cloud/aiplatform/v1/job_service.rb +94 -57
  80. data/proto_docs/google/cloud/aiplatform/v1/machine_resources.rb +53 -43
  81. data/proto_docs/google/cloud/aiplatform/v1/manual_batch_tuning_parameters.rb +5 -5
  82. data/proto_docs/google/cloud/aiplatform/v1/metadata_schema.rb +3 -3
  83. data/proto_docs/google/cloud/aiplatform/v1/metadata_service.rb +141 -91
  84. data/proto_docs/google/cloud/aiplatform/v1/migratable_resource.rb +3 -3
  85. data/proto_docs/google/cloud/aiplatform/v1/migration_service.rb +18 -15
  86. data/proto_docs/google/cloud/aiplatform/v1/model.rb +180 -133
  87. data/proto_docs/google/cloud/aiplatform/v1/model_deployment_monitoring_job.rb +22 -18
  88. data/proto_docs/google/cloud/aiplatform/v1/model_evaluation.rb +7 -6
  89. data/proto_docs/google/cloud/aiplatform/v1/model_evaluation_slice.rb +11 -7
  90. data/proto_docs/google/cloud/aiplatform/v1/model_monitoring.rb +4 -2
  91. data/proto_docs/google/cloud/aiplatform/v1/model_service.rb +119 -54
  92. data/proto_docs/google/cloud/aiplatform/v1/operation.rb +1 -1
  93. data/proto_docs/google/cloud/aiplatform/v1/pipeline_job.rb +59 -45
  94. data/proto_docs/google/cloud/aiplatform/v1/pipeline_service.rb +42 -22
  95. data/proto_docs/google/cloud/aiplatform/v1/prediction_service.rb +42 -28
  96. data/proto_docs/google/cloud/aiplatform/v1/saved_query.rb +2 -2
  97. data/proto_docs/google/cloud/aiplatform/v1/specialist_pool.rb +1 -1
  98. data/proto_docs/google/cloud/aiplatform/v1/specialist_pool_service.rb +19 -11
  99. data/proto_docs/google/cloud/aiplatform/v1/study.rb +42 -25
  100. data/proto_docs/google/cloud/aiplatform/v1/tensorboard.rb +3 -3
  101. data/proto_docs/google/cloud/aiplatform/v1/tensorboard_data.rb +8 -7
  102. data/proto_docs/google/cloud/aiplatform/v1/tensorboard_experiment.rb +2 -1
  103. data/proto_docs/google/cloud/aiplatform/v1/tensorboard_service.rb +203 -109
  104. data/proto_docs/google/cloud/aiplatform/v1/tensorboard_time_series.rb +8 -6
  105. data/proto_docs/google/cloud/aiplatform/v1/training_pipeline.rb +84 -59
  106. data/proto_docs/google/cloud/aiplatform/v1/unmanaged_container_model.rb +2 -2
  107. data/proto_docs/google/cloud/aiplatform/v1/vizier_service.rb +40 -20
  108. data/proto_docs/google/rpc/status.rb +4 -2
  109. metadata +5 -5
@@ -21,13 +21,13 @@ module Google
21
21
  module Cloud
22
22
  module AIPlatform
23
23
  module V1
24
- # Request message for {::Google::Cloud::AIPlatform::V1::MigrationService::Client#search_migratable_resources MigrationService.SearchMigratableResources}.
24
+ # Request message for
25
+ # {::Google::Cloud::AIPlatform::V1::MigrationService::Client#search_migratable_resources MigrationService.SearchMigratableResources}.
25
26
  # @!attribute [rw] parent
26
27
  # @return [::String]
27
- # Required. The location that the migratable resources should be searched from.
28
- # It's the Vertex AI location that the resources can be migrated to, not
29
- # the resources' original location.
30
- # Format:
28
+ # Required. The location that the migratable resources should be searched
29
+ # from. It's the Vertex AI location that the resources can be migrated to,
30
+ # not the resources' original location. Format:
31
31
  # `projects/{project}/locations/{location}`
32
32
  # @!attribute [rw] page_size
33
33
  # @return [::Integer]
@@ -55,7 +55,8 @@ module Google
55
55
  extend ::Google::Protobuf::MessageExts::ClassMethods
56
56
  end
57
57
 
58
- # Response message for {::Google::Cloud::AIPlatform::V1::MigrationService::Client#search_migratable_resources MigrationService.SearchMigratableResources}.
58
+ # Response message for
59
+ # {::Google::Cloud::AIPlatform::V1::MigrationService::Client#search_migratable_resources MigrationService.SearchMigratableResources}.
59
60
  # @!attribute [rw] migratable_resources
60
61
  # @return [::Array<::Google::Cloud::AIPlatform::V1::MigratableResource>]
61
62
  # All migratable resources that can be migrated to the
@@ -70,7 +71,8 @@ module Google
70
71
  extend ::Google::Protobuf::MessageExts::ClassMethods
71
72
  end
72
73
 
73
- # Request message for {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
74
+ # Request message for
75
+ # {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
74
76
  # @!attribute [rw] parent
75
77
  # @return [::String]
76
78
  # Required. The location of the migrated resource will live in.
@@ -109,9 +111,8 @@ module Google
109
111
  # Config for migrating version in ml.googleapis.com to Vertex AI's Model.
110
112
  # @!attribute [rw] endpoint
111
113
  # @return [::String]
112
- # Required. The ml.googleapis.com endpoint that this model version should be migrated
113
- # from.
114
- # Example values:
114
+ # Required. The ml.googleapis.com endpoint that this model version should
115
+ # be migrated from. Example values:
115
116
  #
116
117
  # * ml.googleapis.com
117
118
  #
@@ -177,9 +178,9 @@ module Google
177
178
  # System will pick a display name if unspecified.
178
179
  # @!attribute [rw] migrate_data_labeling_annotated_dataset_configs
179
180
  # @return [::Array<::Google::Cloud::AIPlatform::V1::MigrateResourceRequest::MigrateDataLabelingDatasetConfig::MigrateDataLabelingAnnotatedDatasetConfig>]
180
- # Optional. Configs for migrating AnnotatedDataset in datalabeling.googleapis.com to
181
- # Vertex AI's SavedQuery. The specified AnnotatedDatasets have to belong
182
- # to the datalabeling Dataset.
181
+ # Optional. Configs for migrating AnnotatedDataset in
182
+ # datalabeling.googleapis.com to Vertex AI's SavedQuery. The specified
183
+ # AnnotatedDatasets have to belong to the datalabeling Dataset.
183
184
  class MigrateDataLabelingDatasetConfig
184
185
  include ::Google::Protobuf::MessageExts
185
186
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -198,7 +199,8 @@ module Google
198
199
  end
199
200
  end
200
201
 
201
- # Response message for {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
202
+ # Response message for
203
+ # {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
202
204
  # @!attribute [rw] migrate_resource_responses
203
205
  # @return [::Array<::Google::Cloud::AIPlatform::V1::MigrateResourceResponse>]
204
206
  # Successfully migrated resources.
@@ -223,7 +225,8 @@ module Google
223
225
  extend ::Google::Protobuf::MessageExts::ClassMethods
224
226
  end
225
227
 
226
- # Runtime operation information for {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
228
+ # Runtime operation information for
229
+ # {::Google::Cloud::AIPlatform::V1::MigrationService::Client#batch_migrate_resources MigrationService.BatchMigrateResources}.
227
230
  # @!attribute [rw] generic_metadata
228
231
  # @return [::Google::Cloud::AIPlatform::V1::GenericOperationMetadata]
229
232
  # The common part of the operation metadata.
@@ -51,7 +51,7 @@ module Google
51
51
  # @!attribute [rw] display_name
52
52
  # @return [::String]
53
53
  # Required. The display name of the Model.
54
- # The name can be up to 128 characters long and can be consist of any UTF-8
54
+ # The name can be up to 128 characters long and can consist of any UTF-8
55
55
  # characters.
56
56
  # @!attribute [rw] description
57
57
  # @return [::String]
@@ -63,13 +63,15 @@ module Google
63
63
  # @return [::Google::Cloud::AIPlatform::V1::PredictSchemata]
64
64
  # The schemata that describe formats of the Model's predictions and
65
65
  # explanations as given and returned via
66
- # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict} and {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
66
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
67
+ # and
68
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
67
69
  # @!attribute [rw] metadata_schema_uri
68
70
  # @return [::String]
69
- # Immutable. Points to a YAML file stored on Google Cloud Storage describing additional
70
- # information about the Model, that is specific to it. Unset if the Model
71
- # does not have any additional information.
72
- # The schema is defined as an OpenAPI 3.0.2 [Schema
71
+ # Immutable. Points to a YAML file stored on Google Cloud Storage describing
72
+ # additional information about the Model, that is specific to it. Unset if
73
+ # the Model does not have any additional information. The schema is defined
74
+ # as an OpenAPI 3.0.2 [Schema
73
75
  # Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
74
76
  # AutoML Models always have this field populated by Vertex AI, if no
75
77
  # additional metadata is needed, this field is set to an empty string.
@@ -78,48 +80,55 @@ module Google
78
80
  # point to a location where the user only has a read access.
79
81
  # @!attribute [rw] metadata
80
82
  # @return [::Google::Protobuf::Value]
81
- # Immutable. An additional information about the Model; the schema of the metadata can
82
- # be found in {::Google::Cloud::AIPlatform::V1::Model#metadata_schema_uri metadata_schema}.
83
+ # Immutable. An additional information about the Model; the schema of the
84
+ # metadata can be found in
85
+ # {::Google::Cloud::AIPlatform::V1::Model#metadata_schema_uri metadata_schema}.
83
86
  # Unset if the Model does not have any additional information.
84
87
  # @!attribute [r] supported_export_formats
85
88
  # @return [::Array<::Google::Cloud::AIPlatform::V1::Model::ExportFormat>]
86
- # Output only. The formats in which this Model may be exported. If empty, this Model is
87
- # not available for export.
89
+ # Output only. The formats in which this Model may be exported. If empty,
90
+ # this Model is not available for export.
88
91
  # @!attribute [r] training_pipeline
89
92
  # @return [::String]
90
- # Output only. The resource name of the TrainingPipeline that uploaded this Model, if
91
- # any.
93
+ # Output only. The resource name of the TrainingPipeline that uploaded this
94
+ # Model, if any.
92
95
  # @!attribute [rw] container_spec
93
96
  # @return [::Google::Cloud::AIPlatform::V1::ModelContainerSpec]
94
- # Input only. The specification of the container that is to be used when deploying
95
- # this Model. The specification is ingested upon
96
- # {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel}, and all binaries it contains are copied
97
- # and stored internally by Vertex AI.
97
+ # Input only. The specification of the container that is to be used when
98
+ # deploying this Model. The specification is ingested upon
99
+ # {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel},
100
+ # and all binaries it contains are copied and stored internally by Vertex AI.
98
101
  # Not present for AutoML Models.
99
102
  # @!attribute [rw] artifact_uri
100
103
  # @return [::String]
101
- # Immutable. The path to the directory containing the Model artifact and any of its
102
- # supporting files.
103
- # Not present for AutoML Models.
104
+ # Immutable. The path to the directory containing the Model artifact and any
105
+ # of its supporting files. Not present for AutoML Models.
104
106
  # @!attribute [r] supported_deployment_resources_types
105
107
  # @return [::Array<::Google::Cloud::AIPlatform::V1::Model::DeploymentResourcesType>]
106
- # Output only. When this Model is deployed, its prediction resources are described by the
107
- # `prediction_resources` field of the {::Google::Cloud::AIPlatform::V1::Endpoint#deployed_models Endpoint.deployed_models} object.
108
- # Because not all Models support all resource configuration types, the
109
- # configuration types this Model supports are listed here. If no
108
+ # Output only. When this Model is deployed, its prediction resources are
109
+ # described by the `prediction_resources` field of the
110
+ # {::Google::Cloud::AIPlatform::V1::Endpoint#deployed_models Endpoint.deployed_models}
111
+ # object. Because not all Models support all resource configuration types,
112
+ # the configuration types this Model supports are listed here. If no
110
113
  # configuration types are listed, the Model cannot be deployed to an
111
114
  # {::Google::Cloud::AIPlatform::V1::Endpoint Endpoint} and does not support
112
- # online predictions ({::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict} or
113
- # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}). Such a Model can serve predictions by
114
- # using a {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}, if it has at least one entry each in
115
- # {::Google::Cloud::AIPlatform::V1::Model#supported_input_storage_formats supported_input_storage_formats} and
115
+ # online predictions
116
+ # ({::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
117
+ # or
118
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}).
119
+ # Such a Model can serve predictions by using a
120
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}, if it
121
+ # has at least one entry each in
122
+ # {::Google::Cloud::AIPlatform::V1::Model#supported_input_storage_formats supported_input_storage_formats}
123
+ # and
116
124
  # {::Google::Cloud::AIPlatform::V1::Model#supported_output_storage_formats supported_output_storage_formats}.
117
125
  # @!attribute [r] supported_input_storage_formats
118
126
  # @return [::Array<::String>]
119
127
  # Output only. The formats this Model supports in
120
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#input_config BatchPredictionJob.input_config}. If
121
- # {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri PredictSchemata.instance_schema_uri} exists, the instances
122
- # should be given as per that schema.
128
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#input_config BatchPredictionJob.input_config}.
129
+ # If
130
+ # {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri PredictSchemata.instance_schema_uri}
131
+ # exists, the instances should be given as per that schema.
123
132
  #
124
133
  # The possible formats are:
125
134
  #
@@ -130,11 +139,13 @@ module Google
130
139
  # * `csv`
131
140
  # The CSV format, where each instance is a single comma-separated line.
132
141
  # The first line in the file is the header, containing comma-separated field
133
- # names. Uses {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig#gcs_source GcsSource}.
142
+ # names. Uses
143
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig#gcs_source GcsSource}.
134
144
  #
135
145
  # * `tf-record`
136
146
  # The TFRecord format, where each instance is a single record in tfrecord
137
- # syntax. Uses {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig#gcs_source GcsSource}.
147
+ # syntax. Uses
148
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig#gcs_source GcsSource}.
138
149
  #
139
150
  # * `tf-record-gzip`
140
151
  # Similar to `tf-record`, but the file is gzipped. Uses
@@ -147,23 +158,30 @@ module Google
147
158
  # * `file-list`
148
159
  # Each line of the file is the location of an instance to process, uses
149
160
  # `gcs_source` field of the
150
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig InputConfig} object.
161
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob::InputConfig InputConfig}
162
+ # object.
151
163
  #
152
164
  #
153
165
  # If this Model doesn't support any of these formats it means it cannot be
154
- # used with a {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}. However, if it has
155
- # {::Google::Cloud::AIPlatform::V1::Model#supported_deployment_resources_types supported_deployment_resources_types}, it could serve online
156
- # predictions by using {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict} or
166
+ # used with a
167
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
168
+ # However, if it has
169
+ # {::Google::Cloud::AIPlatform::V1::Model#supported_deployment_resources_types supported_deployment_resources_types},
170
+ # it could serve online predictions by using
171
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
172
+ # or
157
173
  # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
158
174
  # @!attribute [r] supported_output_storage_formats
159
175
  # @return [::Array<::String>]
160
176
  # Output only. The formats this Model supports in
161
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#output_config BatchPredictionJob.output_config}. If both
162
- # {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri PredictSchemata.instance_schema_uri} and
163
- # {::Google::Cloud::AIPlatform::V1::PredictSchemata#prediction_schema_uri PredictSchemata.prediction_schema_uri} exist, the predictions
164
- # are returned together with their instances. In other words, the
165
- # prediction has the original instance data first, followed
166
- # by the actual prediction content (as per the schema).
177
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#output_config BatchPredictionJob.output_config}.
178
+ # If both
179
+ # {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri PredictSchemata.instance_schema_uri}
180
+ # and
181
+ # {::Google::Cloud::AIPlatform::V1::PredictSchemata#prediction_schema_uri PredictSchemata.prediction_schema_uri}
182
+ # exist, the predictions are returned together with their instances. In other
183
+ # words, the prediction has the original instance data first, followed by the
184
+ # actual prediction content (as per the schema).
167
185
  #
168
186
  # The possible formats are:
169
187
  #
@@ -184,9 +202,13 @@ module Google
184
202
  #
185
203
  #
186
204
  # If this Model doesn't support any of these formats it means it cannot be
187
- # used with a {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}. However, if it has
188
- # {::Google::Cloud::AIPlatform::V1::Model#supported_deployment_resources_types supported_deployment_resources_types}, it could serve online
189
- # predictions by using {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict} or
205
+ # used with a
206
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
207
+ # However, if it has
208
+ # {::Google::Cloud::AIPlatform::V1::Model#supported_deployment_resources_types supported_deployment_resources_types},
209
+ # it could serve online predictions by using
210
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}
211
+ # or
190
212
  # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}.
191
213
  # @!attribute [r] create_time
192
214
  # @return [::Google::Protobuf::Timestamp]
@@ -196,32 +218,36 @@ module Google
196
218
  # Output only. Timestamp when this Model was most recently updated.
197
219
  # @!attribute [r] deployed_models
198
220
  # @return [::Array<::Google::Cloud::AIPlatform::V1::DeployedModelRef>]
199
- # Output only. The pointers to DeployedModels created from this Model. Note that
200
- # Model could have been deployed to Endpoints in different Locations.
221
+ # Output only. The pointers to DeployedModels created from this Model. Note
222
+ # that Model could have been deployed to Endpoints in different Locations.
201
223
  # @!attribute [rw] explanation_spec
202
224
  # @return [::Google::Cloud::AIPlatform::V1::ExplanationSpec]
203
225
  # The default explanation specification for this Model.
204
226
  #
205
227
  # The Model can be used for [requesting
206
228
  # explanation][PredictionService.Explain] after being
207
- # {::Google::Cloud::AIPlatform::V1::EndpointService::Client#deploy_model deployed} if it is populated.
208
- # The Model can be used for [batch
229
+ # {::Google::Cloud::AIPlatform::V1::EndpointService::Client#deploy_model deployed} if it is
230
+ # populated. The Model can be used for [batch
209
231
  # explanation][BatchPredictionJob.generate_explanation] if it is populated.
210
232
  #
211
233
  # All fields of the explanation_spec can be overridden by
212
- # {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec} of
213
- # {::Google::Cloud::AIPlatform::V1::DeployModelRequest#deployed_model DeployModelRequest.deployed_model}, or
214
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#explanation_spec explanation_spec} of
215
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
234
+ # {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
235
+ # of
236
+ # {::Google::Cloud::AIPlatform::V1::DeployModelRequest#deployed_model DeployModelRequest.deployed_model},
237
+ # or
238
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#explanation_spec explanation_spec}
239
+ # of {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
216
240
  #
217
241
  # If the default explanation specification is not set for this Model, this
218
242
  # Model can still be used for [requesting
219
243
  # explanation][PredictionService.Explain] by setting
220
- # {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec} of
221
- # {::Google::Cloud::AIPlatform::V1::DeployModelRequest#deployed_model DeployModelRequest.deployed_model} and for [batch
222
- # explanation][BatchPredictionJob.generate_explanation] by setting
223
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#explanation_spec explanation_spec} of
224
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
244
+ # {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
245
+ # of
246
+ # {::Google::Cloud::AIPlatform::V1::DeployModelRequest#deployed_model DeployModelRequest.deployed_model}
247
+ # and for [batch explanation][BatchPredictionJob.generate_explanation] by
248
+ # setting
249
+ # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#explanation_spec explanation_spec}
250
+ # of {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
225
251
  # @!attribute [rw] etag
226
252
  # @return [::String]
227
253
  # Used to perform consistent read-modify-write updates. If not set, a blind
@@ -241,12 +267,13 @@ module Google
241
267
  # Model and all sub-resources of this Model will be secured by this key.
242
268
  # @!attribute [r] model_source_info
243
269
  # @return [::Google::Cloud::AIPlatform::V1::ModelSourceInfo]
244
- # Output only. Source of a model. It can either be automl training pipeline, custom
245
- # training pipeline, BigQuery ML, or existing Vertex AI Model.
270
+ # Output only. Source of a model. It can either be automl training pipeline,
271
+ # custom training pipeline, BigQuery ML, or existing Vertex AI Model.
246
272
  # @!attribute [r] metadata_artifact
247
273
  # @return [::String]
248
- # Output only. The resource name of the Artifact that was created in MetadataStore when
249
- # creating the Model. The Artifact resource name pattern is
274
+ # Output only. The resource name of the Artifact that was created in
275
+ # MetadataStore when creating the Model. The Artifact resource name pattern
276
+ # is
250
277
  # `projects/{project}/locations/{location}/metadataStores/{metadata_store}/artifacts/{artifact}`.
251
278
  class Model
252
279
  include ::Google::Protobuf::MessageExts
@@ -291,12 +318,15 @@ module Google
291
318
 
292
319
  # Model artifact and any of its supported files. Will be exported to the
293
320
  # location specified by the `artifactDestination` field of the
294
- # {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config} object.
321
+ # {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config}
322
+ # object.
295
323
  ARTIFACT = 1
296
324
 
297
325
  # The container image that is to be used when deploying this Model. Will
298
326
  # be exported to the location specified by the `imageDestination` field
299
- # of the {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config} object.
327
+ # of the
328
+ # {::Google::Cloud::AIPlatform::V1::ExportModelRequest#output_config ExportModelRequest.output_config}
329
+ # object.
300
330
  IMAGE = 2
301
331
  end
302
332
  end
@@ -315,28 +345,33 @@ module Google
315
345
  # Should not be used.
316
346
  DEPLOYMENT_RESOURCES_TYPE_UNSPECIFIED = 0
317
347
 
318
- # Resources that are dedicated to the {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModel}, and that need a
319
- # higher degree of manual configuration.
348
+ # Resources that are dedicated to the
349
+ # {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModel}, and that need
350
+ # a higher degree of manual configuration.
320
351
  DEDICATED_RESOURCES = 1
321
352
 
322
353
  # Resources that to large degree are decided by Vertex AI, and require
323
354
  # only a modest additional configuration.
324
355
  AUTOMATIC_RESOURCES = 2
325
356
 
326
- # Resources that can be shared by multiple {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModels}.
327
- # A pre-configured [DeploymentResourcePool][] is required.
357
+ # Resources that can be shared by multiple
358
+ # {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModels}. A
359
+ # pre-configured [DeploymentResourcePool][] is required.
328
360
  SHARED_RESOURCES = 3
329
361
  end
330
362
  end
331
363
 
332
364
  # Contains the schemata used in Model's predictions and explanations via
333
- # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict}, {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain} and
334
- # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
365
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict PredictionService.Predict},
366
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#explain PredictionService.Explain}
367
+ # and {::Google::Cloud::AIPlatform::V1::BatchPredictionJob BatchPredictionJob}.
335
368
  # @!attribute [rw] instance_schema_uri
336
369
  # @return [::String]
337
- # Immutable. Points to a YAML file stored on Google Cloud Storage describing the format
338
- # of a single instance, which are used in {::Google::Cloud::AIPlatform::V1::PredictRequest#instances PredictRequest.instances},
339
- # {::Google::Cloud::AIPlatform::V1::ExplainRequest#instances ExplainRequest.instances} and
370
+ # Immutable. Points to a YAML file stored on Google Cloud Storage describing
371
+ # the format of a single instance, which are used in
372
+ # {::Google::Cloud::AIPlatform::V1::PredictRequest#instances PredictRequest.instances},
373
+ # {::Google::Cloud::AIPlatform::V1::ExplainRequest#instances ExplainRequest.instances}
374
+ # and
340
375
  # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#input_config BatchPredictionJob.input_config}.
341
376
  # The schema is defined as an OpenAPI 3.0.2 [Schema
342
377
  # Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
@@ -346,9 +381,11 @@ module Google
346
381
  # point to a location where the user only has a read access.
347
382
  # @!attribute [rw] parameters_schema_uri
348
383
  # @return [::String]
349
- # Immutable. Points to a YAML file stored on Google Cloud Storage describing the
350
- # parameters of prediction and explanation via
351
- # {::Google::Cloud::AIPlatform::V1::PredictRequest#parameters PredictRequest.parameters}, {::Google::Cloud::AIPlatform::V1::ExplainRequest#parameters ExplainRequest.parameters} and
384
+ # Immutable. Points to a YAML file stored on Google Cloud Storage describing
385
+ # the parameters of prediction and explanation via
386
+ # {::Google::Cloud::AIPlatform::V1::PredictRequest#parameters PredictRequest.parameters},
387
+ # {::Google::Cloud::AIPlatform::V1::ExplainRequest#parameters ExplainRequest.parameters}
388
+ # and
352
389
  # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#model_parameters BatchPredictionJob.model_parameters}.
353
390
  # The schema is defined as an OpenAPI 3.0.2 [Schema
354
391
  # Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
@@ -359,9 +396,12 @@ module Google
359
396
  # point to a location where the user only has a read access.
360
397
  # @!attribute [rw] prediction_schema_uri
361
398
  # @return [::String]
362
- # Immutable. Points to a YAML file stored on Google Cloud Storage describing the format
363
- # of a single prediction produced by this Model, which are returned via
364
- # {::Google::Cloud::AIPlatform::V1::PredictResponse#predictions PredictResponse.predictions}, {::Google::Cloud::AIPlatform::V1::ExplainResponse#explanations ExplainResponse.explanations}, and
399
+ # Immutable. Points to a YAML file stored on Google Cloud Storage describing
400
+ # the format of a single prediction produced by this Model, which are
401
+ # returned via
402
+ # {::Google::Cloud::AIPlatform::V1::PredictResponse#predictions PredictResponse.predictions},
403
+ # {::Google::Cloud::AIPlatform::V1::ExplainResponse#explanations ExplainResponse.explanations},
404
+ # and
365
405
  # {::Google::Cloud::AIPlatform::V1::BatchPredictionJob#output_config BatchPredictionJob.output_config}.
366
406
  # The schema is defined as an OpenAPI 3.0.2 [Schema
367
407
  # Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
@@ -379,14 +419,16 @@ module Google
379
419
  # specification](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
380
420
  # @!attribute [rw] image_uri
381
421
  # @return [::String]
382
- # Required. Immutable. URI of the Docker image to be used as the custom container for serving
383
- # predictions. This URI must identify an image in Artifact Registry or
384
- # Container Registry. Learn more about the [container publishing
422
+ # Required. Immutable. URI of the Docker image to be used as the custom
423
+ # container for serving predictions. This URI must identify an image in
424
+ # Artifact Registry or Container Registry. Learn more about the [container
425
+ # publishing
385
426
  # requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing),
386
427
  # including permissions requirements for the Vertex AI Service Agent.
387
428
  #
388
- # The container image is ingested upon {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel}, stored
389
- # internally, and this original path is afterwards not used.
429
+ # The container image is ingested upon
430
+ # {::Google::Cloud::AIPlatform::V1::ModelService::Client#upload_model ModelService.UploadModel},
431
+ # stored internally, and this original path is afterwards not used.
390
432
  #
391
433
  # To learn about the requirements for the Docker image itself, see
392
434
  # [Custom container
@@ -397,14 +439,15 @@ module Google
397
439
  # in this field.
398
440
  # @!attribute [rw] command
399
441
  # @return [::Array<::String>]
400
- # Immutable. Specifies the command that runs when the container starts. This overrides
401
- # the container's
442
+ # Immutable. Specifies the command that runs when the container starts. This
443
+ # overrides the container's
402
444
  # [ENTRYPOINT](https://docs.docker.com/engine/reference/builder/#entrypoint).
403
445
  # Specify this field as an array of executable and arguments, similar to a
404
446
  # Docker `ENTRYPOINT`'s "exec" form, not its "shell" form.
405
447
  #
406
448
  # If you do not specify this field, then the container's `ENTRYPOINT` runs,
407
- # in conjunction with the {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#args args} field or the
449
+ # in conjunction with the
450
+ # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#args args} field or the
408
451
  # container's [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd),
409
452
  # if either exists. If this field is not specified and the container does not
410
453
  # have an `ENTRYPOINT`, then refer to the Docker documentation about [how
@@ -420,32 +463,30 @@ module Google
420
463
  #
421
464
  # In this field, you can reference [environment variables set by Vertex
422
465
  # AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables)
423
- # and environment variables set in the {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#env env} field.
424
- # You cannot reference environment variables set in the Docker image. In
425
- # order for environment variables to be expanded, reference them by using the
426
- # following syntax:
427
- # <code>$(<var>VARIABLE_NAME</var>)</code>
428
- # Note that this differs from Bash variable expansion, which does not use
429
- # parentheses. If a variable cannot be resolved, the reference in the input
430
- # string is used unchanged. To avoid variable expansion, you can escape this
431
- # syntax with `$$`; for example:
432
- # <code>$$(<var>VARIABLE_NAME</var>)</code>
433
- # This field corresponds to the `command` field of the Kubernetes Containers
434
- # [v1 core
466
+ # and environment variables set in the
467
+ # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#env env} field. You cannot
468
+ # reference environment variables set in the Docker image. In order for
469
+ # environment variables to be expanded, reference them by using the following
470
+ # syntax: <code>$(<var>VARIABLE_NAME</var>)</code> Note that this differs
471
+ # from Bash variable expansion, which does not use parentheses. If a variable
472
+ # cannot be resolved, the reference in the input string is used unchanged. To
473
+ # avoid variable expansion, you can escape this syntax with `$$`; for
474
+ # example: <code>$$(<var>VARIABLE_NAME</var>)</code> This field corresponds
475
+ # to the `command` field of the Kubernetes Containers [v1 core
435
476
  # API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
436
477
  # @!attribute [rw] args
437
478
  # @return [::Array<::String>]
438
- # Immutable. Specifies arguments for the command that runs when the container starts.
439
- # This overrides the container's
479
+ # Immutable. Specifies arguments for the command that runs when the container
480
+ # starts. This overrides the container's
440
481
  # [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd). Specify
441
482
  # this field as an array of executable and arguments, similar to a Docker
442
483
  # `CMD`'s "default parameters" form.
443
484
  #
444
485
  # If you don't specify this field but do specify the
445
- # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} field, then the command from the
446
- # `command` field runs without any additional arguments. See the
447
- # [Kubernetes documentation about how the
448
- # `command` and `args` fields interact with a container's `ENTRYPOINT` and
486
+ # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} field,
487
+ # then the command from the `command` field runs without any additional
488
+ # arguments. See the [Kubernetes documentation about how the `command` and
489
+ # `args` fields interact with a container's `ENTRYPOINT` and
449
490
  # `CMD`](https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#notes).
450
491
  #
451
492
  # If you don't specify this field and don't specify the `command` field,
@@ -458,29 +499,29 @@ module Google
458
499
  # In this field, you can reference [environment variables
459
500
  # set by Vertex
460
501
  # AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables)
461
- # and environment variables set in the {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#env env} field.
462
- # You cannot reference environment variables set in the Docker image. In
463
- # order for environment variables to be expanded, reference them by using the
464
- # following syntax:
465
- # <code>$(<var>VARIABLE_NAME</var>)</code>
466
- # Note that this differs from Bash variable expansion, which does not use
467
- # parentheses. If a variable cannot be resolved, the reference in the input
468
- # string is used unchanged. To avoid variable expansion, you can escape this
469
- # syntax with `$$`; for example:
470
- # <code>$$(<var>VARIABLE_NAME</var>)</code>
471
- # This field corresponds to the `args` field of the Kubernetes Containers
472
- # [v1 core
502
+ # and environment variables set in the
503
+ # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#env env} field. You cannot
504
+ # reference environment variables set in the Docker image. In order for
505
+ # environment variables to be expanded, reference them by using the following
506
+ # syntax: <code>$(<var>VARIABLE_NAME</var>)</code> Note that this differs
507
+ # from Bash variable expansion, which does not use parentheses. If a variable
508
+ # cannot be resolved, the reference in the input string is used unchanged. To
509
+ # avoid variable expansion, you can escape this syntax with `$$`; for
510
+ # example: <code>$$(<var>VARIABLE_NAME</var>)</code> This field corresponds
511
+ # to the `args` field of the Kubernetes Containers [v1 core
473
512
  # API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
474
513
  # @!attribute [rw] env
475
514
  # @return [::Array<::Google::Cloud::AIPlatform::V1::EnvVar>]
476
- # Immutable. List of environment variables to set in the container. After the container
477
- # starts running, code running in the container can read these environment
478
- # variables.
515
+ # Immutable. List of environment variables to set in the container. After the
516
+ # container starts running, code running in the container can read these
517
+ # environment variables.
479
518
  #
480
- # Additionally, the {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} and
481
- # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#args args} fields can reference these variables. Later
482
- # entries in this list can also reference earlier entries. For example, the
483
- # following example sets the variable `VAR_2` to have the value `foo bar`:
519
+ # Additionally, the
520
+ # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#command command} and
521
+ # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#args args} fields can
522
+ # reference these variables. Later entries in this list can also reference
523
+ # earlier entries. For example, the following example sets the variable
524
+ # `VAR_2` to have the value `foo bar`:
484
525
  #
485
526
  # ```json
486
527
  # [
@@ -526,11 +567,11 @@ module Google
526
567
  # API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
527
568
  # @!attribute [rw] predict_route
528
569
  # @return [::String]
529
- # Immutable. HTTP path on the container to send prediction requests to. Vertex AI
530
- # forwards requests sent using
531
- # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict projects.locations.endpoints.predict} to this
532
- # path on the container's IP address and port. Vertex AI then returns the
533
- # container's response in the API response.
570
+ # Immutable. HTTP path on the container to send prediction requests to.
571
+ # Vertex AI forwards requests sent using
572
+ # {::Google::Cloud::AIPlatform::V1::PredictionService::Client#predict projects.locations.endpoints.predict}
573
+ # to this path on the container's IP address and port. Vertex AI then returns
574
+ # the container's response in the API response.
534
575
  #
535
576
  # For example, if you set this field to `/foo`, then when Vertex AI
536
577
  # receives a prediction request, it forwards the request body in a POST
@@ -539,7 +580,8 @@ module Google
539
580
  # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#ports ports} field.
540
581
  #
541
582
  # If you don't specify this field, it defaults to the following value when
542
- # you {::Google::Cloud::AIPlatform::V1::EndpointService::Client#deploy_model deploy this Model to an Endpoint}:
583
+ # you [deploy this Model to an
584
+ # Endpoint][google.cloud.aiplatform.v1.EndpointService.DeployModel]:
543
585
  # <code>/v1/endpoints/<var>ENDPOINT</var>/deployedModels/<var>DEPLOYED_MODEL</var>:predict</code>
544
586
  # The placeholders in this value are replaced as follows:
545
587
  #
@@ -549,7 +591,9 @@ module Google
549
591
  # as the [`AIP_ENDPOINT_ID` environment
550
592
  # variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
551
593
  #
552
- # * <var>DEPLOYED_MODEL</var>: {::Google::Cloud::AIPlatform::V1::DeployedModel#id DeployedModel.id} of the `DeployedModel`.
594
+ # * <var>DEPLOYED_MODEL</var>:
595
+ # {::Google::Cloud::AIPlatform::V1::DeployedModel#id DeployedModel.id} of the
596
+ # `DeployedModel`.
553
597
  # (Vertex AI makes this value available to your container code
554
598
  # as the [`AIP_DEPLOYED_MODEL_ID` environment
555
599
  # variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
@@ -567,7 +611,8 @@ module Google
567
611
  # {::Google::Cloud::AIPlatform::V1::ModelContainerSpec#ports ports} field.
568
612
  #
569
613
  # If you don't specify this field, it defaults to the following value when
570
- # you {::Google::Cloud::AIPlatform::V1::EndpointService::Client#deploy_model deploy this Model to an Endpoint}:
614
+ # you [deploy this Model to an
615
+ # Endpoint][google.cloud.aiplatform.v1.EndpointService.DeployModel]:
571
616
  # <code>/v1/endpoints/<var>ENDPOINT</var>/deployedModels/<var>DEPLOYED_MODEL</var>:predict</code>
572
617
  # The placeholders in this value are replaced as follows:
573
618
  #
@@ -577,7 +622,9 @@ module Google
577
622
  # as the [`AIP_ENDPOINT_ID` environment
578
623
  # variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
579
624
  #
580
- # * <var>DEPLOYED_MODEL</var>: {::Google::Cloud::AIPlatform::V1::DeployedModel#id DeployedModel.id} of the `DeployedModel`.
625
+ # * <var>DEPLOYED_MODEL</var>:
626
+ # {::Google::Cloud::AIPlatform::V1::DeployedModel#id DeployedModel.id} of the
627
+ # `DeployedModel`.
581
628
  # (Vertex AI makes this value available to your container code as the
582
629
  # [`AIP_DEPLOYED_MODEL_ID` environment
583
630
  # variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)