google-cloud-ai_platform-v1 0.16.0 → 0.18.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/google/cloud/ai_platform/v1/dataset_service/client.rb +10 -7
- data/lib/google/cloud/ai_platform/v1/dataset_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/endpoint_service/client.rb +28 -21
- data/lib/google/cloud/ai_platform/v1/endpoint_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/client.rb +10 -9
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/featurestore_service/client.rb +53 -50
- data/lib/google/cloud/ai_platform/v1/featurestore_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/client.rb +18 -14
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/index_service/client.rb +8 -4
- data/lib/google/cloud/ai_platform/v1/index_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/job_service/client.rb +53 -36
- data/lib/google/cloud/ai_platform/v1/job_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/metadata_service/client.rb +33 -36
- data/lib/google/cloud/ai_platform/v1/metadata_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/migration_service/client.rb +3 -4
- data/lib/google/cloud/ai_platform/v1/migration_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/model_service/client.rb +44 -22
- data/lib/google/cloud/ai_platform/v1/model_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/pipeline_service/client.rb +28 -17
- data/lib/google/cloud/ai_platform/v1/pipeline_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/prediction_service/client.rb +28 -19
- data/lib/google/cloud/ai_platform/v1/prediction_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/client.rb +5 -3
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/client.rb +165 -71
- data/lib/google/cloud/ai_platform/v1/tensorboard_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/version.rb +1 -1
- data/lib/google/cloud/ai_platform/v1/vizier_service.rb +1 -1
- data/lib/google/cloud/ai_platform/v1.rb +2 -2
- data/lib/google/cloud/aiplatform/v1/batch_prediction_job_pb.rb +8 -0
- data/lib/google/cloud/aiplatform/v1/custom_job_pb.rb +1 -0
- data/lib/google/cloud/aiplatform/v1/dataset_pb.rb +2 -0
- data/lib/google/cloud/aiplatform/v1/index_service_services_pb.rb +2 -1
- data/lib/google/cloud/aiplatform/v1/job_service_services_pb.rb +24 -14
- data/lib/google/cloud/aiplatform/v1/model_service_pb.rb +1 -0
- data/lib/google/cloud/aiplatform/v1/model_service_services_pb.rb +10 -5
- data/lib/google/cloud/aiplatform/v1/pipeline_service_services_pb.rb +16 -9
- data/lib/google/cloud/aiplatform/v1/prediction_service_services_pb.rb +10 -6
- data/lib/google/cloud/aiplatform/v1/study_pb.rb +1 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_pb.rb +17 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_services_pb.rb +8 -8
- data/proto_docs/google/cloud/aiplatform/v1/annotation.rb +9 -6
- data/proto_docs/google/cloud/aiplatform/v1/annotation_spec.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/batch_prediction_job.rb +183 -50
- data/proto_docs/google/cloud/aiplatform/v1/completion_stats.rb +8 -7
- data/proto_docs/google/cloud/aiplatform/v1/context.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/custom_job.rb +34 -17
- data/proto_docs/google/cloud/aiplatform/v1/data_item.rb +6 -5
- data/proto_docs/google/cloud/aiplatform/v1/data_labeling_job.rb +16 -16
- data/proto_docs/google/cloud/aiplatform/v1/dataset.rb +33 -17
- data/proto_docs/google/cloud/aiplatform/v1/dataset_service.rb +59 -32
- data/proto_docs/google/cloud/aiplatform/v1/encryption_spec.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/endpoint.rb +38 -23
- data/proto_docs/google/cloud/aiplatform/v1/endpoint_service.rb +57 -35
- data/proto_docs/google/cloud/aiplatform/v1/entity_type.rb +8 -6
- data/proto_docs/google/cloud/aiplatform/v1/explanation.rb +111 -75
- data/proto_docs/google/cloud/aiplatform/v1/explanation_metadata.rb +58 -38
- data/proto_docs/google/cloud/aiplatform/v1/feature.rb +9 -6
- data/proto_docs/google/cloud/aiplatform/v1/feature_monitoring_stats.rb +2 -1
- data/proto_docs/google/cloud/aiplatform/v1/featurestore.rb +12 -11
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_monitoring.rb +6 -2
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_online_service.rb +27 -19
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_service.rb +138 -99
- data/proto_docs/google/cloud/aiplatform/v1/hyperparameter_tuning_job.rb +9 -7
- data/proto_docs/google/cloud/aiplatform/v1/index.rb +18 -17
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint.rb +36 -24
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint_service.rb +49 -29
- data/proto_docs/google/cloud/aiplatform/v1/index_service.rb +35 -17
- data/proto_docs/google/cloud/aiplatform/v1/io.rb +2 -1
- data/proto_docs/google/cloud/aiplatform/v1/job_service.rb +94 -57
- data/proto_docs/google/cloud/aiplatform/v1/machine_resources.rb +53 -43
- data/proto_docs/google/cloud/aiplatform/v1/manual_batch_tuning_parameters.rb +5 -5
- data/proto_docs/google/cloud/aiplatform/v1/metadata_schema.rb +3 -3
- data/proto_docs/google/cloud/aiplatform/v1/metadata_service.rb +141 -91
- data/proto_docs/google/cloud/aiplatform/v1/migratable_resource.rb +3 -3
- data/proto_docs/google/cloud/aiplatform/v1/migration_service.rb +18 -15
- data/proto_docs/google/cloud/aiplatform/v1/model.rb +179 -132
- data/proto_docs/google/cloud/aiplatform/v1/model_deployment_monitoring_job.rb +21 -17
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation.rb +7 -6
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation_slice.rb +11 -7
- data/proto_docs/google/cloud/aiplatform/v1/model_monitoring.rb +4 -2
- data/proto_docs/google/cloud/aiplatform/v1/model_service.rb +110 -54
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_job.rb +56 -42
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_service.rb +42 -22
- data/proto_docs/google/cloud/aiplatform/v1/prediction_service.rb +42 -28
- data/proto_docs/google/cloud/aiplatform/v1/saved_query.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/specialist_pool_service.rb +19 -11
- data/proto_docs/google/cloud/aiplatform/v1/study.rb +42 -25
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_data.rb +8 -7
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_experiment.rb +2 -1
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_service.rb +203 -109
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_time_series.rb +8 -6
- data/proto_docs/google/cloud/aiplatform/v1/training_pipeline.rb +84 -59
- data/proto_docs/google/cloud/aiplatform/v1/unmanaged_container_model.rb +2 -2
- data/proto_docs/google/cloud/aiplatform/v1/vizier_service.rb +40 -20
- data/proto_docs/google/rpc/status.rb +4 -2
- metadata +9 -9
@@ -207,8 +207,8 @@ module Google
|
|
207
207
|
# the default parameter values, pass an empty Hash as a request object (see above).
|
208
208
|
#
|
209
209
|
# @param parent [::String]
|
210
|
-
# Required. The resource name of the Location to create the TrainingPipeline
|
211
|
-
# Format: `projects/{project}/locations/{location}`
|
210
|
+
# Required. The resource name of the Location to create the TrainingPipeline
|
211
|
+
# in. Format: `projects/{project}/locations/{location}`
|
212
212
|
# @param training_pipeline [::Google::Cloud::AIPlatform::V1::TrainingPipeline, ::Hash]
|
213
213
|
# Required. The TrainingPipeline to create.
|
214
214
|
#
|
@@ -382,8 +382,8 @@ module Google
|
|
382
382
|
# the default parameter values, pass an empty Hash as a request object (see above).
|
383
383
|
#
|
384
384
|
# @param parent [::String]
|
385
|
-
# Required. The resource name of the Location to list the TrainingPipelines
|
386
|
-
# Format: `projects/{project}/locations/{location}`
|
385
|
+
# Required. The resource name of the Location to list the TrainingPipelines
|
386
|
+
# from. Format: `projects/{project}/locations/{location}`
|
387
387
|
# @param filter [::String]
|
388
388
|
# The standard list filter.
|
389
389
|
#
|
@@ -410,8 +410,10 @@ module Google
|
|
410
410
|
# @param page_token [::String]
|
411
411
|
# The standard list page token.
|
412
412
|
# Typically obtained via
|
413
|
-
# {::Google::Cloud::AIPlatform::V1::ListTrainingPipelinesResponse#next_page_token ListTrainingPipelinesResponse.next_page_token}
|
414
|
-
#
|
413
|
+
# {::Google::Cloud::AIPlatform::V1::ListTrainingPipelinesResponse#next_page_token ListTrainingPipelinesResponse.next_page_token}
|
414
|
+
# of the previous
|
415
|
+
# {::Google::Cloud::AIPlatform::V1::PipelineService::Client#list_training_pipelines PipelineService.ListTrainingPipelines}
|
416
|
+
# call.
|
415
417
|
# @param read_mask [::Google::Protobuf::FieldMask, ::Hash]
|
416
418
|
# Mask specifying which fields to read.
|
417
419
|
#
|
@@ -585,13 +587,17 @@ module Google
|
|
585
587
|
# Cancels a TrainingPipeline.
|
586
588
|
# Starts asynchronous cancellation on the TrainingPipeline. The server
|
587
589
|
# makes a best effort to cancel the pipeline, but success is not
|
588
|
-
# guaranteed. Clients can use
|
589
|
-
#
|
590
|
+
# guaranteed. Clients can use
|
591
|
+
# {::Google::Cloud::AIPlatform::V1::PipelineService::Client#get_training_pipeline PipelineService.GetTrainingPipeline}
|
592
|
+
# or other methods to check whether the cancellation succeeded or whether the
|
590
593
|
# pipeline completed despite cancellation. On successful cancellation,
|
591
594
|
# the TrainingPipeline is not deleted; instead it becomes a pipeline with
|
592
|
-
# a
|
593
|
-
#
|
594
|
-
#
|
595
|
+
# a
|
596
|
+
# {::Google::Cloud::AIPlatform::V1::TrainingPipeline#error TrainingPipeline.error}
|
597
|
+
# value with a {::Google::Rpc::Status#code google.rpc.Status.code} of 1,
|
598
|
+
# corresponding to `Code.CANCELLED`, and
|
599
|
+
# {::Google::Cloud::AIPlatform::V1::TrainingPipeline#state TrainingPipeline.state}
|
600
|
+
# is set to `CANCELLED`.
|
595
601
|
#
|
596
602
|
# @overload cancel_training_pipeline(request, options = nil)
|
597
603
|
# Pass arguments to `cancel_training_pipeline` via a request object, either of type
|
@@ -920,8 +926,10 @@ module Google
|
|
920
926
|
# @param page_token [::String]
|
921
927
|
# The standard list page token.
|
922
928
|
# Typically obtained via
|
923
|
-
# {::Google::Cloud::AIPlatform::V1::ListPipelineJobsResponse#next_page_token ListPipelineJobsResponse.next_page_token}
|
924
|
-
#
|
929
|
+
# {::Google::Cloud::AIPlatform::V1::ListPipelineJobsResponse#next_page_token ListPipelineJobsResponse.next_page_token}
|
930
|
+
# of the previous
|
931
|
+
# {::Google::Cloud::AIPlatform::V1::PipelineService::Client#list_pipeline_jobs PipelineService.ListPipelineJobs}
|
932
|
+
# call.
|
925
933
|
# @param order_by [::String]
|
926
934
|
# A comma-separated list of fields to order by. The default sort order is in
|
927
935
|
# ascending order. Use "desc" after a field name for descending. You can have
|
@@ -1109,12 +1117,15 @@ module Google
|
|
1109
1117
|
# Cancels a PipelineJob.
|
1110
1118
|
# Starts asynchronous cancellation on the PipelineJob. The server
|
1111
1119
|
# makes a best effort to cancel the pipeline, but success is not
|
1112
|
-
# guaranteed. Clients can use
|
1113
|
-
#
|
1120
|
+
# guaranteed. Clients can use
|
1121
|
+
# {::Google::Cloud::AIPlatform::V1::PipelineService::Client#get_pipeline_job PipelineService.GetPipelineJob}
|
1122
|
+
# or other methods to check whether the cancellation succeeded or whether the
|
1114
1123
|
# pipeline completed despite cancellation. On successful cancellation,
|
1115
1124
|
# the PipelineJob is not deleted; instead it becomes a pipeline with
|
1116
|
-
# a {::Google::Cloud::AIPlatform::V1::PipelineJob#error PipelineJob.error} value
|
1117
|
-
#
|
1125
|
+
# a {::Google::Cloud::AIPlatform::V1::PipelineJob#error PipelineJob.error} value
|
1126
|
+
# with a {::Google::Rpc::Status#code google.rpc.Status.code} of 1, corresponding
|
1127
|
+
# to `Code.CANCELLED`, and
|
1128
|
+
# {::Google::Cloud::AIPlatform::V1::PipelineJob#state PipelineJob.state} is set to
|
1118
1129
|
# `CANCELLED`.
|
1119
1130
|
#
|
1120
1131
|
# @overload cancel_pipeline_job(request, options = nil)
|
@@ -36,7 +36,7 @@ module Google
|
|
36
36
|
# `TrainingPipeline` resources (used for AutoML and custom training) and
|
37
37
|
# `PipelineJob` resources (used for Vertex AI Pipelines).
|
38
38
|
#
|
39
|
-
#
|
39
|
+
# @example Load this service and instantiate a gRPC client
|
40
40
|
#
|
41
41
|
# require "google/cloud/ai_platform/v1/pipeline_service"
|
42
42
|
# client = ::Google::Cloud::AIPlatform::V1::PipelineService::Client.new
|
@@ -206,7 +206,8 @@ module Google
|
|
206
206
|
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri instance_schema_uri}.
|
207
207
|
# @param parameters [::Google::Protobuf::Value, ::Hash]
|
208
208
|
# The parameters that govern the prediction. The schema of the parameters may
|
209
|
-
# be specified via Endpoint's DeployedModels' [Model's
|
209
|
+
# be specified via Endpoint's DeployedModels' [Model's
|
210
|
+
# ][google.cloud.aiplatform.v1.DeployedModel.model]
|
210
211
|
# [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata]
|
211
212
|
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#parameters_schema_uri parameters_schema_uri}.
|
212
213
|
#
|
@@ -279,11 +280,13 @@ module Google
|
|
279
280
|
#
|
280
281
|
# The response includes the following HTTP headers:
|
281
282
|
#
|
282
|
-
# * `X-Vertex-AI-Endpoint-Id`: ID of the
|
283
|
+
# * `X-Vertex-AI-Endpoint-Id`: ID of the
|
284
|
+
# {::Google::Cloud::AIPlatform::V1::Endpoint Endpoint} that served this
|
283
285
|
# prediction.
|
284
286
|
#
|
285
|
-
# * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's
|
286
|
-
# that served this
|
287
|
+
# * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's
|
288
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModel} that served this
|
289
|
+
# prediction.
|
287
290
|
#
|
288
291
|
# @overload raw_predict(request, options = nil)
|
289
292
|
# Pass arguments to `raw_predict` via a request object, either of type
|
@@ -307,16 +310,18 @@ module Google
|
|
307
310
|
# @param http_body [::Google::Api::HttpBody, ::Hash]
|
308
311
|
# The prediction input. Supports HTTP headers and arbitrary data payload.
|
309
312
|
#
|
310
|
-
# A {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModel} may have an
|
311
|
-
#
|
312
|
-
#
|
313
|
-
#
|
314
|
-
#
|
313
|
+
# A {::Google::Cloud::AIPlatform::V1::DeployedModel DeployedModel} may have an
|
314
|
+
# upper limit on the number of instances it supports per request. When this
|
315
|
+
# limit it is exceeded for an AutoML model, the
|
316
|
+
# {::Google::Cloud::AIPlatform::V1::PredictionService::Client#raw_predict RawPredict}
|
317
|
+
# method returns an error. When this limit is exceeded for a custom-trained
|
318
|
+
# model, the behavior varies depending on the model.
|
315
319
|
#
|
316
320
|
# You can specify the schema for each instance in the
|
317
321
|
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri predict_schemata.instance_schema_uri}
|
318
|
-
# field when you create a {::Google::Cloud::AIPlatform::V1::Model Model}. This
|
319
|
-
# `Model` as a `DeployedModel` to an
|
322
|
+
# field when you create a {::Google::Cloud::AIPlatform::V1::Model Model}. This
|
323
|
+
# schema applies when you deploy the `Model` as a `DeployedModel` to an
|
324
|
+
# {::Google::Cloud::AIPlatform::V1::Endpoint Endpoint} and use the `RawPredict`
|
320
325
|
# method.
|
321
326
|
#
|
322
327
|
# @yield [response, operation] Access the result along with the RPC operation
|
@@ -386,10 +391,12 @@ module Google
|
|
386
391
|
##
|
387
392
|
# Perform an online explanation.
|
388
393
|
#
|
389
|
-
# If
|
390
|
-
#
|
394
|
+
# If
|
395
|
+
# {::Google::Cloud::AIPlatform::V1::ExplainRequest#deployed_model_id deployed_model_id}
|
396
|
+
# is specified, the corresponding DeployModel must have
|
391
397
|
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
392
|
-
# populated. If
|
398
|
+
# populated. If
|
399
|
+
# {::Google::Cloud::AIPlatform::V1::ExplainRequest#deployed_model_id deployed_model_id}
|
393
400
|
# is not specified, all DeployedModels must have
|
394
401
|
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
395
402
|
# populated. Only deployed AutoML tabular Models have
|
@@ -426,21 +433,23 @@ module Google
|
|
426
433
|
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#instance_schema_uri instance_schema_uri}.
|
427
434
|
# @param parameters [::Google::Protobuf::Value, ::Hash]
|
428
435
|
# The parameters that govern the prediction. The schema of the parameters may
|
429
|
-
# be specified via Endpoint's DeployedModels' [Model's
|
436
|
+
# be specified via Endpoint's DeployedModels' [Model's
|
437
|
+
# ][google.cloud.aiplatform.v1.DeployedModel.model]
|
430
438
|
# [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata]
|
431
439
|
# {::Google::Cloud::AIPlatform::V1::PredictSchemata#parameters_schema_uri parameters_schema_uri}.
|
432
440
|
# @param explanation_spec_override [::Google::Cloud::AIPlatform::V1::ExplanationSpecOverride, ::Hash]
|
433
441
|
# If specified, overrides the
|
434
|
-
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
435
|
-
# Can be used for explaining prediction results with
|
436
|
-
# configurations, such as:
|
442
|
+
# {::Google::Cloud::AIPlatform::V1::DeployedModel#explanation_spec explanation_spec}
|
443
|
+
# of the DeployedModel. Can be used for explaining prediction results with
|
444
|
+
# different configurations, such as:
|
437
445
|
# - Explaining top-5 predictions results as opposed to top-1;
|
438
446
|
# - Increasing path count or step count of the attribution methods to reduce
|
439
447
|
# approximate errors;
|
440
448
|
# - Using different baselines for explaining the prediction results.
|
441
449
|
# @param deployed_model_id [::String]
|
442
450
|
# If specified, this ExplainRequest will be served by the chosen
|
443
|
-
# DeployedModel, overriding
|
451
|
+
# DeployedModel, overriding
|
452
|
+
# {::Google::Cloud::AIPlatform::V1::Endpoint#traffic_split Endpoint.traffic_split}.
|
444
453
|
#
|
445
454
|
# @yield [response, operation] Access the result along with the RPC operation
|
446
455
|
# @yieldparam response [::Google::Cloud::AIPlatform::V1::ExplainResponse]
|
@@ -33,7 +33,7 @@ module Google
|
|
33
33
|
##
|
34
34
|
# A service for online predictions and explanations.
|
35
35
|
#
|
36
|
-
#
|
36
|
+
# @example Load this service and instantiate a gRPC client
|
37
37
|
#
|
38
38
|
# require "google/cloud/ai_platform/v1/prediction_service"
|
39
39
|
# client = ::Google::Cloud::AIPlatform::V1::PredictionService::Client.new
|
@@ -398,9 +398,11 @@ module Google
|
|
398
398
|
# The standard list page size.
|
399
399
|
# @param page_token [::String]
|
400
400
|
# The standard list page token.
|
401
|
-
# Typically obtained by
|
402
|
-
#
|
403
|
-
#
|
401
|
+
# Typically obtained by
|
402
|
+
# {::Google::Cloud::AIPlatform::V1::ListSpecialistPoolsResponse#next_page_token ListSpecialistPoolsResponse.next_page_token}
|
403
|
+
# of the previous
|
404
|
+
# {::Google::Cloud::AIPlatform::V1::SpecialistPoolService::Client#list_specialist_pools SpecialistPoolService.ListSpecialistPools}
|
405
|
+
# call. Return first page if empty.
|
404
406
|
# @param read_mask [::Google::Protobuf::FieldMask, ::Hash]
|
405
407
|
# Mask specifying which fields to read. FieldMask represents a set of
|
406
408
|
#
|
@@ -39,7 +39,7 @@ module Google
|
|
39
39
|
# then Managers will get email notifications to manage Specialists and tasks on
|
40
40
|
# CrowdCompute console.
|
41
41
|
#
|
42
|
-
#
|
42
|
+
# @example Load this service and instantiate a gRPC client
|
43
43
|
#
|
44
44
|
# require "google/cloud/ai_platform/v1/specialist_pool_service"
|
45
45
|
# client = ::Google::Cloud::AIPlatform::V1::SpecialistPoolService::Client.new
|