google-cloud-ai_platform-v1 0.1.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/.yardopts +12 -0
- data/AUTHENTICATION.md +149 -0
- data/LICENSE.md +201 -0
- data/README.md +139 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/client.rb +1364 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/paths.rb +111 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service.rb +51 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/client.rb +1076 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/paths.rb +124 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/client.rb +508 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/paths.rb +54 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service.rb +49 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/client.rb +2707 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/paths.rb +113 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/client.rb +1146 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/paths.rb +88 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/index_service/client.rb +823 -0
- data/lib/google/cloud/ai_platform/v1/index_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/index_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/index_service/paths.rb +88 -0
- data/lib/google/cloud/ai_platform/v1/index_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/job_service/client.rb +3236 -0
- data/lib/google/cloud/ai_platform/v1/job_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/job_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/job_service/paths.rb +259 -0
- data/lib/google/cloud/ai_platform/v1/job_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/client.rb +3654 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/paths.rb +153 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/client.rb +538 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/paths.rb +148 -0
- data/lib/google/cloud/ai_platform/v1/migration_service.rb +51 -0
- data/lib/google/cloud/ai_platform/v1/model_service/client.rb +1355 -0
- data/lib/google/cloud/ai_platform/v1/model_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/model_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/model_service/paths.rb +151 -0
- data/lib/google/cloud/ai_platform/v1/model_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/client.rb +1384 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/paths.rb +225 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service.rb +52 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service/client.rb +650 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service/paths.rb +52 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service.rb +49 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/client.rb +826 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/paths.rb +69 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service.rb +55 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/client.rb +3224 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/credentials.rb +48 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/paths.rb +138 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/version.rb +28 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/client.rb +1793 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/paths.rb +109 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service.rb +54 -0
- data/lib/google/cloud/ai_platform/v1.rb +52 -0
- data/lib/google/cloud/aiplatform/v1/accelerator_type_pb.rb +29 -0
- data/lib/google/cloud/aiplatform/v1/annotation_pb.rb +35 -0
- data/lib/google/cloud/aiplatform/v1/annotation_spec_pb.rb +30 -0
- data/lib/google/cloud/aiplatform/v1/artifact_pb.rb +45 -0
- data/lib/google/cloud/aiplatform/v1/batch_prediction_job_pb.rb +83 -0
- data/lib/google/cloud/aiplatform/v1/completion_stats_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/context_pb.rb +38 -0
- data/lib/google/cloud/aiplatform/v1/custom_job_pb.rb +84 -0
- data/lib/google/cloud/aiplatform/v1/data_item_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/data_labeling_job_pb.rb +78 -0
- data/lib/google/cloud/aiplatform/v1/dataset_pb.rb +53 -0
- data/lib/google/cloud/aiplatform/v1/dataset_service_pb.rb +126 -0
- data/lib/google/cloud/aiplatform/v1/dataset_service_services_pb.rb +64 -0
- data/lib/google/cloud/aiplatform/v1/deployed_index_ref_pb.rb +26 -0
- data/lib/google/cloud/aiplatform/v1/deployed_model_ref_pb.rb +26 -0
- data/lib/google/cloud/aiplatform/v1/encryption_spec_pb.rb +24 -0
- data/lib/google/cloud/aiplatform/v1/endpoint_pb.rb +67 -0
- data/lib/google/cloud/aiplatform/v1/endpoint_service_pb.rb +90 -0
- data/lib/google/cloud/aiplatform/v1/endpoint_service_services_pb.rb +58 -0
- data/lib/google/cloud/aiplatform/v1/entity_type_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/env_var_pb.rb +25 -0
- data/lib/google/cloud/aiplatform/v1/event_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/execution_pb.rb +48 -0
- data/lib/google/cloud/aiplatform/v1/explanation_metadata_pb.rb +107 -0
- data/lib/google/cloud/aiplatform/v1/explanation_pb.rb +106 -0
- data/lib/google/cloud/aiplatform/v1/feature_monitoring_stats_pb.rb +30 -0
- data/lib/google/cloud/aiplatform/v1/feature_pb.rb +46 -0
- data/lib/google/cloud/aiplatform/v1/feature_selector_pb.rb +28 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_online_service_pb.rb +86 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_online_service_services_pb.rb +51 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_pb.rb +44 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_service_pb.rb +280 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_service_services_pb.rb +109 -0
- data/lib/google/cloud/aiplatform/v1/hyperparameter_tuning_job_pb.rb +46 -0
- data/lib/google/cloud/aiplatform/v1/index_endpoint_pb.rb +66 -0
- data/lib/google/cloud/aiplatform/v1/index_endpoint_service_pb.rb +101 -0
- data/lib/google/cloud/aiplatform/v1/index_endpoint_service_services_pb.rb +62 -0
- data/lib/google/cloud/aiplatform/v1/index_pb.rb +38 -0
- data/lib/google/cloud/aiplatform/v1/index_service_pb.rb +98 -0
- data/lib/google/cloud/aiplatform/v1/index_service_services_pb.rb +55 -0
- data/lib/google/cloud/aiplatform/v1/io_pb.rb +56 -0
- data/lib/google/cloud/aiplatform/v1/job_service_pb.rb +217 -0
- data/lib/google/cloud/aiplatform/v1/job_service_services_pb.rb +134 -0
- data/lib/google/cloud/aiplatform/v1/job_state_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/lineage_subgraph_pb.rb +28 -0
- data/lib/google/cloud/aiplatform/v1/machine_resources_pb.rb +59 -0
- data/lib/google/cloud/aiplatform/v1/manual_batch_tuning_parameters_pb.rb +24 -0
- data/lib/google/cloud/aiplatform/v1/metadata_schema_pb.rb +38 -0
- data/lib/google/cloud/aiplatform/v1/metadata_service_pb.rb +272 -0
- data/lib/google/cloud/aiplatform/v1/metadata_service_services_pb.rb +119 -0
- data/lib/google/cloud/aiplatform/v1/metadata_store_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/migratable_resource_pb.rb +59 -0
- data/lib/google/cloud/aiplatform/v1/migration_service_pb.rb +106 -0
- data/lib/google/cloud/aiplatform/v1/migration_service_services_pb.rb +51 -0
- data/lib/google/cloud/aiplatform/v1/model_deployment_monitoring_job_pb.rb +111 -0
- data/lib/google/cloud/aiplatform/v1/model_evaluation_pb.rb +33 -0
- data/lib/google/cloud/aiplatform/v1/model_evaluation_slice_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/model_monitoring_pb.rb +93 -0
- data/lib/google/cloud/aiplatform/v1/model_pb.rb +88 -0
- data/lib/google/cloud/aiplatform/v1/model_service_pb.rb +129 -0
- data/lib/google/cloud/aiplatform/v1/model_service_services_pb.rb +69 -0
- data/lib/google/cloud/aiplatform/v1/operation_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_job_pb.rb +115 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_service_pb.rb +88 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_service_services_pb.rb +84 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_state_pb.rb +31 -0
- data/lib/google/cloud/aiplatform/v1/prediction_service_pb.rb +57 -0
- data/lib/google/cloud/aiplatform/v1/prediction_service_services_pb.rb +66 -0
- data/lib/google/cloud/aiplatform/v1/specialist_pool_pb.rb +30 -0
- data/lib/google/cloud/aiplatform/v1/specialist_pool_service_pb.rb +66 -0
- data/lib/google/cloud/aiplatform/v1/specialist_pool_service_services_pb.rb +58 -0
- data/lib/google/cloud/aiplatform/v1/study_pb.rb +191 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_data_pb.rb +56 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_experiment_pb.rb +33 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_run_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_pb.rb +244 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_services_pb.rb +115 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_time_series_pb.rb +48 -0
- data/lib/google/cloud/aiplatform/v1/training_pipeline_pb.rb +95 -0
- data/lib/google/cloud/aiplatform/v1/types_pb.rb +35 -0
- data/lib/google/cloud/aiplatform/v1/unmanaged_container_model_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/user_action_reference_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/value_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/vizier_service_pb.rb +136 -0
- data/lib/google/cloud/aiplatform/v1/vizier_service_services_pb.rb +90 -0
- data/lib/google-cloud-ai_platform-v1.rb +21 -0
- data/proto_docs/README.md +4 -0
- data/proto_docs/google/api/field_behavior.rb +71 -0
- data/proto_docs/google/api/httpbody.rb +80 -0
- data/proto_docs/google/api/resource.rb +222 -0
- data/proto_docs/google/cloud/aiplatform/v1/accelerator_type.rb +50 -0
- data/proto_docs/google/cloud/aiplatform/v1/annotation.rb +92 -0
- data/proto_docs/google/cloud/aiplatform/v1/annotation_spec.rb +50 -0
- data/proto_docs/google/cloud/aiplatform/v1/artifact.rb +112 -0
- data/proto_docs/google/cloud/aiplatform/v1/batch_prediction_job.rb +278 -0
- data/proto_docs/google/cloud/aiplatform/v1/completion_stats.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/context.rb +92 -0
- data/proto_docs/google/cloud/aiplatform/v1/custom_job.rb +272 -0
- data/proto_docs/google/cloud/aiplatform/v1/data_item.rb +73 -0
- data/proto_docs/google/cloud/aiplatform/v1/data_labeling_job.rb +207 -0
- data/proto_docs/google/cloud/aiplatform/v1/dataset.rb +154 -0
- data/proto_docs/google/cloud/aiplatform/v1/dataset_service.rb +301 -0
- data/proto_docs/google/cloud/aiplatform/v1/deployed_index_ref.rb +38 -0
- data/proto_docs/google/cloud/aiplatform/v1/deployed_model_ref.rb +38 -0
- data/proto_docs/google/cloud/aiplatform/v1/encryption_spec.rb +40 -0
- data/proto_docs/google/cloud/aiplatform/v1/endpoint.rb +227 -0
- data/proto_docs/google/cloud/aiplatform/v1/endpoint_service.rb +258 -0
- data/proto_docs/google/cloud/aiplatform/v1/entity_type.rb +79 -0
- data/proto_docs/google/cloud/aiplatform/v1/env_var.rb +44 -0
- data/proto_docs/google/cloud/aiplatform/v1/event.rb +79 -0
- data/proto_docs/google/cloud/aiplatform/v1/execution.rb +118 -0
- data/proto_docs/google/cloud/aiplatform/v1/explanation.rb +445 -0
- data/proto_docs/google/cloud/aiplatform/v1/explanation_metadata.rb +419 -0
- data/proto_docs/google/cloud/aiplatform/v1/feature.rb +115 -0
- data/proto_docs/google/cloud/aiplatform/v1/feature_monitoring_stats.rb +88 -0
- data/proto_docs/google/cloud/aiplatform/v1/feature_selector.rb +49 -0
- data/proto_docs/google/cloud/aiplatform/v1/featurestore.rb +115 -0
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_online_service.rb +203 -0
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_service.rb +978 -0
- data/proto_docs/google/cloud/aiplatform/v1/hyperparameter_tuning_job.rb +109 -0
- data/proto_docs/google/cloud/aiplatform/v1/index.rb +98 -0
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint.rb +252 -0
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint_service.rb +240 -0
- data/proto_docs/google/cloud/aiplatform/v1/index_service.rb +220 -0
- data/proto_docs/google/cloud/aiplatform/v1/io.rb +134 -0
- data/proto_docs/google/cloud/aiplatform/v1/job_service.rb +660 -0
- data/proto_docs/google/cloud/aiplatform/v1/job_state.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/lineage_subgraph.rb +42 -0
- data/proto_docs/google/cloud/aiplatform/v1/machine_resources.rb +194 -0
- data/proto_docs/google/cloud/aiplatform/v1/manual_batch_tuning_parameters.rb +41 -0
- data/proto_docs/google/cloud/aiplatform/v1/metadata_schema.rb +74 -0
- data/proto_docs/google/cloud/aiplatform/v1/metadata_service.rb +912 -0
- data/proto_docs/google/cloud/aiplatform/v1/metadata_store.rb +62 -0
- data/proto_docs/google/cloud/aiplatform/v1/migratable_resource.rb +133 -0
- data/proto_docs/google/cloud/aiplatform/v1/migration_service.rb +260 -0
- data/proto_docs/google/cloud/aiplatform/v1/model.rb +562 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_deployment_monitoring_job.rb +293 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation_slice.rb +68 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_monitoring.rb +257 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_service.rb +329 -0
- data/proto_docs/google/cloud/aiplatform/v1/operation.rb +55 -0
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_job.rb +347 -0
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_service.rb +258 -0
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_state.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/prediction_service.rb +165 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/image_classification.rb +52 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/image_object_detection.rb +52 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/image_segmentation.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/text_classification.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/text_extraction.rb +53 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/text_sentiment.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/video_action_recognition.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/video_classification.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/video_object_tracking.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/image_classification.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/image_object_detection.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/image_segmentation.rb +44 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/video_action_recognition.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/video_classification.rb +72 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/video_object_tracking.rb +51 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/classification.rb +49 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/image_object_detection.rb +58 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/image_segmentation.rb +53 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/tabular_classification.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/tabular_regression.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/text_extraction.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/text_sentiment.rb +45 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_action_recognition.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_classification.rb +73 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_object_tracking.rb +91 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_image_classification.rb +142 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_image_object_detection.rb +134 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_image_segmentation.rb +120 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_tables.rb +315 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_text_classification.rb +48 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_text_extraction.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_text_sentiment.rb +55 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_video_action_recognition.rb +73 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_video_classification.rb +67 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_video_object_tracking.rb +78 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/export_evaluated_data_items_config.rb +51 -0
- data/proto_docs/google/cloud/aiplatform/v1/specialist_pool.rb +58 -0
- data/proto_docs/google/cloud/aiplatform/v1/specialist_pool_service.rb +136 -0
- data/proto_docs/google/cloud/aiplatform/v1/study.rb +543 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard.rb +89 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_data.rb +110 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_experiment.rb +82 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_run.rb +85 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_service.rb +706 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_time_series.rb +101 -0
- data/proto_docs/google/cloud/aiplatform/v1/training_pipeline.rb +381 -0
- data/proto_docs/google/cloud/aiplatform/v1/types.rb +62 -0
- data/proto_docs/google/cloud/aiplatform/v1/unmanaged_container_model.rb +44 -0
- data/proto_docs/google/cloud/aiplatform/v1/user_action_reference.rb +49 -0
- data/proto_docs/google/cloud/aiplatform/v1/value.rb +41 -0
- data/proto_docs/google/cloud/aiplatform/v1/vizier_service.rb +332 -0
- data/proto_docs/google/longrunning/operations.rb +164 -0
- data/proto_docs/google/protobuf/any.rb +141 -0
- data/proto_docs/google/protobuf/duration.rb +98 -0
- data/proto_docs/google/protobuf/empty.rb +36 -0
- data/proto_docs/google/protobuf/field_mask.rb +229 -0
- data/proto_docs/google/protobuf/struct.rb +96 -0
- data/proto_docs/google/protobuf/timestamp.rb +129 -0
- data/proto_docs/google/protobuf/wrappers.rb +121 -0
- data/proto_docs/google/rpc/status.rb +46 -0
- data/proto_docs/google/type/money.rb +43 -0
- metadata +479 -0
data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_action_recognition.rb
ADDED
@@ -0,0 +1,60 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2022 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module AIPlatform
|
23
|
+
module V1
|
24
|
+
module Schema
|
25
|
+
module Predict
|
26
|
+
module Prediction
|
27
|
+
# Prediction output format for Video Action Recognition.
|
28
|
+
# @!attribute [rw] id
|
29
|
+
# @return [::String]
|
30
|
+
# The resource ID of the AnnotationSpec that had been identified.
|
31
|
+
# @!attribute [rw] display_name
|
32
|
+
# @return [::String]
|
33
|
+
# The display name of the AnnotationSpec that had been identified.
|
34
|
+
# @!attribute [rw] time_segment_start
|
35
|
+
# @return [::Google::Protobuf::Duration]
|
36
|
+
# The beginning, inclusive, of the video's time segment in which the
|
37
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
38
|
+
# measured from the start of the video, with fractions up to a microsecond
|
39
|
+
# precision, and with "s" appended at the end.
|
40
|
+
# @!attribute [rw] time_segment_end
|
41
|
+
# @return [::Google::Protobuf::Duration]
|
42
|
+
# The end, exclusive, of the video's time segment in which the
|
43
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
44
|
+
# measured from the start of the video, with fractions up to a microsecond
|
45
|
+
# precision, and with "s" appended at the end.
|
46
|
+
# @!attribute [rw] confidence
|
47
|
+
# @return [::Google::Protobuf::FloatValue]
|
48
|
+
# The Model's confidence in correction of this prediction, higher
|
49
|
+
# value means higher confidence.
|
50
|
+
class VideoActionRecognitionPredictionResult
|
51
|
+
include ::Google::Protobuf::MessageExts
|
52
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
53
|
+
end
|
54
|
+
end
|
55
|
+
end
|
56
|
+
end
|
57
|
+
end
|
58
|
+
end
|
59
|
+
end
|
60
|
+
end
|
@@ -0,0 +1,73 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2022 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module AIPlatform
|
23
|
+
module V1
|
24
|
+
module Schema
|
25
|
+
module Predict
|
26
|
+
module Prediction
|
27
|
+
# Prediction output format for Video Classification.
|
28
|
+
# @!attribute [rw] id
|
29
|
+
# @return [::String]
|
30
|
+
# The resource ID of the AnnotationSpec that had been identified.
|
31
|
+
# @!attribute [rw] display_name
|
32
|
+
# @return [::String]
|
33
|
+
# The display name of the AnnotationSpec that had been identified.
|
34
|
+
# @!attribute [rw] type
|
35
|
+
# @return [::String]
|
36
|
+
# The type of the prediction. The requested types can be configured
|
37
|
+
# via parameters. This will be one of
|
38
|
+
# - segment-classification
|
39
|
+
# - shot-classification
|
40
|
+
# - one-sec-interval-classification
|
41
|
+
# @!attribute [rw] time_segment_start
|
42
|
+
# @return [::Google::Protobuf::Duration]
|
43
|
+
# The beginning, inclusive, of the video's time segment in which the
|
44
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
45
|
+
# measured from the start of the video, with fractions up to a microsecond
|
46
|
+
# precision, and with "s" appended at the end. Note that for
|
47
|
+
# 'segment-classification' prediction type, this equals the original
|
48
|
+
# 'timeSegmentStart' from the input instance, for other types it is the
|
49
|
+
# start of a shot or a 1 second interval respectively.
|
50
|
+
# @!attribute [rw] time_segment_end
|
51
|
+
# @return [::Google::Protobuf::Duration]
|
52
|
+
# The end, exclusive, of the video's time segment in which the
|
53
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
54
|
+
# measured from the start of the video, with fractions up to a microsecond
|
55
|
+
# precision, and with "s" appended at the end. Note that for
|
56
|
+
# 'segment-classification' prediction type, this equals the original
|
57
|
+
# 'timeSegmentEnd' from the input instance, for other types it is the end
|
58
|
+
# of a shot or a 1 second interval respectively.
|
59
|
+
# @!attribute [rw] confidence
|
60
|
+
# @return [::Google::Protobuf::FloatValue]
|
61
|
+
# The Model's confidence in correction of this prediction, higher
|
62
|
+
# value means higher confidence.
|
63
|
+
class VideoClassificationPredictionResult
|
64
|
+
include ::Google::Protobuf::MessageExts
|
65
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
66
|
+
end
|
67
|
+
end
|
68
|
+
end
|
69
|
+
end
|
70
|
+
end
|
71
|
+
end
|
72
|
+
end
|
73
|
+
end
|
@@ -0,0 +1,91 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2022 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module AIPlatform
|
23
|
+
module V1
|
24
|
+
module Schema
|
25
|
+
module Predict
|
26
|
+
module Prediction
|
27
|
+
# Prediction output format for Video Object Tracking.
|
28
|
+
# @!attribute [rw] id
|
29
|
+
# @return [::String]
|
30
|
+
# The resource ID of the AnnotationSpec that had been identified.
|
31
|
+
# @!attribute [rw] display_name
|
32
|
+
# @return [::String]
|
33
|
+
# The display name of the AnnotationSpec that had been identified.
|
34
|
+
# @!attribute [rw] time_segment_start
|
35
|
+
# @return [::Google::Protobuf::Duration]
|
36
|
+
# The beginning, inclusive, of the video's time segment in which the
|
37
|
+
# object instance has been detected. Expressed as a number of seconds as
|
38
|
+
# measured from the start of the video, with fractions up to a microsecond
|
39
|
+
# precision, and with "s" appended at the end.
|
40
|
+
# @!attribute [rw] time_segment_end
|
41
|
+
# @return [::Google::Protobuf::Duration]
|
42
|
+
# The end, inclusive, of the video's time segment in which the
|
43
|
+
# object instance has been detected. Expressed as a number of seconds as
|
44
|
+
# measured from the start of the video, with fractions up to a microsecond
|
45
|
+
# precision, and with "s" appended at the end.
|
46
|
+
# @!attribute [rw] confidence
|
47
|
+
# @return [::Google::Protobuf::FloatValue]
|
48
|
+
# The Model's confidence in correction of this prediction, higher
|
49
|
+
# value means higher confidence.
|
50
|
+
# @!attribute [rw] frames
|
51
|
+
# @return [::Array<::Google::Cloud::AIPlatform::V1::Schema::Predict::Prediction::VideoObjectTrackingPredictionResult::Frame>]
|
52
|
+
# All of the frames of the video in which a single object instance has been
|
53
|
+
# detected. The bounding boxes in the frames identify the same object.
|
54
|
+
class VideoObjectTrackingPredictionResult
|
55
|
+
include ::Google::Protobuf::MessageExts
|
56
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
57
|
+
|
58
|
+
# The fields `xMin`, `xMax`, `yMin`, and `yMax` refer to a bounding box,
|
59
|
+
# i.e. the rectangle over the video frame pinpointing the found
|
60
|
+
# AnnotationSpec. The coordinates are relative to the frame size, and the
|
61
|
+
# point 0,0 is in the top left of the frame.
|
62
|
+
# @!attribute [rw] time_offset
|
63
|
+
# @return [::Google::Protobuf::Duration]
|
64
|
+
# A time (frame) of a video in which the object has been detected.
|
65
|
+
# Expressed as a number of seconds as measured from the
|
66
|
+
# start of the video, with fractions up to a microsecond precision, and
|
67
|
+
# with "s" appended at the end.
|
68
|
+
# @!attribute [rw] x_min
|
69
|
+
# @return [::Google::Protobuf::FloatValue]
|
70
|
+
# The leftmost coordinate of the bounding box.
|
71
|
+
# @!attribute [rw] x_max
|
72
|
+
# @return [::Google::Protobuf::FloatValue]
|
73
|
+
# The rightmost coordinate of the bounding box.
|
74
|
+
# @!attribute [rw] y_min
|
75
|
+
# @return [::Google::Protobuf::FloatValue]
|
76
|
+
# The topmost coordinate of the bounding box.
|
77
|
+
# @!attribute [rw] y_max
|
78
|
+
# @return [::Google::Protobuf::FloatValue]
|
79
|
+
# The bottommost coordinate of the bounding box.
|
80
|
+
class Frame
|
81
|
+
include ::Google::Protobuf::MessageExts
|
82
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
83
|
+
end
|
84
|
+
end
|
85
|
+
end
|
86
|
+
end
|
87
|
+
end
|
88
|
+
end
|
89
|
+
end
|
90
|
+
end
|
91
|
+
end
|
@@ -0,0 +1,142 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2022 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module AIPlatform
|
23
|
+
module V1
|
24
|
+
module Schema
|
25
|
+
module TrainingJob
|
26
|
+
module Definition
|
27
|
+
# A TrainingJob that trains and uploads an AutoML Image Classification Model.
|
28
|
+
# @!attribute [rw] inputs
|
29
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationInputs]
|
30
|
+
# The input parameters of this TrainingJob.
|
31
|
+
# @!attribute [rw] metadata
|
32
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationMetadata]
|
33
|
+
# The metadata information.
|
34
|
+
class AutoMlImageClassification
|
35
|
+
include ::Google::Protobuf::MessageExts
|
36
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
37
|
+
end
|
38
|
+
|
39
|
+
# @!attribute [rw] model_type
|
40
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationInputs::ModelType]
|
41
|
+
# @!attribute [rw] base_model_id
|
42
|
+
# @return [::String]
|
43
|
+
# The ID of the `base` model. If it is specified, the new model will be
|
44
|
+
# trained based on the `base` model. Otherwise, the new model will be
|
45
|
+
# trained from scratch. The `base` model must be in the same
|
46
|
+
# Project and Location as the new Model to train, and have the same
|
47
|
+
# modelType.
|
48
|
+
# @!attribute [rw] budget_milli_node_hours
|
49
|
+
# @return [::Integer]
|
50
|
+
# The training budget of creating this model, expressed in milli node
|
51
|
+
# hours i.e. 1,000 value in this field means 1 node hour. The actual
|
52
|
+
# metadata.costMilliNodeHours will be equal or less than this value.
|
53
|
+
# If further model training ceases to provide any improvements, it will
|
54
|
+
# stop without using the full budget and the metadata.successfulStopReason
|
55
|
+
# will be `model-converged`.
|
56
|
+
# Note, node_hour = actual_hour * number_of_nodes_involved.
|
57
|
+
# For modelType `cloud`(default), the budget must be between 8,000
|
58
|
+
# and 800,000 milli node hours, inclusive. The default value is 192,000
|
59
|
+
# which represents one day in wall time, considering 8 nodes are used.
|
60
|
+
# For model types `mobile-tf-low-latency-1`, `mobile-tf-versatile-1`,
|
61
|
+
# `mobile-tf-high-accuracy-1`, the training budget must be between
|
62
|
+
# 1,000 and 100,000 milli node hours, inclusive.
|
63
|
+
# The default value is 24,000 which represents one day in wall time on a
|
64
|
+
# single node that is used.
|
65
|
+
# @!attribute [rw] disable_early_stopping
|
66
|
+
# @return [::Boolean]
|
67
|
+
# Use the entire training budget. This disables the early stopping feature.
|
68
|
+
# When false the early stopping feature is enabled, which means that
|
69
|
+
# AutoML Image Classification might stop training before the entire
|
70
|
+
# training budget has been used.
|
71
|
+
# @!attribute [rw] multi_label
|
72
|
+
# @return [::Boolean]
|
73
|
+
# If false, a single-label (multi-class) Model will be trained (i.e.
|
74
|
+
# assuming that for each image just up to one annotation may be
|
75
|
+
# applicable). If true, a multi-label Model will be trained (i.e.
|
76
|
+
# assuming that for each image multiple annotations may be applicable).
|
77
|
+
class AutoMlImageClassificationInputs
|
78
|
+
include ::Google::Protobuf::MessageExts
|
79
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
80
|
+
|
81
|
+
module ModelType
|
82
|
+
# Should not be set.
|
83
|
+
MODEL_TYPE_UNSPECIFIED = 0
|
84
|
+
|
85
|
+
# A Model best tailored to be used within Google Cloud, and which cannot
|
86
|
+
# be exported.
|
87
|
+
# Default.
|
88
|
+
CLOUD = 1
|
89
|
+
|
90
|
+
# A model that, in addition to being available within Google
|
91
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
92
|
+
# or Core ML model and used on a mobile or edge device afterwards.
|
93
|
+
# Expected to have low latency, but may have lower prediction
|
94
|
+
# quality than other mobile models.
|
95
|
+
MOBILE_TF_LOW_LATENCY_1 = 2
|
96
|
+
|
97
|
+
# A model that, in addition to being available within Google
|
98
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
99
|
+
# or Core ML model and used on a mobile or edge device with afterwards.
|
100
|
+
MOBILE_TF_VERSATILE_1 = 3
|
101
|
+
|
102
|
+
# A model that, in addition to being available within Google
|
103
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
104
|
+
# or Core ML model and used on a mobile or edge device afterwards.
|
105
|
+
# Expected to have a higher latency, but should also have a higher
|
106
|
+
# prediction quality than other mobile models.
|
107
|
+
MOBILE_TF_HIGH_ACCURACY_1 = 4
|
108
|
+
end
|
109
|
+
end
|
110
|
+
|
111
|
+
# @!attribute [rw] cost_milli_node_hours
|
112
|
+
# @return [::Integer]
|
113
|
+
# The actual training cost of creating this model, expressed in
|
114
|
+
# milli node hours, i.e. 1,000 value in this field means 1 node hour.
|
115
|
+
# Guaranteed to not exceed inputs.budgetMilliNodeHours.
|
116
|
+
# @!attribute [rw] successful_stop_reason
|
117
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationMetadata::SuccessfulStopReason]
|
118
|
+
# For successful job completions, this is the reason why the job has
|
119
|
+
# finished.
|
120
|
+
class AutoMlImageClassificationMetadata
|
121
|
+
include ::Google::Protobuf::MessageExts
|
122
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
123
|
+
|
124
|
+
module SuccessfulStopReason
|
125
|
+
# Should not be set.
|
126
|
+
SUCCESSFUL_STOP_REASON_UNSPECIFIED = 0
|
127
|
+
|
128
|
+
# The inputs.budgetMilliNodeHours had been reached.
|
129
|
+
BUDGET_REACHED = 1
|
130
|
+
|
131
|
+
# Further training of the Model ceased to increase its quality, since it
|
132
|
+
# already has converged.
|
133
|
+
MODEL_CONVERGED = 2
|
134
|
+
end
|
135
|
+
end
|
136
|
+
end
|
137
|
+
end
|
138
|
+
end
|
139
|
+
end
|
140
|
+
end
|
141
|
+
end
|
142
|
+
end
|
@@ -0,0 +1,134 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2022 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module AIPlatform
|
23
|
+
module V1
|
24
|
+
module Schema
|
25
|
+
module TrainingJob
|
26
|
+
module Definition
|
27
|
+
# A TrainingJob that trains and uploads an AutoML Image Object Detection Model.
|
28
|
+
# @!attribute [rw] inputs
|
29
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionInputs]
|
30
|
+
# The input parameters of this TrainingJob.
|
31
|
+
# @!attribute [rw] metadata
|
32
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionMetadata]
|
33
|
+
# The metadata information
|
34
|
+
class AutoMlImageObjectDetection
|
35
|
+
include ::Google::Protobuf::MessageExts
|
36
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
37
|
+
end
|
38
|
+
|
39
|
+
# @!attribute [rw] model_type
|
40
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionInputs::ModelType]
|
41
|
+
# @!attribute [rw] budget_milli_node_hours
|
42
|
+
# @return [::Integer]
|
43
|
+
# The training budget of creating this model, expressed in milli node
|
44
|
+
# hours i.e. 1,000 value in this field means 1 node hour. The actual
|
45
|
+
# metadata.costMilliNodeHours will be equal or less than this value.
|
46
|
+
# If further model training ceases to provide any improvements, it will
|
47
|
+
# stop without using the full budget and the metadata.successfulStopReason
|
48
|
+
# will be `model-converged`.
|
49
|
+
# Note, node_hour = actual_hour * number_of_nodes_involved.
|
50
|
+
# For modelType `cloud`(default), the budget must be between 20,000
|
51
|
+
# and 900,000 milli node hours, inclusive. The default value is 216,000
|
52
|
+
# which represents one day in wall time, considering 9 nodes are used.
|
53
|
+
# For model types `mobile-tf-low-latency-1`, `mobile-tf-versatile-1`,
|
54
|
+
# `mobile-tf-high-accuracy-1`
|
55
|
+
# the training budget must be between 1,000 and 100,000 milli node hours,
|
56
|
+
# inclusive. The default value is 24,000 which represents one day in
|
57
|
+
# wall time on a single node that is used.
|
58
|
+
# @!attribute [rw] disable_early_stopping
|
59
|
+
# @return [::Boolean]
|
60
|
+
# Use the entire training budget. This disables the early stopping feature.
|
61
|
+
# When false the early stopping feature is enabled, which means that AutoML
|
62
|
+
# Image Object Detection might stop training before the entire training
|
63
|
+
# budget has been used.
|
64
|
+
class AutoMlImageObjectDetectionInputs
|
65
|
+
include ::Google::Protobuf::MessageExts
|
66
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
67
|
+
|
68
|
+
module ModelType
|
69
|
+
# Should not be set.
|
70
|
+
MODEL_TYPE_UNSPECIFIED = 0
|
71
|
+
|
72
|
+
# A model best tailored to be used within Google Cloud, and which cannot
|
73
|
+
# be exported. Expected to have a higher latency, but should also have a
|
74
|
+
# higher prediction quality than other cloud models.
|
75
|
+
CLOUD_HIGH_ACCURACY_1 = 1
|
76
|
+
|
77
|
+
# A model best tailored to be used within Google Cloud, and which cannot
|
78
|
+
# be exported. Expected to have a low latency, but may have lower
|
79
|
+
# prediction quality than other cloud models.
|
80
|
+
CLOUD_LOW_LATENCY_1 = 2
|
81
|
+
|
82
|
+
# A model that, in addition to being available within Google
|
83
|
+
# Cloud can also be exported (see ModelService.ExportModel) and
|
84
|
+
# used on a mobile or edge device with TensorFlow afterwards.
|
85
|
+
# Expected to have low latency, but may have lower prediction
|
86
|
+
# quality than other mobile models.
|
87
|
+
MOBILE_TF_LOW_LATENCY_1 = 3
|
88
|
+
|
89
|
+
# A model that, in addition to being available within Google
|
90
|
+
# Cloud can also be exported (see ModelService.ExportModel) and
|
91
|
+
# used on a mobile or edge device with TensorFlow afterwards.
|
92
|
+
MOBILE_TF_VERSATILE_1 = 4
|
93
|
+
|
94
|
+
# A model that, in addition to being available within Google
|
95
|
+
# Cloud, can also be exported (see ModelService.ExportModel) and
|
96
|
+
# used on a mobile or edge device with TensorFlow afterwards.
|
97
|
+
# Expected to have a higher latency, but should also have a higher
|
98
|
+
# prediction quality than other mobile models.
|
99
|
+
MOBILE_TF_HIGH_ACCURACY_1 = 5
|
100
|
+
end
|
101
|
+
end
|
102
|
+
|
103
|
+
# @!attribute [rw] cost_milli_node_hours
|
104
|
+
# @return [::Integer]
|
105
|
+
# The actual training cost of creating this model, expressed in
|
106
|
+
# milli node hours, i.e. 1,000 value in this field means 1 node hour.
|
107
|
+
# Guaranteed to not exceed inputs.budgetMilliNodeHours.
|
108
|
+
# @!attribute [rw] successful_stop_reason
|
109
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionMetadata::SuccessfulStopReason]
|
110
|
+
# For successful job completions, this is the reason why the job has
|
111
|
+
# finished.
|
112
|
+
class AutoMlImageObjectDetectionMetadata
|
113
|
+
include ::Google::Protobuf::MessageExts
|
114
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
115
|
+
|
116
|
+
module SuccessfulStopReason
|
117
|
+
# Should not be set.
|
118
|
+
SUCCESSFUL_STOP_REASON_UNSPECIFIED = 0
|
119
|
+
|
120
|
+
# The inputs.budgetMilliNodeHours had been reached.
|
121
|
+
BUDGET_REACHED = 1
|
122
|
+
|
123
|
+
# Further training of the Model ceased to increase its quality, since it
|
124
|
+
# already has converged.
|
125
|
+
MODEL_CONVERGED = 2
|
126
|
+
end
|
127
|
+
end
|
128
|
+
end
|
129
|
+
end
|
130
|
+
end
|
131
|
+
end
|
132
|
+
end
|
133
|
+
end
|
134
|
+
end
|
@@ -0,0 +1,120 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2022 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module AIPlatform
|
23
|
+
module V1
|
24
|
+
module Schema
|
25
|
+
module TrainingJob
|
26
|
+
module Definition
|
27
|
+
# A TrainingJob that trains and uploads an AutoML Image Segmentation Model.
|
28
|
+
# @!attribute [rw] inputs
|
29
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationInputs]
|
30
|
+
# The input parameters of this TrainingJob.
|
31
|
+
# @!attribute [rw] metadata
|
32
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationMetadata]
|
33
|
+
# The metadata information.
|
34
|
+
class AutoMlImageSegmentation
|
35
|
+
include ::Google::Protobuf::MessageExts
|
36
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
37
|
+
end
|
38
|
+
|
39
|
+
# @!attribute [rw] model_type
|
40
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationInputs::ModelType]
|
41
|
+
# @!attribute [rw] budget_milli_node_hours
|
42
|
+
# @return [::Integer]
|
43
|
+
# The training budget of creating this model, expressed in milli node
|
44
|
+
# hours i.e. 1,000 value in this field means 1 node hour. The actual
|
45
|
+
# metadata.costMilliNodeHours will be equal or less than this value.
|
46
|
+
# If further model training ceases to provide any improvements, it will
|
47
|
+
# stop without using the full budget and the metadata.successfulStopReason
|
48
|
+
# will be `model-converged`.
|
49
|
+
# Note, node_hour = actual_hour * number_of_nodes_involved. Or
|
50
|
+
# actaul_wall_clock_hours = train_budget_milli_node_hours /
|
51
|
+
# (number_of_nodes_involved * 1000)
|
52
|
+
# For modelType `cloud-high-accuracy-1`(default), the budget must be between
|
53
|
+
# 20,000 and 2,000,000 milli node hours, inclusive. The default value is
|
54
|
+
# 192,000 which represents one day in wall time
|
55
|
+
# (1000 milli * 24 hours * 8 nodes).
|
56
|
+
# @!attribute [rw] base_model_id
|
57
|
+
# @return [::String]
|
58
|
+
# The ID of the `base` model. If it is specified, the new model will be
|
59
|
+
# trained based on the `base` model. Otherwise, the new model will be
|
60
|
+
# trained from scratch. The `base` model must be in the same
|
61
|
+
# Project and Location as the new Model to train, and have the same
|
62
|
+
# modelType.
|
63
|
+
class AutoMlImageSegmentationInputs
|
64
|
+
include ::Google::Protobuf::MessageExts
|
65
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
66
|
+
|
67
|
+
module ModelType
|
68
|
+
# Should not be set.
|
69
|
+
MODEL_TYPE_UNSPECIFIED = 0
|
70
|
+
|
71
|
+
# A model to be used via prediction calls to uCAIP API. Expected
|
72
|
+
# to have a higher latency, but should also have a higher prediction
|
73
|
+
# quality than other models.
|
74
|
+
CLOUD_HIGH_ACCURACY_1 = 1
|
75
|
+
|
76
|
+
# A model to be used via prediction calls to uCAIP API. Expected
|
77
|
+
# to have a lower latency but relatively lower prediction quality.
|
78
|
+
CLOUD_LOW_ACCURACY_1 = 2
|
79
|
+
|
80
|
+
# A model that, in addition to being available within Google
|
81
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
82
|
+
# model and used on a mobile or edge device afterwards.
|
83
|
+
# Expected to have low latency, but may have lower prediction
|
84
|
+
# quality than other mobile models.
|
85
|
+
MOBILE_TF_LOW_LATENCY_1 = 3
|
86
|
+
end
|
87
|
+
end
|
88
|
+
|
89
|
+
# @!attribute [rw] cost_milli_node_hours
|
90
|
+
# @return [::Integer]
|
91
|
+
# The actual training cost of creating this model, expressed in
|
92
|
+
# milli node hours, i.e. 1,000 value in this field means 1 node hour.
|
93
|
+
# Guaranteed to not exceed inputs.budgetMilliNodeHours.
|
94
|
+
# @!attribute [rw] successful_stop_reason
|
95
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationMetadata::SuccessfulStopReason]
|
96
|
+
# For successful job completions, this is the reason why the job has
|
97
|
+
# finished.
|
98
|
+
class AutoMlImageSegmentationMetadata
|
99
|
+
include ::Google::Protobuf::MessageExts
|
100
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
101
|
+
|
102
|
+
module SuccessfulStopReason
|
103
|
+
# Should not be set.
|
104
|
+
SUCCESSFUL_STOP_REASON_UNSPECIFIED = 0
|
105
|
+
|
106
|
+
# The inputs.budgetMilliNodeHours had been reached.
|
107
|
+
BUDGET_REACHED = 1
|
108
|
+
|
109
|
+
# Further training of the Model ceased to increase its quality, since it
|
110
|
+
# already has converged.
|
111
|
+
MODEL_CONVERGED = 2
|
112
|
+
end
|
113
|
+
end
|
114
|
+
end
|
115
|
+
end
|
116
|
+
end
|
117
|
+
end
|
118
|
+
end
|
119
|
+
end
|
120
|
+
end
|