google-cloud-ai_platform-v1 0.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +7 -0
- data/.yardopts +12 -0
- data/AUTHENTICATION.md +149 -0
- data/LICENSE.md +201 -0
- data/README.md +139 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/client.rb +1364 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service/paths.rb +111 -0
- data/lib/google/cloud/ai_platform/v1/dataset_service.rb +51 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/client.rb +1076 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service/paths.rb +124 -0
- data/lib/google/cloud/ai_platform/v1/endpoint_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/client.rb +508 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service/paths.rb +54 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_online_serving_service.rb +49 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/client.rb +2707 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service/paths.rb +113 -0
- data/lib/google/cloud/ai_platform/v1/featurestore_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/client.rb +1146 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service/paths.rb +88 -0
- data/lib/google/cloud/ai_platform/v1/index_endpoint_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/index_service/client.rb +823 -0
- data/lib/google/cloud/ai_platform/v1/index_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/index_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/index_service/paths.rb +88 -0
- data/lib/google/cloud/ai_platform/v1/index_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/job_service/client.rb +3236 -0
- data/lib/google/cloud/ai_platform/v1/job_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/job_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/job_service/paths.rb +259 -0
- data/lib/google/cloud/ai_platform/v1/job_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/client.rb +3654 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service/paths.rb +153 -0
- data/lib/google/cloud/ai_platform/v1/metadata_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/client.rb +538 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/migration_service/paths.rb +148 -0
- data/lib/google/cloud/ai_platform/v1/migration_service.rb +51 -0
- data/lib/google/cloud/ai_platform/v1/model_service/client.rb +1355 -0
- data/lib/google/cloud/ai_platform/v1/model_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/model_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/model_service/paths.rb +151 -0
- data/lib/google/cloud/ai_platform/v1/model_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/client.rb +1384 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service/paths.rb +225 -0
- data/lib/google/cloud/ai_platform/v1/pipeline_service.rb +52 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service/client.rb +650 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service/paths.rb +52 -0
- data/lib/google/cloud/ai_platform/v1/prediction_service.rb +49 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/client.rb +826 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service/paths.rb +69 -0
- data/lib/google/cloud/ai_platform/v1/specialist_pool_service.rb +55 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/client.rb +3224 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/credentials.rb +48 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service/paths.rb +138 -0
- data/lib/google/cloud/ai_platform/v1/tensorboard_service.rb +50 -0
- data/lib/google/cloud/ai_platform/v1/version.rb +28 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/client.rb +1793 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/credentials.rb +47 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/operations.rb +767 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service/paths.rb +109 -0
- data/lib/google/cloud/ai_platform/v1/vizier_service.rb +54 -0
- data/lib/google/cloud/ai_platform/v1.rb +52 -0
- data/lib/google/cloud/aiplatform/v1/accelerator_type_pb.rb +29 -0
- data/lib/google/cloud/aiplatform/v1/annotation_pb.rb +35 -0
- data/lib/google/cloud/aiplatform/v1/annotation_spec_pb.rb +30 -0
- data/lib/google/cloud/aiplatform/v1/artifact_pb.rb +45 -0
- data/lib/google/cloud/aiplatform/v1/batch_prediction_job_pb.rb +83 -0
- data/lib/google/cloud/aiplatform/v1/completion_stats_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/context_pb.rb +38 -0
- data/lib/google/cloud/aiplatform/v1/custom_job_pb.rb +84 -0
- data/lib/google/cloud/aiplatform/v1/data_item_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/data_labeling_job_pb.rb +78 -0
- data/lib/google/cloud/aiplatform/v1/dataset_pb.rb +53 -0
- data/lib/google/cloud/aiplatform/v1/dataset_service_pb.rb +126 -0
- data/lib/google/cloud/aiplatform/v1/dataset_service_services_pb.rb +64 -0
- data/lib/google/cloud/aiplatform/v1/deployed_index_ref_pb.rb +26 -0
- data/lib/google/cloud/aiplatform/v1/deployed_model_ref_pb.rb +26 -0
- data/lib/google/cloud/aiplatform/v1/encryption_spec_pb.rb +24 -0
- data/lib/google/cloud/aiplatform/v1/endpoint_pb.rb +67 -0
- data/lib/google/cloud/aiplatform/v1/endpoint_service_pb.rb +90 -0
- data/lib/google/cloud/aiplatform/v1/endpoint_service_services_pb.rb +58 -0
- data/lib/google/cloud/aiplatform/v1/entity_type_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/env_var_pb.rb +25 -0
- data/lib/google/cloud/aiplatform/v1/event_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/execution_pb.rb +48 -0
- data/lib/google/cloud/aiplatform/v1/explanation_metadata_pb.rb +107 -0
- data/lib/google/cloud/aiplatform/v1/explanation_pb.rb +106 -0
- data/lib/google/cloud/aiplatform/v1/feature_monitoring_stats_pb.rb +30 -0
- data/lib/google/cloud/aiplatform/v1/feature_pb.rb +46 -0
- data/lib/google/cloud/aiplatform/v1/feature_selector_pb.rb +28 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_online_service_pb.rb +86 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_online_service_services_pb.rb +51 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_pb.rb +44 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_service_pb.rb +280 -0
- data/lib/google/cloud/aiplatform/v1/featurestore_service_services_pb.rb +109 -0
- data/lib/google/cloud/aiplatform/v1/hyperparameter_tuning_job_pb.rb +46 -0
- data/lib/google/cloud/aiplatform/v1/index_endpoint_pb.rb +66 -0
- data/lib/google/cloud/aiplatform/v1/index_endpoint_service_pb.rb +101 -0
- data/lib/google/cloud/aiplatform/v1/index_endpoint_service_services_pb.rb +62 -0
- data/lib/google/cloud/aiplatform/v1/index_pb.rb +38 -0
- data/lib/google/cloud/aiplatform/v1/index_service_pb.rb +98 -0
- data/lib/google/cloud/aiplatform/v1/index_service_services_pb.rb +55 -0
- data/lib/google/cloud/aiplatform/v1/io_pb.rb +56 -0
- data/lib/google/cloud/aiplatform/v1/job_service_pb.rb +217 -0
- data/lib/google/cloud/aiplatform/v1/job_service_services_pb.rb +134 -0
- data/lib/google/cloud/aiplatform/v1/job_state_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/lineage_subgraph_pb.rb +28 -0
- data/lib/google/cloud/aiplatform/v1/machine_resources_pb.rb +59 -0
- data/lib/google/cloud/aiplatform/v1/manual_batch_tuning_parameters_pb.rb +24 -0
- data/lib/google/cloud/aiplatform/v1/metadata_schema_pb.rb +38 -0
- data/lib/google/cloud/aiplatform/v1/metadata_service_pb.rb +272 -0
- data/lib/google/cloud/aiplatform/v1/metadata_service_services_pb.rb +119 -0
- data/lib/google/cloud/aiplatform/v1/metadata_store_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/migratable_resource_pb.rb +59 -0
- data/lib/google/cloud/aiplatform/v1/migration_service_pb.rb +106 -0
- data/lib/google/cloud/aiplatform/v1/migration_service_services_pb.rb +51 -0
- data/lib/google/cloud/aiplatform/v1/model_deployment_monitoring_job_pb.rb +111 -0
- data/lib/google/cloud/aiplatform/v1/model_evaluation_pb.rb +33 -0
- data/lib/google/cloud/aiplatform/v1/model_evaluation_slice_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/model_monitoring_pb.rb +93 -0
- data/lib/google/cloud/aiplatform/v1/model_pb.rb +88 -0
- data/lib/google/cloud/aiplatform/v1/model_service_pb.rb +129 -0
- data/lib/google/cloud/aiplatform/v1/model_service_services_pb.rb +69 -0
- data/lib/google/cloud/aiplatform/v1/operation_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_job_pb.rb +115 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_service_pb.rb +88 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_service_services_pb.rb +84 -0
- data/lib/google/cloud/aiplatform/v1/pipeline_state_pb.rb +31 -0
- data/lib/google/cloud/aiplatform/v1/prediction_service_pb.rb +57 -0
- data/lib/google/cloud/aiplatform/v1/prediction_service_services_pb.rb +66 -0
- data/lib/google/cloud/aiplatform/v1/specialist_pool_pb.rb +30 -0
- data/lib/google/cloud/aiplatform/v1/specialist_pool_service_pb.rb +66 -0
- data/lib/google/cloud/aiplatform/v1/specialist_pool_service_services_pb.rb +58 -0
- data/lib/google/cloud/aiplatform/v1/study_pb.rb +191 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_data_pb.rb +56 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_experiment_pb.rb +33 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_pb.rb +36 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_run_pb.rb +32 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_pb.rb +244 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_service_services_pb.rb +115 -0
- data/lib/google/cloud/aiplatform/v1/tensorboard_time_series_pb.rb +48 -0
- data/lib/google/cloud/aiplatform/v1/training_pipeline_pb.rb +95 -0
- data/lib/google/cloud/aiplatform/v1/types_pb.rb +35 -0
- data/lib/google/cloud/aiplatform/v1/unmanaged_container_model_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/user_action_reference_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/value_pb.rb +27 -0
- data/lib/google/cloud/aiplatform/v1/vizier_service_pb.rb +136 -0
- data/lib/google/cloud/aiplatform/v1/vizier_service_services_pb.rb +90 -0
- data/lib/google-cloud-ai_platform-v1.rb +21 -0
- data/proto_docs/README.md +4 -0
- data/proto_docs/google/api/field_behavior.rb +71 -0
- data/proto_docs/google/api/httpbody.rb +80 -0
- data/proto_docs/google/api/resource.rb +222 -0
- data/proto_docs/google/cloud/aiplatform/v1/accelerator_type.rb +50 -0
- data/proto_docs/google/cloud/aiplatform/v1/annotation.rb +92 -0
- data/proto_docs/google/cloud/aiplatform/v1/annotation_spec.rb +50 -0
- data/proto_docs/google/cloud/aiplatform/v1/artifact.rb +112 -0
- data/proto_docs/google/cloud/aiplatform/v1/batch_prediction_job.rb +278 -0
- data/proto_docs/google/cloud/aiplatform/v1/completion_stats.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/context.rb +92 -0
- data/proto_docs/google/cloud/aiplatform/v1/custom_job.rb +272 -0
- data/proto_docs/google/cloud/aiplatform/v1/data_item.rb +73 -0
- data/proto_docs/google/cloud/aiplatform/v1/data_labeling_job.rb +207 -0
- data/proto_docs/google/cloud/aiplatform/v1/dataset.rb +154 -0
- data/proto_docs/google/cloud/aiplatform/v1/dataset_service.rb +301 -0
- data/proto_docs/google/cloud/aiplatform/v1/deployed_index_ref.rb +38 -0
- data/proto_docs/google/cloud/aiplatform/v1/deployed_model_ref.rb +38 -0
- data/proto_docs/google/cloud/aiplatform/v1/encryption_spec.rb +40 -0
- data/proto_docs/google/cloud/aiplatform/v1/endpoint.rb +227 -0
- data/proto_docs/google/cloud/aiplatform/v1/endpoint_service.rb +258 -0
- data/proto_docs/google/cloud/aiplatform/v1/entity_type.rb +79 -0
- data/proto_docs/google/cloud/aiplatform/v1/env_var.rb +44 -0
- data/proto_docs/google/cloud/aiplatform/v1/event.rb +79 -0
- data/proto_docs/google/cloud/aiplatform/v1/execution.rb +118 -0
- data/proto_docs/google/cloud/aiplatform/v1/explanation.rb +445 -0
- data/proto_docs/google/cloud/aiplatform/v1/explanation_metadata.rb +419 -0
- data/proto_docs/google/cloud/aiplatform/v1/feature.rb +115 -0
- data/proto_docs/google/cloud/aiplatform/v1/feature_monitoring_stats.rb +88 -0
- data/proto_docs/google/cloud/aiplatform/v1/feature_selector.rb +49 -0
- data/proto_docs/google/cloud/aiplatform/v1/featurestore.rb +115 -0
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_online_service.rb +203 -0
- data/proto_docs/google/cloud/aiplatform/v1/featurestore_service.rb +978 -0
- data/proto_docs/google/cloud/aiplatform/v1/hyperparameter_tuning_job.rb +109 -0
- data/proto_docs/google/cloud/aiplatform/v1/index.rb +98 -0
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint.rb +252 -0
- data/proto_docs/google/cloud/aiplatform/v1/index_endpoint_service.rb +240 -0
- data/proto_docs/google/cloud/aiplatform/v1/index_service.rb +220 -0
- data/proto_docs/google/cloud/aiplatform/v1/io.rb +134 -0
- data/proto_docs/google/cloud/aiplatform/v1/job_service.rb +660 -0
- data/proto_docs/google/cloud/aiplatform/v1/job_state.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/lineage_subgraph.rb +42 -0
- data/proto_docs/google/cloud/aiplatform/v1/machine_resources.rb +194 -0
- data/proto_docs/google/cloud/aiplatform/v1/manual_batch_tuning_parameters.rb +41 -0
- data/proto_docs/google/cloud/aiplatform/v1/metadata_schema.rb +74 -0
- data/proto_docs/google/cloud/aiplatform/v1/metadata_service.rb +912 -0
- data/proto_docs/google/cloud/aiplatform/v1/metadata_store.rb +62 -0
- data/proto_docs/google/cloud/aiplatform/v1/migratable_resource.rb +133 -0
- data/proto_docs/google/cloud/aiplatform/v1/migration_service.rb +260 -0
- data/proto_docs/google/cloud/aiplatform/v1/model.rb +562 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_deployment_monitoring_job.rb +293 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_evaluation_slice.rb +68 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_monitoring.rb +257 -0
- data/proto_docs/google/cloud/aiplatform/v1/model_service.rb +329 -0
- data/proto_docs/google/cloud/aiplatform/v1/operation.rb +55 -0
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_job.rb +347 -0
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_service.rb +258 -0
- data/proto_docs/google/cloud/aiplatform/v1/pipeline_state.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/prediction_service.rb +165 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/image_classification.rb +52 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/image_object_detection.rb +52 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/image_segmentation.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/text_classification.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/text_extraction.rb +53 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/text_sentiment.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/video_action_recognition.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/video_classification.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/instance/video_object_tracking.rb +59 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/image_classification.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/image_object_detection.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/image_segmentation.rb +44 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/video_action_recognition.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/video_classification.rb +72 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/params/video_object_tracking.rb +51 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/classification.rb +49 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/image_object_detection.rb +58 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/image_segmentation.rb +53 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/tabular_classification.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/tabular_regression.rb +47 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/text_extraction.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/text_sentiment.rb +45 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_action_recognition.rb +60 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_classification.rb +73 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_object_tracking.rb +91 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_image_classification.rb +142 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_image_object_detection.rb +134 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_image_segmentation.rb +120 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_tables.rb +315 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_text_classification.rb +48 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_text_extraction.rb +46 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_text_sentiment.rb +55 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_video_action_recognition.rb +73 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_video_classification.rb +67 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/automl_video_object_tracking.rb +78 -0
- data/proto_docs/google/cloud/aiplatform/v1/schema/trainingjob/definition/export_evaluated_data_items_config.rb +51 -0
- data/proto_docs/google/cloud/aiplatform/v1/specialist_pool.rb +58 -0
- data/proto_docs/google/cloud/aiplatform/v1/specialist_pool_service.rb +136 -0
- data/proto_docs/google/cloud/aiplatform/v1/study.rb +543 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard.rb +89 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_data.rb +110 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_experiment.rb +82 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_run.rb +85 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_service.rb +706 -0
- data/proto_docs/google/cloud/aiplatform/v1/tensorboard_time_series.rb +101 -0
- data/proto_docs/google/cloud/aiplatform/v1/training_pipeline.rb +381 -0
- data/proto_docs/google/cloud/aiplatform/v1/types.rb +62 -0
- data/proto_docs/google/cloud/aiplatform/v1/unmanaged_container_model.rb +44 -0
- data/proto_docs/google/cloud/aiplatform/v1/user_action_reference.rb +49 -0
- data/proto_docs/google/cloud/aiplatform/v1/value.rb +41 -0
- data/proto_docs/google/cloud/aiplatform/v1/vizier_service.rb +332 -0
- data/proto_docs/google/longrunning/operations.rb +164 -0
- data/proto_docs/google/protobuf/any.rb +141 -0
- data/proto_docs/google/protobuf/duration.rb +98 -0
- data/proto_docs/google/protobuf/empty.rb +36 -0
- data/proto_docs/google/protobuf/field_mask.rb +229 -0
- data/proto_docs/google/protobuf/struct.rb +96 -0
- data/proto_docs/google/protobuf/timestamp.rb +129 -0
- data/proto_docs/google/protobuf/wrappers.rb +121 -0
- data/proto_docs/google/rpc/status.rb +46 -0
- data/proto_docs/google/type/money.rb +43 -0
- metadata +479 -0
data/proto_docs/google/cloud/aiplatform/v1/schema/predict/prediction/video_action_recognition.rb
ADDED
|
@@ -0,0 +1,60 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
# Copyright 2022 Google LLC
|
|
4
|
+
#
|
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
6
|
+
# you may not use this file except in compliance with the License.
|
|
7
|
+
# You may obtain a copy of the License at
|
|
8
|
+
#
|
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
|
10
|
+
#
|
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
14
|
+
# See the License for the specific language governing permissions and
|
|
15
|
+
# limitations under the License.
|
|
16
|
+
|
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
module Google
|
|
21
|
+
module Cloud
|
|
22
|
+
module AIPlatform
|
|
23
|
+
module V1
|
|
24
|
+
module Schema
|
|
25
|
+
module Predict
|
|
26
|
+
module Prediction
|
|
27
|
+
# Prediction output format for Video Action Recognition.
|
|
28
|
+
# @!attribute [rw] id
|
|
29
|
+
# @return [::String]
|
|
30
|
+
# The resource ID of the AnnotationSpec that had been identified.
|
|
31
|
+
# @!attribute [rw] display_name
|
|
32
|
+
# @return [::String]
|
|
33
|
+
# The display name of the AnnotationSpec that had been identified.
|
|
34
|
+
# @!attribute [rw] time_segment_start
|
|
35
|
+
# @return [::Google::Protobuf::Duration]
|
|
36
|
+
# The beginning, inclusive, of the video's time segment in which the
|
|
37
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
|
38
|
+
# measured from the start of the video, with fractions up to a microsecond
|
|
39
|
+
# precision, and with "s" appended at the end.
|
|
40
|
+
# @!attribute [rw] time_segment_end
|
|
41
|
+
# @return [::Google::Protobuf::Duration]
|
|
42
|
+
# The end, exclusive, of the video's time segment in which the
|
|
43
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
|
44
|
+
# measured from the start of the video, with fractions up to a microsecond
|
|
45
|
+
# precision, and with "s" appended at the end.
|
|
46
|
+
# @!attribute [rw] confidence
|
|
47
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
48
|
+
# The Model's confidence in correction of this prediction, higher
|
|
49
|
+
# value means higher confidence.
|
|
50
|
+
class VideoActionRecognitionPredictionResult
|
|
51
|
+
include ::Google::Protobuf::MessageExts
|
|
52
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
53
|
+
end
|
|
54
|
+
end
|
|
55
|
+
end
|
|
56
|
+
end
|
|
57
|
+
end
|
|
58
|
+
end
|
|
59
|
+
end
|
|
60
|
+
end
|
|
@@ -0,0 +1,73 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
# Copyright 2022 Google LLC
|
|
4
|
+
#
|
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
6
|
+
# you may not use this file except in compliance with the License.
|
|
7
|
+
# You may obtain a copy of the License at
|
|
8
|
+
#
|
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
|
10
|
+
#
|
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
14
|
+
# See the License for the specific language governing permissions and
|
|
15
|
+
# limitations under the License.
|
|
16
|
+
|
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
module Google
|
|
21
|
+
module Cloud
|
|
22
|
+
module AIPlatform
|
|
23
|
+
module V1
|
|
24
|
+
module Schema
|
|
25
|
+
module Predict
|
|
26
|
+
module Prediction
|
|
27
|
+
# Prediction output format for Video Classification.
|
|
28
|
+
# @!attribute [rw] id
|
|
29
|
+
# @return [::String]
|
|
30
|
+
# The resource ID of the AnnotationSpec that had been identified.
|
|
31
|
+
# @!attribute [rw] display_name
|
|
32
|
+
# @return [::String]
|
|
33
|
+
# The display name of the AnnotationSpec that had been identified.
|
|
34
|
+
# @!attribute [rw] type
|
|
35
|
+
# @return [::String]
|
|
36
|
+
# The type of the prediction. The requested types can be configured
|
|
37
|
+
# via parameters. This will be one of
|
|
38
|
+
# - segment-classification
|
|
39
|
+
# - shot-classification
|
|
40
|
+
# - one-sec-interval-classification
|
|
41
|
+
# @!attribute [rw] time_segment_start
|
|
42
|
+
# @return [::Google::Protobuf::Duration]
|
|
43
|
+
# The beginning, inclusive, of the video's time segment in which the
|
|
44
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
|
45
|
+
# measured from the start of the video, with fractions up to a microsecond
|
|
46
|
+
# precision, and with "s" appended at the end. Note that for
|
|
47
|
+
# 'segment-classification' prediction type, this equals the original
|
|
48
|
+
# 'timeSegmentStart' from the input instance, for other types it is the
|
|
49
|
+
# start of a shot or a 1 second interval respectively.
|
|
50
|
+
# @!attribute [rw] time_segment_end
|
|
51
|
+
# @return [::Google::Protobuf::Duration]
|
|
52
|
+
# The end, exclusive, of the video's time segment in which the
|
|
53
|
+
# AnnotationSpec has been identified. Expressed as a number of seconds as
|
|
54
|
+
# measured from the start of the video, with fractions up to a microsecond
|
|
55
|
+
# precision, and with "s" appended at the end. Note that for
|
|
56
|
+
# 'segment-classification' prediction type, this equals the original
|
|
57
|
+
# 'timeSegmentEnd' from the input instance, for other types it is the end
|
|
58
|
+
# of a shot or a 1 second interval respectively.
|
|
59
|
+
# @!attribute [rw] confidence
|
|
60
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
61
|
+
# The Model's confidence in correction of this prediction, higher
|
|
62
|
+
# value means higher confidence.
|
|
63
|
+
class VideoClassificationPredictionResult
|
|
64
|
+
include ::Google::Protobuf::MessageExts
|
|
65
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
66
|
+
end
|
|
67
|
+
end
|
|
68
|
+
end
|
|
69
|
+
end
|
|
70
|
+
end
|
|
71
|
+
end
|
|
72
|
+
end
|
|
73
|
+
end
|
|
@@ -0,0 +1,91 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
# Copyright 2022 Google LLC
|
|
4
|
+
#
|
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
6
|
+
# you may not use this file except in compliance with the License.
|
|
7
|
+
# You may obtain a copy of the License at
|
|
8
|
+
#
|
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
|
10
|
+
#
|
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
14
|
+
# See the License for the specific language governing permissions and
|
|
15
|
+
# limitations under the License.
|
|
16
|
+
|
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
module Google
|
|
21
|
+
module Cloud
|
|
22
|
+
module AIPlatform
|
|
23
|
+
module V1
|
|
24
|
+
module Schema
|
|
25
|
+
module Predict
|
|
26
|
+
module Prediction
|
|
27
|
+
# Prediction output format for Video Object Tracking.
|
|
28
|
+
# @!attribute [rw] id
|
|
29
|
+
# @return [::String]
|
|
30
|
+
# The resource ID of the AnnotationSpec that had been identified.
|
|
31
|
+
# @!attribute [rw] display_name
|
|
32
|
+
# @return [::String]
|
|
33
|
+
# The display name of the AnnotationSpec that had been identified.
|
|
34
|
+
# @!attribute [rw] time_segment_start
|
|
35
|
+
# @return [::Google::Protobuf::Duration]
|
|
36
|
+
# The beginning, inclusive, of the video's time segment in which the
|
|
37
|
+
# object instance has been detected. Expressed as a number of seconds as
|
|
38
|
+
# measured from the start of the video, with fractions up to a microsecond
|
|
39
|
+
# precision, and with "s" appended at the end.
|
|
40
|
+
# @!attribute [rw] time_segment_end
|
|
41
|
+
# @return [::Google::Protobuf::Duration]
|
|
42
|
+
# The end, inclusive, of the video's time segment in which the
|
|
43
|
+
# object instance has been detected. Expressed as a number of seconds as
|
|
44
|
+
# measured from the start of the video, with fractions up to a microsecond
|
|
45
|
+
# precision, and with "s" appended at the end.
|
|
46
|
+
# @!attribute [rw] confidence
|
|
47
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
48
|
+
# The Model's confidence in correction of this prediction, higher
|
|
49
|
+
# value means higher confidence.
|
|
50
|
+
# @!attribute [rw] frames
|
|
51
|
+
# @return [::Array<::Google::Cloud::AIPlatform::V1::Schema::Predict::Prediction::VideoObjectTrackingPredictionResult::Frame>]
|
|
52
|
+
# All of the frames of the video in which a single object instance has been
|
|
53
|
+
# detected. The bounding boxes in the frames identify the same object.
|
|
54
|
+
class VideoObjectTrackingPredictionResult
|
|
55
|
+
include ::Google::Protobuf::MessageExts
|
|
56
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
57
|
+
|
|
58
|
+
# The fields `xMin`, `xMax`, `yMin`, and `yMax` refer to a bounding box,
|
|
59
|
+
# i.e. the rectangle over the video frame pinpointing the found
|
|
60
|
+
# AnnotationSpec. The coordinates are relative to the frame size, and the
|
|
61
|
+
# point 0,0 is in the top left of the frame.
|
|
62
|
+
# @!attribute [rw] time_offset
|
|
63
|
+
# @return [::Google::Protobuf::Duration]
|
|
64
|
+
# A time (frame) of a video in which the object has been detected.
|
|
65
|
+
# Expressed as a number of seconds as measured from the
|
|
66
|
+
# start of the video, with fractions up to a microsecond precision, and
|
|
67
|
+
# with "s" appended at the end.
|
|
68
|
+
# @!attribute [rw] x_min
|
|
69
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
70
|
+
# The leftmost coordinate of the bounding box.
|
|
71
|
+
# @!attribute [rw] x_max
|
|
72
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
73
|
+
# The rightmost coordinate of the bounding box.
|
|
74
|
+
# @!attribute [rw] y_min
|
|
75
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
76
|
+
# The topmost coordinate of the bounding box.
|
|
77
|
+
# @!attribute [rw] y_max
|
|
78
|
+
# @return [::Google::Protobuf::FloatValue]
|
|
79
|
+
# The bottommost coordinate of the bounding box.
|
|
80
|
+
class Frame
|
|
81
|
+
include ::Google::Protobuf::MessageExts
|
|
82
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
83
|
+
end
|
|
84
|
+
end
|
|
85
|
+
end
|
|
86
|
+
end
|
|
87
|
+
end
|
|
88
|
+
end
|
|
89
|
+
end
|
|
90
|
+
end
|
|
91
|
+
end
|
|
@@ -0,0 +1,142 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
# Copyright 2022 Google LLC
|
|
4
|
+
#
|
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
6
|
+
# you may not use this file except in compliance with the License.
|
|
7
|
+
# You may obtain a copy of the License at
|
|
8
|
+
#
|
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
|
10
|
+
#
|
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
14
|
+
# See the License for the specific language governing permissions and
|
|
15
|
+
# limitations under the License.
|
|
16
|
+
|
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
module Google
|
|
21
|
+
module Cloud
|
|
22
|
+
module AIPlatform
|
|
23
|
+
module V1
|
|
24
|
+
module Schema
|
|
25
|
+
module TrainingJob
|
|
26
|
+
module Definition
|
|
27
|
+
# A TrainingJob that trains and uploads an AutoML Image Classification Model.
|
|
28
|
+
# @!attribute [rw] inputs
|
|
29
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationInputs]
|
|
30
|
+
# The input parameters of this TrainingJob.
|
|
31
|
+
# @!attribute [rw] metadata
|
|
32
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationMetadata]
|
|
33
|
+
# The metadata information.
|
|
34
|
+
class AutoMlImageClassification
|
|
35
|
+
include ::Google::Protobuf::MessageExts
|
|
36
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
37
|
+
end
|
|
38
|
+
|
|
39
|
+
# @!attribute [rw] model_type
|
|
40
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationInputs::ModelType]
|
|
41
|
+
# @!attribute [rw] base_model_id
|
|
42
|
+
# @return [::String]
|
|
43
|
+
# The ID of the `base` model. If it is specified, the new model will be
|
|
44
|
+
# trained based on the `base` model. Otherwise, the new model will be
|
|
45
|
+
# trained from scratch. The `base` model must be in the same
|
|
46
|
+
# Project and Location as the new Model to train, and have the same
|
|
47
|
+
# modelType.
|
|
48
|
+
# @!attribute [rw] budget_milli_node_hours
|
|
49
|
+
# @return [::Integer]
|
|
50
|
+
# The training budget of creating this model, expressed in milli node
|
|
51
|
+
# hours i.e. 1,000 value in this field means 1 node hour. The actual
|
|
52
|
+
# metadata.costMilliNodeHours will be equal or less than this value.
|
|
53
|
+
# If further model training ceases to provide any improvements, it will
|
|
54
|
+
# stop without using the full budget and the metadata.successfulStopReason
|
|
55
|
+
# will be `model-converged`.
|
|
56
|
+
# Note, node_hour = actual_hour * number_of_nodes_involved.
|
|
57
|
+
# For modelType `cloud`(default), the budget must be between 8,000
|
|
58
|
+
# and 800,000 milli node hours, inclusive. The default value is 192,000
|
|
59
|
+
# which represents one day in wall time, considering 8 nodes are used.
|
|
60
|
+
# For model types `mobile-tf-low-latency-1`, `mobile-tf-versatile-1`,
|
|
61
|
+
# `mobile-tf-high-accuracy-1`, the training budget must be between
|
|
62
|
+
# 1,000 and 100,000 milli node hours, inclusive.
|
|
63
|
+
# The default value is 24,000 which represents one day in wall time on a
|
|
64
|
+
# single node that is used.
|
|
65
|
+
# @!attribute [rw] disable_early_stopping
|
|
66
|
+
# @return [::Boolean]
|
|
67
|
+
# Use the entire training budget. This disables the early stopping feature.
|
|
68
|
+
# When false the early stopping feature is enabled, which means that
|
|
69
|
+
# AutoML Image Classification might stop training before the entire
|
|
70
|
+
# training budget has been used.
|
|
71
|
+
# @!attribute [rw] multi_label
|
|
72
|
+
# @return [::Boolean]
|
|
73
|
+
# If false, a single-label (multi-class) Model will be trained (i.e.
|
|
74
|
+
# assuming that for each image just up to one annotation may be
|
|
75
|
+
# applicable). If true, a multi-label Model will be trained (i.e.
|
|
76
|
+
# assuming that for each image multiple annotations may be applicable).
|
|
77
|
+
class AutoMlImageClassificationInputs
|
|
78
|
+
include ::Google::Protobuf::MessageExts
|
|
79
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
80
|
+
|
|
81
|
+
module ModelType
|
|
82
|
+
# Should not be set.
|
|
83
|
+
MODEL_TYPE_UNSPECIFIED = 0
|
|
84
|
+
|
|
85
|
+
# A Model best tailored to be used within Google Cloud, and which cannot
|
|
86
|
+
# be exported.
|
|
87
|
+
# Default.
|
|
88
|
+
CLOUD = 1
|
|
89
|
+
|
|
90
|
+
# A model that, in addition to being available within Google
|
|
91
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
|
92
|
+
# or Core ML model and used on a mobile or edge device afterwards.
|
|
93
|
+
# Expected to have low latency, but may have lower prediction
|
|
94
|
+
# quality than other mobile models.
|
|
95
|
+
MOBILE_TF_LOW_LATENCY_1 = 2
|
|
96
|
+
|
|
97
|
+
# A model that, in addition to being available within Google
|
|
98
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
|
99
|
+
# or Core ML model and used on a mobile or edge device with afterwards.
|
|
100
|
+
MOBILE_TF_VERSATILE_1 = 3
|
|
101
|
+
|
|
102
|
+
# A model that, in addition to being available within Google
|
|
103
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
|
104
|
+
# or Core ML model and used on a mobile or edge device afterwards.
|
|
105
|
+
# Expected to have a higher latency, but should also have a higher
|
|
106
|
+
# prediction quality than other mobile models.
|
|
107
|
+
MOBILE_TF_HIGH_ACCURACY_1 = 4
|
|
108
|
+
end
|
|
109
|
+
end
|
|
110
|
+
|
|
111
|
+
# @!attribute [rw] cost_milli_node_hours
|
|
112
|
+
# @return [::Integer]
|
|
113
|
+
# The actual training cost of creating this model, expressed in
|
|
114
|
+
# milli node hours, i.e. 1,000 value in this field means 1 node hour.
|
|
115
|
+
# Guaranteed to not exceed inputs.budgetMilliNodeHours.
|
|
116
|
+
# @!attribute [rw] successful_stop_reason
|
|
117
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageClassificationMetadata::SuccessfulStopReason]
|
|
118
|
+
# For successful job completions, this is the reason why the job has
|
|
119
|
+
# finished.
|
|
120
|
+
class AutoMlImageClassificationMetadata
|
|
121
|
+
include ::Google::Protobuf::MessageExts
|
|
122
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
123
|
+
|
|
124
|
+
module SuccessfulStopReason
|
|
125
|
+
# Should not be set.
|
|
126
|
+
SUCCESSFUL_STOP_REASON_UNSPECIFIED = 0
|
|
127
|
+
|
|
128
|
+
# The inputs.budgetMilliNodeHours had been reached.
|
|
129
|
+
BUDGET_REACHED = 1
|
|
130
|
+
|
|
131
|
+
# Further training of the Model ceased to increase its quality, since it
|
|
132
|
+
# already has converged.
|
|
133
|
+
MODEL_CONVERGED = 2
|
|
134
|
+
end
|
|
135
|
+
end
|
|
136
|
+
end
|
|
137
|
+
end
|
|
138
|
+
end
|
|
139
|
+
end
|
|
140
|
+
end
|
|
141
|
+
end
|
|
142
|
+
end
|
|
@@ -0,0 +1,134 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
# Copyright 2022 Google LLC
|
|
4
|
+
#
|
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
6
|
+
# you may not use this file except in compliance with the License.
|
|
7
|
+
# You may obtain a copy of the License at
|
|
8
|
+
#
|
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
|
10
|
+
#
|
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
14
|
+
# See the License for the specific language governing permissions and
|
|
15
|
+
# limitations under the License.
|
|
16
|
+
|
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
module Google
|
|
21
|
+
module Cloud
|
|
22
|
+
module AIPlatform
|
|
23
|
+
module V1
|
|
24
|
+
module Schema
|
|
25
|
+
module TrainingJob
|
|
26
|
+
module Definition
|
|
27
|
+
# A TrainingJob that trains and uploads an AutoML Image Object Detection Model.
|
|
28
|
+
# @!attribute [rw] inputs
|
|
29
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionInputs]
|
|
30
|
+
# The input parameters of this TrainingJob.
|
|
31
|
+
# @!attribute [rw] metadata
|
|
32
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionMetadata]
|
|
33
|
+
# The metadata information
|
|
34
|
+
class AutoMlImageObjectDetection
|
|
35
|
+
include ::Google::Protobuf::MessageExts
|
|
36
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
37
|
+
end
|
|
38
|
+
|
|
39
|
+
# @!attribute [rw] model_type
|
|
40
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionInputs::ModelType]
|
|
41
|
+
# @!attribute [rw] budget_milli_node_hours
|
|
42
|
+
# @return [::Integer]
|
|
43
|
+
# The training budget of creating this model, expressed in milli node
|
|
44
|
+
# hours i.e. 1,000 value in this field means 1 node hour. The actual
|
|
45
|
+
# metadata.costMilliNodeHours will be equal or less than this value.
|
|
46
|
+
# If further model training ceases to provide any improvements, it will
|
|
47
|
+
# stop without using the full budget and the metadata.successfulStopReason
|
|
48
|
+
# will be `model-converged`.
|
|
49
|
+
# Note, node_hour = actual_hour * number_of_nodes_involved.
|
|
50
|
+
# For modelType `cloud`(default), the budget must be between 20,000
|
|
51
|
+
# and 900,000 milli node hours, inclusive. The default value is 216,000
|
|
52
|
+
# which represents one day in wall time, considering 9 nodes are used.
|
|
53
|
+
# For model types `mobile-tf-low-latency-1`, `mobile-tf-versatile-1`,
|
|
54
|
+
# `mobile-tf-high-accuracy-1`
|
|
55
|
+
# the training budget must be between 1,000 and 100,000 milli node hours,
|
|
56
|
+
# inclusive. The default value is 24,000 which represents one day in
|
|
57
|
+
# wall time on a single node that is used.
|
|
58
|
+
# @!attribute [rw] disable_early_stopping
|
|
59
|
+
# @return [::Boolean]
|
|
60
|
+
# Use the entire training budget. This disables the early stopping feature.
|
|
61
|
+
# When false the early stopping feature is enabled, which means that AutoML
|
|
62
|
+
# Image Object Detection might stop training before the entire training
|
|
63
|
+
# budget has been used.
|
|
64
|
+
class AutoMlImageObjectDetectionInputs
|
|
65
|
+
include ::Google::Protobuf::MessageExts
|
|
66
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
67
|
+
|
|
68
|
+
module ModelType
|
|
69
|
+
# Should not be set.
|
|
70
|
+
MODEL_TYPE_UNSPECIFIED = 0
|
|
71
|
+
|
|
72
|
+
# A model best tailored to be used within Google Cloud, and which cannot
|
|
73
|
+
# be exported. Expected to have a higher latency, but should also have a
|
|
74
|
+
# higher prediction quality than other cloud models.
|
|
75
|
+
CLOUD_HIGH_ACCURACY_1 = 1
|
|
76
|
+
|
|
77
|
+
# A model best tailored to be used within Google Cloud, and which cannot
|
|
78
|
+
# be exported. Expected to have a low latency, but may have lower
|
|
79
|
+
# prediction quality than other cloud models.
|
|
80
|
+
CLOUD_LOW_LATENCY_1 = 2
|
|
81
|
+
|
|
82
|
+
# A model that, in addition to being available within Google
|
|
83
|
+
# Cloud can also be exported (see ModelService.ExportModel) and
|
|
84
|
+
# used on a mobile or edge device with TensorFlow afterwards.
|
|
85
|
+
# Expected to have low latency, but may have lower prediction
|
|
86
|
+
# quality than other mobile models.
|
|
87
|
+
MOBILE_TF_LOW_LATENCY_1 = 3
|
|
88
|
+
|
|
89
|
+
# A model that, in addition to being available within Google
|
|
90
|
+
# Cloud can also be exported (see ModelService.ExportModel) and
|
|
91
|
+
# used on a mobile or edge device with TensorFlow afterwards.
|
|
92
|
+
MOBILE_TF_VERSATILE_1 = 4
|
|
93
|
+
|
|
94
|
+
# A model that, in addition to being available within Google
|
|
95
|
+
# Cloud, can also be exported (see ModelService.ExportModel) and
|
|
96
|
+
# used on a mobile or edge device with TensorFlow afterwards.
|
|
97
|
+
# Expected to have a higher latency, but should also have a higher
|
|
98
|
+
# prediction quality than other mobile models.
|
|
99
|
+
MOBILE_TF_HIGH_ACCURACY_1 = 5
|
|
100
|
+
end
|
|
101
|
+
end
|
|
102
|
+
|
|
103
|
+
# @!attribute [rw] cost_milli_node_hours
|
|
104
|
+
# @return [::Integer]
|
|
105
|
+
# The actual training cost of creating this model, expressed in
|
|
106
|
+
# milli node hours, i.e. 1,000 value in this field means 1 node hour.
|
|
107
|
+
# Guaranteed to not exceed inputs.budgetMilliNodeHours.
|
|
108
|
+
# @!attribute [rw] successful_stop_reason
|
|
109
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageObjectDetectionMetadata::SuccessfulStopReason]
|
|
110
|
+
# For successful job completions, this is the reason why the job has
|
|
111
|
+
# finished.
|
|
112
|
+
class AutoMlImageObjectDetectionMetadata
|
|
113
|
+
include ::Google::Protobuf::MessageExts
|
|
114
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
115
|
+
|
|
116
|
+
module SuccessfulStopReason
|
|
117
|
+
# Should not be set.
|
|
118
|
+
SUCCESSFUL_STOP_REASON_UNSPECIFIED = 0
|
|
119
|
+
|
|
120
|
+
# The inputs.budgetMilliNodeHours had been reached.
|
|
121
|
+
BUDGET_REACHED = 1
|
|
122
|
+
|
|
123
|
+
# Further training of the Model ceased to increase its quality, since it
|
|
124
|
+
# already has converged.
|
|
125
|
+
MODEL_CONVERGED = 2
|
|
126
|
+
end
|
|
127
|
+
end
|
|
128
|
+
end
|
|
129
|
+
end
|
|
130
|
+
end
|
|
131
|
+
end
|
|
132
|
+
end
|
|
133
|
+
end
|
|
134
|
+
end
|
|
@@ -0,0 +1,120 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
# Copyright 2022 Google LLC
|
|
4
|
+
#
|
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
6
|
+
# you may not use this file except in compliance with the License.
|
|
7
|
+
# You may obtain a copy of the License at
|
|
8
|
+
#
|
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
|
10
|
+
#
|
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
14
|
+
# See the License for the specific language governing permissions and
|
|
15
|
+
# limitations under the License.
|
|
16
|
+
|
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
module Google
|
|
21
|
+
module Cloud
|
|
22
|
+
module AIPlatform
|
|
23
|
+
module V1
|
|
24
|
+
module Schema
|
|
25
|
+
module TrainingJob
|
|
26
|
+
module Definition
|
|
27
|
+
# A TrainingJob that trains and uploads an AutoML Image Segmentation Model.
|
|
28
|
+
# @!attribute [rw] inputs
|
|
29
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationInputs]
|
|
30
|
+
# The input parameters of this TrainingJob.
|
|
31
|
+
# @!attribute [rw] metadata
|
|
32
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationMetadata]
|
|
33
|
+
# The metadata information.
|
|
34
|
+
class AutoMlImageSegmentation
|
|
35
|
+
include ::Google::Protobuf::MessageExts
|
|
36
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
37
|
+
end
|
|
38
|
+
|
|
39
|
+
# @!attribute [rw] model_type
|
|
40
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationInputs::ModelType]
|
|
41
|
+
# @!attribute [rw] budget_milli_node_hours
|
|
42
|
+
# @return [::Integer]
|
|
43
|
+
# The training budget of creating this model, expressed in milli node
|
|
44
|
+
# hours i.e. 1,000 value in this field means 1 node hour. The actual
|
|
45
|
+
# metadata.costMilliNodeHours will be equal or less than this value.
|
|
46
|
+
# If further model training ceases to provide any improvements, it will
|
|
47
|
+
# stop without using the full budget and the metadata.successfulStopReason
|
|
48
|
+
# will be `model-converged`.
|
|
49
|
+
# Note, node_hour = actual_hour * number_of_nodes_involved. Or
|
|
50
|
+
# actaul_wall_clock_hours = train_budget_milli_node_hours /
|
|
51
|
+
# (number_of_nodes_involved * 1000)
|
|
52
|
+
# For modelType `cloud-high-accuracy-1`(default), the budget must be between
|
|
53
|
+
# 20,000 and 2,000,000 milli node hours, inclusive. The default value is
|
|
54
|
+
# 192,000 which represents one day in wall time
|
|
55
|
+
# (1000 milli * 24 hours * 8 nodes).
|
|
56
|
+
# @!attribute [rw] base_model_id
|
|
57
|
+
# @return [::String]
|
|
58
|
+
# The ID of the `base` model. If it is specified, the new model will be
|
|
59
|
+
# trained based on the `base` model. Otherwise, the new model will be
|
|
60
|
+
# trained from scratch. The `base` model must be in the same
|
|
61
|
+
# Project and Location as the new Model to train, and have the same
|
|
62
|
+
# modelType.
|
|
63
|
+
class AutoMlImageSegmentationInputs
|
|
64
|
+
include ::Google::Protobuf::MessageExts
|
|
65
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
66
|
+
|
|
67
|
+
module ModelType
|
|
68
|
+
# Should not be set.
|
|
69
|
+
MODEL_TYPE_UNSPECIFIED = 0
|
|
70
|
+
|
|
71
|
+
# A model to be used via prediction calls to uCAIP API. Expected
|
|
72
|
+
# to have a higher latency, but should also have a higher prediction
|
|
73
|
+
# quality than other models.
|
|
74
|
+
CLOUD_HIGH_ACCURACY_1 = 1
|
|
75
|
+
|
|
76
|
+
# A model to be used via prediction calls to uCAIP API. Expected
|
|
77
|
+
# to have a lower latency but relatively lower prediction quality.
|
|
78
|
+
CLOUD_LOW_ACCURACY_1 = 2
|
|
79
|
+
|
|
80
|
+
# A model that, in addition to being available within Google
|
|
81
|
+
# Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow
|
|
82
|
+
# model and used on a mobile or edge device afterwards.
|
|
83
|
+
# Expected to have low latency, but may have lower prediction
|
|
84
|
+
# quality than other mobile models.
|
|
85
|
+
MOBILE_TF_LOW_LATENCY_1 = 3
|
|
86
|
+
end
|
|
87
|
+
end
|
|
88
|
+
|
|
89
|
+
# @!attribute [rw] cost_milli_node_hours
|
|
90
|
+
# @return [::Integer]
|
|
91
|
+
# The actual training cost of creating this model, expressed in
|
|
92
|
+
# milli node hours, i.e. 1,000 value in this field means 1 node hour.
|
|
93
|
+
# Guaranteed to not exceed inputs.budgetMilliNodeHours.
|
|
94
|
+
# @!attribute [rw] successful_stop_reason
|
|
95
|
+
# @return [::Google::Cloud::AIPlatform::V1::Schema::TrainingJob::Definition::AutoMlImageSegmentationMetadata::SuccessfulStopReason]
|
|
96
|
+
# For successful job completions, this is the reason why the job has
|
|
97
|
+
# finished.
|
|
98
|
+
class AutoMlImageSegmentationMetadata
|
|
99
|
+
include ::Google::Protobuf::MessageExts
|
|
100
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
101
|
+
|
|
102
|
+
module SuccessfulStopReason
|
|
103
|
+
# Should not be set.
|
|
104
|
+
SUCCESSFUL_STOP_REASON_UNSPECIFIED = 0
|
|
105
|
+
|
|
106
|
+
# The inputs.budgetMilliNodeHours had been reached.
|
|
107
|
+
BUDGET_REACHED = 1
|
|
108
|
+
|
|
109
|
+
# Further training of the Model ceased to increase its quality, since it
|
|
110
|
+
# already has converged.
|
|
111
|
+
MODEL_CONVERGED = 2
|
|
112
|
+
end
|
|
113
|
+
end
|
|
114
|
+
end
|
|
115
|
+
end
|
|
116
|
+
end
|
|
117
|
+
end
|
|
118
|
+
end
|
|
119
|
+
end
|
|
120
|
+
end
|