google-api-client 0.24.3 → 0.25.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.travis.yml +0 -2
- data/CHANGELOG.md +99 -0
- data/Gemfile +1 -0
- data/generated/google/apis/abusiveexperiencereport_v1.rb +2 -2
- data/generated/google/apis/abusiveexperiencereport_v1/classes.rb +0 -1
- data/generated/google/apis/abusiveexperiencereport_v1/service.rb +1 -1
- data/generated/google/apis/{content_v2sandbox.rb → accesscontextmanager_v1beta.rb} +11 -12
- data/generated/google/apis/accesscontextmanager_v1beta/classes.rb +697 -0
- data/generated/google/apis/accesscontextmanager_v1beta/representations.rb +240 -0
- data/generated/google/apis/accesscontextmanager_v1beta/service.rb +666 -0
- data/generated/google/apis/adexchangebuyer2_v2beta1.rb +1 -1
- data/generated/google/apis/adexchangebuyer2_v2beta1/classes.rb +16 -0
- data/generated/google/apis/adexchangebuyer2_v2beta1/representations.rb +2 -0
- data/generated/google/apis/adexchangebuyer2_v2beta1/service.rb +2 -2
- data/generated/google/apis/adexperiencereport_v1.rb +2 -2
- data/generated/google/apis/adexperiencereport_v1/service.rb +1 -1
- data/generated/google/apis/alertcenter_v1beta1.rb +4 -3
- data/generated/google/apis/alertcenter_v1beta1/classes.rb +9 -2
- data/generated/google/apis/alertcenter_v1beta1/representations.rb +1 -0
- data/generated/google/apis/alertcenter_v1beta1/service.rb +3 -2
- data/generated/google/apis/androiddeviceprovisioning_v1.rb +1 -1
- data/generated/google/apis/androiddeviceprovisioning_v1/service.rb +3 -2
- data/generated/google/apis/androidenterprise_v1.rb +1 -1
- data/generated/google/apis/androidenterprise_v1/classes.rb +68 -6
- data/generated/google/apis/androidenterprise_v1/representations.rb +20 -0
- data/generated/google/apis/androidmanagement_v1.rb +1 -1
- data/generated/google/apis/androidmanagement_v1/classes.rb +92 -0
- data/generated/google/apis/androidmanagement_v1/representations.rb +36 -0
- data/generated/google/apis/androidpublisher_v1.rb +2 -2
- data/generated/google/apis/androidpublisher_v1/service.rb +1 -1
- data/generated/google/apis/androidpublisher_v1_1.rb +2 -2
- data/generated/google/apis/androidpublisher_v1_1/service.rb +1 -1
- data/generated/google/apis/androidpublisher_v2.rb +2 -2
- data/generated/google/apis/androidpublisher_v2/service.rb +3 -2
- data/generated/google/apis/androidpublisher_v3.rb +2 -2
- data/generated/google/apis/androidpublisher_v3/service.rb +3 -2
- data/generated/google/apis/appengine_v1.rb +2 -3
- data/generated/google/apis/appengine_v1/service.rb +1 -2
- data/generated/google/apis/appengine_v1alpha.rb +2 -3
- data/generated/google/apis/appengine_v1alpha/service.rb +1 -2
- data/generated/google/apis/appengine_v1beta.rb +2 -3
- data/generated/google/apis/appengine_v1beta/service.rb +1 -2
- data/generated/google/apis/appengine_v1beta4.rb +2 -3
- data/generated/google/apis/appengine_v1beta4/service.rb +1 -2
- data/generated/google/apis/appengine_v1beta5.rb +2 -3
- data/generated/google/apis/appengine_v1beta5/service.rb +1 -2
- data/generated/google/apis/bigquery_v2.rb +1 -1
- data/generated/google/apis/bigquery_v2/classes.rb +98 -7
- data/generated/google/apis/bigquery_v2/representations.rb +35 -0
- data/generated/google/apis/bigquerydatatransfer_v1.rb +1 -1
- data/generated/google/apis/civicinfo_v2.rb +1 -1
- data/generated/google/apis/civicinfo_v2/classes.rb +18 -0
- data/generated/google/apis/civicinfo_v2/representations.rb +2 -0
- data/generated/google/apis/cloudasset_v1beta1.rb +1 -1
- data/generated/google/apis/cloudasset_v1beta1/classes.rb +5 -5
- data/generated/google/apis/cloudbuild_v1.rb +1 -1
- data/generated/google/apis/cloudbuild_v1/classes.rb +10 -1
- data/generated/google/apis/cloudbuild_v1/representations.rb +1 -0
- data/generated/google/apis/cloudfunctions_v1.rb +1 -1
- data/generated/google/apis/cloudfunctions_v1/classes.rb +418 -0
- data/generated/google/apis/cloudfunctions_v1/representations.rb +121 -0
- data/generated/google/apis/cloudfunctions_v1/service.rb +107 -0
- data/generated/google/apis/cloudfunctions_v1beta2.rb +1 -1
- data/generated/google/apis/cloudfunctions_v1beta2/classes.rb +16 -0
- data/generated/google/apis/cloudfunctions_v1beta2/representations.rb +1 -0
- data/generated/google/apis/cloudkms_v1.rb +1 -1
- data/generated/google/apis/cloudresourcemanager_v1.rb +3 -3
- data/generated/google/apis/cloudresourcemanager_v1/classes.rb +2 -1
- data/generated/google/apis/cloudresourcemanager_v1/service.rb +2 -2
- data/generated/google/apis/cloudresourcemanager_v1beta1.rb +3 -3
- data/generated/google/apis/cloudresourcemanager_v1beta1/classes.rb +2 -1
- data/generated/google/apis/cloudresourcemanager_v1beta1/service.rb +2 -2
- data/generated/google/apis/cloudresourcemanager_v2.rb +3 -3
- data/generated/google/apis/cloudresourcemanager_v2/service.rb +2 -2
- data/generated/google/apis/cloudresourcemanager_v2beta1.rb +3 -3
- data/generated/google/apis/cloudresourcemanager_v2beta1/service.rb +2 -2
- data/generated/google/apis/cloudsearch_v1.rb +59 -0
- data/generated/google/apis/cloudsearch_v1/classes.rb +4697 -0
- data/generated/google/apis/cloudsearch_v1/representations.rb +2151 -0
- data/generated/google/apis/cloudsearch_v1/service.rb +1351 -0
- data/generated/google/apis/cloudtasks_v2beta2.rb +1 -1
- data/generated/google/apis/cloudtasks_v2beta2/classes.rb +34 -0
- data/generated/google/apis/cloudtasks_v2beta3.rb +1 -1
- data/generated/google/apis/cloudtasks_v2beta3/classes.rb +34 -30
- data/generated/google/apis/composer_v1beta1.rb +1 -1
- data/generated/google/apis/composer_v1beta1/classes.rb +5 -8
- data/generated/google/apis/compute_alpha.rb +1 -1
- data/generated/google/apis/compute_beta.rb +1 -1
- data/generated/google/apis/compute_v1.rb +1 -1
- data/generated/google/apis/content_v2.rb +1 -1
- data/generated/google/apis/content_v2/classes.rb +197 -69
- data/generated/google/apis/content_v2/representations.rb +64 -33
- data/generated/google/apis/content_v2/service.rb +1 -1
- data/generated/google/apis/dataflow_v1b3.rb +1 -1
- data/generated/google/apis/dataflow_v1b3/classes.rb +42 -0
- data/generated/google/apis/dataflow_v1b3/representations.rb +17 -0
- data/generated/google/apis/dataproc_v1.rb +1 -1
- data/generated/google/apis/dataproc_v1/classes.rb +19 -19
- data/generated/google/apis/dataproc_v1beta2.rb +1 -1
- data/generated/google/apis/dataproc_v1beta2/classes.rb +19 -19
- data/generated/google/apis/dialogflow_v2.rb +6 -3
- data/generated/google/apis/dialogflow_v2/classes.rb +112 -8
- data/generated/google/apis/dialogflow_v2/representations.rb +30 -0
- data/generated/google/apis/dialogflow_v2/service.rb +7 -2
- data/generated/google/apis/dialogflow_v2beta1.rb +6 -3
- data/generated/google/apis/dialogflow_v2beta1/classes.rb +116 -8
- data/generated/google/apis/dialogflow_v2beta1/representations.rb +30 -0
- data/generated/google/apis/dialogflow_v2beta1/service.rb +19 -8
- data/generated/google/apis/digitalassetlinks_v1.rb +2 -3
- data/generated/google/apis/digitalassetlinks_v1/service.rb +1 -2
- data/generated/google/apis/dlp_v2.rb +1 -1
- data/generated/google/apis/dlp_v2/classes.rb +17 -5
- data/generated/google/apis/dlp_v2/representations.rb +2 -0
- data/generated/google/apis/dns_v1beta2.rb +1 -1
- data/generated/google/apis/dns_v1beta2/classes.rb +81 -0
- data/generated/google/apis/dns_v1beta2/representations.rb +34 -0
- data/generated/google/apis/doubleclicksearch_v2.rb +1 -1
- data/generated/google/apis/doubleclicksearch_v2/classes.rb +1 -1
- data/generated/google/apis/drive_v2.rb +1 -1
- data/generated/google/apis/drive_v2/classes.rb +3 -1
- data/generated/google/apis/firebaserules_v1.rb +1 -1
- data/generated/google/apis/firestore_v1.rb +1 -1
- data/generated/google/apis/firestore_v1/classes.rb +637 -0
- data/generated/google/apis/firestore_v1/representations.rb +260 -0
- data/generated/google/apis/firestore_v1/service.rb +377 -0
- data/generated/google/apis/firestore_v1beta2.rb +1 -1
- data/generated/google/apis/firestore_v1beta2/classes.rb +205 -0
- data/generated/google/apis/firestore_v1beta2/representations.rb +83 -0
- data/generated/google/apis/firestore_v1beta2/service.rb +79 -0
- data/generated/google/apis/genomics_v1.rb +2 -2
- data/generated/google/apis/genomics_v1/service.rb +1 -1
- data/generated/google/apis/genomics_v1alpha2.rb +2 -2
- data/generated/google/apis/genomics_v1alpha2/service.rb +1 -1
- data/generated/google/apis/genomics_v2alpha1.rb +2 -2
- data/generated/google/apis/genomics_v2alpha1/service.rb +1 -1
- data/generated/google/apis/iam_v1.rb +1 -1
- data/generated/google/apis/iam_v1/classes.rb +4 -3
- data/generated/google/apis/iamcredentials_v1.rb +2 -2
- data/generated/google/apis/iamcredentials_v1/service.rb +1 -1
- data/generated/google/apis/indexing_v3.rb +2 -3
- data/generated/google/apis/indexing_v3/service.rb +1 -2
- data/generated/google/apis/jobs_v2.rb +1 -1
- data/generated/google/apis/jobs_v2/classes.rb +2 -2
- data/generated/google/apis/jobs_v3.rb +1 -1
- data/generated/google/apis/jobs_v3/classes.rb +3 -3
- data/generated/google/apis/jobs_v3/service.rb +20 -2
- data/generated/google/apis/jobs_v3p1beta1.rb +1 -1
- data/generated/google/apis/jobs_v3p1beta1/classes.rb +3 -3
- data/generated/google/apis/jobs_v3p1beta1/service.rb +20 -2
- data/generated/google/apis/language_v1.rb +4 -4
- data/generated/google/apis/language_v1/service.rb +3 -3
- data/generated/google/apis/language_v1beta1.rb +4 -4
- data/generated/google/apis/language_v1beta1/service.rb +3 -3
- data/generated/google/apis/language_v1beta2.rb +4 -4
- data/generated/google/apis/language_v1beta2/service.rb +3 -3
- data/generated/google/apis/logging_v2.rb +1 -1
- data/generated/google/apis/logging_v2/classes.rb +2 -13
- data/generated/google/apis/logging_v2/representations.rb +0 -2
- data/generated/google/apis/logging_v2beta1.rb +1 -1
- data/generated/google/apis/logging_v2beta1/classes.rb +2 -41
- data/generated/google/apis/logging_v2beta1/representations.rb +0 -16
- data/generated/google/apis/logging_v2beta1/service.rb +0 -255
- data/generated/google/apis/monitoring_v3.rb +1 -1
- data/generated/google/apis/monitoring_v3/classes.rb +25 -15
- data/generated/google/apis/monitoring_v3/representations.rb +1 -0
- data/generated/google/apis/monitoring_v3/service.rb +366 -1
- data/generated/google/apis/oslogin_v1.rb +1 -1
- data/generated/google/apis/oslogin_v1/service.rb +7 -1
- data/generated/google/apis/oslogin_v1alpha.rb +1 -1
- data/generated/google/apis/oslogin_v1alpha/service.rb +10 -1
- data/generated/google/apis/oslogin_v1beta.rb +1 -1
- data/generated/google/apis/oslogin_v1beta/service.rb +7 -1
- data/generated/google/apis/redis_v1.rb +1 -1
- data/generated/google/apis/redis_v1/classes.rb +6 -0
- data/generated/google/apis/redis_v1/representations.rb +1 -0
- data/generated/google/apis/redis_v1beta1.rb +1 -1
- data/generated/google/apis/redis_v1beta1/classes.rb +6 -0
- data/generated/google/apis/redis_v1beta1/representations.rb +1 -0
- data/generated/google/apis/runtimeconfig_v1beta1.rb +1 -1
- data/generated/google/apis/runtimeconfig_v1beta1/classes.rb +7 -6
- data/generated/google/apis/script_v1.rb +2 -2
- data/generated/google/apis/script_v1/service.rb +7 -3
- data/generated/google/apis/serviceconsumermanagement_v1.rb +2 -3
- data/generated/google/apis/serviceconsumermanagement_v1/classes.rb +7 -0
- data/generated/google/apis/serviceconsumermanagement_v1/representations.rb +1 -0
- data/generated/google/apis/serviceconsumermanagement_v1/service.rb +1 -2
- data/generated/google/apis/servicecontrol_v1.rb +3 -3
- data/generated/google/apis/servicecontrol_v1/classes.rb +18 -230
- data/generated/google/apis/servicecontrol_v1/representations.rb +2 -99
- data/generated/google/apis/servicecontrol_v1/service.rb +2 -140
- data/generated/google/apis/servicemanagement_v1.rb +1 -1
- data/generated/google/apis/servicemanagement_v1/classes.rb +1 -1
- data/generated/google/apis/servicenetworking_v1beta.rb +1 -1
- data/generated/google/apis/servicenetworking_v1beta/classes.rb +24 -0
- data/generated/google/apis/servicenetworking_v1beta/representations.rb +3 -0
- data/generated/google/apis/serviceusage_v1.rb +1 -1
- data/generated/google/apis/serviceusage_v1/classes.rb +7 -0
- data/generated/google/apis/serviceusage_v1/representations.rb +1 -0
- data/generated/google/apis/serviceusage_v1beta1.rb +1 -1
- data/generated/google/apis/serviceusage_v1beta1/classes.rb +7 -0
- data/generated/google/apis/serviceusage_v1beta1/representations.rb +1 -0
- data/generated/google/apis/slides_v1.rb +2 -2
- data/generated/google/apis/slides_v1/classes.rb +4 -3
- data/generated/google/apis/slides_v1/service.rb +1 -1
- data/generated/google/apis/storage_v1.rb +1 -1
- data/generated/google/apis/storage_v1/service.rb +2 -3
- data/generated/google/apis/testing_v1.rb +1 -1
- data/generated/google/apis/testing_v1/classes.rb +9 -0
- data/generated/google/apis/testing_v1/representations.rb +1 -0
- data/generated/google/apis/toolresults_v1beta3.rb +1 -1
- data/generated/google/apis/toolresults_v1beta3/service.rb +1 -1
- data/generated/google/apis/tpu_v1.rb +1 -1
- data/generated/google/apis/tpu_v1/classes.rb +6 -0
- data/generated/google/apis/tpu_v1/representations.rb +1 -0
- data/generated/google/apis/tpu_v1alpha1.rb +1 -1
- data/generated/google/apis/tpu_v1alpha1/classes.rb +6 -0
- data/generated/google/apis/tpu_v1alpha1/representations.rb +1 -0
- data/generated/google/apis/vault_v1.rb +1 -1
- data/generated/google/apis/vault_v1/classes.rb +7 -0
- data/generated/google/apis/vault_v1/representations.rb +1 -0
- data/generated/google/apis/videointelligence_v1.rb +3 -2
- data/generated/google/apis/videointelligence_v1/classes.rb +547 -12
- data/generated/google/apis/videointelligence_v1/representations.rb +190 -2
- data/generated/google/apis/videointelligence_v1/service.rb +2 -1
- data/generated/google/apis/videointelligence_v1beta2.rb +3 -2
- data/generated/google/apis/videointelligence_v1beta2/classes.rb +547 -12
- data/generated/google/apis/videointelligence_v1beta2/representations.rb +190 -2
- data/generated/google/apis/videointelligence_v1beta2/service.rb +2 -1
- data/generated/google/apis/videointelligence_v1p1beta1.rb +3 -2
- data/generated/google/apis/videointelligence_v1p1beta1/classes.rb +418 -13
- data/generated/google/apis/videointelligence_v1p1beta1/representations.rb +153 -2
- data/generated/google/apis/videointelligence_v1p1beta1/service.rb +2 -1
- data/generated/google/apis/vision_v1.rb +1 -1
- data/generated/google/apis/vision_v1/classes.rb +1139 -54
- data/generated/google/apis/vision_v1/representations.rb +478 -0
- data/generated/google/apis/vision_v1/service.rb +744 -0
- data/generated/google/apis/vision_v1p1beta1.rb +1 -1
- data/generated/google/apis/vision_v1p1beta1/classes.rb +761 -16
- data/generated/google/apis/vision_v1p1beta1/representations.rb +333 -0
- data/generated/google/apis/vision_v1p2beta1.rb +1 -1
- data/generated/google/apis/vision_v1p2beta1/classes.rb +758 -13
- data/generated/google/apis/vision_v1p2beta1/representations.rb +333 -0
- data/generated/google/apis/websecurityscanner_v1alpha.rb +2 -2
- data/generated/google/apis/websecurityscanner_v1alpha/service.rb +1 -1
- data/generated/google/apis/youtube_partner_v1.rb +1 -1
- data/generated/google/apis/youtube_v3.rb +1 -1
- data/lib/google/apis/core/http_command.rb +83 -10
- data/lib/google/apis/options.rb +5 -1
- data/lib/google/apis/version.rb +1 -1
- metadata +10 -14
- data/generated/google/apis/content_v2sandbox/classes.rb +0 -4436
- data/generated/google/apis/content_v2sandbox/representations.rb +0 -1956
- data/generated/google/apis/content_v2sandbox/service.rb +0 -1296
- data/generated/google/apis/dfareporting_v2_8.rb +0 -40
- data/generated/google/apis/dfareporting_v2_8/classes.rb +0 -12043
- data/generated/google/apis/dfareporting_v2_8/representations.rb +0 -4311
- data/generated/google/apis/dfareporting_v2_8/service.rb +0 -8727
- data/generated/google/apis/serviceuser_v1.rb +0 -42
- data/generated/google/apis/serviceuser_v1/classes.rb +0 -3472
- data/generated/google/apis/serviceuser_v1/representations.rb +0 -1030
- data/generated/google/apis/serviceuser_v1/service.rb +0 -215
|
@@ -94,6 +94,30 @@ module Google
|
|
|
94
94
|
include Google::Apis::Core::JsonObjectSupport
|
|
95
95
|
end
|
|
96
96
|
|
|
97
|
+
class GoogleCloudVideointelligenceV1SpeechContext
|
|
98
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
99
|
+
|
|
100
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
101
|
+
end
|
|
102
|
+
|
|
103
|
+
class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative
|
|
104
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
105
|
+
|
|
106
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
107
|
+
end
|
|
108
|
+
|
|
109
|
+
class GoogleCloudVideointelligenceV1SpeechTranscription
|
|
110
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
111
|
+
|
|
112
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
113
|
+
end
|
|
114
|
+
|
|
115
|
+
class GoogleCloudVideointelligenceV1SpeechTranscriptionConfig
|
|
116
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
117
|
+
|
|
118
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
119
|
+
end
|
|
120
|
+
|
|
97
121
|
class GoogleCloudVideointelligenceV1VideoAnnotationProgress
|
|
98
122
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
99
123
|
|
|
@@ -118,6 +142,12 @@ module Google
|
|
|
118
142
|
include Google::Apis::Core::JsonObjectSupport
|
|
119
143
|
end
|
|
120
144
|
|
|
145
|
+
class GoogleCloudVideointelligenceV1WordInfo
|
|
146
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
147
|
+
|
|
148
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
149
|
+
end
|
|
150
|
+
|
|
121
151
|
class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress
|
|
122
152
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
123
153
|
|
|
@@ -166,6 +196,18 @@ module Google
|
|
|
166
196
|
include Google::Apis::Core::JsonObjectSupport
|
|
167
197
|
end
|
|
168
198
|
|
|
199
|
+
class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative
|
|
200
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
201
|
+
|
|
202
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
203
|
+
end
|
|
204
|
+
|
|
205
|
+
class GoogleCloudVideointelligenceV1beta2SpeechTranscription
|
|
206
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
207
|
+
|
|
208
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
209
|
+
end
|
|
210
|
+
|
|
169
211
|
class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress
|
|
170
212
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
171
213
|
|
|
@@ -184,6 +226,12 @@ module Google
|
|
|
184
226
|
include Google::Apis::Core::JsonObjectSupport
|
|
185
227
|
end
|
|
186
228
|
|
|
229
|
+
class GoogleCloudVideointelligenceV1beta2WordInfo
|
|
230
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
231
|
+
|
|
232
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
233
|
+
end
|
|
234
|
+
|
|
187
235
|
class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoProgress
|
|
188
236
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
189
237
|
|
|
@@ -346,6 +394,18 @@ module Google
|
|
|
346
394
|
include Google::Apis::Core::JsonObjectSupport
|
|
347
395
|
end
|
|
348
396
|
|
|
397
|
+
class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative
|
|
398
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
399
|
+
|
|
400
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
401
|
+
end
|
|
402
|
+
|
|
403
|
+
class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription
|
|
404
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
405
|
+
|
|
406
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
407
|
+
end
|
|
408
|
+
|
|
349
409
|
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
|
350
410
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
351
411
|
|
|
@@ -382,6 +442,12 @@ module Google
|
|
|
382
442
|
include Google::Apis::Core::JsonObjectSupport
|
|
383
443
|
end
|
|
384
444
|
|
|
445
|
+
class GoogleCloudVideointelligenceV1p2beta1WordInfo
|
|
446
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
447
|
+
|
|
448
|
+
include Google::Apis::Core::JsonObjectSupport
|
|
449
|
+
end
|
|
450
|
+
|
|
385
451
|
class GoogleLongrunningCancelOperationRequest
|
|
386
452
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
|
387
453
|
|
|
@@ -520,6 +586,48 @@ module Google
|
|
|
520
586
|
end
|
|
521
587
|
end
|
|
522
588
|
|
|
589
|
+
class GoogleCloudVideointelligenceV1SpeechContext
|
|
590
|
+
# @private
|
|
591
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
592
|
+
collection :phrases, as: 'phrases'
|
|
593
|
+
end
|
|
594
|
+
end
|
|
595
|
+
|
|
596
|
+
class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative
|
|
597
|
+
# @private
|
|
598
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
599
|
+
property :confidence, as: 'confidence'
|
|
600
|
+
property :transcript, as: 'transcript'
|
|
601
|
+
collection :words, as: 'words', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1WordInfo, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1WordInfo::Representation
|
|
602
|
+
|
|
603
|
+
end
|
|
604
|
+
end
|
|
605
|
+
|
|
606
|
+
class GoogleCloudVideointelligenceV1SpeechTranscription
|
|
607
|
+
# @private
|
|
608
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
609
|
+
collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechRecognitionAlternative::Representation
|
|
610
|
+
|
|
611
|
+
property :language_code, as: 'languageCode'
|
|
612
|
+
end
|
|
613
|
+
end
|
|
614
|
+
|
|
615
|
+
class GoogleCloudVideointelligenceV1SpeechTranscriptionConfig
|
|
616
|
+
# @private
|
|
617
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
618
|
+
collection :audio_tracks, as: 'audioTracks'
|
|
619
|
+
property :diarization_speaker_count, as: 'diarizationSpeakerCount'
|
|
620
|
+
property :enable_automatic_punctuation, as: 'enableAutomaticPunctuation'
|
|
621
|
+
property :enable_speaker_diarization, as: 'enableSpeakerDiarization'
|
|
622
|
+
property :enable_word_confidence, as: 'enableWordConfidence'
|
|
623
|
+
property :filter_profanity, as: 'filterProfanity'
|
|
624
|
+
property :language_code, as: 'languageCode'
|
|
625
|
+
property :max_alternatives, as: 'maxAlternatives'
|
|
626
|
+
collection :speech_contexts, as: 'speechContexts', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechContext, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechContext::Representation
|
|
627
|
+
|
|
628
|
+
end
|
|
629
|
+
end
|
|
630
|
+
|
|
523
631
|
class GoogleCloudVideointelligenceV1VideoAnnotationProgress
|
|
524
632
|
# @private
|
|
525
633
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
@@ -546,6 +654,8 @@ module Google
|
|
|
546
654
|
|
|
547
655
|
collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1LabelAnnotation::Representation
|
|
548
656
|
|
|
657
|
+
collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechTranscription, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechTranscription::Representation
|
|
658
|
+
|
|
549
659
|
end
|
|
550
660
|
end
|
|
551
661
|
|
|
@@ -560,6 +670,8 @@ module Google
|
|
|
560
670
|
|
|
561
671
|
property :shot_change_detection_config, as: 'shotChangeDetectionConfig', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1ShotChangeDetectionConfig, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1ShotChangeDetectionConfig::Representation
|
|
562
672
|
|
|
673
|
+
property :speech_transcription_config, as: 'speechTranscriptionConfig', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechTranscriptionConfig, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1SpeechTranscriptionConfig::Representation
|
|
674
|
+
|
|
563
675
|
end
|
|
564
676
|
end
|
|
565
677
|
|
|
@@ -571,6 +683,17 @@ module Google
|
|
|
571
683
|
end
|
|
572
684
|
end
|
|
573
685
|
|
|
686
|
+
class GoogleCloudVideointelligenceV1WordInfo
|
|
687
|
+
# @private
|
|
688
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
689
|
+
property :confidence, as: 'confidence'
|
|
690
|
+
property :end_time, as: 'endTime'
|
|
691
|
+
property :speaker_tag, as: 'speakerTag'
|
|
692
|
+
property :start_time, as: 'startTime'
|
|
693
|
+
property :word, as: 'word'
|
|
694
|
+
end
|
|
695
|
+
end
|
|
696
|
+
|
|
574
697
|
class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress
|
|
575
698
|
# @private
|
|
576
699
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
@@ -643,6 +766,25 @@ module Google
|
|
|
643
766
|
end
|
|
644
767
|
end
|
|
645
768
|
|
|
769
|
+
class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative
|
|
770
|
+
# @private
|
|
771
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
772
|
+
property :confidence, as: 'confidence'
|
|
773
|
+
property :transcript, as: 'transcript'
|
|
774
|
+
collection :words, as: 'words', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2WordInfo, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2WordInfo::Representation
|
|
775
|
+
|
|
776
|
+
end
|
|
777
|
+
end
|
|
778
|
+
|
|
779
|
+
class GoogleCloudVideointelligenceV1beta2SpeechTranscription
|
|
780
|
+
# @private
|
|
781
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
782
|
+
collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative::Representation
|
|
783
|
+
|
|
784
|
+
property :language_code, as: 'languageCode'
|
|
785
|
+
end
|
|
786
|
+
end
|
|
787
|
+
|
|
646
788
|
class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress
|
|
647
789
|
# @private
|
|
648
790
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
@@ -669,6 +811,8 @@ module Google
|
|
|
669
811
|
|
|
670
812
|
collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2LabelAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2LabelAnnotation::Representation
|
|
671
813
|
|
|
814
|
+
collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2SpeechTranscription, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1beta2SpeechTranscription::Representation
|
|
815
|
+
|
|
672
816
|
end
|
|
673
817
|
end
|
|
674
818
|
|
|
@@ -680,6 +824,17 @@ module Google
|
|
|
680
824
|
end
|
|
681
825
|
end
|
|
682
826
|
|
|
827
|
+
class GoogleCloudVideointelligenceV1beta2WordInfo
|
|
828
|
+
# @private
|
|
829
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
830
|
+
property :confidence, as: 'confidence'
|
|
831
|
+
property :end_time, as: 'endTime'
|
|
832
|
+
property :speaker_tag, as: 'speakerTag'
|
|
833
|
+
property :start_time, as: 'startTime'
|
|
834
|
+
property :word, as: 'word'
|
|
835
|
+
end
|
|
836
|
+
end
|
|
837
|
+
|
|
683
838
|
class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoProgress
|
|
684
839
|
# @private
|
|
685
840
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
@@ -767,6 +922,7 @@ module Google
|
|
|
767
922
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
768
923
|
collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative::Representation
|
|
769
924
|
|
|
925
|
+
property :language_code, as: 'languageCode'
|
|
770
926
|
end
|
|
771
927
|
end
|
|
772
928
|
|
|
@@ -940,12 +1096,28 @@ module Google
|
|
|
940
1096
|
end
|
|
941
1097
|
end
|
|
942
1098
|
|
|
943
|
-
class
|
|
1099
|
+
class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative
|
|
944
1100
|
# @private
|
|
945
1101
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
946
1102
|
property :confidence, as: 'confidence'
|
|
947
|
-
|
|
1103
|
+
property :transcript, as: 'transcript'
|
|
1104
|
+
collection :words, as: 'words', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1WordInfo, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1WordInfo::Representation
|
|
948
1105
|
|
|
1106
|
+
end
|
|
1107
|
+
end
|
|
1108
|
+
|
|
1109
|
+
class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription
|
|
1110
|
+
# @private
|
|
1111
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
1112
|
+
collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative::Representation
|
|
1113
|
+
|
|
1114
|
+
property :language_code, as: 'languageCode'
|
|
1115
|
+
end
|
|
1116
|
+
end
|
|
1117
|
+
|
|
1118
|
+
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
|
1119
|
+
# @private
|
|
1120
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
949
1121
|
collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextSegment::Representation
|
|
950
1122
|
|
|
951
1123
|
property :text, as: 'text'
|
|
@@ -964,6 +1136,9 @@ module Google
|
|
|
964
1136
|
class GoogleCloudVideointelligenceV1p2beta1TextSegment
|
|
965
1137
|
# @private
|
|
966
1138
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
1139
|
+
property :confidence, as: 'confidence'
|
|
1140
|
+
collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextFrame, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextFrame::Representation
|
|
1141
|
+
|
|
967
1142
|
property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation
|
|
968
1143
|
|
|
969
1144
|
end
|
|
@@ -997,6 +1172,8 @@ module Google
|
|
|
997
1172
|
|
|
998
1173
|
collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation
|
|
999
1174
|
|
|
1175
|
+
collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscription, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscription::Representation
|
|
1176
|
+
|
|
1000
1177
|
collection :text_annotations, as: 'textAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation::Representation
|
|
1001
1178
|
|
|
1002
1179
|
end
|
|
@@ -1010,6 +1187,17 @@ module Google
|
|
|
1010
1187
|
end
|
|
1011
1188
|
end
|
|
1012
1189
|
|
|
1190
|
+
class GoogleCloudVideointelligenceV1p2beta1WordInfo
|
|
1191
|
+
# @private
|
|
1192
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
|
1193
|
+
property :confidence, as: 'confidence'
|
|
1194
|
+
property :end_time, as: 'endTime'
|
|
1195
|
+
property :speaker_tag, as: 'speakerTag'
|
|
1196
|
+
property :start_time, as: 'startTime'
|
|
1197
|
+
property :word, as: 'word'
|
|
1198
|
+
end
|
|
1199
|
+
end
|
|
1200
|
+
|
|
1013
1201
|
class GoogleLongrunningCancelOperationRequest
|
|
1014
1202
|
# @private
|
|
1015
1203
|
class Representation < Google::Apis::Core::JsonRepresentation
|
|
@@ -22,7 +22,8 @@ module Google
|
|
|
22
22
|
module VideointelligenceV1
|
|
23
23
|
# Cloud Video Intelligence API
|
|
24
24
|
#
|
|
25
|
-
#
|
|
25
|
+
# Detects objects, explicit content, and scene changes in videos. It also
|
|
26
|
+
# specifies the region for annotation and transcribes speech to text.
|
|
26
27
|
#
|
|
27
28
|
# @example
|
|
28
29
|
# require 'google/apis/videointelligence_v1'
|
|
@@ -20,12 +20,13 @@ module Google
|
|
|
20
20
|
module Apis
|
|
21
21
|
# Cloud Video Intelligence API
|
|
22
22
|
#
|
|
23
|
-
#
|
|
23
|
+
# Detects objects, explicit content, and scene changes in videos. It also
|
|
24
|
+
# specifies the region for annotation and transcribes speech to text.
|
|
24
25
|
#
|
|
25
26
|
# @see https://cloud.google.com/video-intelligence/docs/
|
|
26
27
|
module VideointelligenceV1beta2
|
|
27
28
|
VERSION = 'V1beta2'
|
|
28
|
-
REVISION = '
|
|
29
|
+
REVISION = '20181027'
|
|
29
30
|
|
|
30
31
|
# View and manage your data across Google Cloud Platform services
|
|
31
32
|
AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform'
|
|
@@ -235,6 +235,73 @@ module Google
|
|
|
235
235
|
end
|
|
236
236
|
end
|
|
237
237
|
|
|
238
|
+
# Alternative hypotheses (a.k.a. n-best list).
|
|
239
|
+
class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative
|
|
240
|
+
include Google::Apis::Core::Hashable
|
|
241
|
+
|
|
242
|
+
# The confidence estimate between 0.0 and 1.0. A higher number
|
|
243
|
+
# indicates an estimated greater likelihood that the recognized words are
|
|
244
|
+
# correct. This field is typically provided only for the top hypothesis, and
|
|
245
|
+
# only for `is_final=true` results. Clients should not rely on the
|
|
246
|
+
# `confidence` field as it is not guaranteed to be accurate or consistent.
|
|
247
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
|
248
|
+
# Corresponds to the JSON property `confidence`
|
|
249
|
+
# @return [Float]
|
|
250
|
+
attr_accessor :confidence
|
|
251
|
+
|
|
252
|
+
# Transcript text representing the words that the user spoke.
|
|
253
|
+
# Corresponds to the JSON property `transcript`
|
|
254
|
+
# @return [String]
|
|
255
|
+
attr_accessor :transcript
|
|
256
|
+
|
|
257
|
+
# A list of word-specific information for each recognized word.
|
|
258
|
+
# Corresponds to the JSON property `words`
|
|
259
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1WordInfo>]
|
|
260
|
+
attr_accessor :words
|
|
261
|
+
|
|
262
|
+
def initialize(**args)
|
|
263
|
+
update!(**args)
|
|
264
|
+
end
|
|
265
|
+
|
|
266
|
+
# Update properties of this object
|
|
267
|
+
def update!(**args)
|
|
268
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
269
|
+
@transcript = args[:transcript] if args.key?(:transcript)
|
|
270
|
+
@words = args[:words] if args.key?(:words)
|
|
271
|
+
end
|
|
272
|
+
end
|
|
273
|
+
|
|
274
|
+
# A speech recognition result corresponding to a portion of the audio.
|
|
275
|
+
class GoogleCloudVideointelligenceV1SpeechTranscription
|
|
276
|
+
include Google::Apis::Core::Hashable
|
|
277
|
+
|
|
278
|
+
# May contain one or more recognition hypotheses (up to the maximum specified
|
|
279
|
+
# in `max_alternatives`). These alternatives are ordered in terms of
|
|
280
|
+
# accuracy, with the top (first) alternative being the most probable, as
|
|
281
|
+
# ranked by the recognizer.
|
|
282
|
+
# Corresponds to the JSON property `alternatives`
|
|
283
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1SpeechRecognitionAlternative>]
|
|
284
|
+
attr_accessor :alternatives
|
|
285
|
+
|
|
286
|
+
# Output only. The
|
|
287
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the
|
|
288
|
+
# language in this result. This language code was detected to have the most
|
|
289
|
+
# likelihood of being spoken in the audio.
|
|
290
|
+
# Corresponds to the JSON property `languageCode`
|
|
291
|
+
# @return [String]
|
|
292
|
+
attr_accessor :language_code
|
|
293
|
+
|
|
294
|
+
def initialize(**args)
|
|
295
|
+
update!(**args)
|
|
296
|
+
end
|
|
297
|
+
|
|
298
|
+
# Update properties of this object
|
|
299
|
+
def update!(**args)
|
|
300
|
+
@alternatives = args[:alternatives] if args.key?(:alternatives)
|
|
301
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
|
302
|
+
end
|
|
303
|
+
end
|
|
304
|
+
|
|
238
305
|
# Annotation progress for a single video.
|
|
239
306
|
class GoogleCloudVideointelligenceV1VideoAnnotationProgress
|
|
240
307
|
include Google::Apis::Core::Hashable
|
|
@@ -357,6 +424,11 @@ module Google
|
|
|
357
424
|
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1LabelAnnotation>]
|
|
358
425
|
attr_accessor :shot_label_annotations
|
|
359
426
|
|
|
427
|
+
# Speech transcription.
|
|
428
|
+
# Corresponds to the JSON property `speechTranscriptions`
|
|
429
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1SpeechTranscription>]
|
|
430
|
+
attr_accessor :speech_transcriptions
|
|
431
|
+
|
|
360
432
|
def initialize(**args)
|
|
361
433
|
update!(**args)
|
|
362
434
|
end
|
|
@@ -370,6 +442,7 @@ module Google
|
|
|
370
442
|
@segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
|
|
371
443
|
@shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations)
|
|
372
444
|
@shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations)
|
|
445
|
+
@speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions)
|
|
373
446
|
end
|
|
374
447
|
end
|
|
375
448
|
|
|
@@ -400,6 +473,65 @@ module Google
|
|
|
400
473
|
end
|
|
401
474
|
end
|
|
402
475
|
|
|
476
|
+
# Word-specific information for recognized words. Word information is only
|
|
477
|
+
# included in the response when certain request parameters are set, such
|
|
478
|
+
# as `enable_word_time_offsets`.
|
|
479
|
+
class GoogleCloudVideointelligenceV1WordInfo
|
|
480
|
+
include Google::Apis::Core::Hashable
|
|
481
|
+
|
|
482
|
+
# Output only. The confidence estimate between 0.0 and 1.0. A higher number
|
|
483
|
+
# indicates an estimated greater likelihood that the recognized words are
|
|
484
|
+
# correct. This field is set only for the top alternative.
|
|
485
|
+
# This field is not guaranteed to be accurate and users should not rely on it
|
|
486
|
+
# to be always provided.
|
|
487
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
|
488
|
+
# Corresponds to the JSON property `confidence`
|
|
489
|
+
# @return [Float]
|
|
490
|
+
attr_accessor :confidence
|
|
491
|
+
|
|
492
|
+
# Time offset relative to the beginning of the audio, and
|
|
493
|
+
# corresponding to the end of the spoken word. This field is only set if
|
|
494
|
+
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
|
495
|
+
# experimental feature and the accuracy of the time offset can vary.
|
|
496
|
+
# Corresponds to the JSON property `endTime`
|
|
497
|
+
# @return [String]
|
|
498
|
+
attr_accessor :end_time
|
|
499
|
+
|
|
500
|
+
# Output only. A distinct integer value is assigned for every speaker within
|
|
501
|
+
# the audio. This field specifies which one of those speakers was detected to
|
|
502
|
+
# have spoken this word. Value ranges from 1 up to diarization_speaker_count,
|
|
503
|
+
# and is only set if speaker diarization is enabled.
|
|
504
|
+
# Corresponds to the JSON property `speakerTag`
|
|
505
|
+
# @return [Fixnum]
|
|
506
|
+
attr_accessor :speaker_tag
|
|
507
|
+
|
|
508
|
+
# Time offset relative to the beginning of the audio, and
|
|
509
|
+
# corresponding to the start of the spoken word. This field is only set if
|
|
510
|
+
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
|
511
|
+
# experimental feature and the accuracy of the time offset can vary.
|
|
512
|
+
# Corresponds to the JSON property `startTime`
|
|
513
|
+
# @return [String]
|
|
514
|
+
attr_accessor :start_time
|
|
515
|
+
|
|
516
|
+
# The word corresponding to this set of information.
|
|
517
|
+
# Corresponds to the JSON property `word`
|
|
518
|
+
# @return [String]
|
|
519
|
+
attr_accessor :word
|
|
520
|
+
|
|
521
|
+
def initialize(**args)
|
|
522
|
+
update!(**args)
|
|
523
|
+
end
|
|
524
|
+
|
|
525
|
+
# Update properties of this object
|
|
526
|
+
def update!(**args)
|
|
527
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
528
|
+
@end_time = args[:end_time] if args.key?(:end_time)
|
|
529
|
+
@speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag)
|
|
530
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
|
531
|
+
@word = args[:word] if args.key?(:word)
|
|
532
|
+
end
|
|
533
|
+
end
|
|
534
|
+
|
|
403
535
|
# Video annotation progress. Included in the `metadata`
|
|
404
536
|
# field of the `Operation` returned by the `GetOperation`
|
|
405
537
|
# call of the `google::longrunning::Operations` service.
|
|
@@ -761,6 +893,197 @@ module Google
|
|
|
761
893
|
end
|
|
762
894
|
end
|
|
763
895
|
|
|
896
|
+
# Provides "hints" to the speech recognizer to favor specific words and phrases
|
|
897
|
+
# in the results.
|
|
898
|
+
class GoogleCloudVideointelligenceV1beta2SpeechContext
|
|
899
|
+
include Google::Apis::Core::Hashable
|
|
900
|
+
|
|
901
|
+
# *Optional* A list of strings containing words and phrases "hints" so that
|
|
902
|
+
# the speech recognition is more likely to recognize them. This can be used
|
|
903
|
+
# to improve the accuracy for specific words and phrases, for example, if
|
|
904
|
+
# specific commands are typically spoken by the user. This can also be used
|
|
905
|
+
# to add additional words to the vocabulary of the recognizer. See
|
|
906
|
+
# [usage limits](https://cloud.google.com/speech/limits#content).
|
|
907
|
+
# Corresponds to the JSON property `phrases`
|
|
908
|
+
# @return [Array<String>]
|
|
909
|
+
attr_accessor :phrases
|
|
910
|
+
|
|
911
|
+
def initialize(**args)
|
|
912
|
+
update!(**args)
|
|
913
|
+
end
|
|
914
|
+
|
|
915
|
+
# Update properties of this object
|
|
916
|
+
def update!(**args)
|
|
917
|
+
@phrases = args[:phrases] if args.key?(:phrases)
|
|
918
|
+
end
|
|
919
|
+
end
|
|
920
|
+
|
|
921
|
+
# Alternative hypotheses (a.k.a. n-best list).
|
|
922
|
+
class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative
|
|
923
|
+
include Google::Apis::Core::Hashable
|
|
924
|
+
|
|
925
|
+
# The confidence estimate between 0.0 and 1.0. A higher number
|
|
926
|
+
# indicates an estimated greater likelihood that the recognized words are
|
|
927
|
+
# correct. This field is typically provided only for the top hypothesis, and
|
|
928
|
+
# only for `is_final=true` results. Clients should not rely on the
|
|
929
|
+
# `confidence` field as it is not guaranteed to be accurate or consistent.
|
|
930
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
|
931
|
+
# Corresponds to the JSON property `confidence`
|
|
932
|
+
# @return [Float]
|
|
933
|
+
attr_accessor :confidence
|
|
934
|
+
|
|
935
|
+
# Transcript text representing the words that the user spoke.
|
|
936
|
+
# Corresponds to the JSON property `transcript`
|
|
937
|
+
# @return [String]
|
|
938
|
+
attr_accessor :transcript
|
|
939
|
+
|
|
940
|
+
# A list of word-specific information for each recognized word.
|
|
941
|
+
# Corresponds to the JSON property `words`
|
|
942
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2WordInfo>]
|
|
943
|
+
attr_accessor :words
|
|
944
|
+
|
|
945
|
+
def initialize(**args)
|
|
946
|
+
update!(**args)
|
|
947
|
+
end
|
|
948
|
+
|
|
949
|
+
# Update properties of this object
|
|
950
|
+
def update!(**args)
|
|
951
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
952
|
+
@transcript = args[:transcript] if args.key?(:transcript)
|
|
953
|
+
@words = args[:words] if args.key?(:words)
|
|
954
|
+
end
|
|
955
|
+
end
|
|
956
|
+
|
|
957
|
+
# A speech recognition result corresponding to a portion of the audio.
|
|
958
|
+
class GoogleCloudVideointelligenceV1beta2SpeechTranscription
|
|
959
|
+
include Google::Apis::Core::Hashable
|
|
960
|
+
|
|
961
|
+
# May contain one or more recognition hypotheses (up to the maximum specified
|
|
962
|
+
# in `max_alternatives`). These alternatives are ordered in terms of
|
|
963
|
+
# accuracy, with the top (first) alternative being the most probable, as
|
|
964
|
+
# ranked by the recognizer.
|
|
965
|
+
# Corresponds to the JSON property `alternatives`
|
|
966
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative>]
|
|
967
|
+
attr_accessor :alternatives
|
|
968
|
+
|
|
969
|
+
# Output only. The
|
|
970
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the
|
|
971
|
+
# language in this result. This language code was detected to have the most
|
|
972
|
+
# likelihood of being spoken in the audio.
|
|
973
|
+
# Corresponds to the JSON property `languageCode`
|
|
974
|
+
# @return [String]
|
|
975
|
+
attr_accessor :language_code
|
|
976
|
+
|
|
977
|
+
def initialize(**args)
|
|
978
|
+
update!(**args)
|
|
979
|
+
end
|
|
980
|
+
|
|
981
|
+
# Update properties of this object
|
|
982
|
+
def update!(**args)
|
|
983
|
+
@alternatives = args[:alternatives] if args.key?(:alternatives)
|
|
984
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
|
985
|
+
end
|
|
986
|
+
end
|
|
987
|
+
|
|
988
|
+
# Config for SPEECH_TRANSCRIPTION.
|
|
989
|
+
class GoogleCloudVideointelligenceV1beta2SpeechTranscriptionConfig
|
|
990
|
+
include Google::Apis::Core::Hashable
|
|
991
|
+
|
|
992
|
+
# *Optional* For file formats, such as MXF or MKV, supporting multiple audio
|
|
993
|
+
# tracks, specify up to two tracks. Default: track 0.
|
|
994
|
+
# Corresponds to the JSON property `audioTracks`
|
|
995
|
+
# @return [Array<Fixnum>]
|
|
996
|
+
attr_accessor :audio_tracks
|
|
997
|
+
|
|
998
|
+
# *Optional*
|
|
999
|
+
# If set, specifies the estimated number of speakers in the conversation.
|
|
1000
|
+
# If not set, defaults to '2'.
|
|
1001
|
+
# Ignored unless enable_speaker_diarization is set to true.
|
|
1002
|
+
# Corresponds to the JSON property `diarizationSpeakerCount`
|
|
1003
|
+
# @return [Fixnum]
|
|
1004
|
+
attr_accessor :diarization_speaker_count
|
|
1005
|
+
|
|
1006
|
+
# *Optional* If 'true', adds punctuation to recognition result hypotheses.
|
|
1007
|
+
# This feature is only available in select languages. Setting this for
|
|
1008
|
+
# requests in other languages has no effect at all. The default 'false' value
|
|
1009
|
+
# does not add punctuation to result hypotheses. NOTE: "This is currently
|
|
1010
|
+
# offered as an experimental service, complimentary to all users. In the
|
|
1011
|
+
# future this may be exclusively available as a premium feature."
|
|
1012
|
+
# Corresponds to the JSON property `enableAutomaticPunctuation`
|
|
1013
|
+
# @return [Boolean]
|
|
1014
|
+
attr_accessor :enable_automatic_punctuation
|
|
1015
|
+
alias_method :enable_automatic_punctuation?, :enable_automatic_punctuation
|
|
1016
|
+
|
|
1017
|
+
# *Optional* If 'true', enables speaker detection for each recognized word in
|
|
1018
|
+
# the top alternative of the recognition result using a speaker_tag provided
|
|
1019
|
+
# in the WordInfo.
|
|
1020
|
+
# Note: When this is true, we send all the words from the beginning of the
|
|
1021
|
+
# audio for the top alternative in every consecutive responses.
|
|
1022
|
+
# This is done in order to improve our speaker tags as our models learn to
|
|
1023
|
+
# identify the speakers in the conversation over time.
|
|
1024
|
+
# Corresponds to the JSON property `enableSpeakerDiarization`
|
|
1025
|
+
# @return [Boolean]
|
|
1026
|
+
attr_accessor :enable_speaker_diarization
|
|
1027
|
+
alias_method :enable_speaker_diarization?, :enable_speaker_diarization
|
|
1028
|
+
|
|
1029
|
+
# *Optional* If `true`, the top result includes a list of words and the
|
|
1030
|
+
# confidence for those words. If `false`, no word-level confidence
|
|
1031
|
+
# information is returned. The default is `false`.
|
|
1032
|
+
# Corresponds to the JSON property `enableWordConfidence`
|
|
1033
|
+
# @return [Boolean]
|
|
1034
|
+
attr_accessor :enable_word_confidence
|
|
1035
|
+
alias_method :enable_word_confidence?, :enable_word_confidence
|
|
1036
|
+
|
|
1037
|
+
# *Optional* If set to `true`, the server will attempt to filter out
|
|
1038
|
+
# profanities, replacing all but the initial character in each filtered word
|
|
1039
|
+
# with asterisks, e.g. "f***". If set to `false` or omitted, profanities
|
|
1040
|
+
# won't be filtered out.
|
|
1041
|
+
# Corresponds to the JSON property `filterProfanity`
|
|
1042
|
+
# @return [Boolean]
|
|
1043
|
+
attr_accessor :filter_profanity
|
|
1044
|
+
alias_method :filter_profanity?, :filter_profanity
|
|
1045
|
+
|
|
1046
|
+
# *Required* The language of the supplied audio as a
|
|
1047
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
|
|
1048
|
+
# Example: "en-US".
|
|
1049
|
+
# See [Language Support](https://cloud.google.com/speech/docs/languages)
|
|
1050
|
+
# for a list of the currently supported language codes.
|
|
1051
|
+
# Corresponds to the JSON property `languageCode`
|
|
1052
|
+
# @return [String]
|
|
1053
|
+
attr_accessor :language_code
|
|
1054
|
+
|
|
1055
|
+
# *Optional* Maximum number of recognition hypotheses to be returned.
|
|
1056
|
+
# Specifically, the maximum number of `SpeechRecognitionAlternative` messages
|
|
1057
|
+
# within each `SpeechTranscription`. The server may return fewer than
|
|
1058
|
+
# `max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will
|
|
1059
|
+
# return a maximum of one. If omitted, will return a maximum of one.
|
|
1060
|
+
# Corresponds to the JSON property `maxAlternatives`
|
|
1061
|
+
# @return [Fixnum]
|
|
1062
|
+
attr_accessor :max_alternatives
|
|
1063
|
+
|
|
1064
|
+
# *Optional* A means to provide context to assist the speech recognition.
|
|
1065
|
+
# Corresponds to the JSON property `speechContexts`
|
|
1066
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2SpeechContext>]
|
|
1067
|
+
attr_accessor :speech_contexts
|
|
1068
|
+
|
|
1069
|
+
def initialize(**args)
|
|
1070
|
+
update!(**args)
|
|
1071
|
+
end
|
|
1072
|
+
|
|
1073
|
+
# Update properties of this object
|
|
1074
|
+
def update!(**args)
|
|
1075
|
+
@audio_tracks = args[:audio_tracks] if args.key?(:audio_tracks)
|
|
1076
|
+
@diarization_speaker_count = args[:diarization_speaker_count] if args.key?(:diarization_speaker_count)
|
|
1077
|
+
@enable_automatic_punctuation = args[:enable_automatic_punctuation] if args.key?(:enable_automatic_punctuation)
|
|
1078
|
+
@enable_speaker_diarization = args[:enable_speaker_diarization] if args.key?(:enable_speaker_diarization)
|
|
1079
|
+
@enable_word_confidence = args[:enable_word_confidence] if args.key?(:enable_word_confidence)
|
|
1080
|
+
@filter_profanity = args[:filter_profanity] if args.key?(:filter_profanity)
|
|
1081
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
|
1082
|
+
@max_alternatives = args[:max_alternatives] if args.key?(:max_alternatives)
|
|
1083
|
+
@speech_contexts = args[:speech_contexts] if args.key?(:speech_contexts)
|
|
1084
|
+
end
|
|
1085
|
+
end
|
|
1086
|
+
|
|
764
1087
|
# Annotation progress for a single video.
|
|
765
1088
|
class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress
|
|
766
1089
|
include Google::Apis::Core::Hashable
|
|
@@ -883,6 +1206,11 @@ module Google
|
|
|
883
1206
|
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2LabelAnnotation>]
|
|
884
1207
|
attr_accessor :shot_label_annotations
|
|
885
1208
|
|
|
1209
|
+
# Speech transcription.
|
|
1210
|
+
# Corresponds to the JSON property `speechTranscriptions`
|
|
1211
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2SpeechTranscription>]
|
|
1212
|
+
attr_accessor :speech_transcriptions
|
|
1213
|
+
|
|
886
1214
|
def initialize(**args)
|
|
887
1215
|
update!(**args)
|
|
888
1216
|
end
|
|
@@ -896,6 +1224,7 @@ module Google
|
|
|
896
1224
|
@segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
|
|
897
1225
|
@shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations)
|
|
898
1226
|
@shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations)
|
|
1227
|
+
@speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions)
|
|
899
1228
|
end
|
|
900
1229
|
end
|
|
901
1230
|
|
|
@@ -925,6 +1254,11 @@ module Google
|
|
|
925
1254
|
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2ShotChangeDetectionConfig]
|
|
926
1255
|
attr_accessor :shot_change_detection_config
|
|
927
1256
|
|
|
1257
|
+
# Config for SPEECH_TRANSCRIPTION.
|
|
1258
|
+
# Corresponds to the JSON property `speechTranscriptionConfig`
|
|
1259
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1beta2SpeechTranscriptionConfig]
|
|
1260
|
+
attr_accessor :speech_transcription_config
|
|
1261
|
+
|
|
928
1262
|
def initialize(**args)
|
|
929
1263
|
update!(**args)
|
|
930
1264
|
end
|
|
@@ -935,6 +1269,7 @@ module Google
|
|
|
935
1269
|
@label_detection_config = args[:label_detection_config] if args.key?(:label_detection_config)
|
|
936
1270
|
@segments = args[:segments] if args.key?(:segments)
|
|
937
1271
|
@shot_change_detection_config = args[:shot_change_detection_config] if args.key?(:shot_change_detection_config)
|
|
1272
|
+
@speech_transcription_config = args[:speech_transcription_config] if args.key?(:speech_transcription_config)
|
|
938
1273
|
end
|
|
939
1274
|
end
|
|
940
1275
|
|
|
@@ -965,6 +1300,65 @@ module Google
|
|
|
965
1300
|
end
|
|
966
1301
|
end
|
|
967
1302
|
|
|
1303
|
+
# Word-specific information for recognized words. Word information is only
|
|
1304
|
+
# included in the response when certain request parameters are set, such
|
|
1305
|
+
# as `enable_word_time_offsets`.
|
|
1306
|
+
class GoogleCloudVideointelligenceV1beta2WordInfo
|
|
1307
|
+
include Google::Apis::Core::Hashable
|
|
1308
|
+
|
|
1309
|
+
# Output only. The confidence estimate between 0.0 and 1.0. A higher number
|
|
1310
|
+
# indicates an estimated greater likelihood that the recognized words are
|
|
1311
|
+
# correct. This field is set only for the top alternative.
|
|
1312
|
+
# This field is not guaranteed to be accurate and users should not rely on it
|
|
1313
|
+
# to be always provided.
|
|
1314
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
|
1315
|
+
# Corresponds to the JSON property `confidence`
|
|
1316
|
+
# @return [Float]
|
|
1317
|
+
attr_accessor :confidence
|
|
1318
|
+
|
|
1319
|
+
# Time offset relative to the beginning of the audio, and
|
|
1320
|
+
# corresponding to the end of the spoken word. This field is only set if
|
|
1321
|
+
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
|
1322
|
+
# experimental feature and the accuracy of the time offset can vary.
|
|
1323
|
+
# Corresponds to the JSON property `endTime`
|
|
1324
|
+
# @return [String]
|
|
1325
|
+
attr_accessor :end_time
|
|
1326
|
+
|
|
1327
|
+
# Output only. A distinct integer value is assigned for every speaker within
|
|
1328
|
+
# the audio. This field specifies which one of those speakers was detected to
|
|
1329
|
+
# have spoken this word. Value ranges from 1 up to diarization_speaker_count,
|
|
1330
|
+
# and is only set if speaker diarization is enabled.
|
|
1331
|
+
# Corresponds to the JSON property `speakerTag`
|
|
1332
|
+
# @return [Fixnum]
|
|
1333
|
+
attr_accessor :speaker_tag
|
|
1334
|
+
|
|
1335
|
+
# Time offset relative to the beginning of the audio, and
|
|
1336
|
+
# corresponding to the start of the spoken word. This field is only set if
|
|
1337
|
+
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
|
1338
|
+
# experimental feature and the accuracy of the time offset can vary.
|
|
1339
|
+
# Corresponds to the JSON property `startTime`
|
|
1340
|
+
# @return [String]
|
|
1341
|
+
attr_accessor :start_time
|
|
1342
|
+
|
|
1343
|
+
# The word corresponding to this set of information.
|
|
1344
|
+
# Corresponds to the JSON property `word`
|
|
1345
|
+
# @return [String]
|
|
1346
|
+
attr_accessor :word
|
|
1347
|
+
|
|
1348
|
+
def initialize(**args)
|
|
1349
|
+
update!(**args)
|
|
1350
|
+
end
|
|
1351
|
+
|
|
1352
|
+
# Update properties of this object
|
|
1353
|
+
def update!(**args)
|
|
1354
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
1355
|
+
@end_time = args[:end_time] if args.key?(:end_time)
|
|
1356
|
+
@speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag)
|
|
1357
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
|
1358
|
+
@word = args[:word] if args.key?(:word)
|
|
1359
|
+
end
|
|
1360
|
+
end
|
|
1361
|
+
|
|
968
1362
|
# Video annotation progress. Included in the `metadata`
|
|
969
1363
|
# field of the `Operation` returned by the `GetOperation`
|
|
970
1364
|
# call of the `google::longrunning::Operations` service.
|
|
@@ -1226,6 +1620,14 @@ module Google
|
|
|
1226
1620
|
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative>]
|
|
1227
1621
|
attr_accessor :alternatives
|
|
1228
1622
|
|
|
1623
|
+
# Output only. The
|
|
1624
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the
|
|
1625
|
+
# language in this result. This language code was detected to have the most
|
|
1626
|
+
# likelihood of being spoken in the audio.
|
|
1627
|
+
# Corresponds to the JSON property `languageCode`
|
|
1628
|
+
# @return [String]
|
|
1629
|
+
attr_accessor :language_code
|
|
1630
|
+
|
|
1229
1631
|
def initialize(**args)
|
|
1230
1632
|
update!(**args)
|
|
1231
1633
|
end
|
|
@@ -1233,6 +1635,7 @@ module Google
|
|
|
1233
1635
|
# Update properties of this object
|
|
1234
1636
|
def update!(**args)
|
|
1235
1637
|
@alternatives = args[:alternatives] if args.key?(:alternatives)
|
|
1638
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
|
1236
1639
|
end
|
|
1237
1640
|
end
|
|
1238
1641
|
|
|
@@ -1843,22 +2246,78 @@ module Google
|
|
|
1843
2246
|
end
|
|
1844
2247
|
end
|
|
1845
2248
|
|
|
1846
|
-
#
|
|
1847
|
-
|
|
1848
|
-
# detection.
|
|
1849
|
-
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
|
2249
|
+
# Alternative hypotheses (a.k.a. n-best list).
|
|
2250
|
+
class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative
|
|
1850
2251
|
include Google::Apis::Core::Hashable
|
|
1851
2252
|
|
|
1852
|
-
#
|
|
1853
|
-
#
|
|
2253
|
+
# The confidence estimate between 0.0 and 1.0. A higher number
|
|
2254
|
+
# indicates an estimated greater likelihood that the recognized words are
|
|
2255
|
+
# correct. This field is typically provided only for the top hypothesis, and
|
|
2256
|
+
# only for `is_final=true` results. Clients should not rely on the
|
|
2257
|
+
# `confidence` field as it is not guaranteed to be accurate or consistent.
|
|
2258
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
|
1854
2259
|
# Corresponds to the JSON property `confidence`
|
|
1855
2260
|
# @return [Float]
|
|
1856
2261
|
attr_accessor :confidence
|
|
1857
2262
|
|
|
1858
|
-
#
|
|
1859
|
-
# Corresponds to the JSON property `
|
|
1860
|
-
# @return [
|
|
1861
|
-
attr_accessor :
|
|
2263
|
+
# Transcript text representing the words that the user spoke.
|
|
2264
|
+
# Corresponds to the JSON property `transcript`
|
|
2265
|
+
# @return [String]
|
|
2266
|
+
attr_accessor :transcript
|
|
2267
|
+
|
|
2268
|
+
# A list of word-specific information for each recognized word.
|
|
2269
|
+
# Corresponds to the JSON property `words`
|
|
2270
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1WordInfo>]
|
|
2271
|
+
attr_accessor :words
|
|
2272
|
+
|
|
2273
|
+
def initialize(**args)
|
|
2274
|
+
update!(**args)
|
|
2275
|
+
end
|
|
2276
|
+
|
|
2277
|
+
# Update properties of this object
|
|
2278
|
+
def update!(**args)
|
|
2279
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
2280
|
+
@transcript = args[:transcript] if args.key?(:transcript)
|
|
2281
|
+
@words = args[:words] if args.key?(:words)
|
|
2282
|
+
end
|
|
2283
|
+
end
|
|
2284
|
+
|
|
2285
|
+
# A speech recognition result corresponding to a portion of the audio.
|
|
2286
|
+
class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription
|
|
2287
|
+
include Google::Apis::Core::Hashable
|
|
2288
|
+
|
|
2289
|
+
# May contain one or more recognition hypotheses (up to the maximum specified
|
|
2290
|
+
# in `max_alternatives`). These alternatives are ordered in terms of
|
|
2291
|
+
# accuracy, with the top (first) alternative being the most probable, as
|
|
2292
|
+
# ranked by the recognizer.
|
|
2293
|
+
# Corresponds to the JSON property `alternatives`
|
|
2294
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative>]
|
|
2295
|
+
attr_accessor :alternatives
|
|
2296
|
+
|
|
2297
|
+
# Output only. The
|
|
2298
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the
|
|
2299
|
+
# language in this result. This language code was detected to have the most
|
|
2300
|
+
# likelihood of being spoken in the audio.
|
|
2301
|
+
# Corresponds to the JSON property `languageCode`
|
|
2302
|
+
# @return [String]
|
|
2303
|
+
attr_accessor :language_code
|
|
2304
|
+
|
|
2305
|
+
def initialize(**args)
|
|
2306
|
+
update!(**args)
|
|
2307
|
+
end
|
|
2308
|
+
|
|
2309
|
+
# Update properties of this object
|
|
2310
|
+
def update!(**args)
|
|
2311
|
+
@alternatives = args[:alternatives] if args.key?(:alternatives)
|
|
2312
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
|
2313
|
+
end
|
|
2314
|
+
end
|
|
2315
|
+
|
|
2316
|
+
# Annotations related to one detected OCR text snippet. This will contain the
|
|
2317
|
+
# corresponding text, confidence value, and frame level information for each
|
|
2318
|
+
# detection.
|
|
2319
|
+
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
|
2320
|
+
include Google::Apis::Core::Hashable
|
|
1862
2321
|
|
|
1863
2322
|
# All video segments where OCR detected text appears.
|
|
1864
2323
|
# Corresponds to the JSON property `segments`
|
|
@@ -1876,8 +2335,6 @@ module Google
|
|
|
1876
2335
|
|
|
1877
2336
|
# Update properties of this object
|
|
1878
2337
|
def update!(**args)
|
|
1879
|
-
@confidence = args[:confidence] if args.key?(:confidence)
|
|
1880
|
-
@frames = args[:frames] if args.key?(:frames)
|
|
1881
2338
|
@segments = args[:segments] if args.key?(:segments)
|
|
1882
2339
|
@text = args[:text] if args.key?(:text)
|
|
1883
2340
|
end
|
|
@@ -1928,6 +2385,17 @@ module Google
|
|
|
1928
2385
|
class GoogleCloudVideointelligenceV1p2beta1TextSegment
|
|
1929
2386
|
include Google::Apis::Core::Hashable
|
|
1930
2387
|
|
|
2388
|
+
# Confidence for the track of detected text. It is calculated as the highest
|
|
2389
|
+
# over all frames where OCR detected text appears.
|
|
2390
|
+
# Corresponds to the JSON property `confidence`
|
|
2391
|
+
# @return [Float]
|
|
2392
|
+
attr_accessor :confidence
|
|
2393
|
+
|
|
2394
|
+
# Information related to the frames where OCR detected text appears.
|
|
2395
|
+
# Corresponds to the JSON property `frames`
|
|
2396
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1TextFrame>]
|
|
2397
|
+
attr_accessor :frames
|
|
2398
|
+
|
|
1931
2399
|
# Video segment.
|
|
1932
2400
|
# Corresponds to the JSON property `segment`
|
|
1933
2401
|
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
|
|
@@ -1939,6 +2407,8 @@ module Google
|
|
|
1939
2407
|
|
|
1940
2408
|
# Update properties of this object
|
|
1941
2409
|
def update!(**args)
|
|
2410
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
2411
|
+
@frames = args[:frames] if args.key?(:frames)
|
|
1942
2412
|
@segment = args[:segment] if args.key?(:segment)
|
|
1943
2413
|
end
|
|
1944
2414
|
end
|
|
@@ -2070,6 +2540,11 @@ module Google
|
|
|
2070
2540
|
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
|
|
2071
2541
|
attr_accessor :shot_label_annotations
|
|
2072
2542
|
|
|
2543
|
+
# Speech transcription.
|
|
2544
|
+
# Corresponds to the JSON property `speechTranscriptions`
|
|
2545
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1SpeechTranscription>]
|
|
2546
|
+
attr_accessor :speech_transcriptions
|
|
2547
|
+
|
|
2073
2548
|
# OCR text detection and tracking.
|
|
2074
2549
|
# Annotations for list of detected text snippets. Each will have list of
|
|
2075
2550
|
# frame information associated with it.
|
|
@@ -2091,6 +2566,7 @@ module Google
|
|
|
2091
2566
|
@segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
|
|
2092
2567
|
@shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations)
|
|
2093
2568
|
@shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations)
|
|
2569
|
+
@speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions)
|
|
2094
2570
|
@text_annotations = args[:text_annotations] if args.key?(:text_annotations)
|
|
2095
2571
|
end
|
|
2096
2572
|
end
|
|
@@ -2122,6 +2598,65 @@ module Google
|
|
|
2122
2598
|
end
|
|
2123
2599
|
end
|
|
2124
2600
|
|
|
2601
|
+
# Word-specific information for recognized words. Word information is only
|
|
2602
|
+
# included in the response when certain request parameters are set, such
|
|
2603
|
+
# as `enable_word_time_offsets`.
|
|
2604
|
+
class GoogleCloudVideointelligenceV1p2beta1WordInfo
|
|
2605
|
+
include Google::Apis::Core::Hashable
|
|
2606
|
+
|
|
2607
|
+
# Output only. The confidence estimate between 0.0 and 1.0. A higher number
|
|
2608
|
+
# indicates an estimated greater likelihood that the recognized words are
|
|
2609
|
+
# correct. This field is set only for the top alternative.
|
|
2610
|
+
# This field is not guaranteed to be accurate and users should not rely on it
|
|
2611
|
+
# to be always provided.
|
|
2612
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
|
2613
|
+
# Corresponds to the JSON property `confidence`
|
|
2614
|
+
# @return [Float]
|
|
2615
|
+
attr_accessor :confidence
|
|
2616
|
+
|
|
2617
|
+
# Time offset relative to the beginning of the audio, and
|
|
2618
|
+
# corresponding to the end of the spoken word. This field is only set if
|
|
2619
|
+
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
|
2620
|
+
# experimental feature and the accuracy of the time offset can vary.
|
|
2621
|
+
# Corresponds to the JSON property `endTime`
|
|
2622
|
+
# @return [String]
|
|
2623
|
+
attr_accessor :end_time
|
|
2624
|
+
|
|
2625
|
+
# Output only. A distinct integer value is assigned for every speaker within
|
|
2626
|
+
# the audio. This field specifies which one of those speakers was detected to
|
|
2627
|
+
# have spoken this word. Value ranges from 1 up to diarization_speaker_count,
|
|
2628
|
+
# and is only set if speaker diarization is enabled.
|
|
2629
|
+
# Corresponds to the JSON property `speakerTag`
|
|
2630
|
+
# @return [Fixnum]
|
|
2631
|
+
attr_accessor :speaker_tag
|
|
2632
|
+
|
|
2633
|
+
# Time offset relative to the beginning of the audio, and
|
|
2634
|
+
# corresponding to the start of the spoken word. This field is only set if
|
|
2635
|
+
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
|
2636
|
+
# experimental feature and the accuracy of the time offset can vary.
|
|
2637
|
+
# Corresponds to the JSON property `startTime`
|
|
2638
|
+
# @return [String]
|
|
2639
|
+
attr_accessor :start_time
|
|
2640
|
+
|
|
2641
|
+
# The word corresponding to this set of information.
|
|
2642
|
+
# Corresponds to the JSON property `word`
|
|
2643
|
+
# @return [String]
|
|
2644
|
+
attr_accessor :word
|
|
2645
|
+
|
|
2646
|
+
def initialize(**args)
|
|
2647
|
+
update!(**args)
|
|
2648
|
+
end
|
|
2649
|
+
|
|
2650
|
+
# Update properties of this object
|
|
2651
|
+
def update!(**args)
|
|
2652
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
|
2653
|
+
@end_time = args[:end_time] if args.key?(:end_time)
|
|
2654
|
+
@speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag)
|
|
2655
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
|
2656
|
+
@word = args[:word] if args.key?(:word)
|
|
2657
|
+
end
|
|
2658
|
+
end
|
|
2659
|
+
|
|
2125
2660
|
# This resource represents a long-running operation that is the result of a
|
|
2126
2661
|
# network API call.
|
|
2127
2662
|
class GoogleLongrunningOperation
|