google-api-client 0.25.0 → 0.26.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/{CONTRIBUTING.md → .github/CONTRIBUTING.md} +0 -0
- data/.github/ISSUE_TEMPLATE/bug_report.md +36 -0
- data/.github/ISSUE_TEMPLATE/feature_request.md +21 -0
- data/.github/ISSUE_TEMPLATE/support_request.md +7 -0
- data/.kokoro/build.bat +8 -0
- data/.kokoro/build.sh +36 -0
- data/.kokoro/common.cfg +22 -0
- data/.kokoro/continuous/common.cfg +20 -0
- data/.kokoro/continuous/linux.cfg +15 -0
- data/.kokoro/continuous/osx.cfg +3 -0
- data/.kokoro/continuous/windows.cfg +3 -0
- data/.kokoro/osx.sh +35 -0
- data/.kokoro/presubmit/common.cfg +19 -0
- data/.kokoro/presubmit/linux.cfg +14 -0
- data/.kokoro/presubmit/osx.cfg +3 -0
- data/.kokoro/presubmit/windows.cfg +3 -0
- data/.kokoro/trampoline.sh +24 -0
- data/.kokoro/windows.sh +32 -0
- data/CHANGELOG.md +83 -0
- data/Gemfile +2 -2
- data/generated/google/apis/adexchangebuyer2_v2beta1.rb +1 -1
- data/generated/google/apis/adexchangebuyer2_v2beta1/classes.rb +3 -1
- data/generated/google/apis/alertcenter_v1beta1.rb +5 -3
- data/generated/google/apis/alertcenter_v1beta1/classes.rb +110 -101
- data/generated/google/apis/alertcenter_v1beta1/representations.rb +13 -0
- data/generated/google/apis/alertcenter_v1beta1/service.rb +83 -76
- data/generated/google/apis/androiddeviceprovisioning_v1.rb +1 -1
- data/generated/google/apis/androiddeviceprovisioning_v1/classes.rb +7 -0
- data/generated/google/apis/androiddeviceprovisioning_v1/representations.rb +2 -0
- data/generated/google/apis/androiddeviceprovisioning_v1/service.rb +1 -1
- data/generated/google/apis/androidenterprise_v1.rb +1 -1
- data/generated/google/apis/androidenterprise_v1/classes.rb +129 -44
- data/generated/google/apis/androidenterprise_v1/representations.rb +48 -0
- data/generated/google/apis/androidenterprise_v1/service.rb +218 -0
- data/generated/google/apis/androidmanagement_v1.rb +1 -1
- data/generated/google/apis/androidmanagement_v1/classes.rb +1 -1
- data/generated/google/apis/androidpublisher_v3.rb +1 -1
- data/generated/google/apis/androidpublisher_v3/classes.rb +8 -0
- data/generated/google/apis/androidpublisher_v3/representations.rb +1 -0
- data/generated/google/apis/appengine_v1.rb +1 -1
- data/generated/google/apis/appengine_v1beta.rb +1 -1
- data/generated/google/apis/bigquery_v2.rb +1 -1
- data/generated/google/apis/bigquery_v2/classes.rb +87 -29
- data/generated/google/apis/bigquery_v2/representations.rb +21 -0
- data/generated/google/apis/bigquery_v2/service.rb +7 -9
- data/generated/google/apis/calendar_v3.rb +2 -2
- data/generated/google/apis/calendar_v3/classes.rb +14 -6
- data/generated/google/apis/classroom_v1.rb +1 -1
- data/generated/google/apis/classroom_v1/service.rb +1 -1
- data/generated/google/apis/cloudasset_v1beta1.rb +1 -1
- data/generated/google/apis/cloudasset_v1beta1/classes.rb +14 -10
- data/generated/google/apis/cloudasset_v1beta1/service.rb +4 -4
- data/generated/google/apis/clouddebugger_v2.rb +1 -1
- data/generated/google/apis/clouddebugger_v2/classes.rb +2 -2
- data/generated/google/apis/cloudfunctions_v1.rb +1 -1
- data/generated/google/apis/cloudfunctions_v1/classes.rb +2 -3
- data/generated/google/apis/cloudfunctions_v1beta2.rb +1 -1
- data/generated/google/apis/cloudfunctions_v1beta2/classes.rb +2 -3
- data/generated/google/apis/cloudiot_v1.rb +1 -1
- data/generated/google/apis/cloudiot_v1/classes.rb +127 -0
- data/generated/google/apis/cloudiot_v1/representations.rb +70 -0
- data/generated/google/apis/cloudiot_v1/service.rb +172 -2
- data/generated/google/apis/cloudkms_v1.rb +1 -1
- data/generated/google/apis/cloudkms_v1/service.rb +107 -0
- data/generated/google/apis/{cloudiot_v1beta1.rb → cloudscheduler_v1beta1.rb} +8 -12
- data/generated/google/apis/cloudscheduler_v1beta1/classes.rb +882 -0
- data/generated/google/apis/cloudscheduler_v1beta1/representations.rb +264 -0
- data/generated/google/apis/cloudscheduler_v1beta1/service.rb +445 -0
- data/generated/google/apis/cloudsearch_v1.rb +1 -1
- data/generated/google/apis/cloudsearch_v1/classes.rb +79 -1
- data/generated/google/apis/cloudsearch_v1/representations.rb +44 -0
- data/generated/google/apis/cloudsearch_v1/service.rb +88 -3
- data/generated/google/apis/composer_v1.rb +1 -1
- data/generated/google/apis/composer_v1/classes.rb +9 -0
- data/generated/google/apis/composer_v1/representations.rb +1 -0
- data/generated/google/apis/compute_alpha.rb +1 -1
- data/generated/google/apis/compute_alpha/classes.rb +938 -117
- data/generated/google/apis/compute_alpha/representations.rb +310 -2
- data/generated/google/apis/compute_alpha/service.rb +1203 -245
- data/generated/google/apis/compute_beta.rb +1 -1
- data/generated/google/apis/compute_beta/classes.rb +1175 -219
- data/generated/google/apis/compute_beta/representations.rb +375 -2
- data/generated/google/apis/compute_beta/service.rb +907 -139
- data/generated/google/apis/compute_v1.rb +1 -1
- data/generated/google/apis/compute_v1/classes.rb +1485 -143
- data/generated/google/apis/compute_v1/representations.rb +503 -0
- data/generated/google/apis/compute_v1/service.rb +1375 -85
- data/generated/google/apis/container_v1.rb +3 -3
- data/generated/google/apis/container_v1/classes.rb +99 -6
- data/generated/google/apis/container_v1/representations.rb +39 -0
- data/generated/google/apis/container_v1/service.rb +2 -2
- data/generated/google/apis/container_v1beta1.rb +3 -3
- data/generated/google/apis/container_v1beta1/classes.rb +162 -9
- data/generated/google/apis/container_v1beta1/representations.rb +80 -0
- data/generated/google/apis/container_v1beta1/service.rb +3 -3
- data/generated/google/apis/content_v2.rb +1 -1
- data/generated/google/apis/content_v2/classes.rb +40 -32
- data/generated/google/apis/content_v2/representations.rb +0 -1
- data/generated/google/apis/content_v2/service.rb +3 -3
- data/generated/google/apis/content_v2_1.rb +35 -0
- data/generated/google/apis/content_v2_1/classes.rb +9104 -0
- data/generated/google/apis/content_v2_1/representations.rb +3967 -0
- data/generated/google/apis/content_v2_1/service.rb +3463 -0
- data/generated/google/apis/dataflow_v1b3.rb +1 -1
- data/generated/google/apis/dataflow_v1b3/classes.rb +82 -0
- data/generated/google/apis/dataflow_v1b3/representations.rb +33 -0
- data/generated/google/apis/dataflow_v1b3/service.rb +97 -6
- data/generated/google/apis/dataproc_v1.rb +1 -1
- data/generated/google/apis/dataproc_v1/classes.rb +15 -1
- data/generated/google/apis/dataproc_v1/representations.rb +2 -0
- data/generated/google/apis/dataproc_v1beta2.rb +1 -1
- data/generated/google/apis/dataproc_v1beta2/classes.rb +125 -1
- data/generated/google/apis/dataproc_v1beta2/representations.rb +41 -0
- data/generated/google/apis/dialogflow_v2.rb +1 -1
- data/generated/google/apis/dialogflow_v2/classes.rb +13 -6
- data/generated/google/apis/dialogflow_v2/service.rb +2 -0
- data/generated/google/apis/dialogflow_v2beta1.rb +1 -1
- data/generated/google/apis/dialogflow_v2beta1/classes.rb +26 -6
- data/generated/google/apis/dialogflow_v2beta1/representations.rb +12 -0
- data/generated/google/apis/dialogflow_v2beta1/service.rb +261 -12
- data/generated/google/apis/dlp_v2.rb +1 -1
- data/generated/google/apis/dlp_v2/classes.rb +4 -4
- data/generated/google/apis/dns_v1beta2.rb +1 -1
- data/generated/google/apis/dns_v1beta2/classes.rb +326 -0
- data/generated/google/apis/dns_v1beta2/representations.rb +149 -0
- data/generated/google/apis/dns_v1beta2/service.rb +246 -0
- data/generated/google/apis/drive_v2.rb +3 -3
- data/generated/google/apis/drive_v3.rb +3 -3
- data/generated/google/apis/driveactivity_v2.rb +37 -0
- data/generated/google/apis/driveactivity_v2/classes.rb +1388 -0
- data/generated/google/apis/driveactivity_v2/representations.rb +799 -0
- data/generated/google/apis/driveactivity_v2/service.rb +89 -0
- data/generated/google/apis/firebasedynamiclinks_v1.rb +1 -1
- data/generated/google/apis/firebasedynamiclinks_v1/classes.rb +8 -1
- data/generated/google/apis/firebasedynamiclinks_v1/representations.rb +1 -0
- data/generated/google/apis/firebasehosting_v1beta1.rb +3 -2
- data/generated/google/apis/firebasehosting_v1beta1/classes.rb +43 -17
- data/generated/google/apis/firebasehosting_v1beta1/representations.rb +13 -0
- data/generated/google/apis/firebasehosting_v1beta1/service.rb +74 -1
- data/generated/google/apis/firestore_v1.rb +1 -1
- data/generated/google/apis/firestore_v1/classes.rb +2225 -520
- data/generated/google/apis/firestore_v1/representations.rb +800 -33
- data/generated/google/apis/firestore_v1/service.rb +526 -0
- data/generated/google/apis/fitness_v1.rb +1 -1
- data/generated/google/apis/fitness_v1/classes.rb +1 -2
- data/generated/google/apis/fitness_v1/service.rb +7 -3
- data/generated/google/apis/games_management_v1management.rb +1 -4
- data/generated/google/apis/games_v1.rb +1 -4
- data/generated/google/apis/genomics_v2alpha1.rb +1 -1
- data/generated/google/apis/genomics_v2alpha1/classes.rb +6 -1
- data/generated/google/apis/gmail_v1.rb +2 -2
- data/generated/google/apis/iap_v1.rb +34 -0
- data/generated/google/apis/iap_v1/classes.rb +308 -0
- data/generated/google/apis/iap_v1/representations.rb +126 -0
- data/generated/google/apis/iap_v1/service.rb +725 -0
- data/generated/google/apis/iap_v1beta1.rb +1 -1
- data/generated/google/apis/iap_v1beta1/classes.rb +0 -132
- data/generated/google/apis/iap_v1beta1/representations.rb +0 -32
- data/generated/google/apis/jobs_v2.rb +1 -1
- data/generated/google/apis/jobs_v2/classes.rb +8 -9
- data/generated/google/apis/jobs_v3.rb +1 -1
- data/generated/google/apis/jobs_v3/classes.rb +6 -6
- data/generated/google/apis/jobs_v3/service.rb +8 -8
- data/generated/google/apis/jobs_v3p1beta1.rb +1 -1
- data/generated/google/apis/jobs_v3p1beta1/classes.rb +9 -9
- data/generated/google/apis/jobs_v3p1beta1/service.rb +8 -8
- data/generated/google/apis/monitoring_v3.rb +1 -1
- data/generated/google/apis/monitoring_v3/classes.rb +20 -7
- data/generated/google/apis/monitoring_v3/representations.rb +1 -0
- data/generated/google/apis/monitoring_v3/service.rb +0 -365
- data/generated/google/apis/pagespeedonline_v5.rb +32 -0
- data/generated/google/apis/pagespeedonline_v5/classes.rb +724 -0
- data/generated/google/apis/pagespeedonline_v5/representations.rb +315 -0
- data/generated/google/apis/pagespeedonline_v5/service.rb +116 -0
- data/generated/google/apis/people_v1.rb +3 -3
- data/generated/google/apis/pubsub_v1.rb +1 -1
- data/generated/google/apis/pubsub_v1/classes.rb +63 -12
- data/generated/google/apis/pubsub_v1/representations.rb +15 -0
- data/generated/google/apis/pubsub_v1/service.rb +26 -18
- data/generated/google/apis/redis_v1.rb +2 -3
- data/generated/google/apis/redis_v1/service.rb +1 -2
- data/generated/google/apis/redis_v1beta1.rb +2 -3
- data/generated/google/apis/redis_v1beta1/service.rb +1 -2
- data/generated/google/apis/runtimeconfig_v1beta1.rb +1 -1
- data/generated/google/apis/runtimeconfig_v1beta1/classes.rb +2 -2
- data/generated/google/apis/script_v1.rb +24 -6
- data/generated/google/apis/serviceconsumermanagement_v1.rb +1 -1
- data/generated/google/apis/serviceconsumermanagement_v1/classes.rb +66 -1
- data/generated/google/apis/serviceconsumermanagement_v1/representations.rb +30 -0
- data/generated/google/apis/serviceconsumermanagement_v1/service.rb +89 -0
- data/generated/google/apis/servicecontrol_v1.rb +1 -1
- data/generated/google/apis/servicemanagement_v1.rb +1 -1
- data/generated/google/apis/servicemanagement_v1/classes.rb +7 -0
- data/generated/google/apis/servicemanagement_v1/representations.rb +1 -0
- data/generated/google/apis/servicenetworking_v1beta.rb +1 -1
- data/generated/google/apis/servicenetworking_v1beta/classes.rb +84 -37
- data/generated/google/apis/servicenetworking_v1beta/representations.rb +15 -0
- data/generated/google/apis/servicenetworking_v1beta/service.rb +132 -43
- data/generated/google/apis/serviceusage_v1.rb +1 -1
- data/generated/google/apis/serviceusage_v1/classes.rb +1 -1
- data/generated/google/apis/serviceusage_v1beta1.rb +1 -1
- data/generated/google/apis/serviceusage_v1beta1/classes.rb +1 -1
- data/generated/google/apis/sheets_v4.rb +4 -4
- data/generated/google/apis/slides_v1.rb +4 -4
- data/generated/google/apis/slides_v1/classes.rb +187 -2
- data/generated/google/apis/slides_v1/representations.rb +67 -0
- data/generated/google/apis/slides_v1/service.rb +4 -3
- data/generated/google/apis/sourcerepo_v1.rb +2 -2
- data/generated/google/apis/sourcerepo_v1/service.rb +1 -1
- data/generated/google/apis/speech_v1.rb +1 -1
- data/generated/google/apis/speech_v1/classes.rb +58 -10
- data/generated/google/apis/speech_v1/representations.rb +29 -1
- data/generated/google/apis/{speech_v1beta1.rb → speech_v1p1beta1.rb} +6 -6
- data/generated/google/apis/speech_v1p1beta1/classes.rb +922 -0
- data/generated/google/apis/speech_v1p1beta1/representations.rb +294 -0
- data/generated/google/apis/{speech_v1beta1 → speech_v1p1beta1}/service.rb +33 -35
- data/generated/google/apis/sqladmin_v1beta4.rb +1 -1
- data/generated/google/apis/sqladmin_v1beta4/classes.rb +64 -3
- data/generated/google/apis/sqladmin_v1beta4/representations.rb +31 -0
- data/generated/google/apis/storage_v1.rb +1 -1
- data/generated/google/apis/storage_v1/classes.rb +54 -0
- data/generated/google/apis/storage_v1/representations.rb +31 -0
- data/generated/google/apis/streetviewpublish_v1.rb +1 -1
- data/generated/google/apis/streetviewpublish_v1/classes.rb +26 -0
- data/generated/google/apis/streetviewpublish_v1/representations.rb +4 -0
- data/generated/google/apis/streetviewpublish_v1/service.rb +29 -3
- data/generated/google/apis/tasks_v1.rb +1 -1
- data/generated/google/apis/testing_v1.rb +1 -1
- data/generated/google/apis/testing_v1/classes.rb +49 -19
- data/generated/google/apis/testing_v1/representations.rb +20 -2
- data/generated/google/apis/toolresults_v1beta3.rb +1 -1
- data/generated/google/apis/toolresults_v1beta3/classes.rb +77 -70
- data/generated/google/apis/vault_v1.rb +1 -1
- data/generated/google/apis/vault_v1/classes.rb +78 -7
- data/generated/google/apis/vault_v1/representations.rb +34 -1
- data/generated/google/apis/vault_v1/service.rb +141 -0
- data/generated/google/apis/vision_v1.rb +1 -1
- data/generated/google/apis/vision_v1/classes.rb +33 -0
- data/generated/google/apis/vision_v1/representations.rb +16 -0
- data/generated/google/apis/vision_v1/service.rb +2 -2
- data/generated/google/apis/vision_v1p1beta1.rb +1 -1
- data/generated/google/apis/vision_v1p1beta1/classes.rb +33 -0
- data/generated/google/apis/vision_v1p1beta1/representations.rb +16 -0
- data/generated/google/apis/vision_v1p2beta1.rb +1 -1
- data/generated/google/apis/vision_v1p2beta1/classes.rb +33 -0
- data/generated/google/apis/vision_v1p2beta1/representations.rb +16 -0
- data/generated/google/apis/youtube_partner_v1.rb +1 -1
- data/google-api-client.gemspec +1 -1
- data/lib/google/api_client/auth/installed_app.rb +17 -2
- data/lib/google/api_client/client_secrets.rb +1 -1
- data/lib/google/apis/core/download.rb +2 -2
- data/lib/google/apis/errors.rb +9 -0
- data/lib/google/apis/generator/annotator.rb +2 -2
- data/lib/google/apis/version.rb +1 -1
- metadata +46 -17
- data/generated/google/apis/cloudiot_v1beta1/classes.rb +0 -959
- data/generated/google/apis/cloudiot_v1beta1/representations.rb +0 -375
- data/generated/google/apis/cloudiot_v1beta1/service.rb +0 -618
- data/generated/google/apis/dfareporting_v3_0.rb +0 -40
- data/generated/google/apis/dfareporting_v3_0/classes.rb +0 -12119
- data/generated/google/apis/dfareporting_v3_0/representations.rb +0 -4336
- data/generated/google/apis/dfareporting_v3_0/service.rb +0 -8701
- data/generated/google/apis/speech_v1beta1/classes.rb +0 -480
- data/generated/google/apis/speech_v1beta1/representations.rb +0 -194
@@ -280,6 +280,12 @@ module Google
|
|
280
280
|
include Google::Apis::Core::JsonObjectSupport
|
281
281
|
end
|
282
282
|
|
283
|
+
class LineConnection
|
284
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
285
|
+
|
286
|
+
include Google::Apis::Core::JsonObjectSupport
|
287
|
+
end
|
288
|
+
|
283
289
|
class LineFill
|
284
290
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
285
291
|
|
@@ -472,6 +478,12 @@ module Google
|
|
472
478
|
include Google::Apis::Core::JsonObjectSupport
|
473
479
|
end
|
474
480
|
|
481
|
+
class RerouteLineRequest
|
482
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
483
|
+
|
484
|
+
include Google::Apis::Core::JsonObjectSupport
|
485
|
+
end
|
486
|
+
|
475
487
|
class Response
|
476
488
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
477
489
|
|
@@ -682,6 +694,12 @@ module Google
|
|
682
694
|
include Google::Apis::Core::JsonObjectSupport
|
683
695
|
end
|
684
696
|
|
697
|
+
class UpdateLineCategoryRequest
|
698
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
699
|
+
|
700
|
+
include Google::Apis::Core::JsonObjectSupport
|
701
|
+
end
|
702
|
+
|
685
703
|
class UpdateLinePropertiesRequest
|
686
704
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
687
705
|
|
@@ -700,6 +718,12 @@ module Google
|
|
700
718
|
include Google::Apis::Core::JsonObjectSupport
|
701
719
|
end
|
702
720
|
|
721
|
+
class UpdatePageElementsZOrderRequest
|
722
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
723
|
+
|
724
|
+
include Google::Apis::Core::JsonObjectSupport
|
725
|
+
end
|
726
|
+
|
703
727
|
class UpdatePagePropertiesRequest
|
704
728
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
705
729
|
|
@@ -883,6 +907,7 @@ module Google
|
|
883
907
|
class CreateLineRequest
|
884
908
|
# @private
|
885
909
|
class Representation < Google::Apis::Core::JsonRepresentation
|
910
|
+
property :category, as: 'category'
|
886
911
|
property :element_properties, as: 'elementProperties', class: Google::Apis::SlidesV1::PageElementProperties, decorator: Google::Apis::SlidesV1::PageElementProperties::Representation
|
887
912
|
|
888
913
|
property :line_category, as: 'lineCategory'
|
@@ -1196,12 +1221,21 @@ module Google
|
|
1196
1221
|
class Line
|
1197
1222
|
# @private
|
1198
1223
|
class Representation < Google::Apis::Core::JsonRepresentation
|
1224
|
+
property :line_category, as: 'lineCategory'
|
1199
1225
|
property :line_properties, as: 'lineProperties', class: Google::Apis::SlidesV1::LineProperties, decorator: Google::Apis::SlidesV1::LineProperties::Representation
|
1200
1226
|
|
1201
1227
|
property :line_type, as: 'lineType'
|
1202
1228
|
end
|
1203
1229
|
end
|
1204
1230
|
|
1231
|
+
class LineConnection
|
1232
|
+
# @private
|
1233
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1234
|
+
property :connected_object_id, as: 'connectedObjectId'
|
1235
|
+
property :connection_site_index, as: 'connectionSiteIndex'
|
1236
|
+
end
|
1237
|
+
end
|
1238
|
+
|
1205
1239
|
class LineFill
|
1206
1240
|
# @private
|
1207
1241
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1215,11 +1249,15 @@ module Google
|
|
1215
1249
|
class Representation < Google::Apis::Core::JsonRepresentation
|
1216
1250
|
property :dash_style, as: 'dashStyle'
|
1217
1251
|
property :end_arrow, as: 'endArrow'
|
1252
|
+
property :end_connection, as: 'endConnection', class: Google::Apis::SlidesV1::LineConnection, decorator: Google::Apis::SlidesV1::LineConnection::Representation
|
1253
|
+
|
1218
1254
|
property :line_fill, as: 'lineFill', class: Google::Apis::SlidesV1::LineFill, decorator: Google::Apis::SlidesV1::LineFill::Representation
|
1219
1255
|
|
1220
1256
|
property :link, as: 'link', class: Google::Apis::SlidesV1::Link, decorator: Google::Apis::SlidesV1::Link::Representation
|
1221
1257
|
|
1222
1258
|
property :start_arrow, as: 'startArrow'
|
1259
|
+
property :start_connection, as: 'startConnection', class: Google::Apis::SlidesV1::LineConnection, decorator: Google::Apis::SlidesV1::LineConnection::Representation
|
1260
|
+
|
1223
1261
|
property :weight, as: 'weight', class: Google::Apis::SlidesV1::Dimension, decorator: Google::Apis::SlidesV1::Dimension::Representation
|
1224
1262
|
|
1225
1263
|
end
|
@@ -1593,18 +1631,24 @@ module Google
|
|
1593
1631
|
|
1594
1632
|
property :replace_image, as: 'replaceImage', class: Google::Apis::SlidesV1::ReplaceImageRequest, decorator: Google::Apis::SlidesV1::ReplaceImageRequest::Representation
|
1595
1633
|
|
1634
|
+
property :reroute_line, as: 'rerouteLine', class: Google::Apis::SlidesV1::RerouteLineRequest, decorator: Google::Apis::SlidesV1::RerouteLineRequest::Representation
|
1635
|
+
|
1596
1636
|
property :ungroup_objects, as: 'ungroupObjects', class: Google::Apis::SlidesV1::UngroupObjectsRequest, decorator: Google::Apis::SlidesV1::UngroupObjectsRequest::Representation
|
1597
1637
|
|
1598
1638
|
property :unmerge_table_cells, as: 'unmergeTableCells', class: Google::Apis::SlidesV1::UnmergeTableCellsRequest, decorator: Google::Apis::SlidesV1::UnmergeTableCellsRequest::Representation
|
1599
1639
|
|
1600
1640
|
property :update_image_properties, as: 'updateImageProperties', class: Google::Apis::SlidesV1::UpdateImagePropertiesRequest, decorator: Google::Apis::SlidesV1::UpdateImagePropertiesRequest::Representation
|
1601
1641
|
|
1642
|
+
property :update_line_category, as: 'updateLineCategory', class: Google::Apis::SlidesV1::UpdateLineCategoryRequest, decorator: Google::Apis::SlidesV1::UpdateLineCategoryRequest::Representation
|
1643
|
+
|
1602
1644
|
property :update_line_properties, as: 'updateLineProperties', class: Google::Apis::SlidesV1::UpdateLinePropertiesRequest, decorator: Google::Apis::SlidesV1::UpdateLinePropertiesRequest::Representation
|
1603
1645
|
|
1604
1646
|
property :update_page_element_alt_text, as: 'updatePageElementAltText', class: Google::Apis::SlidesV1::UpdatePageElementAltTextRequest, decorator: Google::Apis::SlidesV1::UpdatePageElementAltTextRequest::Representation
|
1605
1647
|
|
1606
1648
|
property :update_page_element_transform, as: 'updatePageElementTransform', class: Google::Apis::SlidesV1::UpdatePageElementTransformRequest, decorator: Google::Apis::SlidesV1::UpdatePageElementTransformRequest::Representation
|
1607
1649
|
|
1650
|
+
property :update_page_elements_z_order, as: 'updatePageElementsZOrder', class: Google::Apis::SlidesV1::UpdatePageElementsZOrderRequest, decorator: Google::Apis::SlidesV1::UpdatePageElementsZOrderRequest::Representation
|
1651
|
+
|
1608
1652
|
property :update_page_properties, as: 'updatePageProperties', class: Google::Apis::SlidesV1::UpdatePagePropertiesRequest, decorator: Google::Apis::SlidesV1::UpdatePagePropertiesRequest::Representation
|
1609
1653
|
|
1610
1654
|
property :update_paragraph_style, as: 'updateParagraphStyle', class: Google::Apis::SlidesV1::UpdateParagraphStyleRequest, decorator: Google::Apis::SlidesV1::UpdateParagraphStyleRequest::Representation
|
@@ -1628,6 +1672,13 @@ module Google
|
|
1628
1672
|
end
|
1629
1673
|
end
|
1630
1674
|
|
1675
|
+
class RerouteLineRequest
|
1676
|
+
# @private
|
1677
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1678
|
+
property :object_id_prop, as: 'objectId'
|
1679
|
+
end
|
1680
|
+
end
|
1681
|
+
|
1631
1682
|
class Response
|
1632
1683
|
# @private
|
1633
1684
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -2017,6 +2068,14 @@ module Google
|
|
2017
2068
|
end
|
2018
2069
|
end
|
2019
2070
|
|
2071
|
+
class UpdateLineCategoryRequest
|
2072
|
+
# @private
|
2073
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
2074
|
+
property :line_category, as: 'lineCategory'
|
2075
|
+
property :object_id_prop, as: 'objectId'
|
2076
|
+
end
|
2077
|
+
end
|
2078
|
+
|
2020
2079
|
class UpdateLinePropertiesRequest
|
2021
2080
|
# @private
|
2022
2081
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -2046,6 +2105,14 @@ module Google
|
|
2046
2105
|
end
|
2047
2106
|
end
|
2048
2107
|
|
2108
|
+
class UpdatePageElementsZOrderRequest
|
2109
|
+
# @private
|
2110
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
2111
|
+
property :operation, as: 'operation'
|
2112
|
+
collection :page_element_object_ids, as: 'pageElementObjectIds'
|
2113
|
+
end
|
2114
|
+
end
|
2115
|
+
|
2049
2116
|
class UpdatePagePropertiesRequest
|
2050
2117
|
# @private
|
2051
2118
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -96,9 +96,10 @@ module Google
|
|
96
96
|
execute_or_queue_command(command, &block)
|
97
97
|
end
|
98
98
|
|
99
|
-
# Creates a
|
100
|
-
# presentationId is provided,
|
101
|
-
# Otherwise, a new
|
99
|
+
# Creates a blank presentation using the title given in the request. If a
|
100
|
+
# `presentationId` is provided, it is used as the ID of the new presentation.
|
101
|
+
# Otherwise, a new ID is generated. Other fields in the request, including
|
102
|
+
# any provided content, are ignored.
|
102
103
|
# Returns the created presentation.
|
103
104
|
# @param [Google::Apis::SlidesV1::Presentation] presentation_object
|
104
105
|
# @param [String] fields
|
@@ -20,12 +20,12 @@ module Google
|
|
20
20
|
module Apis
|
21
21
|
# Cloud Source Repositories API
|
22
22
|
#
|
23
|
-
#
|
23
|
+
# Accesses source code repositories hosted by Google.
|
24
24
|
#
|
25
25
|
# @see https://cloud.google.com/source-repositories/docs/apis
|
26
26
|
module SourcerepoV1
|
27
27
|
VERSION = 'V1'
|
28
|
-
REVISION = '
|
28
|
+
REVISION = '20181106'
|
29
29
|
|
30
30
|
# View and manage your data across Google Cloud Platform services
|
31
31
|
AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform'
|
@@ -25,7 +25,7 @@ module Google
|
|
25
25
|
# @see https://cloud.google.com/speech-to-text/docs/quickstart-protocol
|
26
26
|
module SpeechV1
|
27
27
|
VERSION = 'V1'
|
28
|
-
REVISION = '
|
28
|
+
REVISION = '20181127'
|
29
29
|
|
30
30
|
# View and manage your data across Google Cloud Platform services
|
31
31
|
AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform'
|
@@ -47,6 +47,40 @@ module Google
|
|
47
47
|
end
|
48
48
|
end
|
49
49
|
|
50
|
+
# Describes the progress of a long-running `LongRunningRecognize` call. It is
|
51
|
+
# included in the `metadata` field of the `Operation` returned by the
|
52
|
+
# `GetOperation` call of the `google::longrunning::Operations` service.
|
53
|
+
class LongRunningRecognizeMetadata
|
54
|
+
include Google::Apis::Core::Hashable
|
55
|
+
|
56
|
+
# Time of the most recent processing update.
|
57
|
+
# Corresponds to the JSON property `lastUpdateTime`
|
58
|
+
# @return [String]
|
59
|
+
attr_accessor :last_update_time
|
60
|
+
|
61
|
+
# Approximate percentage of audio processed thus far. Guaranteed to be 100
|
62
|
+
# when the audio is fully processed and the results are available.
|
63
|
+
# Corresponds to the JSON property `progressPercent`
|
64
|
+
# @return [Fixnum]
|
65
|
+
attr_accessor :progress_percent
|
66
|
+
|
67
|
+
# Time when the request was received.
|
68
|
+
# Corresponds to the JSON property `startTime`
|
69
|
+
# @return [String]
|
70
|
+
attr_accessor :start_time
|
71
|
+
|
72
|
+
def initialize(**args)
|
73
|
+
update!(**args)
|
74
|
+
end
|
75
|
+
|
76
|
+
# Update properties of this object
|
77
|
+
def update!(**args)
|
78
|
+
@last_update_time = args[:last_update_time] if args.key?(:last_update_time)
|
79
|
+
@progress_percent = args[:progress_percent] if args.key?(:progress_percent)
|
80
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
81
|
+
end
|
82
|
+
end
|
83
|
+
|
50
84
|
# The top-level message sent by the client for the `LongRunningRecognize`
|
51
85
|
# method.
|
52
86
|
class LongRunningRecognizeRequest
|
@@ -77,6 +111,30 @@ module Google
|
|
77
111
|
end
|
78
112
|
end
|
79
113
|
|
114
|
+
# The only message returned to the client by the `LongRunningRecognize` method.
|
115
|
+
# It contains the result as zero or more sequential `SpeechRecognitionResult`
|
116
|
+
# messages. It is included in the `result.response` field of the `Operation`
|
117
|
+
# returned by the `GetOperation` call of the `google::longrunning::Operations`
|
118
|
+
# service.
|
119
|
+
class LongRunningRecognizeResponse
|
120
|
+
include Google::Apis::Core::Hashable
|
121
|
+
|
122
|
+
# Output only. Sequential list of transcription results corresponding to
|
123
|
+
# sequential portions of audio.
|
124
|
+
# Corresponds to the JSON property `results`
|
125
|
+
# @return [Array<Google::Apis::SpeechV1::SpeechRecognitionResult>]
|
126
|
+
attr_accessor :results
|
127
|
+
|
128
|
+
def initialize(**args)
|
129
|
+
update!(**args)
|
130
|
+
end
|
131
|
+
|
132
|
+
# Update properties of this object
|
133
|
+
def update!(**args)
|
134
|
+
@results = args[:results] if args.key?(:results)
|
135
|
+
end
|
136
|
+
end
|
137
|
+
|
80
138
|
# This resource represents a long-running operation that is the result of a
|
81
139
|
# network API call.
|
82
140
|
class Operation
|
@@ -586,15 +644,6 @@ module Google
|
|
586
644
|
# @return [String]
|
587
645
|
attr_accessor :end_time
|
588
646
|
|
589
|
-
# Output only. A distinct integer value is assigned for every speaker within
|
590
|
-
# the audio. This field specifies which one of those speakers was detected to
|
591
|
-
# have spoken this word. Value ranges from '1' to diarization_speaker_count.
|
592
|
-
# speaker_tag is set if enable_speaker_diarization = 'true' and only in the
|
593
|
-
# top alternative.
|
594
|
-
# Corresponds to the JSON property `speakerTag`
|
595
|
-
# @return [Fixnum]
|
596
|
-
attr_accessor :speaker_tag
|
597
|
-
|
598
647
|
# Output only. Time offset relative to the beginning of the audio,
|
599
648
|
# and corresponding to the start of the spoken word.
|
600
649
|
# This field is only set if `enable_word_time_offsets=true` and only
|
@@ -617,7 +666,6 @@ module Google
|
|
617
666
|
# Update properties of this object
|
618
667
|
def update!(**args)
|
619
668
|
@end_time = args[:end_time] if args.key?(:end_time)
|
620
|
-
@speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag)
|
621
669
|
@start_time = args[:start_time] if args.key?(:start_time)
|
622
670
|
@word = args[:word] if args.key?(:word)
|
623
671
|
end
|
@@ -28,12 +28,24 @@ module Google
|
|
28
28
|
include Google::Apis::Core::JsonObjectSupport
|
29
29
|
end
|
30
30
|
|
31
|
+
class LongRunningRecognizeMetadata
|
32
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
33
|
+
|
34
|
+
include Google::Apis::Core::JsonObjectSupport
|
35
|
+
end
|
36
|
+
|
31
37
|
class LongRunningRecognizeRequest
|
32
38
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
33
39
|
|
34
40
|
include Google::Apis::Core::JsonObjectSupport
|
35
41
|
end
|
36
42
|
|
43
|
+
class LongRunningRecognizeResponse
|
44
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
45
|
+
|
46
|
+
include Google::Apis::Core::JsonObjectSupport
|
47
|
+
end
|
48
|
+
|
37
49
|
class Operation
|
38
50
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
39
51
|
|
@@ -103,6 +115,15 @@ module Google
|
|
103
115
|
end
|
104
116
|
end
|
105
117
|
|
118
|
+
class LongRunningRecognizeMetadata
|
119
|
+
# @private
|
120
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
121
|
+
property :last_update_time, as: 'lastUpdateTime'
|
122
|
+
property :progress_percent, as: 'progressPercent'
|
123
|
+
property :start_time, as: 'startTime'
|
124
|
+
end
|
125
|
+
end
|
126
|
+
|
106
127
|
class LongRunningRecognizeRequest
|
107
128
|
# @private
|
108
129
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -113,6 +134,14 @@ module Google
|
|
113
134
|
end
|
114
135
|
end
|
115
136
|
|
137
|
+
class LongRunningRecognizeResponse
|
138
|
+
# @private
|
139
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
140
|
+
collection :results, as: 'results', class: Google::Apis::SpeechV1::SpeechRecognitionResult, decorator: Google::Apis::SpeechV1::SpeechRecognitionResult::Representation
|
141
|
+
|
142
|
+
end
|
143
|
+
end
|
144
|
+
|
116
145
|
class Operation
|
117
146
|
# @private
|
118
147
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -206,7 +235,6 @@ module Google
|
|
206
235
|
# @private
|
207
236
|
class Representation < Google::Apis::Core::JsonRepresentation
|
208
237
|
property :end_time, as: 'endTime'
|
209
|
-
property :speaker_tag, as: 'speakerTag'
|
210
238
|
property :start_time, as: 'startTime'
|
211
239
|
property :word, as: 'word'
|
212
240
|
end
|
@@ -12,9 +12,9 @@
|
|
12
12
|
# See the License for the specific language governing permissions and
|
13
13
|
# limitations under the License.
|
14
14
|
|
15
|
-
require 'google/apis/
|
16
|
-
require 'google/apis/
|
17
|
-
require 'google/apis/
|
15
|
+
require 'google/apis/speech_v1p1beta1/service.rb'
|
16
|
+
require 'google/apis/speech_v1p1beta1/classes.rb'
|
17
|
+
require 'google/apis/speech_v1p1beta1/representations.rb'
|
18
18
|
|
19
19
|
module Google
|
20
20
|
module Apis
|
@@ -23,9 +23,9 @@ module Google
|
|
23
23
|
# Converts audio to text by applying powerful neural network models.
|
24
24
|
#
|
25
25
|
# @see https://cloud.google.com/speech-to-text/docs/quickstart-protocol
|
26
|
-
module
|
27
|
-
VERSION = '
|
28
|
-
REVISION = '
|
26
|
+
module SpeechV1p1beta1
|
27
|
+
VERSION = 'V1p1beta1'
|
28
|
+
REVISION = '20181127'
|
29
29
|
|
30
30
|
# View and manage your data across Google Cloud Platform services
|
31
31
|
AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform'
|
@@ -0,0 +1,922 @@
|
|
1
|
+
# Copyright 2015 Google Inc.
|
2
|
+
#
|
3
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
|
+
# you may not use this file except in compliance with the License.
|
5
|
+
# You may obtain a copy of the License at
|
6
|
+
#
|
7
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
8
|
+
#
|
9
|
+
# Unless required by applicable law or agreed to in writing, software
|
10
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
11
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12
|
+
# See the License for the specific language governing permissions and
|
13
|
+
# limitations under the License.
|
14
|
+
|
15
|
+
require 'date'
|
16
|
+
require 'google/apis/core/base_service'
|
17
|
+
require 'google/apis/core/json_representation'
|
18
|
+
require 'google/apis/core/hashable'
|
19
|
+
require 'google/apis/errors'
|
20
|
+
|
21
|
+
module Google
|
22
|
+
module Apis
|
23
|
+
module SpeechV1p1beta1
|
24
|
+
|
25
|
+
# The response message for Operations.ListOperations.
|
26
|
+
class ListOperationsResponse
|
27
|
+
include Google::Apis::Core::Hashable
|
28
|
+
|
29
|
+
# The standard List next-page token.
|
30
|
+
# Corresponds to the JSON property `nextPageToken`
|
31
|
+
# @return [String]
|
32
|
+
attr_accessor :next_page_token
|
33
|
+
|
34
|
+
# A list of operations that matches the specified filter in the request.
|
35
|
+
# Corresponds to the JSON property `operations`
|
36
|
+
# @return [Array<Google::Apis::SpeechV1p1beta1::Operation>]
|
37
|
+
attr_accessor :operations
|
38
|
+
|
39
|
+
def initialize(**args)
|
40
|
+
update!(**args)
|
41
|
+
end
|
42
|
+
|
43
|
+
# Update properties of this object
|
44
|
+
def update!(**args)
|
45
|
+
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
|
46
|
+
@operations = args[:operations] if args.key?(:operations)
|
47
|
+
end
|
48
|
+
end
|
49
|
+
|
50
|
+
# Describes the progress of a long-running `LongRunningRecognize` call. It is
|
51
|
+
# included in the `metadata` field of the `Operation` returned by the
|
52
|
+
# `GetOperation` call of the `google::longrunning::Operations` service.
|
53
|
+
class LongRunningRecognizeMetadata
|
54
|
+
include Google::Apis::Core::Hashable
|
55
|
+
|
56
|
+
# Time of the most recent processing update.
|
57
|
+
# Corresponds to the JSON property `lastUpdateTime`
|
58
|
+
# @return [String]
|
59
|
+
attr_accessor :last_update_time
|
60
|
+
|
61
|
+
# Approximate percentage of audio processed thus far. Guaranteed to be 100
|
62
|
+
# when the audio is fully processed and the results are available.
|
63
|
+
# Corresponds to the JSON property `progressPercent`
|
64
|
+
# @return [Fixnum]
|
65
|
+
attr_accessor :progress_percent
|
66
|
+
|
67
|
+
# Time when the request was received.
|
68
|
+
# Corresponds to the JSON property `startTime`
|
69
|
+
# @return [String]
|
70
|
+
attr_accessor :start_time
|
71
|
+
|
72
|
+
def initialize(**args)
|
73
|
+
update!(**args)
|
74
|
+
end
|
75
|
+
|
76
|
+
# Update properties of this object
|
77
|
+
def update!(**args)
|
78
|
+
@last_update_time = args[:last_update_time] if args.key?(:last_update_time)
|
79
|
+
@progress_percent = args[:progress_percent] if args.key?(:progress_percent)
|
80
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
81
|
+
end
|
82
|
+
end
|
83
|
+
|
84
|
+
# The top-level message sent by the client for the `LongRunningRecognize`
|
85
|
+
# method.
|
86
|
+
class LongRunningRecognizeRequest
|
87
|
+
include Google::Apis::Core::Hashable
|
88
|
+
|
89
|
+
# Contains audio data in the encoding specified in the `RecognitionConfig`.
|
90
|
+
# Either `content` or `uri` must be supplied. Supplying both or neither
|
91
|
+
# returns google.rpc.Code.INVALID_ARGUMENT. See
|
92
|
+
# [content limits](/speech-to-text/quotas#content).
|
93
|
+
# Corresponds to the JSON property `audio`
|
94
|
+
# @return [Google::Apis::SpeechV1p1beta1::RecognitionAudio]
|
95
|
+
attr_accessor :audio
|
96
|
+
|
97
|
+
# Provides information to the recognizer that specifies how to process the
|
98
|
+
# request.
|
99
|
+
# Corresponds to the JSON property `config`
|
100
|
+
# @return [Google::Apis::SpeechV1p1beta1::RecognitionConfig]
|
101
|
+
attr_accessor :config
|
102
|
+
|
103
|
+
def initialize(**args)
|
104
|
+
update!(**args)
|
105
|
+
end
|
106
|
+
|
107
|
+
# Update properties of this object
|
108
|
+
def update!(**args)
|
109
|
+
@audio = args[:audio] if args.key?(:audio)
|
110
|
+
@config = args[:config] if args.key?(:config)
|
111
|
+
end
|
112
|
+
end
|
113
|
+
|
114
|
+
# The only message returned to the client by the `LongRunningRecognize` method.
|
115
|
+
# It contains the result as zero or more sequential `SpeechRecognitionResult`
|
116
|
+
# messages. It is included in the `result.response` field of the `Operation`
|
117
|
+
# returned by the `GetOperation` call of the `google::longrunning::Operations`
|
118
|
+
# service.
|
119
|
+
class LongRunningRecognizeResponse
|
120
|
+
include Google::Apis::Core::Hashable
|
121
|
+
|
122
|
+
# Output only. Sequential list of transcription results corresponding to
|
123
|
+
# sequential portions of audio.
|
124
|
+
# Corresponds to the JSON property `results`
|
125
|
+
# @return [Array<Google::Apis::SpeechV1p1beta1::SpeechRecognitionResult>]
|
126
|
+
attr_accessor :results
|
127
|
+
|
128
|
+
def initialize(**args)
|
129
|
+
update!(**args)
|
130
|
+
end
|
131
|
+
|
132
|
+
# Update properties of this object
|
133
|
+
def update!(**args)
|
134
|
+
@results = args[:results] if args.key?(:results)
|
135
|
+
end
|
136
|
+
end
|
137
|
+
|
138
|
+
# This resource represents a long-running operation that is the result of a
|
139
|
+
# network API call.
|
140
|
+
class Operation
|
141
|
+
include Google::Apis::Core::Hashable
|
142
|
+
|
143
|
+
# If the value is `false`, it means the operation is still in progress.
|
144
|
+
# If `true`, the operation is completed, and either `error` or `response` is
|
145
|
+
# available.
|
146
|
+
# Corresponds to the JSON property `done`
|
147
|
+
# @return [Boolean]
|
148
|
+
attr_accessor :done
|
149
|
+
alias_method :done?, :done
|
150
|
+
|
151
|
+
# The `Status` type defines a logical error model that is suitable for different
|
152
|
+
# programming environments, including REST APIs and RPC APIs. It is used by
|
153
|
+
# [gRPC](https://github.com/grpc). The error model is designed to be:
|
154
|
+
# - Simple to use and understand for most users
|
155
|
+
# - Flexible enough to meet unexpected needs
|
156
|
+
# # Overview
|
157
|
+
# The `Status` message contains three pieces of data: error code, error message,
|
158
|
+
# and error details. The error code should be an enum value of
|
159
|
+
# google.rpc.Code, but it may accept additional error codes if needed. The
|
160
|
+
# error message should be a developer-facing English message that helps
|
161
|
+
# developers *understand* and *resolve* the error. If a localized user-facing
|
162
|
+
# error message is needed, put the localized message in the error details or
|
163
|
+
# localize it in the client. The optional error details may contain arbitrary
|
164
|
+
# information about the error. There is a predefined set of error detail types
|
165
|
+
# in the package `google.rpc` that can be used for common error conditions.
|
166
|
+
# # Language mapping
|
167
|
+
# The `Status` message is the logical representation of the error model, but it
|
168
|
+
# is not necessarily the actual wire format. When the `Status` message is
|
169
|
+
# exposed in different client libraries and different wire protocols, it can be
|
170
|
+
# mapped differently. For example, it will likely be mapped to some exceptions
|
171
|
+
# in Java, but more likely mapped to some error codes in C.
|
172
|
+
# # Other uses
|
173
|
+
# The error model and the `Status` message can be used in a variety of
|
174
|
+
# environments, either with or without APIs, to provide a
|
175
|
+
# consistent developer experience across different environments.
|
176
|
+
# Example uses of this error model include:
|
177
|
+
# - Partial errors. If a service needs to return partial errors to the client,
|
178
|
+
# it may embed the `Status` in the normal response to indicate the partial
|
179
|
+
# errors.
|
180
|
+
# - Workflow errors. A typical workflow has multiple steps. Each step may
|
181
|
+
# have a `Status` message for error reporting.
|
182
|
+
# - Batch operations. If a client uses batch request and batch response, the
|
183
|
+
# `Status` message should be used directly inside batch response, one for
|
184
|
+
# each error sub-response.
|
185
|
+
# - Asynchronous operations. If an API call embeds asynchronous operation
|
186
|
+
# results in its response, the status of those operations should be
|
187
|
+
# represented directly using the `Status` message.
|
188
|
+
# - Logging. If some API errors are stored in logs, the message `Status` could
|
189
|
+
# be used directly after any stripping needed for security/privacy reasons.
|
190
|
+
# Corresponds to the JSON property `error`
|
191
|
+
# @return [Google::Apis::SpeechV1p1beta1::Status]
|
192
|
+
attr_accessor :error
|
193
|
+
|
194
|
+
# Service-specific metadata associated with the operation. It typically
|
195
|
+
# contains progress information and common metadata such as create time.
|
196
|
+
# Some services might not provide such metadata. Any method that returns a
|
197
|
+
# long-running operation should document the metadata type, if any.
|
198
|
+
# Corresponds to the JSON property `metadata`
|
199
|
+
# @return [Hash<String,Object>]
|
200
|
+
attr_accessor :metadata
|
201
|
+
|
202
|
+
# The server-assigned name, which is only unique within the same service that
|
203
|
+
# originally returns it. If you use the default HTTP mapping, the
|
204
|
+
# `name` should have the format of `operations/some/unique/name`.
|
205
|
+
# Corresponds to the JSON property `name`
|
206
|
+
# @return [String]
|
207
|
+
attr_accessor :name
|
208
|
+
|
209
|
+
# The normal response of the operation in case of success. If the original
|
210
|
+
# method returns no data on success, such as `Delete`, the response is
|
211
|
+
# `google.protobuf.Empty`. If the original method is standard
|
212
|
+
# `Get`/`Create`/`Update`, the response should be the resource. For other
|
213
|
+
# methods, the response should have the type `XxxResponse`, where `Xxx`
|
214
|
+
# is the original method name. For example, if the original method name
|
215
|
+
# is `TakeSnapshot()`, the inferred response type is
|
216
|
+
# `TakeSnapshotResponse`.
|
217
|
+
# Corresponds to the JSON property `response`
|
218
|
+
# @return [Hash<String,Object>]
|
219
|
+
attr_accessor :response
|
220
|
+
|
221
|
+
def initialize(**args)
|
222
|
+
update!(**args)
|
223
|
+
end
|
224
|
+
|
225
|
+
# Update properties of this object
|
226
|
+
def update!(**args)
|
227
|
+
@done = args[:done] if args.key?(:done)
|
228
|
+
@error = args[:error] if args.key?(:error)
|
229
|
+
@metadata = args[:metadata] if args.key?(:metadata)
|
230
|
+
@name = args[:name] if args.key?(:name)
|
231
|
+
@response = args[:response] if args.key?(:response)
|
232
|
+
end
|
233
|
+
end
|
234
|
+
|
235
|
+
# Contains audio data in the encoding specified in the `RecognitionConfig`.
|
236
|
+
# Either `content` or `uri` must be supplied. Supplying both or neither
|
237
|
+
# returns google.rpc.Code.INVALID_ARGUMENT. See
|
238
|
+
# [content limits](/speech-to-text/quotas#content).
|
239
|
+
class RecognitionAudio
|
240
|
+
include Google::Apis::Core::Hashable
|
241
|
+
|
242
|
+
# The audio data bytes encoded as specified in
|
243
|
+
# `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
|
244
|
+
# pure binary representation, whereas JSON representations use base64.
|
245
|
+
# Corresponds to the JSON property `content`
|
246
|
+
# NOTE: Values are automatically base64 encoded/decoded in the client library.
|
247
|
+
# @return [String]
|
248
|
+
attr_accessor :content
|
249
|
+
|
250
|
+
# URI that points to a file that contains audio data bytes as specified in
|
251
|
+
# `RecognitionConfig`. The file must not be compressed (for example, gzip).
|
252
|
+
# Currently, only Google Cloud Storage URIs are
|
253
|
+
# supported, which must be specified in the following format:
|
254
|
+
# `gs://bucket_name/object_name` (other URI formats return
|
255
|
+
# google.rpc.Code.INVALID_ARGUMENT). For more information, see
|
256
|
+
# [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
|
257
|
+
# Corresponds to the JSON property `uri`
|
258
|
+
# @return [String]
|
259
|
+
attr_accessor :uri
|
260
|
+
|
261
|
+
def initialize(**args)
|
262
|
+
update!(**args)
|
263
|
+
end
|
264
|
+
|
265
|
+
# Update properties of this object
|
266
|
+
def update!(**args)
|
267
|
+
@content = args[:content] if args.key?(:content)
|
268
|
+
@uri = args[:uri] if args.key?(:uri)
|
269
|
+
end
|
270
|
+
end
|
271
|
+
|
272
|
+
# Provides information to the recognizer that specifies how to process the
|
273
|
+
# request.
|
274
|
+
class RecognitionConfig
|
275
|
+
include Google::Apis::Core::Hashable
|
276
|
+
|
277
|
+
# *Optional* A list of up to 3 additional
|
278
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
|
279
|
+
# listing possible alternative languages of the supplied audio.
|
280
|
+
# See [Language Support](/speech-to-text/docs/languages)
|
281
|
+
# for a list of the currently supported language codes.
|
282
|
+
# If alternative languages are listed, recognition result will contain
|
283
|
+
# recognition in the most likely language detected including the main
|
284
|
+
# language_code. The recognition result will include the language tag
|
285
|
+
# of the language detected in the audio.
|
286
|
+
# Note: This feature is only supported for Voice Command and Voice Search
|
287
|
+
# use cases and performance may vary for other use cases (e.g., phone call
|
288
|
+
# transcription).
|
289
|
+
# Corresponds to the JSON property `alternativeLanguageCodes`
|
290
|
+
# @return [Array<String>]
|
291
|
+
attr_accessor :alternative_language_codes
|
292
|
+
|
293
|
+
# *Optional* The number of channels in the input audio data.
|
294
|
+
# ONLY set this for MULTI-CHANNEL recognition.
|
295
|
+
# Valid values for LINEAR16 and FLAC are `1`-`8`.
|
296
|
+
# Valid values for OGG_OPUS are '1'-'254'.
|
297
|
+
# Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
|
298
|
+
# If `0` or omitted, defaults to one channel (mono).
|
299
|
+
# Note: We only recognize the first channel by default.
|
300
|
+
# To perform independent recognition on each channel set
|
301
|
+
# `enable_separate_recognition_per_channel` to 'true'.
|
302
|
+
# Corresponds to the JSON property `audioChannelCount`
|
303
|
+
# @return [Fixnum]
|
304
|
+
attr_accessor :audio_channel_count
|
305
|
+
|
306
|
+
# *Optional* Config to enable speaker diarization and set additional
|
307
|
+
# parameters to make diarization better suited for your application.
|
308
|
+
# Note: When this is enabled, we send all the words from the beginning of the
|
309
|
+
# audio for the top alternative in every consecutive STREAMING responses.
|
310
|
+
# This is done in order to improve our speaker tags as our models learn to
|
311
|
+
# identify the speakers in the conversation over time.
|
312
|
+
# For non-streaming requests, the diarization results will be provided only
|
313
|
+
# in the top alternative of the FINAL SpeechRecognitionResult.
|
314
|
+
# Corresponds to the JSON property `diarizationConfig`
|
315
|
+
# @return [Google::Apis::SpeechV1p1beta1::SpeakerDiarizationConfig]
|
316
|
+
attr_accessor :diarization_config
|
317
|
+
|
318
|
+
# *Optional*
|
319
|
+
# If set, specifies the estimated number of speakers in the conversation.
|
320
|
+
# If not set, defaults to '2'.
|
321
|
+
# Ignored unless enable_speaker_diarization is set to true."
|
322
|
+
# Note: Use diarization_config instead. This field will be DEPRECATED soon.
|
323
|
+
# Corresponds to the JSON property `diarizationSpeakerCount`
|
324
|
+
# @return [Fixnum]
|
325
|
+
attr_accessor :diarization_speaker_count
|
326
|
+
|
327
|
+
# *Optional* If 'true', adds punctuation to recognition result hypotheses.
|
328
|
+
# This feature is only available in select languages. Setting this for
|
329
|
+
# requests in other languages has no effect at all.
|
330
|
+
# The default 'false' value does not add punctuation to result hypotheses.
|
331
|
+
# Note: This is currently offered as an experimental service, complimentary
|
332
|
+
# to all users. In the future this may be exclusively available as a
|
333
|
+
# premium feature.
|
334
|
+
# Corresponds to the JSON property `enableAutomaticPunctuation`
|
335
|
+
# @return [Boolean]
|
336
|
+
attr_accessor :enable_automatic_punctuation
|
337
|
+
alias_method :enable_automatic_punctuation?, :enable_automatic_punctuation
|
338
|
+
|
339
|
+
# This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
|
340
|
+
# to get each channel recognized separately. The recognition result will
|
341
|
+
# contain a `channel_tag` field to state which channel that result belongs
|
342
|
+
# to. If this is not true, we will only recognize the first channel. The
|
343
|
+
# request is billed cumulatively for all channels recognized:
|
344
|
+
# `audio_channel_count` multiplied by the length of the audio.
|
345
|
+
# Corresponds to the JSON property `enableSeparateRecognitionPerChannel`
|
346
|
+
# @return [Boolean]
|
347
|
+
attr_accessor :enable_separate_recognition_per_channel
|
348
|
+
alias_method :enable_separate_recognition_per_channel?, :enable_separate_recognition_per_channel
|
349
|
+
|
350
|
+
# *Optional* If 'true', enables speaker detection for each recognized word in
|
351
|
+
# the top alternative of the recognition result using a speaker_tag provided
|
352
|
+
# in the WordInfo.
|
353
|
+
# Note: Use diarization_config instead. This field will be DEPRECATED soon.
|
354
|
+
# Corresponds to the JSON property `enableSpeakerDiarization`
|
355
|
+
# @return [Boolean]
|
356
|
+
attr_accessor :enable_speaker_diarization
|
357
|
+
alias_method :enable_speaker_diarization?, :enable_speaker_diarization
|
358
|
+
|
359
|
+
# *Optional* If `true`, the top result includes a list of words and the
|
360
|
+
# confidence for those words. If `false`, no word-level confidence
|
361
|
+
# information is returned. The default is `false`.
|
362
|
+
# Corresponds to the JSON property `enableWordConfidence`
|
363
|
+
# @return [Boolean]
|
364
|
+
attr_accessor :enable_word_confidence
|
365
|
+
alias_method :enable_word_confidence?, :enable_word_confidence
|
366
|
+
|
367
|
+
# *Optional* If `true`, the top result includes a list of words and
|
368
|
+
# the start and end time offsets (timestamps) for those words. If
|
369
|
+
# `false`, no word-level time offset information is returned. The default is
|
370
|
+
# `false`.
|
371
|
+
# Corresponds to the JSON property `enableWordTimeOffsets`
|
372
|
+
# @return [Boolean]
|
373
|
+
attr_accessor :enable_word_time_offsets
|
374
|
+
alias_method :enable_word_time_offsets?, :enable_word_time_offsets
|
375
|
+
|
376
|
+
# Encoding of audio data sent in all `RecognitionAudio` messages.
|
377
|
+
# This field is optional for `FLAC` and `WAV` audio files and required
|
378
|
+
# for all other audio formats. For details, see AudioEncoding.
|
379
|
+
# Corresponds to the JSON property `encoding`
|
380
|
+
# @return [String]
|
381
|
+
attr_accessor :encoding
|
382
|
+
|
383
|
+
# *Required* The language of the supplied audio as a
|
384
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
|
385
|
+
# Example: "en-US".
|
386
|
+
# See [Language Support](/speech-to-text/docs/languages)
|
387
|
+
# for a list of the currently supported language codes.
|
388
|
+
# Corresponds to the JSON property `languageCode`
|
389
|
+
# @return [String]
|
390
|
+
attr_accessor :language_code
|
391
|
+
|
392
|
+
# *Optional* Maximum number of recognition hypotheses to be returned.
|
393
|
+
# Specifically, the maximum number of `SpeechRecognitionAlternative` messages
|
394
|
+
# within each `SpeechRecognitionResult`.
|
395
|
+
# The server may return fewer than `max_alternatives`.
|
396
|
+
# Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
|
397
|
+
# one. If omitted, will return a maximum of one.
|
398
|
+
# Corresponds to the JSON property `maxAlternatives`
|
399
|
+
# @return [Fixnum]
|
400
|
+
attr_accessor :max_alternatives
|
401
|
+
|
402
|
+
# Description of audio data to be recognized.
|
403
|
+
# Corresponds to the JSON property `metadata`
|
404
|
+
# @return [Google::Apis::SpeechV1p1beta1::RecognitionMetadata]
|
405
|
+
attr_accessor :metadata
|
406
|
+
|
407
|
+
# *Optional* Which model to select for the given request. Select the model
|
408
|
+
# best suited to your domain to get best results. If a model is not
|
409
|
+
# explicitly specified, then we auto-select a model based on the parameters
|
410
|
+
# in the RecognitionConfig.
|
411
|
+
# <table>
|
412
|
+
# <tr>
|
413
|
+
# <td><b>Model</b></td>
|
414
|
+
# <td><b>Description</b></td>
|
415
|
+
# </tr>
|
416
|
+
# <tr>
|
417
|
+
# <td><code>command_and_search</code></td>
|
418
|
+
# <td>Best for short queries such as voice commands or voice search.</td>
|
419
|
+
# </tr>
|
420
|
+
# <tr>
|
421
|
+
# <td><code>phone_call</code></td>
|
422
|
+
# <td>Best for audio that originated from a phone call (typically
|
423
|
+
# recorded at an 8khz sampling rate).</td>
|
424
|
+
# </tr>
|
425
|
+
# <tr>
|
426
|
+
# <td><code>video</code></td>
|
427
|
+
# <td>Best for audio that originated from from video or includes multiple
|
428
|
+
# speakers. Ideally the audio is recorded at a 16khz or greater
|
429
|
+
# sampling rate. This is a premium model that costs more than the
|
430
|
+
# standard rate.</td>
|
431
|
+
# </tr>
|
432
|
+
# <tr>
|
433
|
+
# <td><code>default</code></td>
|
434
|
+
# <td>Best for audio that is not one of the specific audio models.
|
435
|
+
# For example, long-form audio. Ideally the audio is high-fidelity,
|
436
|
+
# recorded at a 16khz or greater sampling rate.</td>
|
437
|
+
# </tr>
|
438
|
+
# </table>
|
439
|
+
# Corresponds to the JSON property `model`
|
440
|
+
# @return [String]
|
441
|
+
attr_accessor :model
|
442
|
+
|
443
|
+
# *Optional* If set to `true`, the server will attempt to filter out
|
444
|
+
# profanities, replacing all but the initial character in each filtered word
|
445
|
+
# with asterisks, e.g. "f***". If set to `false` or omitted, profanities
|
446
|
+
# won't be filtered out.
|
447
|
+
# Corresponds to the JSON property `profanityFilter`
|
448
|
+
# @return [Boolean]
|
449
|
+
attr_accessor :profanity_filter
|
450
|
+
alias_method :profanity_filter?, :profanity_filter
|
451
|
+
|
452
|
+
# Sample rate in Hertz of the audio data sent in all
|
453
|
+
# `RecognitionAudio` messages. Valid values are: 8000-48000.
|
454
|
+
# 16000 is optimal. For best results, set the sampling rate of the audio
|
455
|
+
# source to 16000 Hz. If that's not possible, use the native sample rate of
|
456
|
+
# the audio source (instead of re-sampling).
|
457
|
+
# This field is optional for `FLAC` and `WAV` audio files and required
|
458
|
+
# for all other audio formats. For details, see AudioEncoding.
|
459
|
+
# Corresponds to the JSON property `sampleRateHertz`
|
460
|
+
# @return [Fixnum]
|
461
|
+
attr_accessor :sample_rate_hertz
|
462
|
+
|
463
|
+
# *Optional* array of SpeechContext.
|
464
|
+
# A means to provide context to assist the speech recognition. For more
|
465
|
+
# information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
|
466
|
+
# Corresponds to the JSON property `speechContexts`
|
467
|
+
# @return [Array<Google::Apis::SpeechV1p1beta1::SpeechContext>]
|
468
|
+
attr_accessor :speech_contexts
|
469
|
+
|
470
|
+
# *Optional* Set to true to use an enhanced model for speech recognition.
|
471
|
+
# If `use_enhanced` is set to true and the `model` field is not set, then
|
472
|
+
# an appropriate enhanced model is chosen if:
|
473
|
+
# 1. project is eligible for requesting enhanced models
|
474
|
+
# 2. an enhanced model exists for the audio
|
475
|
+
# If `use_enhanced` is true and an enhanced version of the specified model
|
476
|
+
# does not exist, then the speech is recognized using the standard version
|
477
|
+
# of the specified model.
|
478
|
+
# Enhanced speech models require that you opt-in to data logging using
|
479
|
+
# instructions in the
|
480
|
+
# [documentation](/speech-to-text/docs/enable-data-logging). If you set
|
481
|
+
# `use_enhanced` to true and you have not enabled audio logging, then you
|
482
|
+
# will receive an error.
|
483
|
+
# Corresponds to the JSON property `useEnhanced`
|
484
|
+
# @return [Boolean]
|
485
|
+
attr_accessor :use_enhanced
|
486
|
+
alias_method :use_enhanced?, :use_enhanced
|
487
|
+
|
488
|
+
def initialize(**args)
|
489
|
+
update!(**args)
|
490
|
+
end
|
491
|
+
|
492
|
+
# Update properties of this object
|
493
|
+
def update!(**args)
|
494
|
+
@alternative_language_codes = args[:alternative_language_codes] if args.key?(:alternative_language_codes)
|
495
|
+
@audio_channel_count = args[:audio_channel_count] if args.key?(:audio_channel_count)
|
496
|
+
@diarization_config = args[:diarization_config] if args.key?(:diarization_config)
|
497
|
+
@diarization_speaker_count = args[:diarization_speaker_count] if args.key?(:diarization_speaker_count)
|
498
|
+
@enable_automatic_punctuation = args[:enable_automatic_punctuation] if args.key?(:enable_automatic_punctuation)
|
499
|
+
@enable_separate_recognition_per_channel = args[:enable_separate_recognition_per_channel] if args.key?(:enable_separate_recognition_per_channel)
|
500
|
+
@enable_speaker_diarization = args[:enable_speaker_diarization] if args.key?(:enable_speaker_diarization)
|
501
|
+
@enable_word_confidence = args[:enable_word_confidence] if args.key?(:enable_word_confidence)
|
502
|
+
@enable_word_time_offsets = args[:enable_word_time_offsets] if args.key?(:enable_word_time_offsets)
|
503
|
+
@encoding = args[:encoding] if args.key?(:encoding)
|
504
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
505
|
+
@max_alternatives = args[:max_alternatives] if args.key?(:max_alternatives)
|
506
|
+
@metadata = args[:metadata] if args.key?(:metadata)
|
507
|
+
@model = args[:model] if args.key?(:model)
|
508
|
+
@profanity_filter = args[:profanity_filter] if args.key?(:profanity_filter)
|
509
|
+
@sample_rate_hertz = args[:sample_rate_hertz] if args.key?(:sample_rate_hertz)
|
510
|
+
@speech_contexts = args[:speech_contexts] if args.key?(:speech_contexts)
|
511
|
+
@use_enhanced = args[:use_enhanced] if args.key?(:use_enhanced)
|
512
|
+
end
|
513
|
+
end
|
514
|
+
|
515
|
+
# Description of audio data to be recognized.
|
516
|
+
class RecognitionMetadata
|
517
|
+
include Google::Apis::Core::Hashable
|
518
|
+
|
519
|
+
# Description of the content. Eg. "Recordings of federal supreme court
|
520
|
+
# hearings from 2012".
|
521
|
+
# Corresponds to the JSON property `audioTopic`
|
522
|
+
# @return [String]
|
523
|
+
attr_accessor :audio_topic
|
524
|
+
|
525
|
+
# The industry vertical to which this speech recognition request most
|
526
|
+
# closely applies. This is most indicative of the topics contained
|
527
|
+
# in the audio. Use the 6-digit NAICS code to identify the industry
|
528
|
+
# vertical - see https://www.naics.com/search/.
|
529
|
+
# Corresponds to the JSON property `industryNaicsCodeOfAudio`
|
530
|
+
# @return [Fixnum]
|
531
|
+
attr_accessor :industry_naics_code_of_audio
|
532
|
+
|
533
|
+
# The use case most closely describing the audio content to be recognized.
|
534
|
+
# Corresponds to the JSON property `interactionType`
|
535
|
+
# @return [String]
|
536
|
+
attr_accessor :interaction_type
|
537
|
+
|
538
|
+
# The audio type that most closely describes the audio being recognized.
|
539
|
+
# Corresponds to the JSON property `microphoneDistance`
|
540
|
+
# @return [String]
|
541
|
+
attr_accessor :microphone_distance
|
542
|
+
|
543
|
+
# Obfuscated (privacy-protected) ID of the user, to identify number of
|
544
|
+
# unique users using the service.
|
545
|
+
# Corresponds to the JSON property `obfuscatedId`
|
546
|
+
# @return [Fixnum]
|
547
|
+
attr_accessor :obfuscated_id
|
548
|
+
|
549
|
+
# The original media the speech was recorded on.
|
550
|
+
# Corresponds to the JSON property `originalMediaType`
|
551
|
+
# @return [String]
|
552
|
+
attr_accessor :original_media_type
|
553
|
+
|
554
|
+
# Mime type of the original audio file. For example `audio/m4a`,
|
555
|
+
# `audio/x-alaw-basic`, `audio/mp3`, `audio/3gpp`.
|
556
|
+
# A list of possible audio mime types is maintained at
|
557
|
+
# http://www.iana.org/assignments/media-types/media-types.xhtml#audio
|
558
|
+
# Corresponds to the JSON property `originalMimeType`
|
559
|
+
# @return [String]
|
560
|
+
attr_accessor :original_mime_type
|
561
|
+
|
562
|
+
# The device used to make the recording. Examples 'Nexus 5X' or
|
563
|
+
# 'Polycom SoundStation IP 6000' or 'POTS' or 'VoIP' or
|
564
|
+
# 'Cardioid Microphone'.
|
565
|
+
# Corresponds to the JSON property `recordingDeviceName`
|
566
|
+
# @return [String]
|
567
|
+
attr_accessor :recording_device_name
|
568
|
+
|
569
|
+
# The type of device the speech was recorded with.
|
570
|
+
# Corresponds to the JSON property `recordingDeviceType`
|
571
|
+
# @return [String]
|
572
|
+
attr_accessor :recording_device_type
|
573
|
+
|
574
|
+
def initialize(**args)
|
575
|
+
update!(**args)
|
576
|
+
end
|
577
|
+
|
578
|
+
# Update properties of this object
|
579
|
+
def update!(**args)
|
580
|
+
@audio_topic = args[:audio_topic] if args.key?(:audio_topic)
|
581
|
+
@industry_naics_code_of_audio = args[:industry_naics_code_of_audio] if args.key?(:industry_naics_code_of_audio)
|
582
|
+
@interaction_type = args[:interaction_type] if args.key?(:interaction_type)
|
583
|
+
@microphone_distance = args[:microphone_distance] if args.key?(:microphone_distance)
|
584
|
+
@obfuscated_id = args[:obfuscated_id] if args.key?(:obfuscated_id)
|
585
|
+
@original_media_type = args[:original_media_type] if args.key?(:original_media_type)
|
586
|
+
@original_mime_type = args[:original_mime_type] if args.key?(:original_mime_type)
|
587
|
+
@recording_device_name = args[:recording_device_name] if args.key?(:recording_device_name)
|
588
|
+
@recording_device_type = args[:recording_device_type] if args.key?(:recording_device_type)
|
589
|
+
end
|
590
|
+
end
|
591
|
+
|
592
|
+
# The top-level message sent by the client for the `Recognize` method.
|
593
|
+
class RecognizeRequest
|
594
|
+
include Google::Apis::Core::Hashable
|
595
|
+
|
596
|
+
# Contains audio data in the encoding specified in the `RecognitionConfig`.
|
597
|
+
# Either `content` or `uri` must be supplied. Supplying both or neither
|
598
|
+
# returns google.rpc.Code.INVALID_ARGUMENT. See
|
599
|
+
# [content limits](/speech-to-text/quotas#content).
|
600
|
+
# Corresponds to the JSON property `audio`
|
601
|
+
# @return [Google::Apis::SpeechV1p1beta1::RecognitionAudio]
|
602
|
+
attr_accessor :audio
|
603
|
+
|
604
|
+
# Provides information to the recognizer that specifies how to process the
|
605
|
+
# request.
|
606
|
+
# Corresponds to the JSON property `config`
|
607
|
+
# @return [Google::Apis::SpeechV1p1beta1::RecognitionConfig]
|
608
|
+
attr_accessor :config
|
609
|
+
|
610
|
+
def initialize(**args)
|
611
|
+
update!(**args)
|
612
|
+
end
|
613
|
+
|
614
|
+
# Update properties of this object
|
615
|
+
def update!(**args)
|
616
|
+
@audio = args[:audio] if args.key?(:audio)
|
617
|
+
@config = args[:config] if args.key?(:config)
|
618
|
+
end
|
619
|
+
end
|
620
|
+
|
621
|
+
# The only message returned to the client by the `Recognize` method. It
|
622
|
+
# contains the result as zero or more sequential `SpeechRecognitionResult`
|
623
|
+
# messages.
|
624
|
+
class RecognizeResponse
|
625
|
+
include Google::Apis::Core::Hashable
|
626
|
+
|
627
|
+
# Output only. Sequential list of transcription results corresponding to
|
628
|
+
# sequential portions of audio.
|
629
|
+
# Corresponds to the JSON property `results`
|
630
|
+
# @return [Array<Google::Apis::SpeechV1p1beta1::SpeechRecognitionResult>]
|
631
|
+
attr_accessor :results
|
632
|
+
|
633
|
+
def initialize(**args)
|
634
|
+
update!(**args)
|
635
|
+
end
|
636
|
+
|
637
|
+
# Update properties of this object
|
638
|
+
def update!(**args)
|
639
|
+
@results = args[:results] if args.key?(:results)
|
640
|
+
end
|
641
|
+
end
|
642
|
+
|
643
|
+
#
|
644
|
+
class SpeakerDiarizationConfig
|
645
|
+
include Google::Apis::Core::Hashable
|
646
|
+
|
647
|
+
# *Optional* If 'true', enables speaker detection for each recognized word in
|
648
|
+
# the top alternative of the recognition result using a speaker_tag provided
|
649
|
+
# in the WordInfo.
|
650
|
+
# Corresponds to the JSON property `enableSpeakerDiarization`
|
651
|
+
# @return [Boolean]
|
652
|
+
attr_accessor :enable_speaker_diarization
|
653
|
+
alias_method :enable_speaker_diarization?, :enable_speaker_diarization
|
654
|
+
|
655
|
+
# *Optional* Only used if diarization_speaker_count is not set.
|
656
|
+
# Maximum number of speakers in the conversation. This range gives you more
|
657
|
+
# flexibility by allowing the system to automatically determine the correct
|
658
|
+
# number of speakers. If not set, the default value is 6.
|
659
|
+
# Corresponds to the JSON property `maxSpeakerCount`
|
660
|
+
# @return [Fixnum]
|
661
|
+
attr_accessor :max_speaker_count
|
662
|
+
|
663
|
+
# *Optional* Only used if diarization_speaker_count is not set.
|
664
|
+
# Minimum number of speakers in the conversation. This range gives you more
|
665
|
+
# flexibility by allowing the system to automatically determine the correct
|
666
|
+
# number of speakers. If not set, the default value is 2.
|
667
|
+
# Corresponds to the JSON property `minSpeakerCount`
|
668
|
+
# @return [Fixnum]
|
669
|
+
attr_accessor :min_speaker_count
|
670
|
+
|
671
|
+
def initialize(**args)
|
672
|
+
update!(**args)
|
673
|
+
end
|
674
|
+
|
675
|
+
# Update properties of this object
|
676
|
+
def update!(**args)
|
677
|
+
@enable_speaker_diarization = args[:enable_speaker_diarization] if args.key?(:enable_speaker_diarization)
|
678
|
+
@max_speaker_count = args[:max_speaker_count] if args.key?(:max_speaker_count)
|
679
|
+
@min_speaker_count = args[:min_speaker_count] if args.key?(:min_speaker_count)
|
680
|
+
end
|
681
|
+
end
|
682
|
+
|
683
|
+
# Provides "hints" to the speech recognizer to favor specific words and phrases
|
684
|
+
# in the results.
|
685
|
+
class SpeechContext
|
686
|
+
include Google::Apis::Core::Hashable
|
687
|
+
|
688
|
+
# *Optional* A list of strings containing words and phrases "hints" so that
|
689
|
+
# the speech recognition is more likely to recognize them. This can be used
|
690
|
+
# to improve the accuracy for specific words and phrases, for example, if
|
691
|
+
# specific commands are typically spoken by the user. This can also be used
|
692
|
+
# to add additional words to the vocabulary of the recognizer. See
|
693
|
+
# [usage limits](/speech-to-text/quotas#content).
|
694
|
+
# Corresponds to the JSON property `phrases`
|
695
|
+
# @return [Array<String>]
|
696
|
+
attr_accessor :phrases
|
697
|
+
|
698
|
+
def initialize(**args)
|
699
|
+
update!(**args)
|
700
|
+
end
|
701
|
+
|
702
|
+
# Update properties of this object
|
703
|
+
def update!(**args)
|
704
|
+
@phrases = args[:phrases] if args.key?(:phrases)
|
705
|
+
end
|
706
|
+
end
|
707
|
+
|
708
|
+
# Alternative hypotheses (a.k.a. n-best list).
|
709
|
+
class SpeechRecognitionAlternative
|
710
|
+
include Google::Apis::Core::Hashable
|
711
|
+
|
712
|
+
# Output only. The confidence estimate between 0.0 and 1.0. A higher number
|
713
|
+
# indicates an estimated greater likelihood that the recognized words are
|
714
|
+
# correct. This field is set only for the top alternative of a non-streaming
|
715
|
+
# result or, of a streaming result where `is_final=true`.
|
716
|
+
# This field is not guaranteed to be accurate and users should not rely on it
|
717
|
+
# to be always provided.
|
718
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
719
|
+
# Corresponds to the JSON property `confidence`
|
720
|
+
# @return [Float]
|
721
|
+
attr_accessor :confidence
|
722
|
+
|
723
|
+
# Output only. Transcript text representing the words that the user spoke.
|
724
|
+
# Corresponds to the JSON property `transcript`
|
725
|
+
# @return [String]
|
726
|
+
attr_accessor :transcript
|
727
|
+
|
728
|
+
# Output only. A list of word-specific information for each recognized word.
|
729
|
+
# Note: When `enable_speaker_diarization` is true, you will see all the words
|
730
|
+
# from the beginning of the audio.
|
731
|
+
# Corresponds to the JSON property `words`
|
732
|
+
# @return [Array<Google::Apis::SpeechV1p1beta1::WordInfo>]
|
733
|
+
attr_accessor :words
|
734
|
+
|
735
|
+
def initialize(**args)
|
736
|
+
update!(**args)
|
737
|
+
end
|
738
|
+
|
739
|
+
# Update properties of this object
|
740
|
+
def update!(**args)
|
741
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
742
|
+
@transcript = args[:transcript] if args.key?(:transcript)
|
743
|
+
@words = args[:words] if args.key?(:words)
|
744
|
+
end
|
745
|
+
end
|
746
|
+
|
747
|
+
# A speech recognition result corresponding to a portion of the audio.
|
748
|
+
class SpeechRecognitionResult
|
749
|
+
include Google::Apis::Core::Hashable
|
750
|
+
|
751
|
+
# Output only. May contain one or more recognition hypotheses (up to the
|
752
|
+
# maximum specified in `max_alternatives`).
|
753
|
+
# These alternatives are ordered in terms of accuracy, with the top (first)
|
754
|
+
# alternative being the most probable, as ranked by the recognizer.
|
755
|
+
# Corresponds to the JSON property `alternatives`
|
756
|
+
# @return [Array<Google::Apis::SpeechV1p1beta1::SpeechRecognitionAlternative>]
|
757
|
+
attr_accessor :alternatives
|
758
|
+
|
759
|
+
# For multi-channel audio, this is the channel number corresponding to the
|
760
|
+
# recognized result for the audio from that channel.
|
761
|
+
# For audio_channel_count = N, its output values can range from '1' to 'N'.
|
762
|
+
# Corresponds to the JSON property `channelTag`
|
763
|
+
# @return [Fixnum]
|
764
|
+
attr_accessor :channel_tag
|
765
|
+
|
766
|
+
# Output only. The
|
767
|
+
# [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the
|
768
|
+
# language in this result. This language code was detected to have the most
|
769
|
+
# likelihood of being spoken in the audio.
|
770
|
+
# Corresponds to the JSON property `languageCode`
|
771
|
+
# @return [String]
|
772
|
+
attr_accessor :language_code
|
773
|
+
|
774
|
+
def initialize(**args)
|
775
|
+
update!(**args)
|
776
|
+
end
|
777
|
+
|
778
|
+
# Update properties of this object
|
779
|
+
def update!(**args)
|
780
|
+
@alternatives = args[:alternatives] if args.key?(:alternatives)
|
781
|
+
@channel_tag = args[:channel_tag] if args.key?(:channel_tag)
|
782
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
783
|
+
end
|
784
|
+
end
|
785
|
+
|
786
|
+
# The `Status` type defines a logical error model that is suitable for different
|
787
|
+
# programming environments, including REST APIs and RPC APIs. It is used by
|
788
|
+
# [gRPC](https://github.com/grpc). The error model is designed to be:
|
789
|
+
# - Simple to use and understand for most users
|
790
|
+
# - Flexible enough to meet unexpected needs
|
791
|
+
# # Overview
|
792
|
+
# The `Status` message contains three pieces of data: error code, error message,
|
793
|
+
# and error details. The error code should be an enum value of
|
794
|
+
# google.rpc.Code, but it may accept additional error codes if needed. The
|
795
|
+
# error message should be a developer-facing English message that helps
|
796
|
+
# developers *understand* and *resolve* the error. If a localized user-facing
|
797
|
+
# error message is needed, put the localized message in the error details or
|
798
|
+
# localize it in the client. The optional error details may contain arbitrary
|
799
|
+
# information about the error. There is a predefined set of error detail types
|
800
|
+
# in the package `google.rpc` that can be used for common error conditions.
|
801
|
+
# # Language mapping
|
802
|
+
# The `Status` message is the logical representation of the error model, but it
|
803
|
+
# is not necessarily the actual wire format. When the `Status` message is
|
804
|
+
# exposed in different client libraries and different wire protocols, it can be
|
805
|
+
# mapped differently. For example, it will likely be mapped to some exceptions
|
806
|
+
# in Java, but more likely mapped to some error codes in C.
|
807
|
+
# # Other uses
|
808
|
+
# The error model and the `Status` message can be used in a variety of
|
809
|
+
# environments, either with or without APIs, to provide a
|
810
|
+
# consistent developer experience across different environments.
|
811
|
+
# Example uses of this error model include:
|
812
|
+
# - Partial errors. If a service needs to return partial errors to the client,
|
813
|
+
# it may embed the `Status` in the normal response to indicate the partial
|
814
|
+
# errors.
|
815
|
+
# - Workflow errors. A typical workflow has multiple steps. Each step may
|
816
|
+
# have a `Status` message for error reporting.
|
817
|
+
# - Batch operations. If a client uses batch request and batch response, the
|
818
|
+
# `Status` message should be used directly inside batch response, one for
|
819
|
+
# each error sub-response.
|
820
|
+
# - Asynchronous operations. If an API call embeds asynchronous operation
|
821
|
+
# results in its response, the status of those operations should be
|
822
|
+
# represented directly using the `Status` message.
|
823
|
+
# - Logging. If some API errors are stored in logs, the message `Status` could
|
824
|
+
# be used directly after any stripping needed for security/privacy reasons.
|
825
|
+
class Status
|
826
|
+
include Google::Apis::Core::Hashable
|
827
|
+
|
828
|
+
# The status code, which should be an enum value of google.rpc.Code.
|
829
|
+
# Corresponds to the JSON property `code`
|
830
|
+
# @return [Fixnum]
|
831
|
+
attr_accessor :code
|
832
|
+
|
833
|
+
# A list of messages that carry the error details. There is a common set of
|
834
|
+
# message types for APIs to use.
|
835
|
+
# Corresponds to the JSON property `details`
|
836
|
+
# @return [Array<Hash<String,Object>>]
|
837
|
+
attr_accessor :details
|
838
|
+
|
839
|
+
# A developer-facing error message, which should be in English. Any
|
840
|
+
# user-facing error message should be localized and sent in the
|
841
|
+
# google.rpc.Status.details field, or localized by the client.
|
842
|
+
# Corresponds to the JSON property `message`
|
843
|
+
# @return [String]
|
844
|
+
attr_accessor :message
|
845
|
+
|
846
|
+
def initialize(**args)
|
847
|
+
update!(**args)
|
848
|
+
end
|
849
|
+
|
850
|
+
# Update properties of this object
|
851
|
+
def update!(**args)
|
852
|
+
@code = args[:code] if args.key?(:code)
|
853
|
+
@details = args[:details] if args.key?(:details)
|
854
|
+
@message = args[:message] if args.key?(:message)
|
855
|
+
end
|
856
|
+
end
|
857
|
+
|
858
|
+
# Word-specific information for recognized words.
|
859
|
+
class WordInfo
|
860
|
+
include Google::Apis::Core::Hashable
|
861
|
+
|
862
|
+
# Output only. The confidence estimate between 0.0 and 1.0. A higher number
|
863
|
+
# indicates an estimated greater likelihood that the recognized words are
|
864
|
+
# correct. This field is set only for the top alternative of a non-streaming
|
865
|
+
# result or, of a streaming result where `is_final=true`.
|
866
|
+
# This field is not guaranteed to be accurate and users should not rely on it
|
867
|
+
# to be always provided.
|
868
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
869
|
+
# Corresponds to the JSON property `confidence`
|
870
|
+
# @return [Float]
|
871
|
+
attr_accessor :confidence
|
872
|
+
|
873
|
+
# Output only. Time offset relative to the beginning of the audio,
|
874
|
+
# and corresponding to the end of the spoken word.
|
875
|
+
# This field is only set if `enable_word_time_offsets=true` and only
|
876
|
+
# in the top hypothesis.
|
877
|
+
# This is an experimental feature and the accuracy of the time offset can
|
878
|
+
# vary.
|
879
|
+
# Corresponds to the JSON property `endTime`
|
880
|
+
# @return [String]
|
881
|
+
attr_accessor :end_time
|
882
|
+
|
883
|
+
# Output only. A distinct integer value is assigned for every speaker within
|
884
|
+
# the audio. This field specifies which one of those speakers was detected to
|
885
|
+
# have spoken this word. Value ranges from '1' to diarization_speaker_count.
|
886
|
+
# speaker_tag is set if enable_speaker_diarization = 'true' and only in the
|
887
|
+
# top alternative.
|
888
|
+
# Corresponds to the JSON property `speakerTag`
|
889
|
+
# @return [Fixnum]
|
890
|
+
attr_accessor :speaker_tag
|
891
|
+
|
892
|
+
# Output only. Time offset relative to the beginning of the audio,
|
893
|
+
# and corresponding to the start of the spoken word.
|
894
|
+
# This field is only set if `enable_word_time_offsets=true` and only
|
895
|
+
# in the top hypothesis.
|
896
|
+
# This is an experimental feature and the accuracy of the time offset can
|
897
|
+
# vary.
|
898
|
+
# Corresponds to the JSON property `startTime`
|
899
|
+
# @return [String]
|
900
|
+
attr_accessor :start_time
|
901
|
+
|
902
|
+
# Output only. The word corresponding to this set of information.
|
903
|
+
# Corresponds to the JSON property `word`
|
904
|
+
# @return [String]
|
905
|
+
attr_accessor :word
|
906
|
+
|
907
|
+
def initialize(**args)
|
908
|
+
update!(**args)
|
909
|
+
end
|
910
|
+
|
911
|
+
# Update properties of this object
|
912
|
+
def update!(**args)
|
913
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
914
|
+
@end_time = args[:end_time] if args.key?(:end_time)
|
915
|
+
@speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag)
|
916
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
917
|
+
@word = args[:word] if args.key?(:word)
|
918
|
+
end
|
919
|
+
end
|
920
|
+
end
|
921
|
+
end
|
922
|
+
end
|