google-api-client 0.24.2 → 0.24.3
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +68 -0
- data/README.md +9 -0
- data/generated/google/apis/adexchangebuyer2_v2beta1.rb +5 -4
- data/generated/google/apis/adexchangebuyer2_v2beta1/classes.rb +90 -87
- data/generated/google/apis/adexchangebuyer2_v2beta1/service.rb +17 -15
- data/generated/google/apis/admin_directory_v1.rb +1 -1
- data/generated/google/apis/admin_directory_v1/classes.rb +155 -0
- data/generated/google/apis/admin_directory_v1/representations.rb +82 -0
- data/generated/google/apis/alertcenter_v1beta1.rb +31 -0
- data/generated/google/apis/alertcenter_v1beta1/classes.rb +835 -0
- data/generated/google/apis/alertcenter_v1beta1/representations.rb +394 -0
- data/generated/google/apis/alertcenter_v1beta1/service.rb +302 -0
- data/generated/google/apis/androiddeviceprovisioning_v1.rb +1 -1
- data/generated/google/apis/androiddeviceprovisioning_v1/classes.rb +37 -0
- data/generated/google/apis/androiddeviceprovisioning_v1/representations.rb +6 -0
- data/generated/google/apis/androiddeviceprovisioning_v1/service.rb +8 -1
- data/generated/google/apis/androidenterprise_v1.rb +1 -1
- data/generated/google/apis/androidenterprise_v1/classes.rb +8 -4
- data/generated/google/apis/androidenterprise_v1/representations.rb +1 -0
- data/generated/google/apis/androidpublisher_v2.rb +1 -1
- data/generated/google/apis/androidpublisher_v2/service.rb +5 -1
- data/generated/google/apis/androidpublisher_v3.rb +1 -1
- data/generated/google/apis/androidpublisher_v3/service.rb +5 -1
- data/generated/google/apis/appengine_v1.rb +1 -1
- data/generated/google/apis/appengine_v1/classes.rb +8 -1
- data/generated/google/apis/appengine_v1/representations.rb +1 -0
- data/generated/google/apis/appengine_v1beta.rb +1 -1
- data/generated/google/apis/appengine_v1beta/classes.rb +1 -1
- data/generated/google/apis/bigquerydatatransfer_v1.rb +1 -1
- data/generated/google/apis/bigquerydatatransfer_v1/classes.rb +6 -5
- data/generated/google/apis/bigquerydatatransfer_v1/service.rb +12 -10
- data/generated/google/apis/calendar_v3.rb +1 -1
- data/generated/google/apis/calendar_v3/service.rb +52 -18
- data/generated/google/apis/cloudasset_v1beta1.rb +34 -0
- data/generated/google/apis/cloudasset_v1beta1/classes.rb +798 -0
- data/generated/google/apis/cloudasset_v1beta1/representations.rb +263 -0
- data/generated/google/apis/cloudasset_v1beta1/service.rb +313 -0
- data/generated/google/apis/cloudbuild_v1.rb +1 -1
- data/generated/google/apis/cloudbuild_v1/classes.rb +42 -5
- data/generated/google/apis/cloudbuild_v1/representations.rb +6 -0
- data/generated/google/apis/cloudiot_v1.rb +1 -1
- data/generated/google/apis/cloudiot_v1/classes.rb +59 -0
- data/generated/google/apis/cloudiot_v1/representations.rb +28 -0
- data/generated/google/apis/cloudiot_v1/service.rb +94 -0
- data/generated/google/apis/composer_v1.rb +1 -1
- data/generated/google/apis/composer_v1/classes.rb +1 -0
- data/generated/google/apis/composer_v1beta1.rb +1 -1
- data/generated/google/apis/composer_v1beta1/classes.rb +34 -5
- data/generated/google/apis/composer_v1beta1/representations.rb +1 -0
- data/generated/google/apis/compute_alpha.rb +1 -1
- data/generated/google/apis/compute_alpha/classes.rb +227 -48
- data/generated/google/apis/compute_alpha/representations.rb +84 -1
- data/generated/google/apis/compute_alpha/service.rb +50 -10
- data/generated/google/apis/compute_beta.rb +1 -1
- data/generated/google/apis/compute_beta/classes.rb +593 -77
- data/generated/google/apis/compute_beta/representations.rb +224 -18
- data/generated/google/apis/compute_beta/service.rb +174 -3
- data/generated/google/apis/compute_v1.rb +1 -1
- data/generated/google/apis/compute_v1/classes.rb +41 -18
- data/generated/google/apis/compute_v1/representations.rb +3 -0
- data/generated/google/apis/content_v2.rb +1 -1
- data/generated/google/apis/content_v2/classes.rb +372 -119
- data/generated/google/apis/content_v2/representations.rb +157 -39
- data/generated/google/apis/content_v2/service.rb +101 -11
- data/generated/google/apis/content_v2sandbox.rb +1 -1
- data/generated/google/apis/content_v2sandbox/classes.rb +372 -119
- data/generated/google/apis/content_v2sandbox/representations.rb +157 -39
- data/generated/google/apis/content_v2sandbox/service.rb +90 -0
- data/generated/google/apis/customsearch_v1.rb +1 -1
- data/generated/google/apis/dataflow_v1b3.rb +1 -1
- data/generated/google/apis/dataflow_v1b3/classes.rb +7 -0
- data/generated/google/apis/dataflow_v1b3/representations.rb +1 -0
- data/generated/google/apis/dataproc_v1.rb +1 -1
- data/generated/google/apis/dataproc_v1/classes.rb +12 -0
- data/generated/google/apis/dataproc_v1/representations.rb +2 -0
- data/generated/google/apis/dataproc_v1beta2.rb +1 -1
- data/generated/google/apis/dataproc_v1beta2/classes.rb +21 -6
- data/generated/google/apis/dataproc_v1beta2/representations.rb +2 -0
- data/generated/google/apis/datastore_v1.rb +1 -1
- data/generated/google/apis/datastore_v1/classes.rb +2 -2
- data/generated/google/apis/datastore_v1beta3.rb +1 -1
- data/generated/google/apis/datastore_v1beta3/classes.rb +2 -2
- data/generated/google/apis/dlp_v2.rb +1 -1
- data/generated/google/apis/dlp_v2/classes.rb +110 -5
- data/generated/google/apis/dlp_v2/representations.rb +17 -0
- data/generated/google/apis/dlp_v2/service.rb +41 -3
- data/generated/google/apis/file_v1beta1.rb +1 -1
- data/generated/google/apis/file_v1beta1/classes.rb +0 -234
- data/generated/google/apis/file_v1beta1/representations.rb +0 -79
- data/generated/google/apis/firebasedynamiclinks_v1.rb +1 -1
- data/generated/google/apis/firebasedynamiclinks_v1/classes.rb +19 -1
- data/generated/google/apis/firebasedynamiclinks_v1/representations.rb +3 -0
- data/generated/google/apis/firebasedynamiclinks_v1/service.rb +4 -1
- data/generated/google/apis/firebasehosting_v1beta1.rb +43 -0
- data/generated/google/apis/firebasehosting_v1beta1/classes.rb +767 -0
- data/generated/google/apis/firebasehosting_v1beta1/representations.rb +337 -0
- data/generated/google/apis/firebasehosting_v1beta1/service.rb +502 -0
- data/generated/google/apis/firebaserules_v1.rb +1 -1
- data/generated/google/apis/firebaserules_v1/classes.rb +8 -0
- data/generated/google/apis/firebaserules_v1/representations.rb +1 -0
- data/generated/google/apis/firebaserules_v1/service.rb +1 -1
- data/generated/google/apis/firestore_v1beta2.rb +1 -1
- data/generated/google/apis/firestore_v1beta2/service.rb +80 -80
- data/generated/google/apis/games_v1.rb +1 -1
- data/generated/google/apis/games_v1/service.rb +4 -1
- data/generated/google/apis/iam_v1.rb +1 -1
- data/generated/google/apis/iam_v1/classes.rb +3 -1
- data/generated/google/apis/iamcredentials_v1.rb +1 -1
- data/generated/google/apis/iamcredentials_v1/service.rb +0 -10
- data/generated/google/apis/iap_v1beta1.rb +1 -1
- data/generated/google/apis/iap_v1beta1/service.rb +339 -0
- data/generated/google/apis/jobs_v2.rb +1 -1
- data/generated/google/apis/jobs_v2/classes.rb +45 -37
- data/generated/google/apis/jobs_v3.rb +1 -1
- data/generated/google/apis/jobs_v3/classes.rb +21 -18
- data/generated/google/apis/jobs_v3p1beta1.rb +1 -1
- data/generated/google/apis/jobs_v3p1beta1/classes.rb +45 -20
- data/generated/google/apis/jobs_v3p1beta1/representations.rb +2 -0
- data/generated/google/apis/language_v1.rb +1 -1
- data/generated/google/apis/language_v1beta1.rb +1 -1
- data/generated/google/apis/language_v1beta2.rb +1 -1
- data/generated/google/apis/logging_v2.rb +1 -1
- data/generated/google/apis/logging_v2/classes.rb +12 -0
- data/generated/google/apis/logging_v2/representations.rb +1 -0
- data/generated/google/apis/logging_v2beta1.rb +1 -1
- data/generated/google/apis/logging_v2beta1/classes.rb +12 -0
- data/generated/google/apis/logging_v2beta1/representations.rb +1 -0
- data/generated/google/apis/ml_v1.rb +1 -1
- data/generated/google/apis/ml_v1/classes.rb +2 -2
- data/generated/google/apis/monitoring_v3.rb +1 -1
- data/generated/google/apis/monitoring_v3/classes.rb +19 -17
- data/generated/google/apis/monitoring_v3/representations.rb +1 -2
- data/generated/google/apis/partners_v2.rb +1 -1
- data/generated/google/apis/partners_v2/classes.rb +18 -15
- data/generated/google/apis/proximitybeacon_v1beta1.rb +1 -1
- data/generated/google/apis/proximitybeacon_v1beta1/classes.rb +18 -15
- data/generated/google/apis/redis_v1.rb +1 -1
- data/generated/google/apis/redis_v1/classes.rb +1 -1
- data/generated/google/apis/serviceconsumermanagement_v1.rb +1 -1
- data/generated/google/apis/serviceconsumermanagement_v1/classes.rb +1 -1
- data/generated/google/apis/servicemanagement_v1.rb +1 -1
- data/generated/google/apis/servicemanagement_v1/classes.rb +2 -150
- data/generated/google/apis/servicemanagement_v1/representations.rb +0 -42
- data/generated/google/apis/servicenetworking_v1beta.rb +38 -0
- data/generated/google/apis/servicenetworking_v1beta/classes.rb +3440 -0
- data/generated/google/apis/servicenetworking_v1beta/representations.rb +992 -0
- data/generated/google/apis/servicenetworking_v1beta/service.rb +227 -0
- data/generated/google/apis/serviceusage_v1.rb +1 -1
- data/generated/google/apis/serviceusage_v1/classes.rb +1 -1
- data/generated/google/apis/serviceusage_v1beta1.rb +1 -1
- data/generated/google/apis/serviceusage_v1beta1/classes.rb +1 -1
- data/generated/google/apis/serviceuser_v1.rb +1 -1
- data/generated/google/apis/serviceuser_v1/classes.rb +2 -150
- data/generated/google/apis/serviceuser_v1/representations.rb +0 -42
- data/generated/google/apis/spanner_v1.rb +1 -1
- data/generated/google/apis/spanner_v1/classes.rb +308 -30
- data/generated/google/apis/spanner_v1/representations.rb +17 -0
- data/generated/google/apis/streetviewpublish_v1.rb +1 -1
- data/generated/google/apis/streetviewpublish_v1/classes.rb +12 -0
- data/generated/google/apis/streetviewpublish_v1/representations.rb +1 -0
- data/generated/google/apis/testing_v1.rb +1 -1
- data/generated/google/apis/testing_v1/classes.rb +47 -0
- data/generated/google/apis/testing_v1/representations.rb +18 -0
- data/generated/google/apis/videointelligence_v1.rb +1 -1
- data/generated/google/apis/videointelligence_v1/classes.rb +676 -0
- data/generated/google/apis/videointelligence_v1/representations.rb +306 -0
- data/generated/google/apis/videointelligence_v1beta2.rb +1 -1
- data/generated/google/apis/videointelligence_v1beta2/classes.rb +676 -0
- data/generated/google/apis/videointelligence_v1beta2/representations.rb +306 -0
- data/generated/google/apis/{videointelligence_v1beta1.rb → videointelligence_v1p1beta1.rb} +6 -6
- data/generated/google/apis/{videointelligence_v1beta1 → videointelligence_v1p1beta1}/classes.rb +885 -489
- data/generated/google/apis/{videointelligence_v1beta1 → videointelligence_v1p1beta1}/representations.rb +357 -194
- data/generated/google/apis/{videointelligence_v1beta1 → videointelligence_v1p1beta1}/service.rb +12 -12
- data/generated/google/apis/vision_v1.rb +1 -1
- data/generated/google/apis/vision_v1/classes.rb +1 -1
- data/generated/google/apis/vision_v1p1beta1.rb +1 -1
- data/generated/google/apis/vision_v1p1beta1/classes.rb +1 -1
- data/generated/google/apis/vision_v1p2beta1.rb +1 -1
- data/generated/google/apis/vision_v1p2beta1/classes.rb +1 -1
- data/generated/google/apis/youtube_partner_v1.rb +2 -2
- data/generated/google/apis/youtube_partner_v1/classes.rb +2 -1
- data/generated/google/apis/youtube_partner_v1/service.rb +1 -1
- data/lib/google/apis/version.rb +1 -1
- metadata +22 -6
@@ -268,6 +268,120 @@ module Google
|
|
268
268
|
include Google::Apis::Core::JsonObjectSupport
|
269
269
|
end
|
270
270
|
|
271
|
+
class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress
|
272
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
273
|
+
|
274
|
+
include Google::Apis::Core::JsonObjectSupport
|
275
|
+
end
|
276
|
+
|
277
|
+
class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse
|
278
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
279
|
+
|
280
|
+
include Google::Apis::Core::JsonObjectSupport
|
281
|
+
end
|
282
|
+
|
283
|
+
class GoogleCloudVideointelligenceV1p2beta1Entity
|
284
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
285
|
+
|
286
|
+
include Google::Apis::Core::JsonObjectSupport
|
287
|
+
end
|
288
|
+
|
289
|
+
class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation
|
290
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
291
|
+
|
292
|
+
include Google::Apis::Core::JsonObjectSupport
|
293
|
+
end
|
294
|
+
|
295
|
+
class GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame
|
296
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
297
|
+
|
298
|
+
include Google::Apis::Core::JsonObjectSupport
|
299
|
+
end
|
300
|
+
|
301
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation
|
302
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
303
|
+
|
304
|
+
include Google::Apis::Core::JsonObjectSupport
|
305
|
+
end
|
306
|
+
|
307
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelFrame
|
308
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
309
|
+
|
310
|
+
include Google::Apis::Core::JsonObjectSupport
|
311
|
+
end
|
312
|
+
|
313
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelSegment
|
314
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
315
|
+
|
316
|
+
include Google::Apis::Core::JsonObjectSupport
|
317
|
+
end
|
318
|
+
|
319
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox
|
320
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
321
|
+
|
322
|
+
include Google::Apis::Core::JsonObjectSupport
|
323
|
+
end
|
324
|
+
|
325
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly
|
326
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
327
|
+
|
328
|
+
include Google::Apis::Core::JsonObjectSupport
|
329
|
+
end
|
330
|
+
|
331
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex
|
332
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
333
|
+
|
334
|
+
include Google::Apis::Core::JsonObjectSupport
|
335
|
+
end
|
336
|
+
|
337
|
+
class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation
|
338
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
339
|
+
|
340
|
+
include Google::Apis::Core::JsonObjectSupport
|
341
|
+
end
|
342
|
+
|
343
|
+
class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame
|
344
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
345
|
+
|
346
|
+
include Google::Apis::Core::JsonObjectSupport
|
347
|
+
end
|
348
|
+
|
349
|
+
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
350
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
351
|
+
|
352
|
+
include Google::Apis::Core::JsonObjectSupport
|
353
|
+
end
|
354
|
+
|
355
|
+
class GoogleCloudVideointelligenceV1p2beta1TextFrame
|
356
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
357
|
+
|
358
|
+
include Google::Apis::Core::JsonObjectSupport
|
359
|
+
end
|
360
|
+
|
361
|
+
class GoogleCloudVideointelligenceV1p2beta1TextSegment
|
362
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
363
|
+
|
364
|
+
include Google::Apis::Core::JsonObjectSupport
|
365
|
+
end
|
366
|
+
|
367
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress
|
368
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
369
|
+
|
370
|
+
include Google::Apis::Core::JsonObjectSupport
|
371
|
+
end
|
372
|
+
|
373
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults
|
374
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
375
|
+
|
376
|
+
include Google::Apis::Core::JsonObjectSupport
|
377
|
+
end
|
378
|
+
|
379
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoSegment
|
380
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
381
|
+
|
382
|
+
include Google::Apis::Core::JsonObjectSupport
|
383
|
+
end
|
384
|
+
|
271
385
|
class GoogleLongrunningCancelOperationRequest
|
272
386
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
273
387
|
|
@@ -698,12 +812,204 @@ module Google
|
|
698
812
|
class GoogleCloudVideointelligenceV1p1beta1WordInfo
|
699
813
|
# @private
|
700
814
|
class Representation < Google::Apis::Core::JsonRepresentation
|
815
|
+
property :confidence, as: 'confidence'
|
701
816
|
property :end_time, as: 'endTime'
|
817
|
+
property :speaker_tag, as: 'speakerTag'
|
702
818
|
property :start_time, as: 'startTime'
|
703
819
|
property :word, as: 'word'
|
704
820
|
end
|
705
821
|
end
|
706
822
|
|
823
|
+
class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress
|
824
|
+
# @private
|
825
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
826
|
+
collection :annotation_progress, as: 'annotationProgress', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress::Representation
|
827
|
+
|
828
|
+
end
|
829
|
+
end
|
830
|
+
|
831
|
+
class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse
|
832
|
+
# @private
|
833
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
834
|
+
collection :annotation_results, as: 'annotationResults', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults::Representation
|
835
|
+
|
836
|
+
end
|
837
|
+
end
|
838
|
+
|
839
|
+
class GoogleCloudVideointelligenceV1p2beta1Entity
|
840
|
+
# @private
|
841
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
842
|
+
property :description, as: 'description'
|
843
|
+
property :entity_id, as: 'entityId'
|
844
|
+
property :language_code, as: 'languageCode'
|
845
|
+
end
|
846
|
+
end
|
847
|
+
|
848
|
+
class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation
|
849
|
+
# @private
|
850
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
851
|
+
collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame::Representation
|
852
|
+
|
853
|
+
end
|
854
|
+
end
|
855
|
+
|
856
|
+
class GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame
|
857
|
+
# @private
|
858
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
859
|
+
property :pornography_likelihood, as: 'pornographyLikelihood'
|
860
|
+
property :time_offset, as: 'timeOffset'
|
861
|
+
end
|
862
|
+
end
|
863
|
+
|
864
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation
|
865
|
+
# @private
|
866
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
867
|
+
collection :category_entities, as: 'categoryEntities', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1Entity, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1Entity::Representation
|
868
|
+
|
869
|
+
property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1Entity, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1Entity::Representation
|
870
|
+
|
871
|
+
collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelFrame, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelFrame::Representation
|
872
|
+
|
873
|
+
collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelSegment::Representation
|
874
|
+
|
875
|
+
end
|
876
|
+
end
|
877
|
+
|
878
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelFrame
|
879
|
+
# @private
|
880
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
881
|
+
property :confidence, as: 'confidence'
|
882
|
+
property :time_offset, as: 'timeOffset'
|
883
|
+
end
|
884
|
+
end
|
885
|
+
|
886
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelSegment
|
887
|
+
# @private
|
888
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
889
|
+
property :confidence, as: 'confidence'
|
890
|
+
property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation
|
891
|
+
|
892
|
+
end
|
893
|
+
end
|
894
|
+
|
895
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox
|
896
|
+
# @private
|
897
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
898
|
+
property :bottom, as: 'bottom'
|
899
|
+
property :left, as: 'left'
|
900
|
+
property :right, as: 'right'
|
901
|
+
property :top, as: 'top'
|
902
|
+
end
|
903
|
+
end
|
904
|
+
|
905
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly
|
906
|
+
# @private
|
907
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
908
|
+
collection :vertices, as: 'vertices', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1NormalizedVertex, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1NormalizedVertex::Representation
|
909
|
+
|
910
|
+
end
|
911
|
+
end
|
912
|
+
|
913
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex
|
914
|
+
# @private
|
915
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
916
|
+
property :x, as: 'x'
|
917
|
+
property :y, as: 'y'
|
918
|
+
end
|
919
|
+
end
|
920
|
+
|
921
|
+
class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation
|
922
|
+
# @private
|
923
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
924
|
+
property :confidence, as: 'confidence'
|
925
|
+
property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1Entity, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1Entity::Representation
|
926
|
+
|
927
|
+
collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame::Representation
|
928
|
+
|
929
|
+
property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation
|
930
|
+
|
931
|
+
end
|
932
|
+
end
|
933
|
+
|
934
|
+
class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame
|
935
|
+
# @private
|
936
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
937
|
+
property :normalized_bounding_box, as: 'normalizedBoundingBox', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox::Representation
|
938
|
+
|
939
|
+
property :time_offset, as: 'timeOffset'
|
940
|
+
end
|
941
|
+
end
|
942
|
+
|
943
|
+
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
944
|
+
# @private
|
945
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
946
|
+
property :confidence, as: 'confidence'
|
947
|
+
collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextFrame, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextFrame::Representation
|
948
|
+
|
949
|
+
collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextSegment::Representation
|
950
|
+
|
951
|
+
property :text, as: 'text'
|
952
|
+
end
|
953
|
+
end
|
954
|
+
|
955
|
+
class GoogleCloudVideointelligenceV1p2beta1TextFrame
|
956
|
+
# @private
|
957
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
958
|
+
property :rotated_bounding_box, as: 'rotatedBoundingBox', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly::Representation
|
959
|
+
|
960
|
+
property :time_offset, as: 'timeOffset'
|
961
|
+
end
|
962
|
+
end
|
963
|
+
|
964
|
+
class GoogleCloudVideointelligenceV1p2beta1TextSegment
|
965
|
+
# @private
|
966
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
967
|
+
property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation
|
968
|
+
|
969
|
+
end
|
970
|
+
end
|
971
|
+
|
972
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress
|
973
|
+
# @private
|
974
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
975
|
+
property :input_uri, as: 'inputUri'
|
976
|
+
property :progress_percent, as: 'progressPercent'
|
977
|
+
property :start_time, as: 'startTime'
|
978
|
+
property :update_time, as: 'updateTime'
|
979
|
+
end
|
980
|
+
end
|
981
|
+
|
982
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults
|
983
|
+
# @private
|
984
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
985
|
+
property :error, as: 'error', class: Google::Apis::VideointelligenceV1::GoogleRpcStatus, decorator: Google::Apis::VideointelligenceV1::GoogleRpcStatus::Representation
|
986
|
+
|
987
|
+
property :explicit_annotation, as: 'explicitAnnotation', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation::Representation
|
988
|
+
|
989
|
+
collection :frame_label_annotations, as: 'frameLabelAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation
|
990
|
+
|
991
|
+
property :input_uri, as: 'inputUri'
|
992
|
+
collection :object_annotations, as: 'objectAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation::Representation
|
993
|
+
|
994
|
+
collection :segment_label_annotations, as: 'segmentLabelAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation
|
995
|
+
|
996
|
+
collection :shot_annotations, as: 'shotAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation
|
997
|
+
|
998
|
+
collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation
|
999
|
+
|
1000
|
+
collection :text_annotations, as: 'textAnnotations', class: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation, decorator: Google::Apis::VideointelligenceV1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation::Representation
|
1001
|
+
|
1002
|
+
end
|
1003
|
+
end
|
1004
|
+
|
1005
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoSegment
|
1006
|
+
# @private
|
1007
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1008
|
+
property :end_time_offset, as: 'endTimeOffset'
|
1009
|
+
property :start_time_offset, as: 'startTimeOffset'
|
1010
|
+
end
|
1011
|
+
end
|
1012
|
+
|
707
1013
|
class GoogleLongrunningCancelOperationRequest
|
708
1014
|
# @private
|
709
1015
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -25,7 +25,7 @@ module Google
|
|
25
25
|
# @see https://cloud.google.com/video-intelligence/docs/
|
26
26
|
module VideointelligenceV1beta2
|
27
27
|
VERSION = 'V1beta2'
|
28
|
-
REVISION = '
|
28
|
+
REVISION = '20180920'
|
29
29
|
|
30
30
|
# View and manage your data across Google Cloud Platform services
|
31
31
|
AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform'
|
@@ -1413,6 +1413,16 @@ module Google
|
|
1413
1413
|
class GoogleCloudVideointelligenceV1p1beta1WordInfo
|
1414
1414
|
include Google::Apis::Core::Hashable
|
1415
1415
|
|
1416
|
+
# Output only. The confidence estimate between 0.0 and 1.0. A higher number
|
1417
|
+
# indicates an estimated greater likelihood that the recognized words are
|
1418
|
+
# correct. This field is set only for the top alternative.
|
1419
|
+
# This field is not guaranteed to be accurate and users should not rely on it
|
1420
|
+
# to be always provided.
|
1421
|
+
# The default of 0.0 is a sentinel value indicating `confidence` was not set.
|
1422
|
+
# Corresponds to the JSON property `confidence`
|
1423
|
+
# @return [Float]
|
1424
|
+
attr_accessor :confidence
|
1425
|
+
|
1416
1426
|
# Time offset relative to the beginning of the audio, and
|
1417
1427
|
# corresponding to the end of the spoken word. This field is only set if
|
1418
1428
|
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
@@ -1421,6 +1431,14 @@ module Google
|
|
1421
1431
|
# @return [String]
|
1422
1432
|
attr_accessor :end_time
|
1423
1433
|
|
1434
|
+
# Output only. A distinct integer value is assigned for every speaker within
|
1435
|
+
# the audio. This field specifies which one of those speakers was detected to
|
1436
|
+
# have spoken this word. Value ranges from 1 up to diarization_speaker_count,
|
1437
|
+
# and is only set if speaker diarization is enabled.
|
1438
|
+
# Corresponds to the JSON property `speakerTag`
|
1439
|
+
# @return [Fixnum]
|
1440
|
+
attr_accessor :speaker_tag
|
1441
|
+
|
1424
1442
|
# Time offset relative to the beginning of the audio, and
|
1425
1443
|
# corresponding to the start of the spoken word. This field is only set if
|
1426
1444
|
# `enable_word_time_offsets=true` and only in the top hypothesis. This is an
|
@@ -1440,12 +1458,670 @@ module Google
|
|
1440
1458
|
|
1441
1459
|
# Update properties of this object
|
1442
1460
|
def update!(**args)
|
1461
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
1443
1462
|
@end_time = args[:end_time] if args.key?(:end_time)
|
1463
|
+
@speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag)
|
1444
1464
|
@start_time = args[:start_time] if args.key?(:start_time)
|
1445
1465
|
@word = args[:word] if args.key?(:word)
|
1446
1466
|
end
|
1447
1467
|
end
|
1448
1468
|
|
1469
|
+
# Video annotation progress. Included in the `metadata`
|
1470
|
+
# field of the `Operation` returned by the `GetOperation`
|
1471
|
+
# call of the `google::longrunning::Operations` service.
|
1472
|
+
class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress
|
1473
|
+
include Google::Apis::Core::Hashable
|
1474
|
+
|
1475
|
+
# Progress metadata for all videos specified in `AnnotateVideoRequest`.
|
1476
|
+
# Corresponds to the JSON property `annotationProgress`
|
1477
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress>]
|
1478
|
+
attr_accessor :annotation_progress
|
1479
|
+
|
1480
|
+
def initialize(**args)
|
1481
|
+
update!(**args)
|
1482
|
+
end
|
1483
|
+
|
1484
|
+
# Update properties of this object
|
1485
|
+
def update!(**args)
|
1486
|
+
@annotation_progress = args[:annotation_progress] if args.key?(:annotation_progress)
|
1487
|
+
end
|
1488
|
+
end
|
1489
|
+
|
1490
|
+
# Video annotation response. Included in the `response`
|
1491
|
+
# field of the `Operation` returned by the `GetOperation`
|
1492
|
+
# call of the `google::longrunning::Operations` service.
|
1493
|
+
class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse
|
1494
|
+
include Google::Apis::Core::Hashable
|
1495
|
+
|
1496
|
+
# Annotation results for all videos specified in `AnnotateVideoRequest`.
|
1497
|
+
# Corresponds to the JSON property `annotationResults`
|
1498
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults>]
|
1499
|
+
attr_accessor :annotation_results
|
1500
|
+
|
1501
|
+
def initialize(**args)
|
1502
|
+
update!(**args)
|
1503
|
+
end
|
1504
|
+
|
1505
|
+
# Update properties of this object
|
1506
|
+
def update!(**args)
|
1507
|
+
@annotation_results = args[:annotation_results] if args.key?(:annotation_results)
|
1508
|
+
end
|
1509
|
+
end
|
1510
|
+
|
1511
|
+
# Detected entity from video analysis.
|
1512
|
+
class GoogleCloudVideointelligenceV1p2beta1Entity
|
1513
|
+
include Google::Apis::Core::Hashable
|
1514
|
+
|
1515
|
+
# Textual description, e.g. `Fixed-gear bicycle`.
|
1516
|
+
# Corresponds to the JSON property `description`
|
1517
|
+
# @return [String]
|
1518
|
+
attr_accessor :description
|
1519
|
+
|
1520
|
+
# Opaque entity ID. Some IDs may be available in
|
1521
|
+
# [Google Knowledge Graph Search
|
1522
|
+
# API](https://developers.google.com/knowledge-graph/).
|
1523
|
+
# Corresponds to the JSON property `entityId`
|
1524
|
+
# @return [String]
|
1525
|
+
attr_accessor :entity_id
|
1526
|
+
|
1527
|
+
# Language code for `description` in BCP-47 format.
|
1528
|
+
# Corresponds to the JSON property `languageCode`
|
1529
|
+
# @return [String]
|
1530
|
+
attr_accessor :language_code
|
1531
|
+
|
1532
|
+
def initialize(**args)
|
1533
|
+
update!(**args)
|
1534
|
+
end
|
1535
|
+
|
1536
|
+
# Update properties of this object
|
1537
|
+
def update!(**args)
|
1538
|
+
@description = args[:description] if args.key?(:description)
|
1539
|
+
@entity_id = args[:entity_id] if args.key?(:entity_id)
|
1540
|
+
@language_code = args[:language_code] if args.key?(:language_code)
|
1541
|
+
end
|
1542
|
+
end
|
1543
|
+
|
1544
|
+
# Explicit content annotation (based on per-frame visual signals only).
|
1545
|
+
# If no explicit content has been detected in a frame, no annotations are
|
1546
|
+
# present for that frame.
|
1547
|
+
class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation
|
1548
|
+
include Google::Apis::Core::Hashable
|
1549
|
+
|
1550
|
+
# All video frames where explicit content was detected.
|
1551
|
+
# Corresponds to the JSON property `frames`
|
1552
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame>]
|
1553
|
+
attr_accessor :frames
|
1554
|
+
|
1555
|
+
def initialize(**args)
|
1556
|
+
update!(**args)
|
1557
|
+
end
|
1558
|
+
|
1559
|
+
# Update properties of this object
|
1560
|
+
def update!(**args)
|
1561
|
+
@frames = args[:frames] if args.key?(:frames)
|
1562
|
+
end
|
1563
|
+
end
|
1564
|
+
|
1565
|
+
# Video frame level annotation results for explicit content.
|
1566
|
+
class GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame
|
1567
|
+
include Google::Apis::Core::Hashable
|
1568
|
+
|
1569
|
+
# Likelihood of the pornography content..
|
1570
|
+
# Corresponds to the JSON property `pornographyLikelihood`
|
1571
|
+
# @return [String]
|
1572
|
+
attr_accessor :pornography_likelihood
|
1573
|
+
|
1574
|
+
# Time-offset, relative to the beginning of the video, corresponding to the
|
1575
|
+
# video frame for this location.
|
1576
|
+
# Corresponds to the JSON property `timeOffset`
|
1577
|
+
# @return [String]
|
1578
|
+
attr_accessor :time_offset
|
1579
|
+
|
1580
|
+
def initialize(**args)
|
1581
|
+
update!(**args)
|
1582
|
+
end
|
1583
|
+
|
1584
|
+
# Update properties of this object
|
1585
|
+
def update!(**args)
|
1586
|
+
@pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood)
|
1587
|
+
@time_offset = args[:time_offset] if args.key?(:time_offset)
|
1588
|
+
end
|
1589
|
+
end
|
1590
|
+
|
1591
|
+
# Label annotation.
|
1592
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation
|
1593
|
+
include Google::Apis::Core::Hashable
|
1594
|
+
|
1595
|
+
# Common categories for the detected entity.
|
1596
|
+
# E.g. when the label is `Terrier` the category is likely `dog`. And in some
|
1597
|
+
# cases there might be more than one categories e.g. `Terrier` could also be
|
1598
|
+
# a `pet`.
|
1599
|
+
# Corresponds to the JSON property `categoryEntities`
|
1600
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1Entity>]
|
1601
|
+
attr_accessor :category_entities
|
1602
|
+
|
1603
|
+
# Detected entity from video analysis.
|
1604
|
+
# Corresponds to the JSON property `entity`
|
1605
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1Entity]
|
1606
|
+
attr_accessor :entity
|
1607
|
+
|
1608
|
+
# All video frames where a label was detected.
|
1609
|
+
# Corresponds to the JSON property `frames`
|
1610
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1LabelFrame>]
|
1611
|
+
attr_accessor :frames
|
1612
|
+
|
1613
|
+
# All video segments where a label was detected.
|
1614
|
+
# Corresponds to the JSON property `segments`
|
1615
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1LabelSegment>]
|
1616
|
+
attr_accessor :segments
|
1617
|
+
|
1618
|
+
def initialize(**args)
|
1619
|
+
update!(**args)
|
1620
|
+
end
|
1621
|
+
|
1622
|
+
# Update properties of this object
|
1623
|
+
def update!(**args)
|
1624
|
+
@category_entities = args[:category_entities] if args.key?(:category_entities)
|
1625
|
+
@entity = args[:entity] if args.key?(:entity)
|
1626
|
+
@frames = args[:frames] if args.key?(:frames)
|
1627
|
+
@segments = args[:segments] if args.key?(:segments)
|
1628
|
+
end
|
1629
|
+
end
|
1630
|
+
|
1631
|
+
# Video frame level annotation results for label detection.
|
1632
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelFrame
|
1633
|
+
include Google::Apis::Core::Hashable
|
1634
|
+
|
1635
|
+
# Confidence that the label is accurate. Range: [0, 1].
|
1636
|
+
# Corresponds to the JSON property `confidence`
|
1637
|
+
# @return [Float]
|
1638
|
+
attr_accessor :confidence
|
1639
|
+
|
1640
|
+
# Time-offset, relative to the beginning of the video, corresponding to the
|
1641
|
+
# video frame for this location.
|
1642
|
+
# Corresponds to the JSON property `timeOffset`
|
1643
|
+
# @return [String]
|
1644
|
+
attr_accessor :time_offset
|
1645
|
+
|
1646
|
+
def initialize(**args)
|
1647
|
+
update!(**args)
|
1648
|
+
end
|
1649
|
+
|
1650
|
+
# Update properties of this object
|
1651
|
+
def update!(**args)
|
1652
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
1653
|
+
@time_offset = args[:time_offset] if args.key?(:time_offset)
|
1654
|
+
end
|
1655
|
+
end
|
1656
|
+
|
1657
|
+
# Video segment level annotation results for label detection.
|
1658
|
+
class GoogleCloudVideointelligenceV1p2beta1LabelSegment
|
1659
|
+
include Google::Apis::Core::Hashable
|
1660
|
+
|
1661
|
+
# Confidence that the label is accurate. Range: [0, 1].
|
1662
|
+
# Corresponds to the JSON property `confidence`
|
1663
|
+
# @return [Float]
|
1664
|
+
attr_accessor :confidence
|
1665
|
+
|
1666
|
+
# Video segment.
|
1667
|
+
# Corresponds to the JSON property `segment`
|
1668
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
|
1669
|
+
attr_accessor :segment
|
1670
|
+
|
1671
|
+
def initialize(**args)
|
1672
|
+
update!(**args)
|
1673
|
+
end
|
1674
|
+
|
1675
|
+
# Update properties of this object
|
1676
|
+
def update!(**args)
|
1677
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
1678
|
+
@segment = args[:segment] if args.key?(:segment)
|
1679
|
+
end
|
1680
|
+
end
|
1681
|
+
|
1682
|
+
# Normalized bounding box.
|
1683
|
+
# The normalized vertex coordinates are relative to the original image.
|
1684
|
+
# Range: [0, 1].
|
1685
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox
|
1686
|
+
include Google::Apis::Core::Hashable
|
1687
|
+
|
1688
|
+
# Bottom Y coordinate.
|
1689
|
+
# Corresponds to the JSON property `bottom`
|
1690
|
+
# @return [Float]
|
1691
|
+
attr_accessor :bottom
|
1692
|
+
|
1693
|
+
# Left X coordinate.
|
1694
|
+
# Corresponds to the JSON property `left`
|
1695
|
+
# @return [Float]
|
1696
|
+
attr_accessor :left
|
1697
|
+
|
1698
|
+
# Right X coordinate.
|
1699
|
+
# Corresponds to the JSON property `right`
|
1700
|
+
# @return [Float]
|
1701
|
+
attr_accessor :right
|
1702
|
+
|
1703
|
+
# Top Y coordinate.
|
1704
|
+
# Corresponds to the JSON property `top`
|
1705
|
+
# @return [Float]
|
1706
|
+
attr_accessor :top
|
1707
|
+
|
1708
|
+
def initialize(**args)
|
1709
|
+
update!(**args)
|
1710
|
+
end
|
1711
|
+
|
1712
|
+
# Update properties of this object
|
1713
|
+
def update!(**args)
|
1714
|
+
@bottom = args[:bottom] if args.key?(:bottom)
|
1715
|
+
@left = args[:left] if args.key?(:left)
|
1716
|
+
@right = args[:right] if args.key?(:right)
|
1717
|
+
@top = args[:top] if args.key?(:top)
|
1718
|
+
end
|
1719
|
+
end
|
1720
|
+
|
1721
|
+
# Normalized bounding polygon for text (that might not be aligned with axis).
|
1722
|
+
# Contains list of the corner points in clockwise order starting from
|
1723
|
+
# top-left corner. For example, for a rectangular bounding box:
|
1724
|
+
# When the text is horizontal it might look like:
|
1725
|
+
# 0----1
|
1726
|
+
# | |
|
1727
|
+
# 3----2
|
1728
|
+
# When it's clockwise rotated 180 degrees around the top-left corner it
|
1729
|
+
# becomes:
|
1730
|
+
# 2----3
|
1731
|
+
# | |
|
1732
|
+
# 1----0
|
1733
|
+
# and the vertex order will still be (0, 1, 2, 3). Note that values can be less
|
1734
|
+
# than 0, or greater than 1 due to trignometric calculations for location of
|
1735
|
+
# the box.
|
1736
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly
|
1737
|
+
include Google::Apis::Core::Hashable
|
1738
|
+
|
1739
|
+
# Normalized vertices of the bounding polygon.
|
1740
|
+
# Corresponds to the JSON property `vertices`
|
1741
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1NormalizedVertex>]
|
1742
|
+
attr_accessor :vertices
|
1743
|
+
|
1744
|
+
def initialize(**args)
|
1745
|
+
update!(**args)
|
1746
|
+
end
|
1747
|
+
|
1748
|
+
# Update properties of this object
|
1749
|
+
def update!(**args)
|
1750
|
+
@vertices = args[:vertices] if args.key?(:vertices)
|
1751
|
+
end
|
1752
|
+
end
|
1753
|
+
|
1754
|
+
# A vertex represents a 2D point in the image.
|
1755
|
+
# NOTE: the normalized vertex coordinates are relative to the original image
|
1756
|
+
# and range from 0 to 1.
|
1757
|
+
class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex
|
1758
|
+
include Google::Apis::Core::Hashable
|
1759
|
+
|
1760
|
+
# X coordinate.
|
1761
|
+
# Corresponds to the JSON property `x`
|
1762
|
+
# @return [Float]
|
1763
|
+
attr_accessor :x
|
1764
|
+
|
1765
|
+
# Y coordinate.
|
1766
|
+
# Corresponds to the JSON property `y`
|
1767
|
+
# @return [Float]
|
1768
|
+
attr_accessor :y
|
1769
|
+
|
1770
|
+
def initialize(**args)
|
1771
|
+
update!(**args)
|
1772
|
+
end
|
1773
|
+
|
1774
|
+
# Update properties of this object
|
1775
|
+
def update!(**args)
|
1776
|
+
@x = args[:x] if args.key?(:x)
|
1777
|
+
@y = args[:y] if args.key?(:y)
|
1778
|
+
end
|
1779
|
+
end
|
1780
|
+
|
1781
|
+
# Annotations corresponding to one tracked object.
|
1782
|
+
class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation
|
1783
|
+
include Google::Apis::Core::Hashable
|
1784
|
+
|
1785
|
+
# Object category's labeling confidence of this track.
|
1786
|
+
# Corresponds to the JSON property `confidence`
|
1787
|
+
# @return [Float]
|
1788
|
+
attr_accessor :confidence
|
1789
|
+
|
1790
|
+
# Detected entity from video analysis.
|
1791
|
+
# Corresponds to the JSON property `entity`
|
1792
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1Entity]
|
1793
|
+
attr_accessor :entity
|
1794
|
+
|
1795
|
+
# Information corresponding to all frames where this object track appears.
|
1796
|
+
# Corresponds to the JSON property `frames`
|
1797
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame>]
|
1798
|
+
attr_accessor :frames
|
1799
|
+
|
1800
|
+
# Video segment.
|
1801
|
+
# Corresponds to the JSON property `segment`
|
1802
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
|
1803
|
+
attr_accessor :segment
|
1804
|
+
|
1805
|
+
def initialize(**args)
|
1806
|
+
update!(**args)
|
1807
|
+
end
|
1808
|
+
|
1809
|
+
# Update properties of this object
|
1810
|
+
def update!(**args)
|
1811
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
1812
|
+
@entity = args[:entity] if args.key?(:entity)
|
1813
|
+
@frames = args[:frames] if args.key?(:frames)
|
1814
|
+
@segment = args[:segment] if args.key?(:segment)
|
1815
|
+
end
|
1816
|
+
end
|
1817
|
+
|
1818
|
+
# Video frame level annotations for object detection and tracking. This field
|
1819
|
+
# stores per frame location, time offset, and confidence.
|
1820
|
+
class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame
|
1821
|
+
include Google::Apis::Core::Hashable
|
1822
|
+
|
1823
|
+
# Normalized bounding box.
|
1824
|
+
# The normalized vertex coordinates are relative to the original image.
|
1825
|
+
# Range: [0, 1].
|
1826
|
+
# Corresponds to the JSON property `normalizedBoundingBox`
|
1827
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox]
|
1828
|
+
attr_accessor :normalized_bounding_box
|
1829
|
+
|
1830
|
+
# The timestamp of the frame in microseconds.
|
1831
|
+
# Corresponds to the JSON property `timeOffset`
|
1832
|
+
# @return [String]
|
1833
|
+
attr_accessor :time_offset
|
1834
|
+
|
1835
|
+
def initialize(**args)
|
1836
|
+
update!(**args)
|
1837
|
+
end
|
1838
|
+
|
1839
|
+
# Update properties of this object
|
1840
|
+
def update!(**args)
|
1841
|
+
@normalized_bounding_box = args[:normalized_bounding_box] if args.key?(:normalized_bounding_box)
|
1842
|
+
@time_offset = args[:time_offset] if args.key?(:time_offset)
|
1843
|
+
end
|
1844
|
+
end
|
1845
|
+
|
1846
|
+
# Annotations related to one detected OCR text snippet. This will contain the
|
1847
|
+
# corresponding text, confidence value, and frame level information for each
|
1848
|
+
# detection.
|
1849
|
+
class GoogleCloudVideointelligenceV1p2beta1TextAnnotation
|
1850
|
+
include Google::Apis::Core::Hashable
|
1851
|
+
|
1852
|
+
# Confidence for the track of detected text. It is calculated as the highest
|
1853
|
+
# over all frames where OCR detected text appears.
|
1854
|
+
# Corresponds to the JSON property `confidence`
|
1855
|
+
# @return [Float]
|
1856
|
+
attr_accessor :confidence
|
1857
|
+
|
1858
|
+
# Information related to the frames where OCR detected text appears.
|
1859
|
+
# Corresponds to the JSON property `frames`
|
1860
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1TextFrame>]
|
1861
|
+
attr_accessor :frames
|
1862
|
+
|
1863
|
+
# All video segments where OCR detected text appears.
|
1864
|
+
# Corresponds to the JSON property `segments`
|
1865
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1TextSegment>]
|
1866
|
+
attr_accessor :segments
|
1867
|
+
|
1868
|
+
# The detected text.
|
1869
|
+
# Corresponds to the JSON property `text`
|
1870
|
+
# @return [String]
|
1871
|
+
attr_accessor :text
|
1872
|
+
|
1873
|
+
def initialize(**args)
|
1874
|
+
update!(**args)
|
1875
|
+
end
|
1876
|
+
|
1877
|
+
# Update properties of this object
|
1878
|
+
def update!(**args)
|
1879
|
+
@confidence = args[:confidence] if args.key?(:confidence)
|
1880
|
+
@frames = args[:frames] if args.key?(:frames)
|
1881
|
+
@segments = args[:segments] if args.key?(:segments)
|
1882
|
+
@text = args[:text] if args.key?(:text)
|
1883
|
+
end
|
1884
|
+
end
|
1885
|
+
|
1886
|
+
# Video frame level annotation results for text annotation (OCR).
|
1887
|
+
# Contains information regarding timestamp and bounding box locations for the
|
1888
|
+
# frames containing detected OCR text snippets.
|
1889
|
+
class GoogleCloudVideointelligenceV1p2beta1TextFrame
|
1890
|
+
include Google::Apis::Core::Hashable
|
1891
|
+
|
1892
|
+
# Normalized bounding polygon for text (that might not be aligned with axis).
|
1893
|
+
# Contains list of the corner points in clockwise order starting from
|
1894
|
+
# top-left corner. For example, for a rectangular bounding box:
|
1895
|
+
# When the text is horizontal it might look like:
|
1896
|
+
# 0----1
|
1897
|
+
# | |
|
1898
|
+
# 3----2
|
1899
|
+
# When it's clockwise rotated 180 degrees around the top-left corner it
|
1900
|
+
# becomes:
|
1901
|
+
# 2----3
|
1902
|
+
# | |
|
1903
|
+
# 1----0
|
1904
|
+
# and the vertex order will still be (0, 1, 2, 3). Note that values can be less
|
1905
|
+
# than 0, or greater than 1 due to trignometric calculations for location of
|
1906
|
+
# the box.
|
1907
|
+
# Corresponds to the JSON property `rotatedBoundingBox`
|
1908
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly]
|
1909
|
+
attr_accessor :rotated_bounding_box
|
1910
|
+
|
1911
|
+
# Timestamp of this frame.
|
1912
|
+
# Corresponds to the JSON property `timeOffset`
|
1913
|
+
# @return [String]
|
1914
|
+
attr_accessor :time_offset
|
1915
|
+
|
1916
|
+
def initialize(**args)
|
1917
|
+
update!(**args)
|
1918
|
+
end
|
1919
|
+
|
1920
|
+
# Update properties of this object
|
1921
|
+
def update!(**args)
|
1922
|
+
@rotated_bounding_box = args[:rotated_bounding_box] if args.key?(:rotated_bounding_box)
|
1923
|
+
@time_offset = args[:time_offset] if args.key?(:time_offset)
|
1924
|
+
end
|
1925
|
+
end
|
1926
|
+
|
1927
|
+
# Video segment level annotation results for text detection.
|
1928
|
+
class GoogleCloudVideointelligenceV1p2beta1TextSegment
|
1929
|
+
include Google::Apis::Core::Hashable
|
1930
|
+
|
1931
|
+
# Video segment.
|
1932
|
+
# Corresponds to the JSON property `segment`
|
1933
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
|
1934
|
+
attr_accessor :segment
|
1935
|
+
|
1936
|
+
def initialize(**args)
|
1937
|
+
update!(**args)
|
1938
|
+
end
|
1939
|
+
|
1940
|
+
# Update properties of this object
|
1941
|
+
def update!(**args)
|
1942
|
+
@segment = args[:segment] if args.key?(:segment)
|
1943
|
+
end
|
1944
|
+
end
|
1945
|
+
|
1946
|
+
# Annotation progress for a single video.
|
1947
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress
|
1948
|
+
include Google::Apis::Core::Hashable
|
1949
|
+
|
1950
|
+
# Video file location in
|
1951
|
+
# [Google Cloud Storage](https://cloud.google.com/storage/).
|
1952
|
+
# Corresponds to the JSON property `inputUri`
|
1953
|
+
# @return [String]
|
1954
|
+
attr_accessor :input_uri
|
1955
|
+
|
1956
|
+
# Approximate percentage processed thus far. Guaranteed to be
|
1957
|
+
# 100 when fully processed.
|
1958
|
+
# Corresponds to the JSON property `progressPercent`
|
1959
|
+
# @return [Fixnum]
|
1960
|
+
attr_accessor :progress_percent
|
1961
|
+
|
1962
|
+
# Time when the request was received.
|
1963
|
+
# Corresponds to the JSON property `startTime`
|
1964
|
+
# @return [String]
|
1965
|
+
attr_accessor :start_time
|
1966
|
+
|
1967
|
+
# Time of the most recent update.
|
1968
|
+
# Corresponds to the JSON property `updateTime`
|
1969
|
+
# @return [String]
|
1970
|
+
attr_accessor :update_time
|
1971
|
+
|
1972
|
+
def initialize(**args)
|
1973
|
+
update!(**args)
|
1974
|
+
end
|
1975
|
+
|
1976
|
+
# Update properties of this object
|
1977
|
+
def update!(**args)
|
1978
|
+
@input_uri = args[:input_uri] if args.key?(:input_uri)
|
1979
|
+
@progress_percent = args[:progress_percent] if args.key?(:progress_percent)
|
1980
|
+
@start_time = args[:start_time] if args.key?(:start_time)
|
1981
|
+
@update_time = args[:update_time] if args.key?(:update_time)
|
1982
|
+
end
|
1983
|
+
end
|
1984
|
+
|
1985
|
+
# Annotation results for a single video.
|
1986
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults
|
1987
|
+
include Google::Apis::Core::Hashable
|
1988
|
+
|
1989
|
+
# The `Status` type defines a logical error model that is suitable for different
|
1990
|
+
# programming environments, including REST APIs and RPC APIs. It is used by
|
1991
|
+
# [gRPC](https://github.com/grpc). The error model is designed to be:
|
1992
|
+
# - Simple to use and understand for most users
|
1993
|
+
# - Flexible enough to meet unexpected needs
|
1994
|
+
# # Overview
|
1995
|
+
# The `Status` message contains three pieces of data: error code, error message,
|
1996
|
+
# and error details. The error code should be an enum value of
|
1997
|
+
# google.rpc.Code, but it may accept additional error codes if needed. The
|
1998
|
+
# error message should be a developer-facing English message that helps
|
1999
|
+
# developers *understand* and *resolve* the error. If a localized user-facing
|
2000
|
+
# error message is needed, put the localized message in the error details or
|
2001
|
+
# localize it in the client. The optional error details may contain arbitrary
|
2002
|
+
# information about the error. There is a predefined set of error detail types
|
2003
|
+
# in the package `google.rpc` that can be used for common error conditions.
|
2004
|
+
# # Language mapping
|
2005
|
+
# The `Status` message is the logical representation of the error model, but it
|
2006
|
+
# is not necessarily the actual wire format. When the `Status` message is
|
2007
|
+
# exposed in different client libraries and different wire protocols, it can be
|
2008
|
+
# mapped differently. For example, it will likely be mapped to some exceptions
|
2009
|
+
# in Java, but more likely mapped to some error codes in C.
|
2010
|
+
# # Other uses
|
2011
|
+
# The error model and the `Status` message can be used in a variety of
|
2012
|
+
# environments, either with or without APIs, to provide a
|
2013
|
+
# consistent developer experience across different environments.
|
2014
|
+
# Example uses of this error model include:
|
2015
|
+
# - Partial errors. If a service needs to return partial errors to the client,
|
2016
|
+
# it may embed the `Status` in the normal response to indicate the partial
|
2017
|
+
# errors.
|
2018
|
+
# - Workflow errors. A typical workflow has multiple steps. Each step may
|
2019
|
+
# have a `Status` message for error reporting.
|
2020
|
+
# - Batch operations. If a client uses batch request and batch response, the
|
2021
|
+
# `Status` message should be used directly inside batch response, one for
|
2022
|
+
# each error sub-response.
|
2023
|
+
# - Asynchronous operations. If an API call embeds asynchronous operation
|
2024
|
+
# results in its response, the status of those operations should be
|
2025
|
+
# represented directly using the `Status` message.
|
2026
|
+
# - Logging. If some API errors are stored in logs, the message `Status` could
|
2027
|
+
# be used directly after any stripping needed for security/privacy reasons.
|
2028
|
+
# Corresponds to the JSON property `error`
|
2029
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleRpcStatus]
|
2030
|
+
attr_accessor :error
|
2031
|
+
|
2032
|
+
# Explicit content annotation (based on per-frame visual signals only).
|
2033
|
+
# If no explicit content has been detected in a frame, no annotations are
|
2034
|
+
# present for that frame.
|
2035
|
+
# Corresponds to the JSON property `explicitAnnotation`
|
2036
|
+
# @return [Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation]
|
2037
|
+
attr_accessor :explicit_annotation
|
2038
|
+
|
2039
|
+
# Label annotations on frame level.
|
2040
|
+
# There is exactly one element for each unique label.
|
2041
|
+
# Corresponds to the JSON property `frameLabelAnnotations`
|
2042
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
|
2043
|
+
attr_accessor :frame_label_annotations
|
2044
|
+
|
2045
|
+
# Video file location in
|
2046
|
+
# [Google Cloud Storage](https://cloud.google.com/storage/).
|
2047
|
+
# Corresponds to the JSON property `inputUri`
|
2048
|
+
# @return [String]
|
2049
|
+
attr_accessor :input_uri
|
2050
|
+
|
2051
|
+
# Annotations for list of objects detected and tracked in video.
|
2052
|
+
# Corresponds to the JSON property `objectAnnotations`
|
2053
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation>]
|
2054
|
+
attr_accessor :object_annotations
|
2055
|
+
|
2056
|
+
# Label annotations on video level or user specified segment level.
|
2057
|
+
# There is exactly one element for each unique label.
|
2058
|
+
# Corresponds to the JSON property `segmentLabelAnnotations`
|
2059
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
|
2060
|
+
attr_accessor :segment_label_annotations
|
2061
|
+
|
2062
|
+
# Shot annotations. Each shot is represented as a video segment.
|
2063
|
+
# Corresponds to the JSON property `shotAnnotations`
|
2064
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1VideoSegment>]
|
2065
|
+
attr_accessor :shot_annotations
|
2066
|
+
|
2067
|
+
# Label annotations on shot level.
|
2068
|
+
# There is exactly one element for each unique label.
|
2069
|
+
# Corresponds to the JSON property `shotLabelAnnotations`
|
2070
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
|
2071
|
+
attr_accessor :shot_label_annotations
|
2072
|
+
|
2073
|
+
# OCR text detection and tracking.
|
2074
|
+
# Annotations for list of detected text snippets. Each will have list of
|
2075
|
+
# frame information associated with it.
|
2076
|
+
# Corresponds to the JSON property `textAnnotations`
|
2077
|
+
# @return [Array<Google::Apis::VideointelligenceV1beta2::GoogleCloudVideointelligenceV1p2beta1TextAnnotation>]
|
2078
|
+
attr_accessor :text_annotations
|
2079
|
+
|
2080
|
+
def initialize(**args)
|
2081
|
+
update!(**args)
|
2082
|
+
end
|
2083
|
+
|
2084
|
+
# Update properties of this object
|
2085
|
+
def update!(**args)
|
2086
|
+
@error = args[:error] if args.key?(:error)
|
2087
|
+
@explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
|
2088
|
+
@frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations)
|
2089
|
+
@input_uri = args[:input_uri] if args.key?(:input_uri)
|
2090
|
+
@object_annotations = args[:object_annotations] if args.key?(:object_annotations)
|
2091
|
+
@segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
|
2092
|
+
@shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations)
|
2093
|
+
@shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations)
|
2094
|
+
@text_annotations = args[:text_annotations] if args.key?(:text_annotations)
|
2095
|
+
end
|
2096
|
+
end
|
2097
|
+
|
2098
|
+
# Video segment.
|
2099
|
+
class GoogleCloudVideointelligenceV1p2beta1VideoSegment
|
2100
|
+
include Google::Apis::Core::Hashable
|
2101
|
+
|
2102
|
+
# Time-offset, relative to the beginning of the video,
|
2103
|
+
# corresponding to the end of the segment (inclusive).
|
2104
|
+
# Corresponds to the JSON property `endTimeOffset`
|
2105
|
+
# @return [String]
|
2106
|
+
attr_accessor :end_time_offset
|
2107
|
+
|
2108
|
+
# Time-offset, relative to the beginning of the video,
|
2109
|
+
# corresponding to the start of the segment (inclusive).
|
2110
|
+
# Corresponds to the JSON property `startTimeOffset`
|
2111
|
+
# @return [String]
|
2112
|
+
attr_accessor :start_time_offset
|
2113
|
+
|
2114
|
+
def initialize(**args)
|
2115
|
+
update!(**args)
|
2116
|
+
end
|
2117
|
+
|
2118
|
+
# Update properties of this object
|
2119
|
+
def update!(**args)
|
2120
|
+
@end_time_offset = args[:end_time_offset] if args.key?(:end_time_offset)
|
2121
|
+
@start_time_offset = args[:start_time_offset] if args.key?(:start_time_offset)
|
2122
|
+
end
|
2123
|
+
end
|
2124
|
+
|
1449
2125
|
# This resource represents a long-running operation that is the result of a
|
1450
2126
|
# network API call.
|
1451
2127
|
class GoogleLongrunningOperation
|