google-cloud-video_intelligence 2.0.1 → 3.0.0

Sign up to get free protection for your applications and to get access to all the features.
Files changed (55) hide show
  1. checksums.yaml +4 -4
  2. data/.yardopts +2 -1
  3. data/AUTHENTICATION.md +51 -59
  4. data/LICENSE.md +203 -0
  5. data/MIGRATING.md +274 -0
  6. data/README.md +36 -62
  7. data/lib/{google/cloud/video_intelligence/credentials.rb → google-cloud-video_intelligence.rb} +5 -13
  8. data/lib/google/cloud/video_intelligence.rb +88 -157
  9. data/lib/google/cloud/video_intelligence/version.rb +6 -2
  10. metadata +99 -83
  11. data/LICENSE +0 -201
  12. data/lib/google/cloud/video_intelligence/v1.rb +0 -186
  13. data/lib/google/cloud/video_intelligence/v1/credentials.rb +0 -41
  14. data/lib/google/cloud/video_intelligence/v1/doc/google/cloud/videointelligence/v1/video_intelligence.rb +0 -707
  15. data/lib/google/cloud/video_intelligence/v1/doc/google/longrunning/operations.rb +0 -51
  16. data/lib/google/cloud/video_intelligence/v1/doc/google/protobuf/any.rb +0 -131
  17. data/lib/google/cloud/video_intelligence/v1/doc/google/protobuf/duration.rb +0 -91
  18. data/lib/google/cloud/video_intelligence/v1/doc/google/rpc/status.rb +0 -87
  19. data/lib/google/cloud/video_intelligence/v1/video_intelligence_service_client.rb +0 -306
  20. data/lib/google/cloud/video_intelligence/v1/video_intelligence_service_client_config.json +0 -31
  21. data/lib/google/cloud/video_intelligence/v1beta2.rb +0 -149
  22. data/lib/google/cloud/video_intelligence/v1beta2/credentials.rb +0 -41
  23. data/lib/google/cloud/video_intelligence/v1beta2/doc/google/cloud/videointelligence/v1beta2/video_intelligence.rb +0 -372
  24. data/lib/google/cloud/video_intelligence/v1beta2/doc/google/longrunning/operations.rb +0 -51
  25. data/lib/google/cloud/video_intelligence/v1beta2/doc/google/protobuf/any.rb +0 -131
  26. data/lib/google/cloud/video_intelligence/v1beta2/doc/google/protobuf/duration.rb +0 -91
  27. data/lib/google/cloud/video_intelligence/v1beta2/doc/google/rpc/status.rb +0 -87
  28. data/lib/google/cloud/video_intelligence/v1beta2/video_intelligence_service_client.rb +0 -306
  29. data/lib/google/cloud/video_intelligence/v1beta2/video_intelligence_service_client_config.json +0 -31
  30. data/lib/google/cloud/video_intelligence/v1p1beta1.rb +0 -187
  31. data/lib/google/cloud/video_intelligence/v1p1beta1/credentials.rb +0 -41
  32. data/lib/google/cloud/video_intelligence/v1p1beta1/doc/google/cloud/videointelligence/v1p1beta1/video_intelligence.rb +0 -410
  33. data/lib/google/cloud/video_intelligence/v1p1beta1/doc/google/longrunning/operations.rb +0 -51
  34. data/lib/google/cloud/video_intelligence/v1p1beta1/doc/google/protobuf/any.rb +0 -131
  35. data/lib/google/cloud/video_intelligence/v1p1beta1/doc/google/protobuf/duration.rb +0 -91
  36. data/lib/google/cloud/video_intelligence/v1p1beta1/doc/google/rpc/status.rb +0 -87
  37. data/lib/google/cloud/video_intelligence/v1p1beta1/video_intelligence_service_client.rb +0 -306
  38. data/lib/google/cloud/video_intelligence/v1p1beta1/video_intelligence_service_client_config.json +0 -31
  39. data/lib/google/cloud/video_intelligence/v1p2beta1.rb +0 -187
  40. data/lib/google/cloud/video_intelligence/v1p2beta1/credentials.rb +0 -41
  41. data/lib/google/cloud/video_intelligence/v1p2beta1/doc/google/cloud/videointelligence/v1p2beta1/video_intelligence.rb +0 -442
  42. data/lib/google/cloud/video_intelligence/v1p2beta1/doc/google/longrunning/operations.rb +0 -51
  43. data/lib/google/cloud/video_intelligence/v1p2beta1/doc/google/protobuf/any.rb +0 -131
  44. data/lib/google/cloud/video_intelligence/v1p2beta1/doc/google/protobuf/duration.rb +0 -91
  45. data/lib/google/cloud/video_intelligence/v1p2beta1/doc/google/rpc/status.rb +0 -87
  46. data/lib/google/cloud/video_intelligence/v1p2beta1/video_intelligence_service_client.rb +0 -306
  47. data/lib/google/cloud/video_intelligence/v1p2beta1/video_intelligence_service_client_config.json +0 -31
  48. data/lib/google/cloud/videointelligence/v1/video_intelligence_pb.rb +0 -270
  49. data/lib/google/cloud/videointelligence/v1/video_intelligence_services_pb.rb +0 -51
  50. data/lib/google/cloud/videointelligence/v1beta2/video_intelligence_pb.rb +0 -168
  51. data/lib/google/cloud/videointelligence/v1beta2/video_intelligence_services_pb.rb +0 -50
  52. data/lib/google/cloud/videointelligence/v1p1beta1/video_intelligence_pb.rb +0 -170
  53. data/lib/google/cloud/videointelligence/v1p1beta1/video_intelligence_services_pb.rb +0 -50
  54. data/lib/google/cloud/videointelligence/v1p2beta1/video_intelligence_pb.rb +0 -191
  55. data/lib/google/cloud/videointelligence/v1p2beta1/video_intelligence_services_pb.rb +0 -51
data/LICENSE DELETED
@@ -1,201 +0,0 @@
1
- Apache License
2
- Version 2.0, January 2004
3
- https://www.apache.org/licenses/
4
-
5
- TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
6
-
7
- 1. Definitions.
8
-
9
- "License" shall mean the terms and conditions for use, reproduction,
10
- and distribution as defined by Sections 1 through 9 of this document.
11
-
12
- "Licensor" shall mean the copyright owner or entity authorized by
13
- the copyright owner that is granting the License.
14
-
15
- "Legal Entity" shall mean the union of the acting entity and all
16
- other entities that control, are controlled by, or are under common
17
- control with that entity. For the purposes of this definition,
18
- "control" means (i) the power, direct or indirect, to cause the
19
- direction or management of such entity, whether by contract or
20
- otherwise, or (ii) ownership of fifty percent (50%) or more of the
21
- outstanding shares, or (iii) beneficial ownership of such entity.
22
-
23
- "You" (or "Your") shall mean an individual or Legal Entity
24
- exercising permissions granted by this License.
25
-
26
- "Source" form shall mean the preferred form for making modifications,
27
- including but not limited to software source code, documentation
28
- source, and configuration files.
29
-
30
- "Object" form shall mean any form resulting from mechanical
31
- transformation or translation of a Source form, including but
32
- not limited to compiled object code, generated documentation,
33
- and conversions to other media types.
34
-
35
- "Work" shall mean the work of authorship, whether in Source or
36
- Object form, made available under the License, as indicated by a
37
- copyright notice that is included in or attached to the work
38
- (an example is provided in the Appendix below).
39
-
40
- "Derivative Works" shall mean any work, whether in Source or Object
41
- form, that is based on (or derived from) the Work and for which the
42
- editorial revisions, annotations, elaborations, or other modifications
43
- represent, as a whole, an original work of authorship. For the purposes
44
- of this License, Derivative Works shall not include works that remain
45
- separable from, or merely link (or bind by name) to the interfaces of,
46
- the Work and Derivative Works thereof.
47
-
48
- "Contribution" shall mean any work of authorship, including
49
- the original version of the Work and any modifications or additions
50
- to that Work or Derivative Works thereof, that is intentionally
51
- submitted to Licensor for inclusion in the Work by the copyright owner
52
- or by an individual or Legal Entity authorized to submit on behalf of
53
- the copyright owner. For the purposes of this definition, "submitted"
54
- means any form of electronic, verbal, or written communication sent
55
- to the Licensor or its representatives, including but not limited to
56
- communication on electronic mailing lists, source code control systems,
57
- and issue tracking systems that are managed by, or on behalf of, the
58
- Licensor for the purpose of discussing and improving the Work, but
59
- excluding communication that is conspicuously marked or otherwise
60
- designated in writing by the copyright owner as "Not a Contribution."
61
-
62
- "Contributor" shall mean Licensor and any individual or Legal Entity
63
- on behalf of whom a Contribution has been received by Licensor and
64
- subsequently incorporated within the Work.
65
-
66
- 2. Grant of Copyright License. Subject to the terms and conditions of
67
- this License, each Contributor hereby grants to You a perpetual,
68
- worldwide, non-exclusive, no-charge, royalty-free, irrevocable
69
- copyright license to reproduce, prepare Derivative Works of,
70
- publicly display, publicly perform, sublicense, and distribute the
71
- Work and such Derivative Works in Source or Object form.
72
-
73
- 3. Grant of Patent License. Subject to the terms and conditions of
74
- this License, each Contributor hereby grants to You a perpetual,
75
- worldwide, non-exclusive, no-charge, royalty-free, irrevocable
76
- (except as stated in this section) patent license to make, have made,
77
- use, offer to sell, sell, import, and otherwise transfer the Work,
78
- where such license applies only to those patent claims licensable
79
- by such Contributor that are necessarily infringed by their
80
- Contribution(s) alone or by combination of their Contribution(s)
81
- with the Work to which such Contribution(s) was submitted. If You
82
- institute patent litigation against any entity (including a
83
- cross-claim or counterclaim in a lawsuit) alleging that the Work
84
- or a Contribution incorporated within the Work constitutes direct
85
- or contributory patent infringement, then any patent licenses
86
- granted to You under this License for that Work shall terminate
87
- as of the date such litigation is filed.
88
-
89
- 4. Redistribution. You may reproduce and distribute copies of the
90
- Work or Derivative Works thereof in any medium, with or without
91
- modifications, and in Source or Object form, provided that You
92
- meet the following conditions:
93
-
94
- (a) You must give any other recipients of the Work or
95
- Derivative Works a copy of this License; and
96
-
97
- (b) You must cause any modified files to carry prominent notices
98
- stating that You changed the files; and
99
-
100
- (c) You must retain, in the Source form of any Derivative Works
101
- that You distribute, all copyright, patent, trademark, and
102
- attribution notices from the Source form of the Work,
103
- excluding those notices that do not pertain to any part of
104
- the Derivative Works; and
105
-
106
- (d) If the Work includes a "NOTICE" text file as part of its
107
- distribution, then any Derivative Works that You distribute must
108
- include a readable copy of the attribution notices contained
109
- within such NOTICE file, excluding those notices that do not
110
- pertain to any part of the Derivative Works, in at least one
111
- of the following places: within a NOTICE text file distributed
112
- as part of the Derivative Works; within the Source form or
113
- documentation, if provided along with the Derivative Works; or,
114
- within a display generated by the Derivative Works, if and
115
- wherever such third-party notices normally appear. The contents
116
- of the NOTICE file are for informational purposes only and
117
- do not modify the License. You may add Your own attribution
118
- notices within Derivative Works that You distribute, alongside
119
- or as an addendum to the NOTICE text from the Work, provided
120
- that such additional attribution notices cannot be construed
121
- as modifying the License.
122
-
123
- You may add Your own copyright statement to Your modifications and
124
- may provide additional or different license terms and conditions
125
- for use, reproduction, or distribution of Your modifications, or
126
- for any such Derivative Works as a whole, provided Your use,
127
- reproduction, and distribution of the Work otherwise complies with
128
- the conditions stated in this License.
129
-
130
- 5. Submission of Contributions. Unless You explicitly state otherwise,
131
- any Contribution intentionally submitted for inclusion in the Work
132
- by You to the Licensor shall be under the terms and conditions of
133
- this License, without any additional terms or conditions.
134
- Notwithstanding the above, nothing herein shall supersede or modify
135
- the terms of any separate license agreement you may have executed
136
- with Licensor regarding such Contributions.
137
-
138
- 6. Trademarks. This License does not grant permission to use the trade
139
- names, trademarks, service marks, or product names of the Licensor,
140
- except as required for reasonable and customary use in describing the
141
- origin of the Work and reproducing the content of the NOTICE file.
142
-
143
- 7. Disclaimer of Warranty. Unless required by applicable law or
144
- agreed to in writing, Licensor provides the Work (and each
145
- Contributor provides its Contributions) on an "AS IS" BASIS,
146
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
147
- implied, including, without limitation, any warranties or conditions
148
- of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
149
- PARTICULAR PURPOSE. You are solely responsible for determining the
150
- appropriateness of using or redistributing the Work and assume any
151
- risks associated with Your exercise of permissions under this License.
152
-
153
- 8. Limitation of Liability. In no event and under no legal theory,
154
- whether in tort (including negligence), contract, or otherwise,
155
- unless required by applicable law (such as deliberate and grossly
156
- negligent acts) or agreed to in writing, shall any Contributor be
157
- liable to You for damages, including any direct, indirect, special,
158
- incidental, or consequential damages of any character arising as a
159
- result of this License or out of the use or inability to use the
160
- Work (including but not limited to damages for loss of goodwill,
161
- work stoppage, computer failure or malfunction, or any and all
162
- other commercial damages or losses), even if such Contributor
163
- has been advised of the possibility of such damages.
164
-
165
- 9. Accepting Warranty or Additional Liability. While redistributing
166
- the Work or Derivative Works thereof, You may choose to offer,
167
- and charge a fee for, acceptance of support, warranty, indemnity,
168
- or other liability obligations and/or rights consistent with this
169
- License. However, in accepting such obligations, You may act only
170
- on Your own behalf and on Your sole responsibility, not on behalf
171
- of any other Contributor, and only if You agree to indemnify,
172
- defend, and hold each Contributor harmless for any liability
173
- incurred by, or claims asserted against, such Contributor by reason
174
- of your accepting any such warranty or additional liability.
175
-
176
- END OF TERMS AND CONDITIONS
177
-
178
- APPENDIX: How to apply the Apache License to your work.
179
-
180
- To apply the Apache License to your work, attach the following
181
- boilerplate notice, with the fields enclosed by brackets "[]"
182
- replaced with your own identifying information. (Don't include
183
- the brackets!) The text should be enclosed in the appropriate
184
- comment syntax for the file format. We also recommend that a
185
- file or class name and description of purpose be included on the
186
- same "printed page" as the copyright notice for easier
187
- identification within third-party archives.
188
-
189
- Copyright [yyyy] [name of copyright owner]
190
-
191
- Licensed under the Apache License, Version 2.0 (the "License");
192
- you may not use this file except in compliance with the License.
193
- You may obtain a copy of the License at
194
-
195
- https://www.apache.org/licenses/LICENSE-2.0
196
-
197
- Unless required by applicable law or agreed to in writing, software
198
- distributed under the License is distributed on an "AS IS" BASIS,
199
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
200
- See the License for the specific language governing permissions and
201
- limitations under the License.
@@ -1,186 +0,0 @@
1
- # Copyright 2019 Google LLC
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # https://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
-
15
-
16
- require "google/cloud/video_intelligence/v1/video_intelligence_service_client"
17
- require "google/cloud/videointelligence/v1/video_intelligence_pb"
18
-
19
- module Google
20
- module Cloud
21
- module VideoIntelligence
22
- # rubocop:disable LineLength
23
-
24
- ##
25
- # # Ruby Client for Cloud Video Intelligence API ([GA](https://github.com/googleapis/google-cloud-ruby#versioning))
26
- #
27
- # [Cloud Video Intelligence API][Product Documentation]:
28
- # Detects objects, explicit content, and scene changes in videos. It also
29
- # specifies the region for annotation and transcribes speech to text.
30
- # Supports both asynchronous API and streaming API.
31
- # - [Product Documentation][]
32
- #
33
- # ## Quick Start
34
- # In order to use this library, you first need to go through the following
35
- # steps:
36
- #
37
- # 1. [Select or create a Cloud Platform project.](https://console.cloud.google.com/project)
38
- # 2. [Enable billing for your project.](https://cloud.google.com/billing/docs/how-to/modify-project#enable_billing_for_a_project)
39
- # 3. [Enable the Cloud Video Intelligence API.](https://console.cloud.google.com/apis/library/videointelligence.googleapis.com)
40
- # 4. [Setup Authentication.](https://googleapis.dev/ruby/google-cloud-video_intelligence/latest/file.AUTHENTICATION.html)
41
- #
42
- # ### Installation
43
- # ```
44
- # $ gem install google-cloud-video_intelligence
45
- # ```
46
- #
47
- # ### Preview
48
- # #### VideoIntelligenceServiceClient
49
- # ```rb
50
- # require "google/cloud/video_intelligence"
51
- #
52
- # video_intelligence_client = Google::Cloud::VideoIntelligence.new(version: :v1)
53
- # input_uri = "gs://cloud-samples-data/video/cat.mp4"
54
- # features_element = :LABEL_DETECTION
55
- # features = [features_element]
56
- #
57
- # # Register a callback during the method call.
58
- # operation = video_intelligence_client.annotate_video(features, input_uri: input_uri) do |op|
59
- # raise op.results.message if op.error?
60
- # op_results = op.results
61
- # # Process the results.
62
- #
63
- # metadata = op.metadata
64
- # # Process the metadata.
65
- # end
66
- #
67
- # # Or use the return value to register a callback.
68
- # operation.on_done do |op|
69
- # raise op.results.message if op.error?
70
- # op_results = op.results
71
- # # Process the results.
72
- #
73
- # metadata = op.metadata
74
- # # Process the metadata.
75
- # end
76
- #
77
- # # Manually reload the operation.
78
- # operation.reload!
79
- #
80
- # # Or block until the operation completes, triggering callbacks on
81
- # # completion.
82
- # operation.wait_until_done!
83
- # ```
84
- #
85
- # ### Next Steps
86
- # - Read the [Cloud Video Intelligence API Product documentation][Product Documentation]
87
- # to learn more about the product and see How-to Guides.
88
- # - View this [repository's main README](https://github.com/googleapis/google-cloud-ruby/blob/master/README.md)
89
- # to see the full list of Cloud APIs that we cover.
90
- #
91
- # [Product Documentation]: https://cloud.google.com/video-intelligence
92
- #
93
- # ## Enabling Logging
94
- #
95
- # To enable logging for this library, set the logger for the underlying [gRPC](https://github.com/grpc/grpc/tree/master/src/ruby) library.
96
- # The logger that you set may be a Ruby stdlib [`Logger`](https://ruby-doc.org/stdlib-2.5.0/libdoc/logger/rdoc/Logger.html) as shown below,
97
- # or a [`Google::Cloud::Logging::Logger`](https://googleapis.dev/ruby/google-cloud-logging/latest)
98
- # that will write logs to [Stackdriver Logging](https://cloud.google.com/logging/). See [grpc/logconfig.rb](https://github.com/grpc/grpc/blob/master/src/ruby/lib/grpc/logconfig.rb)
99
- # and the gRPC [spec_helper.rb](https://github.com/grpc/grpc/blob/master/src/ruby/spec/spec_helper.rb) for additional information.
100
- #
101
- # Configuring a Ruby stdlib logger:
102
- #
103
- # ```ruby
104
- # require "logger"
105
- #
106
- # module MyLogger
107
- # LOGGER = Logger.new $stderr, level: Logger::WARN
108
- # def logger
109
- # LOGGER
110
- # end
111
- # end
112
- #
113
- # # Define a gRPC module-level logger method before grpc/logconfig.rb loads.
114
- # module GRPC
115
- # extend MyLogger
116
- # end
117
- # ```
118
- #
119
- module V1
120
- # rubocop:enable LineLength
121
-
122
- ##
123
- # Service that implements Google Cloud Video Intelligence API.
124
- #
125
- # @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
126
- # Provides the means for authenticating requests made by the client. This parameter can
127
- # be many types.
128
- # A `Google::Auth::Credentials` uses a the properties of its represented keyfile for
129
- # authenticating requests made by this client.
130
- # A `String` will be treated as the path to the keyfile to be used for the construction of
131
- # credentials for this client.
132
- # A `Hash` will be treated as the contents of a keyfile to be used for the construction of
133
- # credentials for this client.
134
- # A `GRPC::Core::Channel` will be used to make calls through.
135
- # A `GRPC::Core::ChannelCredentials` for the setting up the RPC client. The channel credentials
136
- # should already be composed with a `GRPC::Core::CallCredentials` object.
137
- # A `Proc` will be used as an updater_proc for the Grpc channel. The proc transforms the
138
- # metadata for requests, generally, to give OAuth credentials.
139
- # @param scopes [Array<String>]
140
- # The OAuth scopes for this service. This parameter is ignored if
141
- # an updater_proc is supplied.
142
- # @param client_config [Hash]
143
- # A Hash for call options for each method. See
144
- # Google::Gax#construct_settings for the structure of
145
- # this data. Falls back to the default config if not specified
146
- # or the specified config is missing data points.
147
- # @param timeout [Numeric]
148
- # The default timeout, in seconds, for calls made through this client.
149
- # @param metadata [Hash]
150
- # Default metadata to be sent with each request. This can be overridden on a per call basis.
151
- # @param service_address [String]
152
- # Override for the service hostname, or `nil` to leave as the default.
153
- # @param service_port [Integer]
154
- # Override for the service port, or `nil` to leave as the default.
155
- # @param exception_transformer [Proc]
156
- # An optional proc that intercepts any exceptions raised during an API call to inject
157
- # custom error handling.
158
- def self.new \
159
- credentials: nil,
160
- scopes: nil,
161
- client_config: nil,
162
- timeout: nil,
163
- metadata: nil,
164
- service_address: nil,
165
- service_port: nil,
166
- exception_transformer: nil,
167
- lib_name: nil,
168
- lib_version: nil
169
- kwargs = {
170
- credentials: credentials,
171
- scopes: scopes,
172
- client_config: client_config,
173
- timeout: timeout,
174
- metadata: metadata,
175
- exception_transformer: exception_transformer,
176
- lib_name: lib_name,
177
- service_address: service_address,
178
- service_port: service_port,
179
- lib_version: lib_version
180
- }.select { |_, v| v != nil }
181
- Google::Cloud::VideoIntelligence::V1::VideoIntelligenceServiceClient.new(**kwargs)
182
- end
183
- end
184
- end
185
- end
186
- end
@@ -1,41 +0,0 @@
1
- # Copyright 2019 Google LLC
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # https://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
-
15
-
16
- require "googleauth"
17
-
18
- module Google
19
- module Cloud
20
- module VideoIntelligence
21
- module V1
22
- class Credentials < Google::Auth::Credentials
23
- SCOPE = [
24
- "https://www.googleapis.com/auth/cloud-platform"
25
- ].freeze
26
- PATH_ENV_VARS = %w(VIDEO_INTELLIGENCE_CREDENTIALS
27
- VIDEO_INTELLIGENCE_KEYFILE
28
- GOOGLE_CLOUD_CREDENTIALS
29
- GOOGLE_CLOUD_KEYFILE
30
- GCLOUD_KEYFILE)
31
- JSON_ENV_VARS = %w(VIDEO_INTELLIGENCE_CREDENTIALS_JSON
32
- VIDEO_INTELLIGENCE_KEYFILE_JSON
33
- GOOGLE_CLOUD_CREDENTIALS_JSON
34
- GOOGLE_CLOUD_KEYFILE_JSON
35
- GCLOUD_KEYFILE_JSON)
36
- DEFAULT_PATHS = ["~/.config/gcloud/application_default_credentials.json"]
37
- end
38
- end
39
- end
40
- end
41
- end
@@ -1,707 +0,0 @@
1
- # Copyright 2019 Google LLC
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # https://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
-
15
-
16
- module Google
17
- module Cloud
18
- module VideoIntelligence
19
- module V1
20
- # Video annotation request.
21
- # @!attribute [rw] input_uri
22
- # @return [String]
23
- # Input video location. Currently, only
24
- # [Google Cloud Storage](https://cloud.google.com/storage/) URIs are
25
- # supported, which must be specified in the following format:
26
- # `gs://bucket-id/object-id` (other URI formats return
27
- # {Google::Rpc::Code::INVALID_ARGUMENT}). For more information, see
28
- # [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
29
- # A video URI may include wildcards in `object-id`, and thus identify
30
- # multiple videos. Supported wildcards: '*' to match 0 or more characters;
31
- # '?' to match 1 character. If unset, the input video should be embedded
32
- # in the request as `input_content`. If set, `input_content` should be unset.
33
- # @!attribute [rw] input_content
34
- # @return [String]
35
- # The video data bytes.
36
- # If unset, the input video(s) should be specified via `input_uri`.
37
- # If set, `input_uri` should be unset.
38
- # @!attribute [rw] features
39
- # @return [Array<Google::Cloud::VideoIntelligence::V1::Feature>]
40
- # Required. Requested video annotation features.
41
- # @!attribute [rw] video_context
42
- # @return [Google::Cloud::VideoIntelligence::V1::VideoContext]
43
- # Additional video context and/or feature-specific parameters.
44
- # @!attribute [rw] output_uri
45
- # @return [String]
46
- # Optional. Location where the output (in JSON format) should be stored.
47
- # Currently, only [Google Cloud Storage](https://cloud.google.com/storage/)
48
- # URIs are supported, which must be specified in the following format:
49
- # `gs://bucket-id/object-id` (other URI formats return
50
- # {Google::Rpc::Code::INVALID_ARGUMENT}). For more information, see
51
- # [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
52
- # @!attribute [rw] location_id
53
- # @return [String]
54
- # Optional. Cloud region where annotation should take place. Supported cloud
55
- # regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region
56
- # is specified, a region will be determined based on video file location.
57
- class AnnotateVideoRequest; end
58
-
59
- # Video context and/or feature-specific parameters.
60
- # @!attribute [rw] segments
61
- # @return [Array<Google::Cloud::VideoIntelligence::V1::VideoSegment>]
62
- # Video segments to annotate. The segments may overlap and are not required
63
- # to be contiguous or span the whole video. If unspecified, each video is
64
- # treated as a single segment.
65
- # @!attribute [rw] label_detection_config
66
- # @return [Google::Cloud::VideoIntelligence::V1::LabelDetectionConfig]
67
- # Config for LABEL_DETECTION.
68
- # @!attribute [rw] shot_change_detection_config
69
- # @return [Google::Cloud::VideoIntelligence::V1::ShotChangeDetectionConfig]
70
- # Config for SHOT_CHANGE_DETECTION.
71
- # @!attribute [rw] explicit_content_detection_config
72
- # @return [Google::Cloud::VideoIntelligence::V1::ExplicitContentDetectionConfig]
73
- # Config for EXPLICIT_CONTENT_DETECTION.
74
- # @!attribute [rw] face_detection_config
75
- # @return [Google::Cloud::VideoIntelligence::V1::FaceDetectionConfig]
76
- # Config for FACE_DETECTION.
77
- # @!attribute [rw] speech_transcription_config
78
- # @return [Google::Cloud::VideoIntelligence::V1::SpeechTranscriptionConfig]
79
- # Config for SPEECH_TRANSCRIPTION.
80
- # @!attribute [rw] text_detection_config
81
- # @return [Google::Cloud::VideoIntelligence::V1::TextDetectionConfig]
82
- # Config for TEXT_DETECTION.
83
- # @!attribute [rw] object_tracking_config
84
- # @return [Google::Cloud::VideoIntelligence::V1::ObjectTrackingConfig]
85
- # Config for OBJECT_TRACKING.
86
- class VideoContext; end
87
-
88
- # Config for LABEL_DETECTION.
89
- # @!attribute [rw] label_detection_mode
90
- # @return [Google::Cloud::VideoIntelligence::V1::LabelDetectionMode]
91
- # What labels should be detected with LABEL_DETECTION, in addition to
92
- # video-level labels or segment-level labels.
93
- # If unspecified, defaults to `SHOT_MODE`.
94
- # @!attribute [rw] stationary_camera
95
- # @return [true, false]
96
- # Whether the video has been shot from a stationary (i.e. non-moving) camera.
97
- # When set to true, might improve detection accuracy for moving objects.
98
- # Should be used with `SHOT_AND_FRAME_MODE` enabled.
99
- # @!attribute [rw] model
100
- # @return [String]
101
- # Model to use for label detection.
102
- # Supported values: "builtin/stable" (the default if unset) and
103
- # "builtin/latest".
104
- # @!attribute [rw] frame_confidence_threshold
105
- # @return [Float]
106
- # The confidence threshold we perform filtering on the labels from
107
- # frame-level detection. If not set, it is set to 0.4 by default. The valid
108
- # range for this threshold is [0.1, 0.9]. Any value set outside of this
109
- # range will be clipped.
110
- # Note: for best results please follow the default threshold. We will update
111
- # the default threshold everytime when we release a new model.
112
- # @!attribute [rw] video_confidence_threshold
113
- # @return [Float]
114
- # The confidence threshold we perform filtering on the labels from
115
- # video-level and shot-level detections. If not set, it is set to 0.3 by
116
- # default. The valid range for this threshold is [0.1, 0.9]. Any value set
117
- # outside of this range will be clipped.
118
- # Note: for best results please follow the default threshold. We will update
119
- # the default threshold everytime when we release a new model.
120
- class LabelDetectionConfig; end
121
-
122
- # Config for SHOT_CHANGE_DETECTION.
123
- # @!attribute [rw] model
124
- # @return [String]
125
- # Model to use for shot change detection.
126
- # Supported values: "builtin/stable" (the default if unset) and
127
- # "builtin/latest".
128
- class ShotChangeDetectionConfig; end
129
-
130
- # Config for OBJECT_TRACKING.
131
- # @!attribute [rw] model
132
- # @return [String]
133
- # Model to use for object tracking.
134
- # Supported values: "builtin/stable" (the default if unset) and
135
- # "builtin/latest".
136
- class ObjectTrackingConfig; end
137
-
138
- # Config for FACE_DETECTION.
139
- # @!attribute [rw] model
140
- # @return [String]
141
- # Model to use for face detection.
142
- # Supported values: "builtin/stable" (the default if unset) and
143
- # "builtin/latest".
144
- # @!attribute [rw] include_bounding_boxes
145
- # @return [true, false]
146
- # Whether bounding boxes be included in the face annotation output.
147
- class FaceDetectionConfig; end
148
-
149
- # Config for EXPLICIT_CONTENT_DETECTION.
150
- # @!attribute [rw] model
151
- # @return [String]
152
- # Model to use for explicit content detection.
153
- # Supported values: "builtin/stable" (the default if unset) and
154
- # "builtin/latest".
155
- class ExplicitContentDetectionConfig; end
156
-
157
- # Config for TEXT_DETECTION.
158
- # @!attribute [rw] language_hints
159
- # @return [Array<String>]
160
- # Language hint can be specified if the language to be detected is known a
161
- # priori. It can increase the accuracy of the detection. Language hint must
162
- # be language code in BCP-47 format.
163
- #
164
- # Automatic language detection is performed if no hint is provided.
165
- # @!attribute [rw] model
166
- # @return [String]
167
- # Model to use for text detection.
168
- # Supported values: "builtin/stable" (the default if unset) and
169
- # "builtin/latest".
170
- class TextDetectionConfig; end
171
-
172
- # Video segment.
173
- # @!attribute [rw] start_time_offset
174
- # @return [Google::Protobuf::Duration]
175
- # Time-offset, relative to the beginning of the video,
176
- # corresponding to the start of the segment (inclusive).
177
- # @!attribute [rw] end_time_offset
178
- # @return [Google::Protobuf::Duration]
179
- # Time-offset, relative to the beginning of the video,
180
- # corresponding to the end of the segment (inclusive).
181
- class VideoSegment; end
182
-
183
- # Video segment level annotation results for label detection.
184
- # @!attribute [rw] segment
185
- # @return [Google::Cloud::VideoIntelligence::V1::VideoSegment]
186
- # Video segment where a label was detected.
187
- # @!attribute [rw] confidence
188
- # @return [Float]
189
- # Confidence that the label is accurate. Range: [0, 1].
190
- class LabelSegment; end
191
-
192
- # Video frame level annotation results for label detection.
193
- # @!attribute [rw] time_offset
194
- # @return [Google::Protobuf::Duration]
195
- # Time-offset, relative to the beginning of the video, corresponding to the
196
- # video frame for this location.
197
- # @!attribute [rw] confidence
198
- # @return [Float]
199
- # Confidence that the label is accurate. Range: [0, 1].
200
- class LabelFrame; end
201
-
202
- # Detected entity from video analysis.
203
- # @!attribute [rw] entity_id
204
- # @return [String]
205
- # Opaque entity ID. Some IDs may be available in
206
- # [Google Knowledge Graph Search
207
- # API](https://developers.google.com/knowledge-graph/).
208
- # @!attribute [rw] description
209
- # @return [String]
210
- # Textual description, e.g. `Fixed-gear bicycle`.
211
- # @!attribute [rw] language_code
212
- # @return [String]
213
- # Language code for `description` in BCP-47 format.
214
- class Entity; end
215
-
216
- # Label annotation.
217
- # @!attribute [rw] entity
218
- # @return [Google::Cloud::VideoIntelligence::V1::Entity]
219
- # Detected entity.
220
- # @!attribute [rw] category_entities
221
- # @return [Array<Google::Cloud::VideoIntelligence::V1::Entity>]
222
- # Common categories for the detected entity.
223
- # E.g. when the label is `Terrier` the category is likely `dog`. And in some
224
- # cases there might be more than one categories e.g. `Terrier` could also be
225
- # a `pet`.
226
- # @!attribute [rw] segments
227
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelSegment>]
228
- # All video segments where a label was detected.
229
- # @!attribute [rw] frames
230
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelFrame>]
231
- # All video frames where a label was detected.
232
- class LabelAnnotation; end
233
-
234
- # Video frame level annotation results for explicit content.
235
- # @!attribute [rw] time_offset
236
- # @return [Google::Protobuf::Duration]
237
- # Time-offset, relative to the beginning of the video, corresponding to the
238
- # video frame for this location.
239
- # @!attribute [rw] pornography_likelihood
240
- # @return [Google::Cloud::VideoIntelligence::V1::Likelihood]
241
- # Likelihood of the pornography content..
242
- class ExplicitContentFrame; end
243
-
244
- # Explicit content annotation (based on per-frame visual signals only).
245
- # If no explicit content has been detected in a frame, no annotations are
246
- # present for that frame.
247
- # @!attribute [rw] frames
248
- # @return [Array<Google::Cloud::VideoIntelligence::V1::ExplicitContentFrame>]
249
- # All video frames where explicit content was detected.
250
- class ExplicitContentAnnotation; end
251
-
252
- # Normalized bounding box.
253
- # The normalized vertex coordinates are relative to the original image.
254
- # Range: [0, 1].
255
- # @!attribute [rw] left
256
- # @return [Float]
257
- # Left X coordinate.
258
- # @!attribute [rw] top
259
- # @return [Float]
260
- # Top Y coordinate.
261
- # @!attribute [rw] right
262
- # @return [Float]
263
- # Right X coordinate.
264
- # @!attribute [rw] bottom
265
- # @return [Float]
266
- # Bottom Y coordinate.
267
- class NormalizedBoundingBox; end
268
-
269
- # Video segment level annotation results for face detection.
270
- # @!attribute [rw] segment
271
- # @return [Google::Cloud::VideoIntelligence::V1::VideoSegment]
272
- # Video segment where a face was detected.
273
- class FaceSegment; end
274
-
275
- # Video frame level annotation results for face detection.
276
- # @!attribute [rw] normalized_bounding_boxes
277
- # @return [Array<Google::Cloud::VideoIntelligence::V1::NormalizedBoundingBox>]
278
- # Normalized Bounding boxes in a frame.
279
- # There can be more than one boxes if the same face is detected in multiple
280
- # locations within the current frame.
281
- # @!attribute [rw] time_offset
282
- # @return [Google::Protobuf::Duration]
283
- # Time-offset, relative to the beginning of the video,
284
- # corresponding to the video frame for this location.
285
- class FaceFrame; end
286
-
287
- # Face annotation.
288
- # @!attribute [rw] thumbnail
289
- # @return [String]
290
- # Thumbnail of a representative face view (in JPEG format).
291
- # @!attribute [rw] segments
292
- # @return [Array<Google::Cloud::VideoIntelligence::V1::FaceSegment>]
293
- # All video segments where a face was detected.
294
- # @!attribute [rw] frames
295
- # @return [Array<Google::Cloud::VideoIntelligence::V1::FaceFrame>]
296
- # All video frames where a face was detected.
297
- class FaceAnnotation; end
298
-
299
- # Annotation results for a single video.
300
- # @!attribute [rw] input_uri
301
- # @return [String]
302
- # Video file location in
303
- # [Google Cloud Storage](https://cloud.google.com/storage/).
304
- # @!attribute [rw] segment
305
- # @return [Google::Cloud::VideoIntelligence::V1::VideoSegment]
306
- # Video segment on which the annotation is run.
307
- # @!attribute [rw] segment_label_annotations
308
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelAnnotation>]
309
- # Topical label annotations on video level or user specified segment level.
310
- # There is exactly one element for each unique label.
311
- # @!attribute [rw] segment_presence_label_annotations
312
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelAnnotation>]
313
- # Presence label annotations on video level or user specified segment level.
314
- # There is exactly one element for each unique label. Compared to the
315
- # existing topical `segment_label_annotations`, this field presents more
316
- # fine-grained, segment-level labels detected in video content and is made
317
- # available only when the client sets `LabelDetectionConfig.model` to
318
- # "builtin/latest" in the request.
319
- # @!attribute [rw] shot_label_annotations
320
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelAnnotation>]
321
- # Topical label annotations on shot level.
322
- # There is exactly one element for each unique label.
323
- # @!attribute [rw] shot_presence_label_annotations
324
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelAnnotation>]
325
- # Presence label annotations on shot level. There is exactly one element for
326
- # each unique label. Compared to the existing topical
327
- # `shot_label_annotations`, this field presents more fine-grained, shot-level
328
- # labels detected in video content and is made available only when the client
329
- # sets `LabelDetectionConfig.model` to "builtin/latest" in the request.
330
- # @!attribute [rw] frame_label_annotations
331
- # @return [Array<Google::Cloud::VideoIntelligence::V1::LabelAnnotation>]
332
- # Label annotations on frame level.
333
- # There is exactly one element for each unique label.
334
- # @!attribute [rw] face_annotations
335
- # @return [Array<Google::Cloud::VideoIntelligence::V1::FaceAnnotation>]
336
- # Face annotations. There is exactly one element for each unique face.
337
- # @!attribute [rw] shot_annotations
338
- # @return [Array<Google::Cloud::VideoIntelligence::V1::VideoSegment>]
339
- # Shot annotations. Each shot is represented as a video segment.
340
- # @!attribute [rw] explicit_annotation
341
- # @return [Google::Cloud::VideoIntelligence::V1::ExplicitContentAnnotation]
342
- # Explicit content annotation.
343
- # @!attribute [rw] speech_transcriptions
344
- # @return [Array<Google::Cloud::VideoIntelligence::V1::SpeechTranscription>]
345
- # Speech transcription.
346
- # @!attribute [rw] text_annotations
347
- # @return [Array<Google::Cloud::VideoIntelligence::V1::TextAnnotation>]
348
- # OCR text detection and tracking.
349
- # Annotations for list of detected text snippets. Each will have list of
350
- # frame information associated with it.
351
- # @!attribute [rw] object_annotations
352
- # @return [Array<Google::Cloud::VideoIntelligence::V1::ObjectTrackingAnnotation>]
353
- # Annotations for list of objects detected and tracked in video.
354
- # @!attribute [rw] error
355
- # @return [Google::Rpc::Status]
356
- # If set, indicates an error. Note that for a single `AnnotateVideoRequest`
357
- # some videos may succeed and some may fail.
358
- class VideoAnnotationResults; end
359
-
360
- # Video annotation response. Included in the `response`
361
- # field of the `Operation` returned by the `GetOperation`
362
- # call of the `google::longrunning::Operations` service.
363
- # @!attribute [rw] annotation_results
364
- # @return [Array<Google::Cloud::VideoIntelligence::V1::VideoAnnotationResults>]
365
- # Annotation results for all videos specified in `AnnotateVideoRequest`.
366
- class AnnotateVideoResponse; end
367
-
368
- # Annotation progress for a single video.
369
- # @!attribute [rw] input_uri
370
- # @return [String]
371
- # Video file location in
372
- # [Google Cloud Storage](https://cloud.google.com/storage/).
373
- # @!attribute [rw] progress_percent
374
- # @return [Integer]
375
- # Approximate percentage processed thus far. Guaranteed to be
376
- # 100 when fully processed.
377
- # @!attribute [rw] start_time
378
- # @return [Google::Protobuf::Timestamp]
379
- # Time when the request was received.
380
- # @!attribute [rw] update_time
381
- # @return [Google::Protobuf::Timestamp]
382
- # Time of the most recent update.
383
- # @!attribute [rw] feature
384
- # @return [Google::Cloud::VideoIntelligence::V1::Feature]
385
- # Specifies which feature is being tracked if the request contains more than
386
- # one features.
387
- # @!attribute [rw] segment
388
- # @return [Google::Cloud::VideoIntelligence::V1::VideoSegment]
389
- # Specifies which segment is being tracked if the request contains more than
390
- # one segments.
391
- class VideoAnnotationProgress; end
392
-
393
- # Video annotation progress. Included in the `metadata`
394
- # field of the `Operation` returned by the `GetOperation`
395
- # call of the `google::longrunning::Operations` service.
396
- # @!attribute [rw] annotation_progress
397
- # @return [Array<Google::Cloud::VideoIntelligence::V1::VideoAnnotationProgress>]
398
- # Progress metadata for all videos specified in `AnnotateVideoRequest`.
399
- class AnnotateVideoProgress; end
400
-
401
- # Config for SPEECH_TRANSCRIPTION.
402
- # @!attribute [rw] language_code
403
- # @return [String]
404
- # Required. *Required* The language of the supplied audio as a
405
- # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
406
- # Example: "en-US".
407
- # See [Language Support](https://cloud.google.com/speech/docs/languages)
408
- # for a list of the currently supported language codes.
409
- # @!attribute [rw] max_alternatives
410
- # @return [Integer]
411
- # Optional. Maximum number of recognition hypotheses to be returned.
412
- # Specifically, the maximum number of `SpeechRecognitionAlternative` messages
413
- # within each `SpeechTranscription`. The server may return fewer than
414
- # `max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will
415
- # return a maximum of one. If omitted, will return a maximum of one.
416
- # @!attribute [rw] filter_profanity
417
- # @return [true, false]
418
- # Optional. If set to `true`, the server will attempt to filter out
419
- # profanities, replacing all but the initial character in each filtered word
420
- # with asterisks, e.g. "f***". If set to `false` or omitted, profanities
421
- # won't be filtered out.
422
- # @!attribute [rw] speech_contexts
423
- # @return [Array<Google::Cloud::VideoIntelligence::V1::SpeechContext>]
424
- # Optional. A means to provide context to assist the speech recognition.
425
- # @!attribute [rw] enable_automatic_punctuation
426
- # @return [true, false]
427
- # Optional. If 'true', adds punctuation to recognition result hypotheses.
428
- # This feature is only available in select languages. Setting this for
429
- # requests in other languages has no effect at all. The default 'false' value
430
- # does not add punctuation to result hypotheses. NOTE: "This is currently
431
- # offered as an experimental service, complimentary to all users. In the
432
- # future this may be exclusively available as a premium feature."
433
- # @!attribute [rw] audio_tracks
434
- # @return [Array<Integer>]
435
- # Optional. For file formats, such as MXF or MKV, supporting multiple audio
436
- # tracks, specify up to two tracks. Default: track 0.
437
- # @!attribute [rw] enable_speaker_diarization
438
- # @return [true, false]
439
- # Optional. If 'true', enables speaker detection for each recognized word in
440
- # the top alternative of the recognition result using a speaker_tag provided
441
- # in the WordInfo.
442
- # Note: When this is true, we send all the words from the beginning of the
443
- # audio for the top alternative in every consecutive responses.
444
- # This is done in order to improve our speaker tags as our models learn to
445
- # identify the speakers in the conversation over time.
446
- # @!attribute [rw] diarization_speaker_count
447
- # @return [Integer]
448
- # Optional. If set, specifies the estimated number of speakers in the conversation.
449
- # If not set, defaults to '2'.
450
- # Ignored unless enable_speaker_diarization is set to true.
451
- # @!attribute [rw] enable_word_confidence
452
- # @return [true, false]
453
- # Optional. If `true`, the top result includes a list of words and the
454
- # confidence for those words. If `false`, no word-level confidence
455
- # information is returned. The default is `false`.
456
- class SpeechTranscriptionConfig; end
457
-
458
- # Provides "hints" to the speech recognizer to favor specific words and phrases
459
- # in the results.
460
- # @!attribute [rw] phrases
461
- # @return [Array<String>]
462
- # Optional. A list of strings containing words and phrases "hints" so that
463
- # the speech recognition is more likely to recognize them. This can be used
464
- # to improve the accuracy for specific words and phrases, for example, if
465
- # specific commands are typically spoken by the user. This can also be used
466
- # to add additional words to the vocabulary of the recognizer. See
467
- # [usage limits](https://cloud.google.com/speech/limits#content).
468
- class SpeechContext; end
469
-
470
- # A speech recognition result corresponding to a portion of the audio.
471
- # @!attribute [rw] alternatives
472
- # @return [Array<Google::Cloud::VideoIntelligence::V1::SpeechRecognitionAlternative>]
473
- # May contain one or more recognition hypotheses (up to the maximum specified
474
- # in `max_alternatives`). These alternatives are ordered in terms of
475
- # accuracy, with the top (first) alternative being the most probable, as
476
- # ranked by the recognizer.
477
- # @!attribute [rw] language_code
478
- # @return [String]
479
- # Output only. The [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of
480
- # the language in this result. This language code was detected to have the
481
- # most likelihood of being spoken in the audio.
482
- class SpeechTranscription; end
483
-
484
- # Alternative hypotheses (a.k.a. n-best list).
485
- # @!attribute [rw] transcript
486
- # @return [String]
487
- # Transcript text representing the words that the user spoke.
488
- # @!attribute [rw] confidence
489
- # @return [Float]
490
- # Output only. The confidence estimate between 0.0 and 1.0. A higher number
491
- # indicates an estimated greater likelihood that the recognized words are
492
- # correct. This field is set only for the top alternative.
493
- # This field is not guaranteed to be accurate and users should not rely on it
494
- # to be always provided.
495
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
496
- # @!attribute [rw] words
497
- # @return [Array<Google::Cloud::VideoIntelligence::V1::WordInfo>]
498
- # Output only. A list of word-specific information for each recognized word.
499
- # Note: When `enable_speaker_diarization` is true, you will see all the words
500
- # from the beginning of the audio.
501
- class SpeechRecognitionAlternative; end
502
-
503
- # Word-specific information for recognized words. Word information is only
504
- # included in the response when certain request parameters are set, such
505
- # as `enable_word_time_offsets`.
506
- # @!attribute [rw] start_time
507
- # @return [Google::Protobuf::Duration]
508
- # Time offset relative to the beginning of the audio, and
509
- # corresponding to the start of the spoken word. This field is only set if
510
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
511
- # experimental feature and the accuracy of the time offset can vary.
512
- # @!attribute [rw] end_time
513
- # @return [Google::Protobuf::Duration]
514
- # Time offset relative to the beginning of the audio, and
515
- # corresponding to the end of the spoken word. This field is only set if
516
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
517
- # experimental feature and the accuracy of the time offset can vary.
518
- # @!attribute [rw] word
519
- # @return [String]
520
- # The word corresponding to this set of information.
521
- # @!attribute [rw] confidence
522
- # @return [Float]
523
- # Output only. The confidence estimate between 0.0 and 1.0. A higher number
524
- # indicates an estimated greater likelihood that the recognized words are
525
- # correct. This field is set only for the top alternative.
526
- # This field is not guaranteed to be accurate and users should not rely on it
527
- # to be always provided.
528
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
529
- # @!attribute [rw] speaker_tag
530
- # @return [Integer]
531
- # Output only. A distinct integer value is assigned for every speaker within
532
- # the audio. This field specifies which one of those speakers was detected to
533
- # have spoken this word. Value ranges from 1 up to diarization_speaker_count,
534
- # and is only set if speaker diarization is enabled.
535
- class WordInfo; end
536
-
537
- # A vertex represents a 2D point in the image.
538
- # NOTE: the normalized vertex coordinates are relative to the original image
539
- # and range from 0 to 1.
540
- # @!attribute [rw] x
541
- # @return [Float]
542
- # X coordinate.
543
- # @!attribute [rw] y
544
- # @return [Float]
545
- # Y coordinate.
546
- class NormalizedVertex; end
547
-
548
- # Normalized bounding polygon for text (that might not be aligned with axis).
549
- # Contains list of the corner points in clockwise order starting from
550
- # top-left corner. For example, for a rectangular bounding box:
551
- # When the text is horizontal it might look like:
552
- # 0----1
553
- # | |
554
- # 3----2
555
- #
556
- # When it's clockwise rotated 180 degrees around the top-left corner it
557
- # becomes:
558
- # 2----3
559
- # | |
560
- # 1----0
561
- #
562
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
563
- # than 0, or greater than 1 due to trignometric calculations for location of
564
- # the box.
565
- # @!attribute [rw] vertices
566
- # @return [Array<Google::Cloud::VideoIntelligence::V1::NormalizedVertex>]
567
- # Normalized vertices of the bounding polygon.
568
- class NormalizedBoundingPoly; end
569
-
570
- # Video segment level annotation results for text detection.
571
- # @!attribute [rw] segment
572
- # @return [Google::Cloud::VideoIntelligence::V1::VideoSegment]
573
- # Video segment where a text snippet was detected.
574
- # @!attribute [rw] confidence
575
- # @return [Float]
576
- # Confidence for the track of detected text. It is calculated as the highest
577
- # over all frames where OCR detected text appears.
578
- # @!attribute [rw] frames
579
- # @return [Array<Google::Cloud::VideoIntelligence::V1::TextFrame>]
580
- # Information related to the frames where OCR detected text appears.
581
- class TextSegment; end
582
-
583
- # Video frame level annotation results for text annotation (OCR).
584
- # Contains information regarding timestamp and bounding box locations for the
585
- # frames containing detected OCR text snippets.
586
- # @!attribute [rw] rotated_bounding_box
587
- # @return [Google::Cloud::VideoIntelligence::V1::NormalizedBoundingPoly]
588
- # Bounding polygon of the detected text for this frame.
589
- # @!attribute [rw] time_offset
590
- # @return [Google::Protobuf::Duration]
591
- # Timestamp of this frame.
592
- class TextFrame; end
593
-
594
- # Annotations related to one detected OCR text snippet. This will contain the
595
- # corresponding text, confidence value, and frame level information for each
596
- # detection.
597
- # @!attribute [rw] text
598
- # @return [String]
599
- # The detected text.
600
- # @!attribute [rw] segments
601
- # @return [Array<Google::Cloud::VideoIntelligence::V1::TextSegment>]
602
- # All video segments where OCR detected text appears.
603
- class TextAnnotation; end
604
-
605
- # Video frame level annotations for object detection and tracking. This field
606
- # stores per frame location, time offset, and confidence.
607
- # @!attribute [rw] normalized_bounding_box
608
- # @return [Google::Cloud::VideoIntelligence::V1::NormalizedBoundingBox]
609
- # The normalized bounding box location of this object track for the frame.
610
- # @!attribute [rw] time_offset
611
- # @return [Google::Protobuf::Duration]
612
- # The timestamp of the frame in microseconds.
613
- class ObjectTrackingFrame; end
614
-
615
- # Annotations corresponding to one tracked object.
616
- # @!attribute [rw] segment
617
- # @return [Google::Cloud::VideoIntelligence::V1::VideoSegment]
618
- # Non-streaming batch mode ONLY.
619
- # Each object track corresponds to one video segment where it appears.
620
- # @!attribute [rw] track_id
621
- # @return [Integer]
622
- # Streaming mode ONLY.
623
- # In streaming mode, we do not know the end time of a tracked object
624
- # before it is completed. Hence, there is no VideoSegment info returned.
625
- # Instead, we provide a unique identifiable integer track_id so that
626
- # the customers can correlate the results of the ongoing
627
- # ObjectTrackAnnotation of the same track_id over time.
628
- # @!attribute [rw] entity
629
- # @return [Google::Cloud::VideoIntelligence::V1::Entity]
630
- # Entity to specify the object category that this track is labeled as.
631
- # @!attribute [rw] confidence
632
- # @return [Float]
633
- # Object category's labeling confidence of this track.
634
- # @!attribute [rw] frames
635
- # @return [Array<Google::Cloud::VideoIntelligence::V1::ObjectTrackingFrame>]
636
- # Information corresponding to all frames where this object track appears.
637
- # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame
638
- # messages in frames.
639
- # Streaming mode: it can only be one ObjectTrackingFrame message in frames.
640
- class ObjectTrackingAnnotation; end
641
-
642
- # Video annotation feature.
643
- module Feature
644
- # Unspecified.
645
- FEATURE_UNSPECIFIED = 0
646
-
647
- # Label detection. Detect objects, such as dog or flower.
648
- LABEL_DETECTION = 1
649
-
650
- # Shot change detection.
651
- SHOT_CHANGE_DETECTION = 2
652
-
653
- # Explicit content detection.
654
- EXPLICIT_CONTENT_DETECTION = 3
655
-
656
- # Human face detection and tracking.
657
- FACE_DETECTION = 4
658
-
659
- # Speech transcription.
660
- SPEECH_TRANSCRIPTION = 6
661
-
662
- # OCR text detection and tracking.
663
- TEXT_DETECTION = 7
664
-
665
- # Object detection and tracking.
666
- OBJECT_TRACKING = 9
667
- end
668
-
669
- # Label detection mode.
670
- module LabelDetectionMode
671
- # Unspecified.
672
- LABEL_DETECTION_MODE_UNSPECIFIED = 0
673
-
674
- # Detect shot-level labels.
675
- SHOT_MODE = 1
676
-
677
- # Detect frame-level labels.
678
- FRAME_MODE = 2
679
-
680
- # Detect both shot-level and frame-level labels.
681
- SHOT_AND_FRAME_MODE = 3
682
- end
683
-
684
- # Bucketized representation of likelihood.
685
- module Likelihood
686
- # Unspecified likelihood.
687
- LIKELIHOOD_UNSPECIFIED = 0
688
-
689
- # Very unlikely.
690
- VERY_UNLIKELY = 1
691
-
692
- # Unlikely.
693
- UNLIKELY = 2
694
-
695
- # Possible.
696
- POSSIBLE = 3
697
-
698
- # Likely.
699
- LIKELY = 4
700
-
701
- # Very likely.
702
- VERY_LIKELY = 5
703
- end
704
- end
705
- end
706
- end
707
- end