google-cloud-bigquery-storage-v1 0.1.1 → 0.2.1

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 58ebbdaea74b95825c1a5ffa1f6e43f2163f6f025dcdd7bf8e5409d839e1a033
4
- data.tar.gz: cef8388d6488300663dee05b2a1812500a02187eda8d94f91f409330b1ea0f7f
3
+ metadata.gz: 8a2fd2dae50772f81c0d7ac7f9bb91d31a53ae5de85f1446eae9b6a422451ff5
4
+ data.tar.gz: 4dc33ffc503b2132fc39e8949e87cbdea6772d9338e67f3596c16846d1a25e7d
5
5
  SHA512:
6
- metadata.gz: d33ff824fdbab2efc29bc97827555ce7ec7c88690d42bd8fd70cfb1cb5f2c3b54b3a1b7ba6ddb1fad33ecaaa7077d04d1ccfc33587b87780a059a182873980db
7
- data.tar.gz: c28774c60882b4fa804914233cd9d34c3c3fd2b1049ac6a45ee718b83145c434a65c9d2937a8d0cb174e6ff43df8f281afc61c646d6b656d18b5411be8cb8178
6
+ metadata.gz: 7413aad3b26e33f589dfbffacf104af97cda6bdc0a7518d8b4ac33735df32053fcbf1a812b93cc0b8ea300ffff94d2197823db2942faee2a466098dc328bd498
7
+ data.tar.gz: b9a5ce9aebb17d83fdbdfd51f12952137c14b4cd256e354a2f8a1356fcb77ae9bde948aa1ba3fa588932de2a4a2588285f67a2ba7075b6d8b6b89d7e18f3dac3
@@ -27,7 +27,7 @@ export BIGQUERY_STORAGE_CREDENTIALS=path/to/keyfile.json
27
27
  ```ruby
28
28
  require "google/cloud/bigquery/storage/v1"
29
29
 
30
- client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
30
+ client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
31
31
  ```
32
32
 
33
33
  ## Credential Lookup
@@ -64,7 +64,7 @@ containers where writing files is difficult or not encouraged.
64
64
 
65
65
  The environment variables that google-cloud-bigquery-storage-v1
66
66
  checks for credentials are configured on the service Credentials class (such as
67
- {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Credentials}):
67
+ {::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Credentials}):
68
68
 
69
69
  1. `BIGQUERY_STORAGE_CREDENTIALS` - Path to JSON file, or JSON contents
70
70
  2. `BIGQUERY_STORAGE_KEYFILE` - Path to JSON file, or JSON contents
@@ -77,7 +77,7 @@ require "google/cloud/bigquery/storage/v1"
77
77
 
78
78
  ENV["BIGQUERY_STORAGE_CREDENTIALS"] = "path/to/keyfile.json"
79
79
 
80
- client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
80
+ client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
81
81
  ```
82
82
 
83
83
  ### Configuration
@@ -88,7 +88,7 @@ environment variables. Either on an individual client initialization:
88
88
  ```ruby
89
89
  require "google/cloud/bigquery/storage/v1"
90
90
 
91
- client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
91
+ client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
92
92
  config.credentials = "path/to/keyfile.json"
93
93
  end
94
94
  ```
@@ -98,11 +98,11 @@ Or configured globally for all clients:
98
98
  ```ruby
99
99
  require "google/cloud/bigquery/storage/v1"
100
100
 
101
- Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
101
+ ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
102
102
  config.credentials = "path/to/keyfile.json"
103
103
  end
104
104
 
105
- client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
105
+ client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
106
106
  ```
107
107
 
108
108
  ### Cloud SDK
data/README.md CHANGED
@@ -1,4 +1,4 @@
1
- # BigQuery Storage V1
1
+ # Ruby Client for the BigQuery Storage V1 API
2
2
 
3
3
  API Client library for the BigQuery Storage V1 API
4
4
 
@@ -12,6 +12,53 @@ https://github.com/googleapis/google-cloud-ruby
12
12
  $ gem install google-cloud-bigquery-storage-v1
13
13
  ```
14
14
 
15
+ ## Before You Begin
16
+
17
+ In order to use this library, you first need to go through the following steps:
18
+
19
+ 1. [Select or create a Cloud Platform project.](https://console.cloud.google.com/project)
20
+ 1. [Enable billing for your project.](https://cloud.google.com/billing/docs/how-to/modify-project#enable_billing_for_a_project)
21
+ 1. {file:AUTHENTICATION.md Set up authentication.}
22
+
23
+ ## Quick Start
24
+
25
+ ```ruby
26
+ require "google/cloud/bigquery/storage/v1"
27
+
28
+ client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
29
+ request = my_create_request
30
+ response = client.create_read_session request
31
+ ```
32
+
33
+ View the [Client Library Documentation](https://googleapis.dev/ruby/google-cloud-bigquery-storage-v1/latest)
34
+ for class and method documentation.
35
+
36
+ ## Enabling Logging
37
+
38
+ To enable logging for this library, set the logger for the underlying [gRPC](https://github.com/grpc/grpc/tree/master/src/ruby) library.
39
+ The logger that you set may be a Ruby stdlib [`Logger`](https://ruby-doc.org/stdlib/libdoc/logger/rdoc/Logger.html) as shown below,
40
+ or a [`Google::Cloud::Logging::Logger`](https://googleapis.dev/ruby/google-cloud-logging/latest)
41
+ that will write logs to [Cloud Logging](https://cloud.google.com/logging/). See [grpc/logconfig.rb](https://github.com/grpc/grpc/blob/master/src/ruby/lib/grpc/logconfig.rb)
42
+ and the gRPC [spec_helper.rb](https://github.com/grpc/grpc/blob/master/src/ruby/spec/spec_helper.rb) for additional information.
43
+
44
+ Configuring a Ruby stdlib logger:
45
+
46
+ ```ruby
47
+ require "logger"
48
+
49
+ module MyLogger
50
+ LOGGER = Logger.new $stderr, level: Logger::WARN
51
+ def logger
52
+ LOGGER
53
+ end
54
+ end
55
+
56
+ # Define a gRPC module-level logger method before grpc/logconfig.rb loads.
57
+ module GRPC
58
+ extend MyLogger
59
+ end
60
+ ```
61
+
15
62
  ## Supported Ruby Versions
16
63
 
17
64
  This library is supported on Ruby 2.4+.
@@ -1 +1,21 @@
1
- require "google/cloud/bigquery/storage/v1/version"
1
+ # frozen_string_literal: true
2
+
3
+ # Copyright 2020 Google LLC
4
+ #
5
+ # Licensed under the Apache License, Version 2.0 (the "License");
6
+ # you may not use this file except in compliance with the License.
7
+ # You may obtain a copy of the License at
8
+ #
9
+ # https://www.apache.org/licenses/LICENSE-2.0
10
+ #
11
+ # Unless required by applicable law or agreed to in writing, software
12
+ # distributed under the License is distributed on an "AS IS" BASIS,
13
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14
+ # See the License for the specific language governing permissions and
15
+ # limitations under the License.
16
+
17
+ # Auto-generated by gapic-generator-ruby. DO NOT EDIT!
18
+
19
+ # This gem does not autoload during Bundler.require. To load this gem,
20
+ # issue explicit require statements for the packages desired, e.g.:
21
+ # require "google/cloud/bigquery/storage/v1"
@@ -17,3 +17,21 @@
17
17
  # Auto-generated by gapic-generator-ruby. DO NOT EDIT!
18
18
 
19
19
  require "google/cloud/bigquery/storage/v1/big_query_read"
20
+ require "google/cloud/bigquery/storage/v1/version"
21
+
22
+ module Google
23
+ module Cloud
24
+ module Bigquery
25
+ module Storage
26
+ ##
27
+ # To load this package, including all its services, and instantiate a client:
28
+ #
29
+ # require "google/cloud/bigquery/storage/v1"
30
+ # client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
31
+ #
32
+ module V1
33
+ end
34
+ end
35
+ end
36
+ end
37
+ end
@@ -16,5 +16,38 @@
16
16
 
17
17
  # Auto-generated by gapic-generator-ruby. DO NOT EDIT!
18
18
 
19
- require "google/cloud/bigquery/storage/v1/big_query_read/client"
19
+ require "gapic/common"
20
+ require "gapic/config"
21
+ require "gapic/config/method"
22
+
23
+ require "google/cloud/bigquery/storage/v1/version"
24
+
20
25
  require "google/cloud/bigquery/storage/v1/big_query_read/credentials"
26
+ require "google/cloud/bigquery/storage/v1/big_query_read/paths"
27
+ require "google/cloud/bigquery/storage/v1/big_query_read/client"
28
+
29
+ module Google
30
+ module Cloud
31
+ module Bigquery
32
+ module Storage
33
+ module V1
34
+ ##
35
+ # BigQuery Read API.
36
+ #
37
+ # The Read API can be used to read data from BigQuery.
38
+ #
39
+ # To load this service and instantiate a client:
40
+ #
41
+ # require "google/cloud/bigquery/storage/v1/big_query_read"
42
+ # client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
43
+ #
44
+ module BigQueryRead
45
+ end
46
+ end
47
+ end
48
+ end
49
+ end
50
+ end
51
+
52
+ helper_path = ::File.join __dir__, "big_query_read", "helpers.rb"
53
+ require "google/cloud/bigquery/storage/v1/big_query_read/helpers" if ::File.file? helper_path
@@ -16,15 +16,8 @@
16
16
 
17
17
  # Auto-generated by gapic-generator-ruby. DO NOT EDIT!
18
18
 
19
- require "gapic/common"
20
- require "gapic/config"
21
- require "gapic/config/method"
22
-
23
19
  require "google/cloud/errors"
24
- require "google/cloud/bigquery/storage/v1/version"
25
20
  require "google/cloud/bigquery/storage/v1/storage_pb"
26
- require "google/cloud/bigquery/storage/v1/big_query_read/credentials"
27
- require "google/cloud/bigquery/storage/v1/big_query_read/paths"
28
21
 
29
22
  module Google
30
23
  module Cloud
@@ -48,15 +41,15 @@ module Google
48
41
  ##
49
42
  # Configure the BigQueryRead Client class.
50
43
  #
51
- # See {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration}
44
+ # See {::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration}
52
45
  # for a description of the configuration fields.
53
46
  #
54
47
  # ## Example
55
48
  #
56
49
  # To modify the configuration for all BigQueryRead clients:
57
50
  #
58
- # Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
59
- # config.timeout = 10_000
51
+ # ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
52
+ # config.timeout = 10.0
60
53
  # end
61
54
  #
62
55
  # @yield [config] Configure the Client client.
@@ -112,7 +105,7 @@ module Google
112
105
  # but structural changes (adding new fields, etc.) are not allowed. Structural changes
113
106
  # should be made on {Client.configure}.
114
107
  #
115
- # See {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration}
108
+ # See {::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration}
116
109
  # for a description of the configuration fields.
117
110
  #
118
111
  # @yield [config] Configure the Client client.
@@ -133,13 +126,13 @@ module Google
133
126
  # To create a new BigQueryRead client with the default
134
127
  # configuration:
135
128
  #
136
- # client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
129
+ # client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
137
130
  #
138
131
  # To create a new BigQueryRead client with a custom
139
132
  # configuration:
140
133
  #
141
- # client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
142
- # config.timeout = 10_000
134
+ # client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
135
+ # config.timeout = 10.0
143
136
  # end
144
137
  #
145
138
  # @yield [config] Configure the BigQueryRead client.
@@ -164,10 +157,11 @@ module Google
164
157
  if credentials.is_a?(String) || credentials.is_a?(Hash)
165
158
  credentials = Credentials.new credentials, scope: @config.scope
166
159
  end
167
- @quota_project_id = credentials.respond_to?(:quota_project_id) ? credentials.quota_project_id : nil
160
+ @quota_project_id = @config.quota_project
161
+ @quota_project_id ||= credentials.quota_project_id if credentials.respond_to? :quota_project_id
168
162
 
169
- @big_query_read_stub = Gapic::ServiceStub.new(
170
- Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Stub,
163
+ @big_query_read_stub = ::Gapic::ServiceStub.new(
164
+ ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Stub,
171
165
  credentials: credentials,
172
166
  endpoint: @config.endpoint,
173
167
  channel_args: @config.channel_args,
@@ -199,36 +193,26 @@ module Google
199
193
  # not require manual clean-up by the caller.
200
194
  #
201
195
  # @overload create_read_session(request, options = nil)
202
- # @param request [Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest | Hash]
203
- # Creates a new read session. A read session divides the contents of a
204
- # BigQuery table into one or more streams, which can then be used to read
205
- # data from the table. The read session also specifies properties of the
206
- # data to be read, such as a list of columns or a push-down filter describing
207
- # the rows to be returned.
208
- #
209
- # A particular row can be read by at most one stream. When the caller has
210
- # reached the end of each stream in the session, then all the data in the
211
- # table has been read.
212
- #
213
- # Data is assigned to each stream such that roughly the same number of
214
- # rows can be read from each stream. Because the server-side unit for
215
- # assigning data is collections of rows, the API does not guarantee that
216
- # each stream will return the same number or rows. Additionally, the
217
- # limits are enforced based on the number of pre-filtered rows, so some
218
- # filters can lead to lopsided assignments.
219
- #
220
- # Read sessions automatically expire 24 hours after they are created and do
221
- # not require manual clean-up by the caller.
222
- # @param options [Gapic::CallOptions, Hash]
196
+ # Pass arguments to `create_read_session` via a request object, either of type
197
+ # {::Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest} or an equivalent Hash.
198
+ #
199
+ # @param request [::Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest, ::Hash]
200
+ # A request object representing the call parameters. Required. To specify no
201
+ # parameters, or to keep all the default parameter values, pass an empty Hash.
202
+ # @param options [::Gapic::CallOptions, ::Hash]
223
203
  # Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
224
204
  #
225
205
  # @overload create_read_session(parent: nil, read_session: nil, max_stream_count: nil)
226
- # @param parent [String]
206
+ # Pass arguments to `create_read_session` via keyword arguments. Note that at
207
+ # least one keyword argument is required. To specify no parameters, or to keep all
208
+ # the default parameter values, pass an empty Hash as a request object (see above).
209
+ #
210
+ # @param parent [::String]
227
211
  # Required. The request project that owns the session, in the form of
228
212
  # `projects/{project_id}`.
229
- # @param read_session [Google::Cloud::Bigquery::Storage::V1::ReadSession | Hash]
213
+ # @param read_session [::Google::Cloud::Bigquery::Storage::V1::ReadSession, ::Hash]
230
214
  # Required. Session to be created.
231
- # @param max_stream_count [Integer]
215
+ # @param max_stream_count [::Integer]
232
216
  # Max initial number of streams. If unset or zero, the server will
233
217
  # provide a value of streams so as to produce reasonable throughput. Must be
234
218
  # non-negative. The number of streams may be lower than the requested number,
@@ -238,28 +222,27 @@ module Google
238
222
  #
239
223
  # Streams must be read starting from offset 0.
240
224
  #
241
- #
242
225
  # @yield [response, operation] Access the result along with the RPC operation
243
- # @yieldparam response [Google::Cloud::Bigquery::Storage::V1::ReadSession]
244
- # @yieldparam operation [GRPC::ActiveCall::Operation]
226
+ # @yieldparam response [::Google::Cloud::Bigquery::Storage::V1::ReadSession]
227
+ # @yieldparam operation [::GRPC::ActiveCall::Operation]
245
228
  #
246
- # @return [Google::Cloud::Bigquery::Storage::V1::ReadSession]
229
+ # @return [::Google::Cloud::Bigquery::Storage::V1::ReadSession]
247
230
  #
248
- # @raise [Google::Cloud::Error] if the RPC is aborted.
231
+ # @raise [::Google::Cloud::Error] if the RPC is aborted.
249
232
  #
250
233
  def create_read_session request, options = nil
251
- raise ArgumentError, "request must be provided" if request.nil?
234
+ raise ::ArgumentError, "request must be provided" if request.nil?
252
235
 
253
- request = Gapic::Protobuf.coerce request, to: Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest
236
+ request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest
254
237
 
255
238
  # Converts hash and nil to an options object
256
- options = Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
239
+ options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
257
240
 
258
241
  # Customize the options with defaults
259
242
  metadata = @config.rpcs.create_read_session.metadata.to_h
260
243
 
261
244
  # Set x-goog-api-client and x-goog-user-project headers
262
- metadata[:"x-goog-api-client"] ||= Gapic::Headers.x_goog_api_client \
245
+ metadata[:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
263
246
  lib_name: @config.lib_name, lib_version: @config.lib_version,
264
247
  gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
265
248
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
@@ -280,8 +263,8 @@ module Google
280
263
  yield response, operation if block_given?
281
264
  return response
282
265
  end
283
- rescue GRPC::BadStatus => e
284
- raise Google::Cloud::Error.from_error(e)
266
+ rescue ::GRPC::BadStatus => e
267
+ raise ::Google::Cloud::Error.from_error(e)
285
268
  end
286
269
 
287
270
  ##
@@ -294,47 +277,48 @@ module Google
294
277
  # state of the stream.
295
278
  #
296
279
  # @overload read_rows(request, options = nil)
297
- # @param request [Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest | Hash]
298
- # Reads rows from the stream in the format prescribed by the ReadSession.
299
- # Each response contains one or more table rows, up to a maximum of 100 MiB
300
- # per response; read requests which attempt to read individual rows larger
301
- # than 100 MiB will fail.
302
- #
303
- # Each request also returns a set of stream statistics reflecting the current
304
- # state of the stream.
305
- # @param options [Gapic::CallOptions, Hash]
280
+ # Pass arguments to `read_rows` via a request object, either of type
281
+ # {::Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest} or an equivalent Hash.
282
+ #
283
+ # @param request [::Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest, ::Hash]
284
+ # A request object representing the call parameters. Required. To specify no
285
+ # parameters, or to keep all the default parameter values, pass an empty Hash.
286
+ # @param options [::Gapic::CallOptions, ::Hash]
306
287
  # Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
307
288
  #
308
289
  # @overload read_rows(read_stream: nil, offset: nil)
309
- # @param read_stream [String]
290
+ # Pass arguments to `read_rows` via keyword arguments. Note that at
291
+ # least one keyword argument is required. To specify no parameters, or to keep all
292
+ # the default parameter values, pass an empty Hash as a request object (see above).
293
+ #
294
+ # @param read_stream [::String]
310
295
  # Required. Stream to read rows from.
311
- # @param offset [Integer]
296
+ # @param offset [::Integer]
312
297
  # The offset requested must be less than the last row read from Read.
313
298
  # Requesting a larger offset is undefined. If not specified, start reading
314
299
  # from offset zero.
315
300
  #
316
- #
317
301
  # @yield [response, operation] Access the result along with the RPC operation
318
- # @yieldparam response [Enumerable<Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse>]
319
- # @yieldparam operation [GRPC::ActiveCall::Operation]
302
+ # @yieldparam response [::Enumerable<::Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse>]
303
+ # @yieldparam operation [::GRPC::ActiveCall::Operation]
320
304
  #
321
- # @return [Enumerable<Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse>]
305
+ # @return [::Enumerable<::Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse>]
322
306
  #
323
- # @raise [Google::Cloud::Error] if the RPC is aborted.
307
+ # @raise [::Google::Cloud::Error] if the RPC is aborted.
324
308
  #
325
309
  def read_rows request, options = nil
326
- raise ArgumentError, "request must be provided" if request.nil?
310
+ raise ::ArgumentError, "request must be provided" if request.nil?
327
311
 
328
- request = Gapic::Protobuf.coerce request, to: Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest
312
+ request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest
329
313
 
330
314
  # Converts hash and nil to an options object
331
- options = Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
315
+ options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
332
316
 
333
317
  # Customize the options with defaults
334
318
  metadata = @config.rpcs.read_rows.metadata.to_h
335
319
 
336
320
  # Set x-goog-api-client and x-goog-user-project headers
337
- metadata[:"x-goog-api-client"] ||= Gapic::Headers.x_goog_api_client \
321
+ metadata[:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
338
322
  lib_name: @config.lib_name, lib_version: @config.lib_version,
339
323
  gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
340
324
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
@@ -355,8 +339,8 @@ module Google
355
339
  yield response, operation if block_given?
356
340
  return response
357
341
  end
358
- rescue GRPC::BadStatus => e
359
- raise Google::Cloud::Error.from_error(e)
342
+ rescue ::GRPC::BadStatus => e
343
+ raise ::Google::Cloud::Error.from_error(e)
360
344
  end
361
345
 
362
346
  ##
@@ -374,26 +358,23 @@ module Google
374
358
  # completion.
375
359
  #
376
360
  # @overload split_read_stream(request, options = nil)
377
- # @param request [Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest | Hash]
378
- # Splits a given `ReadStream` into two `ReadStream` objects. These
379
- # `ReadStream` objects are referred to as the primary and the residual
380
- # streams of the split. The original `ReadStream` can still be read from in
381
- # the same manner as before. Both of the returned `ReadStream` objects can
382
- # also be read from, and the rows returned by both child streams will be
383
- # the same as the rows read from the original stream.
384
- #
385
- # Moreover, the two child streams will be allocated back-to-back in the
386
- # original `ReadStream`. Concretely, it is guaranteed that for streams
387
- # original, primary, and residual, that original[0-j] = primary[0-j] and
388
- # original[j-n] = residual[0-m] once the streams have been read to
389
- # completion.
390
- # @param options [Gapic::CallOptions, Hash]
361
+ # Pass arguments to `split_read_stream` via a request object, either of type
362
+ # {::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest} or an equivalent Hash.
363
+ #
364
+ # @param request [::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest, ::Hash]
365
+ # A request object representing the call parameters. Required. To specify no
366
+ # parameters, or to keep all the default parameter values, pass an empty Hash.
367
+ # @param options [::Gapic::CallOptions, ::Hash]
391
368
  # Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
392
369
  #
393
370
  # @overload split_read_stream(name: nil, fraction: nil)
394
- # @param name [String]
371
+ # Pass arguments to `split_read_stream` via keyword arguments. Note that at
372
+ # least one keyword argument is required. To specify no parameters, or to keep all
373
+ # the default parameter values, pass an empty Hash as a request object (see above).
374
+ #
375
+ # @param name [::String]
395
376
  # Required. Name of the stream to split.
396
- # @param fraction [Float]
377
+ # @param fraction [::Float]
397
378
  # A value in the range (0.0, 1.0) that specifies the fractional point at
398
379
  # which the original stream should be split. The actual split point is
399
380
  # evaluated on pre-filtered rows, so if a filter is provided, then there is
@@ -402,28 +383,27 @@ module Google
402
383
  # server-side unit for assigning data is collections of rows, this fraction
403
384
  # will always map to a data storage boundary on the server side.
404
385
  #
405
- #
406
386
  # @yield [response, operation] Access the result along with the RPC operation
407
- # @yieldparam response [Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse]
408
- # @yieldparam operation [GRPC::ActiveCall::Operation]
387
+ # @yieldparam response [::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse]
388
+ # @yieldparam operation [::GRPC::ActiveCall::Operation]
409
389
  #
410
- # @return [Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse]
390
+ # @return [::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse]
411
391
  #
412
- # @raise [Google::Cloud::Error] if the RPC is aborted.
392
+ # @raise [::Google::Cloud::Error] if the RPC is aborted.
413
393
  #
414
394
  def split_read_stream request, options = nil
415
- raise ArgumentError, "request must be provided" if request.nil?
395
+ raise ::ArgumentError, "request must be provided" if request.nil?
416
396
 
417
- request = Gapic::Protobuf.coerce request, to: Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest
397
+ request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest
418
398
 
419
399
  # Converts hash and nil to an options object
420
- options = Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
400
+ options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
421
401
 
422
402
  # Customize the options with defaults
423
403
  metadata = @config.rpcs.split_read_stream.metadata.to_h
424
404
 
425
405
  # Set x-goog-api-client and x-goog-user-project headers
426
- metadata[:"x-goog-api-client"] ||= Gapic::Headers.x_goog_api_client \
406
+ metadata[:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
427
407
  lib_name: @config.lib_name, lib_version: @config.lib_version,
428
408
  gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
429
409
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
@@ -444,8 +424,8 @@ module Google
444
424
  yield response, operation if block_given?
445
425
  return response
446
426
  end
447
- rescue GRPC::BadStatus => e
448
- raise Google::Cloud::Error.from_error(e)
427
+ rescue ::GRPC::BadStatus => e
428
+ raise ::Google::Cloud::Error.from_error(e)
449
429
  end
450
430
 
451
431
  ##
@@ -455,7 +435,7 @@ module Google
455
435
  # providing control over timeouts, retry behavior, logging, transport
456
436
  # parameters, and other low-level controls. Certain parameters can also be
457
437
  # applied individually to specific RPCs. See
458
- # {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration::Rpcs}
438
+ # {::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration::Rpcs}
459
439
  # for a list of RPCs that can be configured independently.
460
440
  #
461
441
  # Configuration can be applied globally to all clients, or to a single client
@@ -466,22 +446,22 @@ module Google
466
446
  # To modify the global config, setting the timeout for create_read_session
467
447
  # to 20 seconds, and all remaining timeouts to 10 seconds:
468
448
  #
469
- # Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
470
- # config.timeout = 10_000
471
- # config.rpcs.create_read_session.timeout = 20_000
449
+ # ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
450
+ # config.timeout = 10.0
451
+ # config.rpcs.create_read_session.timeout = 20.0
472
452
  # end
473
453
  #
474
454
  # To apply the above configuration only to a new client:
475
455
  #
476
- # client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
477
- # config.timeout = 10_000
478
- # config.rpcs.create_read_session.timeout = 20_000
456
+ # client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
457
+ # config.timeout = 10.0
458
+ # config.rpcs.create_read_session.timeout = 20.0
479
459
  # end
480
460
  #
481
461
  # @!attribute [rw] endpoint
482
462
  # The hostname or hostname:port of the service endpoint.
483
463
  # Defaults to `"bigquerystorage.googleapis.com"`.
484
- # @return [String]
464
+ # @return [::String]
485
465
  # @!attribute [rw] credentials
486
466
  # Credentials to send with calls. You may provide any of the following types:
487
467
  # * (`String`) The path to a service account key file in JSON format
@@ -493,29 +473,29 @@ module Google
493
473
  # * (`GRPC::Core::Channel`) a gRPC channel with included credentials
494
474
  # * (`GRPC::Core::ChannelCredentials`) a gRPC credentails object
495
475
  # * (`nil`) indicating no credentials
496
- # @return [Object]
476
+ # @return [::Object]
497
477
  # @!attribute [rw] scope
498
478
  # The OAuth scopes
499
- # @return [Array<String>]
479
+ # @return [::Array<::String>]
500
480
  # @!attribute [rw] lib_name
501
481
  # The library name as recorded in instrumentation and logging
502
- # @return [String]
482
+ # @return [::String]
503
483
  # @!attribute [rw] lib_version
504
484
  # The library version as recorded in instrumentation and logging
505
- # @return [String]
485
+ # @return [::String]
506
486
  # @!attribute [rw] channel_args
507
487
  # Extra parameters passed to the gRPC channel. Note: this is ignored if a
508
488
  # `GRPC::Core::Channel` object is provided as the credential.
509
- # @return [Hash]
489
+ # @return [::Hash]
510
490
  # @!attribute [rw] interceptors
511
491
  # An array of interceptors that are run before calls are executed.
512
- # @return [Array<GRPC::ClientInterceptor>]
492
+ # @return [::Array<::GRPC::ClientInterceptor>]
513
493
  # @!attribute [rw] timeout
514
- # The call timeout in milliseconds.
515
- # @return [Numeric]
494
+ # The call timeout in seconds.
495
+ # @return [::Numeric]
516
496
  # @!attribute [rw] metadata
517
497
  # Additional gRPC headers to be sent with the call.
518
- # @return [Hash{Symbol=>String}]
498
+ # @return [::Hash{::Symbol=>::String}]
519
499
  # @!attribute [rw] retry_policy
520
500
  # The retry policy. The value is a hash with the following keys:
521
501
  # * `:initial_delay` (*type:* `Numeric`) - The initial delay in seconds.
@@ -523,25 +503,29 @@ module Google
523
503
  # * `:multiplier` (*type:* `Numeric`) - The incremental backoff multiplier.
524
504
  # * `:retry_codes` (*type:* `Array<String>`) - The error codes that should
525
505
  # trigger a retry.
526
- # @return [Hash]
506
+ # @return [::Hash]
507
+ # @!attribute [rw] quota_project
508
+ # A separate project against which to charge quota.
509
+ # @return [::String]
527
510
  #
528
511
  class Configuration
529
- extend Gapic::Config
512
+ extend ::Gapic::Config
530
513
 
531
- config_attr :endpoint, "bigquerystorage.googleapis.com", String
532
- config_attr :credentials, nil do |value|
514
+ config_attr :endpoint, "bigquerystorage.googleapis.com", ::String
515
+ config_attr :credentials, nil do |value|
533
516
  allowed = [::String, ::Hash, ::Proc, ::Google::Auth::Credentials, ::Signet::OAuth2::Client, nil]
534
517
  allowed += [::GRPC::Core::Channel, ::GRPC::Core::ChannelCredentials] if defined? ::GRPC
535
518
  allowed.any? { |klass| klass === value }
536
519
  end
537
- config_attr :scope, nil, String, Array, nil
538
- config_attr :lib_name, nil, String, nil
539
- config_attr :lib_version, nil, String, nil
540
- config_attr(:channel_args, { "grpc.service_config_disable_resolution"=>1 }, Hash, nil)
541
- config_attr :interceptors, nil, Array, nil
542
- config_attr :timeout, nil, Numeric, nil
543
- config_attr :metadata, nil, Hash, nil
544
- config_attr :retry_policy, nil, Hash, Proc, nil
520
+ config_attr :scope, nil, ::String, ::Array, nil
521
+ config_attr :lib_name, nil, ::String, nil
522
+ config_attr :lib_version, nil, ::String, nil
523
+ config_attr(:channel_args, { "grpc.service_config_disable_resolution"=>1 }, ::Hash, nil)
524
+ config_attr :interceptors, nil, ::Array, nil
525
+ config_attr :timeout, nil, ::Numeric, nil
526
+ config_attr :metadata, nil, ::Hash, nil
527
+ config_attr :retry_policy, nil, ::Hash, ::Proc, nil
528
+ config_attr :quota_project, nil, ::String, nil
545
529
 
546
530
  # @private
547
531
  def initialize parent_config = nil
@@ -582,28 +566,28 @@ module Google
582
566
  class Rpcs
583
567
  ##
584
568
  # RPC-specific configuration for `create_read_session`
585
- # @return [Gapic::Config::Method]
569
+ # @return [::Gapic::Config::Method]
586
570
  #
587
571
  attr_reader :create_read_session
588
572
  ##
589
573
  # RPC-specific configuration for `read_rows`
590
- # @return [Gapic::Config::Method]
574
+ # @return [::Gapic::Config::Method]
591
575
  #
592
576
  attr_reader :read_rows
593
577
  ##
594
578
  # RPC-specific configuration for `split_read_stream`
595
- # @return [Gapic::Config::Method]
579
+ # @return [::Gapic::Config::Method]
596
580
  #
597
581
  attr_reader :split_read_stream
598
582
 
599
583
  # @private
600
584
  def initialize parent_rpcs = nil
601
585
  create_read_session_config = parent_rpcs&.create_read_session if parent_rpcs&.respond_to? :create_read_session
602
- @create_read_session = Gapic::Config::Method.new create_read_session_config
586
+ @create_read_session = ::Gapic::Config::Method.new create_read_session_config
603
587
  read_rows_config = parent_rpcs&.read_rows if parent_rpcs&.respond_to? :read_rows
604
- @read_rows = Gapic::Config::Method.new read_rows_config
588
+ @read_rows = ::Gapic::Config::Method.new read_rows_config
605
589
  split_read_stream_config = parent_rpcs&.split_read_stream if parent_rpcs&.respond_to? :split_read_stream
606
- @split_read_stream = Gapic::Config::Method.new split_read_stream_config
590
+ @split_read_stream = ::Gapic::Config::Method.new split_read_stream_config
607
591
 
608
592
  yield self if block_given?
609
593
  end
@@ -616,13 +600,3 @@ module Google
616
600
  end
617
601
  end
618
602
  end
619
-
620
- # rubocop:disable Lint/HandleExceptions
621
-
622
- # Once client is loaded, load helpers.rb if it exists.
623
- begin
624
- require "google/cloud/bigquery/storage/v1/big_query_read/helpers"
625
- rescue LoadError
626
- end
627
-
628
- # rubocop:enable Lint/HandleExceptions