google-cloud-bigquery-storage-v1 0.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +7 -0
- data/.yardopts +12 -0
- data/AUTHENTICATION.md +169 -0
- data/LICENSE.md +203 -0
- data/README.md +24 -0
- data/lib/google-cloud-bigquery-storage-v1.rb +1 -0
- data/lib/google/cloud/bigquery/storage/v1.rb +19 -0
- data/lib/google/cloud/bigquery/storage/v1/arrow_pb.rb +29 -0
- data/lib/google/cloud/bigquery/storage/v1/avro_pb.rb +29 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_read.rb +20 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/client.rb +628 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/credentials.rb +55 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/paths.rb +108 -0
- data/lib/google/cloud/bigquery/storage/v1/storage_pb.rb +71 -0
- data/lib/google/cloud/bigquery/storage/v1/storage_services_pb.rb +89 -0
- data/lib/google/cloud/bigquery/storage/v1/stream_pb.rb +58 -0
- data/lib/google/cloud/bigquery/storage/v1/version.rb +30 -0
- data/lib/google/cloud/common_resources_pb.rb +15 -0
- data/proto_docs/README.md +4 -0
- data/proto_docs/google/api/field_behavior.rb +59 -0
- data/proto_docs/google/api/resource.rb +247 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/arrow.rb +54 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/avro.rb +50 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/storage.rb +163 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/stream.rb +126 -0
- data/proto_docs/google/protobuf/timestamp.rb +116 -0
- metadata +179 -0
@@ -0,0 +1,29 @@
|
|
1
|
+
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
|
+
# source: google/cloud/bigquery/storage/v1/avro.proto
|
3
|
+
|
4
|
+
require 'google/protobuf'
|
5
|
+
|
6
|
+
Google::Protobuf::DescriptorPool.generated_pool.build do
|
7
|
+
add_file("google/cloud/bigquery/storage/v1/avro.proto", :syntax => :proto3) do
|
8
|
+
add_message "google.cloud.bigquery.storage.v1.AvroSchema" do
|
9
|
+
optional :schema, :string, 1
|
10
|
+
end
|
11
|
+
add_message "google.cloud.bigquery.storage.v1.AvroRows" do
|
12
|
+
optional :serialized_binary_rows, :bytes, 1
|
13
|
+
optional :row_count, :int64, 2
|
14
|
+
end
|
15
|
+
end
|
16
|
+
end
|
17
|
+
|
18
|
+
module Google
|
19
|
+
module Cloud
|
20
|
+
module Bigquery
|
21
|
+
module Storage
|
22
|
+
module V1
|
23
|
+
AvroSchema = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.AvroSchema").msgclass
|
24
|
+
AvroRows = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.AvroRows").msgclass
|
25
|
+
end
|
26
|
+
end
|
27
|
+
end
|
28
|
+
end
|
29
|
+
end
|
@@ -0,0 +1,20 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2020 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
require "google/cloud/bigquery/storage/v1/big_query_read/client"
|
20
|
+
require "google/cloud/bigquery/storage/v1/big_query_read/credentials"
|
@@ -0,0 +1,628 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2020 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
require "gapic/common"
|
20
|
+
require "gapic/config"
|
21
|
+
require "gapic/config/method"
|
22
|
+
|
23
|
+
require "google/cloud/errors"
|
24
|
+
require "google/cloud/bigquery/storage/v1/version"
|
25
|
+
require "google/cloud/bigquery/storage/v1/storage_pb"
|
26
|
+
require "google/cloud/bigquery/storage/v1/big_query_read/credentials"
|
27
|
+
require "google/cloud/bigquery/storage/v1/big_query_read/paths"
|
28
|
+
|
29
|
+
module Google
|
30
|
+
module Cloud
|
31
|
+
module Bigquery
|
32
|
+
module Storage
|
33
|
+
module V1
|
34
|
+
module BigQueryRead
|
35
|
+
##
|
36
|
+
# Client for the BigQueryRead service.
|
37
|
+
#
|
38
|
+
# BigQuery Read API.
|
39
|
+
#
|
40
|
+
# The Read API can be used to read data from BigQuery.
|
41
|
+
#
|
42
|
+
class Client
|
43
|
+
include Paths
|
44
|
+
|
45
|
+
# @private
|
46
|
+
attr_reader :big_query_read_stub
|
47
|
+
|
48
|
+
##
|
49
|
+
# Configure the BigQueryRead Client class.
|
50
|
+
#
|
51
|
+
# See {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration}
|
52
|
+
# for a description of the configuration fields.
|
53
|
+
#
|
54
|
+
# ## Example
|
55
|
+
#
|
56
|
+
# To modify the configuration for all BigQueryRead clients:
|
57
|
+
#
|
58
|
+
# Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
|
59
|
+
# config.timeout = 10_000
|
60
|
+
# end
|
61
|
+
#
|
62
|
+
# @yield [config] Configure the Client client.
|
63
|
+
# @yieldparam config [Client::Configuration]
|
64
|
+
#
|
65
|
+
# @return [Client::Configuration]
|
66
|
+
#
|
67
|
+
def self.configure
|
68
|
+
@configure ||= begin
|
69
|
+
namespace = ["Google", "Cloud", "Bigquery", "Storage", "V1"]
|
70
|
+
parent_config = while namespace.any?
|
71
|
+
parent_name = namespace.join "::"
|
72
|
+
parent_const = const_get parent_name
|
73
|
+
break parent_const.configure if parent_const&.respond_to? :configure
|
74
|
+
namespace.pop
|
75
|
+
end
|
76
|
+
default_config = Client::Configuration.new parent_config
|
77
|
+
|
78
|
+
default_config.rpcs.create_read_session.timeout = 600.0
|
79
|
+
default_config.rpcs.create_read_session.retry_policy = {
|
80
|
+
initial_delay: 0.1,
|
81
|
+
max_delay: 60.0,
|
82
|
+
multiplier: 1.3,
|
83
|
+
retry_codes: ["DEADLINE_EXCEEDED", "UNAVAILABLE"]
|
84
|
+
}
|
85
|
+
|
86
|
+
default_config.rpcs.read_rows.timeout = 86_400.0
|
87
|
+
default_config.rpcs.read_rows.retry_policy = {
|
88
|
+
initial_delay: 0.1,
|
89
|
+
max_delay: 60.0,
|
90
|
+
multiplier: 1.3,
|
91
|
+
retry_codes: ["UNAVAILABLE"]
|
92
|
+
}
|
93
|
+
|
94
|
+
default_config.rpcs.split_read_stream.timeout = 600.0
|
95
|
+
default_config.rpcs.split_read_stream.retry_policy = {
|
96
|
+
initial_delay: 0.1,
|
97
|
+
max_delay: 60.0,
|
98
|
+
multiplier: 1.3,
|
99
|
+
retry_codes: ["DEADLINE_EXCEEDED", "UNAVAILABLE"]
|
100
|
+
}
|
101
|
+
|
102
|
+
default_config
|
103
|
+
end
|
104
|
+
yield @configure if block_given?
|
105
|
+
@configure
|
106
|
+
end
|
107
|
+
|
108
|
+
##
|
109
|
+
# Configure the BigQueryRead Client instance.
|
110
|
+
#
|
111
|
+
# The configuration is set to the derived mode, meaning that values can be changed,
|
112
|
+
# but structural changes (adding new fields, etc.) are not allowed. Structural changes
|
113
|
+
# should be made on {Client.configure}.
|
114
|
+
#
|
115
|
+
# See {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration}
|
116
|
+
# for a description of the configuration fields.
|
117
|
+
#
|
118
|
+
# @yield [config] Configure the Client client.
|
119
|
+
# @yieldparam config [Client::Configuration]
|
120
|
+
#
|
121
|
+
# @return [Client::Configuration]
|
122
|
+
#
|
123
|
+
def configure
|
124
|
+
yield @config if block_given?
|
125
|
+
@config
|
126
|
+
end
|
127
|
+
|
128
|
+
##
|
129
|
+
# Create a new BigQueryRead client object.
|
130
|
+
#
|
131
|
+
# ## Examples
|
132
|
+
#
|
133
|
+
# To create a new BigQueryRead client with the default
|
134
|
+
# configuration:
|
135
|
+
#
|
136
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
|
137
|
+
#
|
138
|
+
# To create a new BigQueryRead client with a custom
|
139
|
+
# configuration:
|
140
|
+
#
|
141
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
|
142
|
+
# config.timeout = 10_000
|
143
|
+
# end
|
144
|
+
#
|
145
|
+
# @yield [config] Configure the BigQueryRead client.
|
146
|
+
# @yieldparam config [Client::Configuration]
|
147
|
+
#
|
148
|
+
def initialize
|
149
|
+
# These require statements are intentionally placed here to initialize
|
150
|
+
# the gRPC module only when it's required.
|
151
|
+
# See https://github.com/googleapis/toolkit/issues/446
|
152
|
+
require "gapic/grpc"
|
153
|
+
require "google/cloud/bigquery/storage/v1/storage_services_pb"
|
154
|
+
|
155
|
+
# Create the configuration object
|
156
|
+
@config = Configuration.new Client.configure
|
157
|
+
|
158
|
+
# Yield the configuration if needed
|
159
|
+
yield @config if block_given?
|
160
|
+
|
161
|
+
# Create credentials
|
162
|
+
credentials = @config.credentials
|
163
|
+
credentials ||= Credentials.default scope: @config.scope
|
164
|
+
if credentials.is_a?(String) || credentials.is_a?(Hash)
|
165
|
+
credentials = Credentials.new credentials, scope: @config.scope
|
166
|
+
end
|
167
|
+
@quota_project_id = credentials.respond_to?(:quota_project_id) ? credentials.quota_project_id : nil
|
168
|
+
|
169
|
+
@big_query_read_stub = Gapic::ServiceStub.new(
|
170
|
+
Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Stub,
|
171
|
+
credentials: credentials,
|
172
|
+
endpoint: @config.endpoint,
|
173
|
+
channel_args: @config.channel_args,
|
174
|
+
interceptors: @config.interceptors
|
175
|
+
)
|
176
|
+
end
|
177
|
+
|
178
|
+
# Service calls
|
179
|
+
|
180
|
+
##
|
181
|
+
# Creates a new read session. A read session divides the contents of a
|
182
|
+
# BigQuery table into one or more streams, which can then be used to read
|
183
|
+
# data from the table. The read session also specifies properties of the
|
184
|
+
# data to be read, such as a list of columns or a push-down filter describing
|
185
|
+
# the rows to be returned.
|
186
|
+
#
|
187
|
+
# A particular row can be read by at most one stream. When the caller has
|
188
|
+
# reached the end of each stream in the session, then all the data in the
|
189
|
+
# table has been read.
|
190
|
+
#
|
191
|
+
# Data is assigned to each stream such that roughly the same number of
|
192
|
+
# rows can be read from each stream. Because the server-side unit for
|
193
|
+
# assigning data is collections of rows, the API does not guarantee that
|
194
|
+
# each stream will return the same number or rows. Additionally, the
|
195
|
+
# limits are enforced based on the number of pre-filtered rows, so some
|
196
|
+
# filters can lead to lopsided assignments.
|
197
|
+
#
|
198
|
+
# Read sessions automatically expire 24 hours after they are created and do
|
199
|
+
# not require manual clean-up by the caller.
|
200
|
+
#
|
201
|
+
# @overload create_read_session(request, options = nil)
|
202
|
+
# @param request [Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest | Hash]
|
203
|
+
# Creates a new read session. A read session divides the contents of a
|
204
|
+
# BigQuery table into one or more streams, which can then be used to read
|
205
|
+
# data from the table. The read session also specifies properties of the
|
206
|
+
# data to be read, such as a list of columns or a push-down filter describing
|
207
|
+
# the rows to be returned.
|
208
|
+
#
|
209
|
+
# A particular row can be read by at most one stream. When the caller has
|
210
|
+
# reached the end of each stream in the session, then all the data in the
|
211
|
+
# table has been read.
|
212
|
+
#
|
213
|
+
# Data is assigned to each stream such that roughly the same number of
|
214
|
+
# rows can be read from each stream. Because the server-side unit for
|
215
|
+
# assigning data is collections of rows, the API does not guarantee that
|
216
|
+
# each stream will return the same number or rows. Additionally, the
|
217
|
+
# limits are enforced based on the number of pre-filtered rows, so some
|
218
|
+
# filters can lead to lopsided assignments.
|
219
|
+
#
|
220
|
+
# Read sessions automatically expire 24 hours after they are created and do
|
221
|
+
# not require manual clean-up by the caller.
|
222
|
+
# @param options [Gapic::CallOptions, Hash]
|
223
|
+
# Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
|
224
|
+
#
|
225
|
+
# @overload create_read_session(parent: nil, read_session: nil, max_stream_count: nil)
|
226
|
+
# @param parent [String]
|
227
|
+
# Required. The request project that owns the session, in the form of
|
228
|
+
# `projects/{project_id}`.
|
229
|
+
# @param read_session [Google::Cloud::Bigquery::Storage::V1::ReadSession | Hash]
|
230
|
+
# Required. Session to be created.
|
231
|
+
# @param max_stream_count [Integer]
|
232
|
+
# Max initial number of streams. If unset or zero, the server will
|
233
|
+
# provide a value of streams so as to produce reasonable throughput. Must be
|
234
|
+
# non-negative. The number of streams may be lower than the requested number,
|
235
|
+
# depending on the amount parallelism that is reasonable for the table. Error
|
236
|
+
# will be returned if the max count is greater than the current system
|
237
|
+
# max limit of 1,000.
|
238
|
+
#
|
239
|
+
# Streams must be read starting from offset 0.
|
240
|
+
#
|
241
|
+
#
|
242
|
+
# @yield [response, operation] Access the result along with the RPC operation
|
243
|
+
# @yieldparam response [Google::Cloud::Bigquery::Storage::V1::ReadSession]
|
244
|
+
# @yieldparam operation [GRPC::ActiveCall::Operation]
|
245
|
+
#
|
246
|
+
# @return [Google::Cloud::Bigquery::Storage::V1::ReadSession]
|
247
|
+
#
|
248
|
+
# @raise [Google::Cloud::Error] if the RPC is aborted.
|
249
|
+
#
|
250
|
+
def create_read_session request, options = nil
|
251
|
+
raise ArgumentError, "request must be provided" if request.nil?
|
252
|
+
|
253
|
+
request = Gapic::Protobuf.coerce request, to: Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest
|
254
|
+
|
255
|
+
# Converts hash and nil to an options object
|
256
|
+
options = Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
|
257
|
+
|
258
|
+
# Customize the options with defaults
|
259
|
+
metadata = @config.rpcs.create_read_session.metadata.to_h
|
260
|
+
|
261
|
+
# Set x-goog-api-client and x-goog-user-project headers
|
262
|
+
metadata[:"x-goog-api-client"] ||= Gapic::Headers.x_goog_api_client \
|
263
|
+
lib_name: @config.lib_name, lib_version: @config.lib_version,
|
264
|
+
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
265
|
+
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
266
|
+
|
267
|
+
header_params = {
|
268
|
+
"read_session.table" => request.read_session.table
|
269
|
+
}
|
270
|
+
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
271
|
+
metadata[:"x-goog-request-params"] ||= request_params_header
|
272
|
+
|
273
|
+
options.apply_defaults timeout: @config.rpcs.create_read_session.timeout,
|
274
|
+
metadata: metadata,
|
275
|
+
retry_policy: @config.rpcs.create_read_session.retry_policy
|
276
|
+
options.apply_defaults metadata: @config.metadata,
|
277
|
+
retry_policy: @config.retry_policy
|
278
|
+
|
279
|
+
@big_query_read_stub.call_rpc :create_read_session, request, options: options do |response, operation|
|
280
|
+
yield response, operation if block_given?
|
281
|
+
return response
|
282
|
+
end
|
283
|
+
rescue GRPC::BadStatus => e
|
284
|
+
raise Google::Cloud::Error.from_error(e)
|
285
|
+
end
|
286
|
+
|
287
|
+
##
|
288
|
+
# Reads rows from the stream in the format prescribed by the ReadSession.
|
289
|
+
# Each response contains one or more table rows, up to a maximum of 100 MiB
|
290
|
+
# per response; read requests which attempt to read individual rows larger
|
291
|
+
# than 100 MiB will fail.
|
292
|
+
#
|
293
|
+
# Each request also returns a set of stream statistics reflecting the current
|
294
|
+
# state of the stream.
|
295
|
+
#
|
296
|
+
# @overload read_rows(request, options = nil)
|
297
|
+
# @param request [Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest | Hash]
|
298
|
+
# Reads rows from the stream in the format prescribed by the ReadSession.
|
299
|
+
# Each response contains one or more table rows, up to a maximum of 100 MiB
|
300
|
+
# per response; read requests which attempt to read individual rows larger
|
301
|
+
# than 100 MiB will fail.
|
302
|
+
#
|
303
|
+
# Each request also returns a set of stream statistics reflecting the current
|
304
|
+
# state of the stream.
|
305
|
+
# @param options [Gapic::CallOptions, Hash]
|
306
|
+
# Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
|
307
|
+
#
|
308
|
+
# @overload read_rows(read_stream: nil, offset: nil)
|
309
|
+
# @param read_stream [String]
|
310
|
+
# Required. Stream to read rows from.
|
311
|
+
# @param offset [Integer]
|
312
|
+
# The offset requested must be less than the last row read from Read.
|
313
|
+
# Requesting a larger offset is undefined. If not specified, start reading
|
314
|
+
# from offset zero.
|
315
|
+
#
|
316
|
+
#
|
317
|
+
# @yield [response, operation] Access the result along with the RPC operation
|
318
|
+
# @yieldparam response [Enumerable<Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse>]
|
319
|
+
# @yieldparam operation [GRPC::ActiveCall::Operation]
|
320
|
+
#
|
321
|
+
# @return [Enumerable<Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse>]
|
322
|
+
#
|
323
|
+
# @raise [Google::Cloud::Error] if the RPC is aborted.
|
324
|
+
#
|
325
|
+
def read_rows request, options = nil
|
326
|
+
raise ArgumentError, "request must be provided" if request.nil?
|
327
|
+
|
328
|
+
request = Gapic::Protobuf.coerce request, to: Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest
|
329
|
+
|
330
|
+
# Converts hash and nil to an options object
|
331
|
+
options = Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
|
332
|
+
|
333
|
+
# Customize the options with defaults
|
334
|
+
metadata = @config.rpcs.read_rows.metadata.to_h
|
335
|
+
|
336
|
+
# Set x-goog-api-client and x-goog-user-project headers
|
337
|
+
metadata[:"x-goog-api-client"] ||= Gapic::Headers.x_goog_api_client \
|
338
|
+
lib_name: @config.lib_name, lib_version: @config.lib_version,
|
339
|
+
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
340
|
+
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
341
|
+
|
342
|
+
header_params = {
|
343
|
+
"read_stream" => request.read_stream
|
344
|
+
}
|
345
|
+
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
346
|
+
metadata[:"x-goog-request-params"] ||= request_params_header
|
347
|
+
|
348
|
+
options.apply_defaults timeout: @config.rpcs.read_rows.timeout,
|
349
|
+
metadata: metadata,
|
350
|
+
retry_policy: @config.rpcs.read_rows.retry_policy
|
351
|
+
options.apply_defaults metadata: @config.metadata,
|
352
|
+
retry_policy: @config.retry_policy
|
353
|
+
|
354
|
+
@big_query_read_stub.call_rpc :read_rows, request, options: options do |response, operation|
|
355
|
+
yield response, operation if block_given?
|
356
|
+
return response
|
357
|
+
end
|
358
|
+
rescue GRPC::BadStatus => e
|
359
|
+
raise Google::Cloud::Error.from_error(e)
|
360
|
+
end
|
361
|
+
|
362
|
+
##
|
363
|
+
# Splits a given `ReadStream` into two `ReadStream` objects. These
|
364
|
+
# `ReadStream` objects are referred to as the primary and the residual
|
365
|
+
# streams of the split. The original `ReadStream` can still be read from in
|
366
|
+
# the same manner as before. Both of the returned `ReadStream` objects can
|
367
|
+
# also be read from, and the rows returned by both child streams will be
|
368
|
+
# the same as the rows read from the original stream.
|
369
|
+
#
|
370
|
+
# Moreover, the two child streams will be allocated back-to-back in the
|
371
|
+
# original `ReadStream`. Concretely, it is guaranteed that for streams
|
372
|
+
# original, primary, and residual, that original[0-j] = primary[0-j] and
|
373
|
+
# original[j-n] = residual[0-m] once the streams have been read to
|
374
|
+
# completion.
|
375
|
+
#
|
376
|
+
# @overload split_read_stream(request, options = nil)
|
377
|
+
# @param request [Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest | Hash]
|
378
|
+
# Splits a given `ReadStream` into two `ReadStream` objects. These
|
379
|
+
# `ReadStream` objects are referred to as the primary and the residual
|
380
|
+
# streams of the split. The original `ReadStream` can still be read from in
|
381
|
+
# the same manner as before. Both of the returned `ReadStream` objects can
|
382
|
+
# also be read from, and the rows returned by both child streams will be
|
383
|
+
# the same as the rows read from the original stream.
|
384
|
+
#
|
385
|
+
# Moreover, the two child streams will be allocated back-to-back in the
|
386
|
+
# original `ReadStream`. Concretely, it is guaranteed that for streams
|
387
|
+
# original, primary, and residual, that original[0-j] = primary[0-j] and
|
388
|
+
# original[j-n] = residual[0-m] once the streams have been read to
|
389
|
+
# completion.
|
390
|
+
# @param options [Gapic::CallOptions, Hash]
|
391
|
+
# Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
|
392
|
+
#
|
393
|
+
# @overload split_read_stream(name: nil, fraction: nil)
|
394
|
+
# @param name [String]
|
395
|
+
# Required. Name of the stream to split.
|
396
|
+
# @param fraction [Float]
|
397
|
+
# A value in the range (0.0, 1.0) that specifies the fractional point at
|
398
|
+
# which the original stream should be split. The actual split point is
|
399
|
+
# evaluated on pre-filtered rows, so if a filter is provided, then there is
|
400
|
+
# no guarantee that the division of the rows between the new child streams
|
401
|
+
# will be proportional to this fractional value. Additionally, because the
|
402
|
+
# server-side unit for assigning data is collections of rows, this fraction
|
403
|
+
# will always map to a data storage boundary on the server side.
|
404
|
+
#
|
405
|
+
#
|
406
|
+
# @yield [response, operation] Access the result along with the RPC operation
|
407
|
+
# @yieldparam response [Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse]
|
408
|
+
# @yieldparam operation [GRPC::ActiveCall::Operation]
|
409
|
+
#
|
410
|
+
# @return [Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse]
|
411
|
+
#
|
412
|
+
# @raise [Google::Cloud::Error] if the RPC is aborted.
|
413
|
+
#
|
414
|
+
def split_read_stream request, options = nil
|
415
|
+
raise ArgumentError, "request must be provided" if request.nil?
|
416
|
+
|
417
|
+
request = Gapic::Protobuf.coerce request, to: Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest
|
418
|
+
|
419
|
+
# Converts hash and nil to an options object
|
420
|
+
options = Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h
|
421
|
+
|
422
|
+
# Customize the options with defaults
|
423
|
+
metadata = @config.rpcs.split_read_stream.metadata.to_h
|
424
|
+
|
425
|
+
# Set x-goog-api-client and x-goog-user-project headers
|
426
|
+
metadata[:"x-goog-api-client"] ||= Gapic::Headers.x_goog_api_client \
|
427
|
+
lib_name: @config.lib_name, lib_version: @config.lib_version,
|
428
|
+
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
429
|
+
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
430
|
+
|
431
|
+
header_params = {
|
432
|
+
"name" => request.name
|
433
|
+
}
|
434
|
+
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
435
|
+
metadata[:"x-goog-request-params"] ||= request_params_header
|
436
|
+
|
437
|
+
options.apply_defaults timeout: @config.rpcs.split_read_stream.timeout,
|
438
|
+
metadata: metadata,
|
439
|
+
retry_policy: @config.rpcs.split_read_stream.retry_policy
|
440
|
+
options.apply_defaults metadata: @config.metadata,
|
441
|
+
retry_policy: @config.retry_policy
|
442
|
+
|
443
|
+
@big_query_read_stub.call_rpc :split_read_stream, request, options: options do |response, operation|
|
444
|
+
yield response, operation if block_given?
|
445
|
+
return response
|
446
|
+
end
|
447
|
+
rescue GRPC::BadStatus => e
|
448
|
+
raise Google::Cloud::Error.from_error(e)
|
449
|
+
end
|
450
|
+
|
451
|
+
##
|
452
|
+
# Configuration class for the BigQueryRead API.
|
453
|
+
#
|
454
|
+
# This class represents the configuration for BigQueryRead,
|
455
|
+
# providing control over timeouts, retry behavior, logging, transport
|
456
|
+
# parameters, and other low-level controls. Certain parameters can also be
|
457
|
+
# applied individually to specific RPCs. See
|
458
|
+
# {Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client::Configuration::Rpcs}
|
459
|
+
# for a list of RPCs that can be configured independently.
|
460
|
+
#
|
461
|
+
# Configuration can be applied globally to all clients, or to a single client
|
462
|
+
# on construction.
|
463
|
+
#
|
464
|
+
# # Examples
|
465
|
+
#
|
466
|
+
# To modify the global config, setting the timeout for create_read_session
|
467
|
+
# to 20 seconds, and all remaining timeouts to 10 seconds:
|
468
|
+
#
|
469
|
+
# Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.configure do |config|
|
470
|
+
# config.timeout = 10_000
|
471
|
+
# config.rpcs.create_read_session.timeout = 20_000
|
472
|
+
# end
|
473
|
+
#
|
474
|
+
# To apply the above configuration only to a new client:
|
475
|
+
#
|
476
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new do |config|
|
477
|
+
# config.timeout = 10_000
|
478
|
+
# config.rpcs.create_read_session.timeout = 20_000
|
479
|
+
# end
|
480
|
+
#
|
481
|
+
# @!attribute [rw] endpoint
|
482
|
+
# The hostname or hostname:port of the service endpoint.
|
483
|
+
# Defaults to `"bigquerystorage.googleapis.com"`.
|
484
|
+
# @return [String]
|
485
|
+
# @!attribute [rw] credentials
|
486
|
+
# Credentials to send with calls. You may provide any of the following types:
|
487
|
+
# * (`String`) The path to a service account key file in JSON format
|
488
|
+
# * (`Hash`) A service account key as a Hash
|
489
|
+
# * (`Google::Auth::Credentials`) A googleauth credentials object
|
490
|
+
# (see the [googleauth docs](https://googleapis.dev/ruby/googleauth/latest/index.html))
|
491
|
+
# * (`Signet::OAuth2::Client`) A signet oauth2 client object
|
492
|
+
# (see the [signet docs](https://googleapis.dev/ruby/signet/latest/Signet/OAuth2/Client.html))
|
493
|
+
# * (`GRPC::Core::Channel`) a gRPC channel with included credentials
|
494
|
+
# * (`GRPC::Core::ChannelCredentials`) a gRPC credentails object
|
495
|
+
# * (`nil`) indicating no credentials
|
496
|
+
# @return [Object]
|
497
|
+
# @!attribute [rw] scope
|
498
|
+
# The OAuth scopes
|
499
|
+
# @return [Array<String>]
|
500
|
+
# @!attribute [rw] lib_name
|
501
|
+
# The library name as recorded in instrumentation and logging
|
502
|
+
# @return [String]
|
503
|
+
# @!attribute [rw] lib_version
|
504
|
+
# The library version as recorded in instrumentation and logging
|
505
|
+
# @return [String]
|
506
|
+
# @!attribute [rw] channel_args
|
507
|
+
# Extra parameters passed to the gRPC channel. Note: this is ignored if a
|
508
|
+
# `GRPC::Core::Channel` object is provided as the credential.
|
509
|
+
# @return [Hash]
|
510
|
+
# @!attribute [rw] interceptors
|
511
|
+
# An array of interceptors that are run before calls are executed.
|
512
|
+
# @return [Array<GRPC::ClientInterceptor>]
|
513
|
+
# @!attribute [rw] timeout
|
514
|
+
# The call timeout in milliseconds.
|
515
|
+
# @return [Numeric]
|
516
|
+
# @!attribute [rw] metadata
|
517
|
+
# Additional gRPC headers to be sent with the call.
|
518
|
+
# @return [Hash{Symbol=>String}]
|
519
|
+
# @!attribute [rw] retry_policy
|
520
|
+
# The retry policy. The value is a hash with the following keys:
|
521
|
+
# * `:initial_delay` (*type:* `Numeric`) - The initial delay in seconds.
|
522
|
+
# * `:max_delay` (*type:* `Numeric`) - The max delay in seconds.
|
523
|
+
# * `:multiplier` (*type:* `Numeric`) - The incremental backoff multiplier.
|
524
|
+
# * `:retry_codes` (*type:* `Array<String>`) - The error codes that should
|
525
|
+
# trigger a retry.
|
526
|
+
# @return [Hash]
|
527
|
+
#
|
528
|
+
class Configuration
|
529
|
+
extend Gapic::Config
|
530
|
+
|
531
|
+
config_attr :endpoint, "bigquerystorage.googleapis.com", String
|
532
|
+
config_attr :credentials, nil do |value|
|
533
|
+
allowed = [::String, ::Hash, ::Proc, ::Google::Auth::Credentials, ::Signet::OAuth2::Client, nil]
|
534
|
+
allowed += [::GRPC::Core::Channel, ::GRPC::Core::ChannelCredentials] if defined? ::GRPC
|
535
|
+
allowed.any? { |klass| klass === value }
|
536
|
+
end
|
537
|
+
config_attr :scope, nil, String, Array, nil
|
538
|
+
config_attr :lib_name, nil, String, nil
|
539
|
+
config_attr :lib_version, nil, String, nil
|
540
|
+
config_attr(:channel_args, { "grpc.service_config_disable_resolution"=>1 }, Hash, nil)
|
541
|
+
config_attr :interceptors, nil, Array, nil
|
542
|
+
config_attr :timeout, nil, Numeric, nil
|
543
|
+
config_attr :metadata, nil, Hash, nil
|
544
|
+
config_attr :retry_policy, nil, Hash, Proc, nil
|
545
|
+
|
546
|
+
# @private
|
547
|
+
def initialize parent_config = nil
|
548
|
+
@parent_config = parent_config unless parent_config.nil?
|
549
|
+
|
550
|
+
yield self if block_given?
|
551
|
+
end
|
552
|
+
|
553
|
+
##
|
554
|
+
# Configurations for individual RPCs
|
555
|
+
# @return [Rpcs]
|
556
|
+
#
|
557
|
+
def rpcs
|
558
|
+
@rpcs ||= begin
|
559
|
+
parent_rpcs = nil
|
560
|
+
parent_rpcs = @parent_config.rpcs if @parent_config&.respond_to? :rpcs
|
561
|
+
Rpcs.new parent_rpcs
|
562
|
+
end
|
563
|
+
end
|
564
|
+
|
565
|
+
##
|
566
|
+
# Configuration RPC class for the BigQueryRead API.
|
567
|
+
#
|
568
|
+
# Includes fields providing the configuration for each RPC in this service.
|
569
|
+
# Each configuration object is of type `Gapic::Config::Method` and includes
|
570
|
+
# the following configuration fields:
|
571
|
+
#
|
572
|
+
# * `timeout` (*type:* `Numeric`) - The call timeout in milliseconds
|
573
|
+
# * `metadata` (*type:* `Hash{Symbol=>String}`) - Additional gRPC headers
|
574
|
+
# * `retry_policy (*type:* `Hash`) - The retry policy. The policy fields
|
575
|
+
# include the following keys:
|
576
|
+
# * `:initial_delay` (*type:* `Numeric`) - The initial delay in seconds.
|
577
|
+
# * `:max_delay` (*type:* `Numeric`) - The max delay in seconds.
|
578
|
+
# * `:multiplier` (*type:* `Numeric`) - The incremental backoff multiplier.
|
579
|
+
# * `:retry_codes` (*type:* `Array<String>`) - The error codes that should
|
580
|
+
# trigger a retry.
|
581
|
+
#
|
582
|
+
class Rpcs
|
583
|
+
##
|
584
|
+
# RPC-specific configuration for `create_read_session`
|
585
|
+
# @return [Gapic::Config::Method]
|
586
|
+
#
|
587
|
+
attr_reader :create_read_session
|
588
|
+
##
|
589
|
+
# RPC-specific configuration for `read_rows`
|
590
|
+
# @return [Gapic::Config::Method]
|
591
|
+
#
|
592
|
+
attr_reader :read_rows
|
593
|
+
##
|
594
|
+
# RPC-specific configuration for `split_read_stream`
|
595
|
+
# @return [Gapic::Config::Method]
|
596
|
+
#
|
597
|
+
attr_reader :split_read_stream
|
598
|
+
|
599
|
+
# @private
|
600
|
+
def initialize parent_rpcs = nil
|
601
|
+
create_read_session_config = parent_rpcs&.create_read_session if parent_rpcs&.respond_to? :create_read_session
|
602
|
+
@create_read_session = Gapic::Config::Method.new create_read_session_config
|
603
|
+
read_rows_config = parent_rpcs&.read_rows if parent_rpcs&.respond_to? :read_rows
|
604
|
+
@read_rows = Gapic::Config::Method.new read_rows_config
|
605
|
+
split_read_stream_config = parent_rpcs&.split_read_stream if parent_rpcs&.respond_to? :split_read_stream
|
606
|
+
@split_read_stream = Gapic::Config::Method.new split_read_stream_config
|
607
|
+
|
608
|
+
yield self if block_given?
|
609
|
+
end
|
610
|
+
end
|
611
|
+
end
|
612
|
+
end
|
613
|
+
end
|
614
|
+
end
|
615
|
+
end
|
616
|
+
end
|
617
|
+
end
|
618
|
+
end
|
619
|
+
|
620
|
+
# rubocop:disable Lint/HandleExceptions
|
621
|
+
|
622
|
+
# Once client is loaded, load helpers.rb if it exists.
|
623
|
+
begin
|
624
|
+
require "google/cloud/bigquery/storage/v1/big_query_read/helpers"
|
625
|
+
rescue LoadError
|
626
|
+
end
|
627
|
+
|
628
|
+
# rubocop:enable Lint/HandleExceptions
|