google-cloud-bigquery-storage-v1 0.10.0 → 0.11.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/README.md +5 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/client.rb +6 -4
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/client.rb +2 -2
- data/lib/google/cloud/bigquery/storage/v1/protobuf_pb.rb +2 -1
- data/lib/google/cloud/bigquery/storage/v1/storage_pb.rb +14 -1
- data/lib/google/cloud/bigquery/storage/v1/stream_pb.rb +2 -1
- data/lib/google/cloud/bigquery/storage/v1/table_pb.rb +2 -1
- data/lib/google/cloud/bigquery/storage/v1/version.rb +1 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/arrow.rb +2 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/avro.rb +2 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/storage.rb +35 -4
- data/proto_docs/google/cloud/bigquery/storage/v1/stream.rb +2 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/table.rb +3 -1
- data/proto_docs/google/protobuf/any.rb +3 -3
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 93523ae607e50f73b09408fdb11c67f28b18a7b23585d65876e3613a71c80393
|
4
|
+
data.tar.gz: 186c06d4d9ec840c84472b7e8cb69680713d28457b81d009ba8bec85bb346ce2
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 30d6f2151678722ff459a97691e99c310d0428e4717ca2123457903ae39cc4c24ef1d4dde6db9225079aad77d8c956a80cf35c9afe96b6c537cb765e717aa53f
|
7
|
+
data.tar.gz: 85b427659ebd8058d344feb0f79eb770e861bc570cc55a0c4711dce350933b82fc5650e83c0aeb9da28fa3f295df5e227f3708f98ea102bf965685082ac9116f
|
data/README.md
CHANGED
@@ -207,11 +207,13 @@ module Google
|
|
207
207
|
# Max initial number of streams. If unset or zero, the server will
|
208
208
|
# provide a value of streams so as to produce reasonable throughput. Must be
|
209
209
|
# non-negative. The number of streams may be lower than the requested number,
|
210
|
-
# depending on the amount parallelism that is reasonable for the table.
|
211
|
-
#
|
212
|
-
# max limit of 1,000.
|
210
|
+
# depending on the amount parallelism that is reasonable for the table.
|
211
|
+
# There is a default system max limit of 1,000.
|
213
212
|
#
|
214
|
-
#
|
213
|
+
# This must be greater than or equal to preferred_min_stream_count.
|
214
|
+
# Typically, clients should either leave this unset to let the system to
|
215
|
+
# determine an upper bound OR set this a size for the maximum "units of work"
|
216
|
+
# it can gracefully handle.
|
215
217
|
#
|
216
218
|
# @yield [response, operation] Access the result along with the RPC operation
|
217
219
|
# @yieldparam response [::Google::Cloud::Bigquery::Storage::V1::ReadSession]
|
@@ -70,9 +70,9 @@ module Google
|
|
70
70
|
end
|
71
71
|
default_config = Client::Configuration.new parent_config
|
72
72
|
|
73
|
-
default_config.rpcs.create_write_stream.timeout =
|
73
|
+
default_config.rpcs.create_write_stream.timeout = 1200.0
|
74
74
|
default_config.rpcs.create_write_stream.retry_policy = {
|
75
|
-
initial_delay: 0
|
75
|
+
initial_delay: 10.0, max_delay: 120.0, multiplier: 1.3, retry_codes: [4, 14, 8]
|
76
76
|
}
|
77
77
|
|
78
78
|
default_config.rpcs.append_rows.timeout = 86_400.0
|
@@ -1,9 +1,10 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/protobuf.proto
|
3
3
|
|
4
|
-
require 'google/protobuf/descriptor_pb'
|
5
4
|
require 'google/protobuf'
|
6
5
|
|
6
|
+
require 'google/protobuf/descriptor_pb'
|
7
|
+
|
7
8
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
8
9
|
add_file("google/cloud/bigquery/storage/v1/protobuf.proto", :syntax => :proto3) do
|
9
10
|
add_message "google.cloud.bigquery.storage.v1.ProtoSchema" do
|
@@ -1,6 +1,8 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/storage.proto
|
3
3
|
|
4
|
+
require 'google/protobuf'
|
5
|
+
|
4
6
|
require 'google/api/annotations_pb'
|
5
7
|
require 'google/api/client_pb'
|
6
8
|
require 'google/api/field_behavior_pb'
|
@@ -13,7 +15,6 @@ require 'google/cloud/bigquery/storage/v1/table_pb'
|
|
13
15
|
require 'google/protobuf/timestamp_pb'
|
14
16
|
require 'google/protobuf/wrappers_pb'
|
15
17
|
require 'google/rpc/status_pb'
|
16
|
-
require 'google/protobuf'
|
17
18
|
|
18
19
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
19
20
|
add_file("google/cloud/bigquery/storage/v1/storage.proto", :syntax => :proto3) do
|
@@ -75,6 +76,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
75
76
|
end
|
76
77
|
add_message "google.cloud.bigquery.storage.v1.AppendRowsResponse" do
|
77
78
|
optional :updated_schema, :message, 3, "google.cloud.bigquery.storage.v1.TableSchema"
|
79
|
+
repeated :row_errors, :message, 4, "google.cloud.bigquery.storage.v1.RowError"
|
78
80
|
oneof :response do
|
79
81
|
optional :append_result, :message, 1, "google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult"
|
80
82
|
optional :error, :message, 2, "google.rpc.Status"
|
@@ -124,6 +126,15 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
124
126
|
value :OFFSET_ALREADY_EXISTS, 8
|
125
127
|
value :OFFSET_OUT_OF_RANGE, 9
|
126
128
|
end
|
129
|
+
add_message "google.cloud.bigquery.storage.v1.RowError" do
|
130
|
+
optional :index, :int64, 1
|
131
|
+
optional :code, :enum, 2, "google.cloud.bigquery.storage.v1.RowError.RowErrorCode"
|
132
|
+
optional :message, :string, 3
|
133
|
+
end
|
134
|
+
add_enum "google.cloud.bigquery.storage.v1.RowError.RowErrorCode" do
|
135
|
+
value :ROW_ERROR_CODE_UNSPECIFIED, 0
|
136
|
+
value :FIELDS_ERROR, 1
|
137
|
+
end
|
127
138
|
end
|
128
139
|
end
|
129
140
|
|
@@ -154,6 +165,8 @@ module Google
|
|
154
165
|
FlushRowsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FlushRowsResponse").msgclass
|
155
166
|
StorageError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError").msgclass
|
156
167
|
StorageError::StorageErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError.StorageErrorCode").enummodule
|
168
|
+
RowError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.RowError").msgclass
|
169
|
+
RowError::RowErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.RowError.RowErrorCode").enummodule
|
157
170
|
end
|
158
171
|
end
|
159
172
|
end
|
@@ -1,13 +1,14 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/stream.proto
|
3
3
|
|
4
|
+
require 'google/protobuf'
|
5
|
+
|
4
6
|
require 'google/api/field_behavior_pb'
|
5
7
|
require 'google/api/resource_pb'
|
6
8
|
require 'google/cloud/bigquery/storage/v1/arrow_pb'
|
7
9
|
require 'google/cloud/bigquery/storage/v1/avro_pb'
|
8
10
|
require 'google/cloud/bigquery/storage/v1/table_pb'
|
9
11
|
require 'google/protobuf/timestamp_pb'
|
10
|
-
require 'google/protobuf'
|
11
12
|
|
12
13
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
13
14
|
add_file("google/cloud/bigquery/storage/v1/stream.proto", :syntax => :proto3) do
|
@@ -1,9 +1,10 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/table.proto
|
3
3
|
|
4
|
-
require 'google/api/field_behavior_pb'
|
5
4
|
require 'google/protobuf'
|
6
5
|
|
6
|
+
require 'google/api/field_behavior_pb'
|
7
|
+
|
7
8
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
8
9
|
add_file("google/cloud/bigquery/storage/v1/table.proto", :syntax => :proto3) do
|
9
10
|
add_message "google.cloud.bigquery.storage.v1.TableSchema" do
|
@@ -42,7 +42,8 @@ module Google
|
|
42
42
|
# IPC-serialized Arrow RecordBatch.
|
43
43
|
# @!attribute [rw] row_count
|
44
44
|
# @return [::Integer]
|
45
|
-
# The count of rows in `serialized_record_batch`.
|
45
|
+
# [Deprecated] The count of rows in `serialized_record_batch`.
|
46
|
+
# Please use the format-independent ReadRowsResponse.row_count instead.
|
46
47
|
class ArrowRecordBatch
|
47
48
|
include ::Google::Protobuf::MessageExts
|
48
49
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -38,7 +38,8 @@ module Google
|
|
38
38
|
# Binary serialized rows in a block.
|
39
39
|
# @!attribute [rw] row_count
|
40
40
|
# @return [::Integer]
|
41
|
-
# The count of rows in the returning block.
|
41
|
+
# [Deprecated] The count of rows in the returning block.
|
42
|
+
# Please use the format-independent ReadRowsResponse.row_count instead.
|
42
43
|
class AvroRows
|
43
44
|
include ::Google::Protobuf::MessageExts
|
44
45
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -35,11 +35,13 @@ module Google
|
|
35
35
|
# Max initial number of streams. If unset or zero, the server will
|
36
36
|
# provide a value of streams so as to produce reasonable throughput. Must be
|
37
37
|
# non-negative. The number of streams may be lower than the requested number,
|
38
|
-
# depending on the amount parallelism that is reasonable for the table.
|
39
|
-
#
|
40
|
-
# max limit of 1,000.
|
38
|
+
# depending on the amount parallelism that is reasonable for the table.
|
39
|
+
# There is a default system max limit of 1,000.
|
41
40
|
#
|
42
|
-
#
|
41
|
+
# This must be greater than or equal to preferred_min_stream_count.
|
42
|
+
# Typically, clients should either leave this unset to let the system to
|
43
|
+
# determine an upper bound OR set this a size for the maximum "units of work"
|
44
|
+
# it can gracefully handle.
|
43
45
|
class CreateReadSessionRequest
|
44
46
|
include ::Google::Protobuf::MessageExts
|
45
47
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -262,6 +264,11 @@ module Google
|
|
262
264
|
# If backend detects a schema update, pass it to user so that user can
|
263
265
|
# use it to input new type of message. It will be empty when no schema
|
264
266
|
# updates have occurred.
|
267
|
+
# @!attribute [rw] row_errors
|
268
|
+
# @return [::Array<::Google::Cloud::Bigquery::Storage::V1::RowError>]
|
269
|
+
# If a request failed due to corrupted rows, no rows in the batch will be
|
270
|
+
# appended. The API will return row level error info, so that the caller can
|
271
|
+
# remove the bad rows and retry the request.
|
265
272
|
class AppendRowsResponse
|
266
273
|
include ::Google::Protobuf::MessageExts
|
267
274
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -413,6 +420,30 @@ module Google
|
|
413
420
|
OFFSET_OUT_OF_RANGE = 9
|
414
421
|
end
|
415
422
|
end
|
423
|
+
|
424
|
+
# The message that presents row level error info in a request.
|
425
|
+
# @!attribute [rw] index
|
426
|
+
# @return [::Integer]
|
427
|
+
# Index of the malformed row in the request.
|
428
|
+
# @!attribute [rw] code
|
429
|
+
# @return [::Google::Cloud::Bigquery::Storage::V1::RowError::RowErrorCode]
|
430
|
+
# Structured error reason for a row error.
|
431
|
+
# @!attribute [rw] message
|
432
|
+
# @return [::String]
|
433
|
+
# Description of the issue encountered when processing the row.
|
434
|
+
class RowError
|
435
|
+
include ::Google::Protobuf::MessageExts
|
436
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
437
|
+
|
438
|
+
# Error code for `RowError`.
|
439
|
+
module RowErrorCode
|
440
|
+
# Default error.
|
441
|
+
ROW_ERROR_CODE_UNSPECIFIED = 0
|
442
|
+
|
443
|
+
# One or more fields in the row has errors.
|
444
|
+
FIELDS_ERROR = 1
|
445
|
+
end
|
446
|
+
end
|
416
447
|
end
|
417
448
|
end
|
418
449
|
end
|
@@ -34,7 +34,7 @@ module Google
|
|
34
34
|
# automatically assigned and currently cannot be specified or updated.
|
35
35
|
# @!attribute [rw] data_format
|
36
36
|
# @return [::Google::Cloud::Bigquery::Storage::V1::DataFormat]
|
37
|
-
# Immutable. Data format of the output data.
|
37
|
+
# Immutable. Data format of the output data. DATA_FORMAT_UNSPECIFIED not supported.
|
38
38
|
# @!attribute [r] avro_schema
|
39
39
|
# @return [::Google::Cloud::Bigquery::Storage::V1::AvroSchema]
|
40
40
|
# Output only. Avro schema.
|
@@ -185,6 +185,7 @@ module Google
|
|
185
185
|
|
186
186
|
# Data format for input or output data.
|
187
187
|
module DataFormat
|
188
|
+
# Data format is unspecified.
|
188
189
|
DATA_FORMAT_UNSPECIFIED = 0
|
189
190
|
|
190
191
|
# Avro is a standard open source row based file format.
|
@@ -22,7 +22,9 @@ module Google
|
|
22
22
|
module Bigquery
|
23
23
|
module Storage
|
24
24
|
module V1
|
25
|
-
# Schema of a table.
|
25
|
+
# Schema of a table. This schema is a subset of
|
26
|
+
# google.cloud.bigquery.v2.TableSchema containing information necessary to
|
27
|
+
# generate valid message to write to BigQuery.
|
26
28
|
# @!attribute [rw] fields
|
27
29
|
# @return [::Array<::Google::Cloud::Bigquery::Storage::V1::TableFieldSchema>]
|
28
30
|
# Describes the fields in a table.
|
@@ -44,7 +44,7 @@ module Google
|
|
44
44
|
# foo = any.unpack(Foo.class);
|
45
45
|
# }
|
46
46
|
#
|
47
|
-
#
|
47
|
+
# Example 3: Pack and unpack a message in Python.
|
48
48
|
#
|
49
49
|
# foo = Foo(...)
|
50
50
|
# any = Any()
|
@@ -54,7 +54,7 @@ module Google
|
|
54
54
|
# any.Unpack(foo)
|
55
55
|
# ...
|
56
56
|
#
|
57
|
-
#
|
57
|
+
# Example 4: Pack and unpack a message in Go
|
58
58
|
#
|
59
59
|
# foo := &pb.Foo{...}
|
60
60
|
# any, err := anypb.New(foo)
|
@@ -75,7 +75,7 @@ module Google
|
|
75
75
|
#
|
76
76
|
#
|
77
77
|
# JSON
|
78
|
-
#
|
78
|
+
#
|
79
79
|
# The JSON representation of an `Any` value uses the regular
|
80
80
|
# representation of the deserialized, embedded message, with an
|
81
81
|
# additional field `@type` which contains the type URL. Example:
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-cloud-bigquery-storage-v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.11.1
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2022-
|
11
|
+
date: 2022-06-28 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: gapic-common
|
@@ -219,7 +219,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
219
219
|
- !ruby/object:Gem::Version
|
220
220
|
version: '0'
|
221
221
|
requirements: []
|
222
|
-
rubygems_version: 3.3.
|
222
|
+
rubygems_version: 3.3.14
|
223
223
|
signing_key:
|
224
224
|
specification_version: 4
|
225
225
|
summary: API Client library for the BigQuery Storage V1 API
|