google-cloud-bigquery-storage-v1 0.9.2 → 0.11.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/README.md +5 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/client.rb +6 -4
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/client.rb +7 -0
- data/lib/google/cloud/bigquery/storage/v1/protobuf_pb.rb +2 -1
- data/lib/google/cloud/bigquery/storage/v1/storage_pb.rb +16 -1
- data/lib/google/cloud/bigquery/storage/v1/storage_services_pb.rb +7 -0
- data/lib/google/cloud/bigquery/storage/v1/stream_pb.rb +3 -1
- data/lib/google/cloud/bigquery/storage/v1/table_pb.rb +2 -1
- data/lib/google/cloud/bigquery/storage/v1/version.rb +1 -1
- data/lib/google/cloud/bigquery/storage/v1.rb +2 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/arrow.rb +2 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/avro.rb +2 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/storage.rb +45 -6
- data/proto_docs/google/cloud/bigquery/storage/v1/stream.rb +10 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/table.rb +3 -1
- data/proto_docs/google/protobuf/any.rb +3 -3
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 9041851c373718175111413b2e1125660d22bcc2ca78bb0d46bf97ad75866661
|
4
|
+
data.tar.gz: 771714ee33b98b3750ff14a38295b1156c59011ce18fc46664aa54152e798a64
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 89db88a25bb273beee8ba6dcedae7231c2ee4906a240c1ec2505fbd81e6196bb737879f37150a235cd7e378ef07b4b2a759b317af9a8e365d776d27025a8cfbb
|
7
|
+
data.tar.gz: 5c3f7a8acc54d45b82d2d5786dc08431d7b090f41ed7681f795899d3a747e3a16332aec2f8fcc2090fa1c060cdbbe003fc5c8e78246dd38c18b452c4358ec8b6
|
data/README.md
CHANGED
@@ -207,11 +207,13 @@ module Google
|
|
207
207
|
# Max initial number of streams. If unset or zero, the server will
|
208
208
|
# provide a value of streams so as to produce reasonable throughput. Must be
|
209
209
|
# non-negative. The number of streams may be lower than the requested number,
|
210
|
-
# depending on the amount parallelism that is reasonable for the table.
|
211
|
-
#
|
212
|
-
# max limit of 1,000.
|
210
|
+
# depending on the amount parallelism that is reasonable for the table.
|
211
|
+
# There is a default system max limit of 1,000.
|
213
212
|
#
|
214
|
-
#
|
213
|
+
# This must be greater than or equal to preferred_min_stream_count.
|
214
|
+
# Typically, clients should either leave this unset to let the system to
|
215
|
+
# determine an upper bound OR set this a size for the maximum "units of work"
|
216
|
+
# it can gracefully handle.
|
215
217
|
#
|
216
218
|
# @yield [response, operation] Access the result along with the RPC operation
|
217
219
|
# @yieldparam response [::Google::Cloud::Bigquery::Storage::V1::ReadSession]
|
@@ -306,6 +306,13 @@ module Google
|
|
306
306
|
# finalized (via the `FinalizeWriteStream` rpc), and the stream is explicitly
|
307
307
|
# committed via the `BatchCommitWriteStreams` rpc.
|
308
308
|
#
|
309
|
+
# Note: For users coding against the gRPC api directly, it may be
|
310
|
+
# necessary to supply the x-goog-request-params system parameter
|
311
|
+
# with `write_stream=<full_write_stream_name>`.
|
312
|
+
#
|
313
|
+
# More information about system parameters:
|
314
|
+
# https://cloud.google.com/apis/docs/system-parameters
|
315
|
+
#
|
309
316
|
# @param request [::Gapic::StreamInput, ::Enumerable<::Google::Cloud::Bigquery::Storage::V1::AppendRowsRequest, ::Hash>]
|
310
317
|
# An enumerable of {::Google::Cloud::Bigquery::Storage::V1::AppendRowsRequest} instances.
|
311
318
|
# @param options [::Gapic::CallOptions, ::Hash]
|
@@ -1,9 +1,10 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/protobuf.proto
|
3
3
|
|
4
|
-
require 'google/protobuf/descriptor_pb'
|
5
4
|
require 'google/protobuf'
|
6
5
|
|
6
|
+
require 'google/protobuf/descriptor_pb'
|
7
|
+
|
7
8
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
8
9
|
add_file("google/cloud/bigquery/storage/v1/protobuf.proto", :syntax => :proto3) do
|
9
10
|
add_message "google.cloud.bigquery.storage.v1.ProtoSchema" do
|
@@ -1,6 +1,8 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/storage.proto
|
3
3
|
|
4
|
+
require 'google/protobuf'
|
5
|
+
|
4
6
|
require 'google/api/annotations_pb'
|
5
7
|
require 'google/api/client_pb'
|
6
8
|
require 'google/api/field_behavior_pb'
|
@@ -13,7 +15,6 @@ require 'google/cloud/bigquery/storage/v1/table_pb'
|
|
13
15
|
require 'google/protobuf/timestamp_pb'
|
14
16
|
require 'google/protobuf/wrappers_pb'
|
15
17
|
require 'google/rpc/status_pb'
|
16
|
-
require 'google/protobuf'
|
17
18
|
|
18
19
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
19
20
|
add_file("google/cloud/bigquery/storage/v1/storage.proto", :syntax => :proto3) do
|
@@ -75,6 +76,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
75
76
|
end
|
76
77
|
add_message "google.cloud.bigquery.storage.v1.AppendRowsResponse" do
|
77
78
|
optional :updated_schema, :message, 3, "google.cloud.bigquery.storage.v1.TableSchema"
|
79
|
+
repeated :row_errors, :message, 4, "google.cloud.bigquery.storage.v1.RowError"
|
78
80
|
oneof :response do
|
79
81
|
optional :append_result, :message, 1, "google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult"
|
80
82
|
optional :error, :message, 2, "google.rpc.Status"
|
@@ -121,6 +123,17 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
121
123
|
value :INVALID_STREAM_STATE, 5
|
122
124
|
value :STREAM_FINALIZED, 6
|
123
125
|
value :SCHEMA_MISMATCH_EXTRA_FIELDS, 7
|
126
|
+
value :OFFSET_ALREADY_EXISTS, 8
|
127
|
+
value :OFFSET_OUT_OF_RANGE, 9
|
128
|
+
end
|
129
|
+
add_message "google.cloud.bigquery.storage.v1.RowError" do
|
130
|
+
optional :index, :int64, 1
|
131
|
+
optional :code, :enum, 2, "google.cloud.bigquery.storage.v1.RowError.RowErrorCode"
|
132
|
+
optional :message, :string, 3
|
133
|
+
end
|
134
|
+
add_enum "google.cloud.bigquery.storage.v1.RowError.RowErrorCode" do
|
135
|
+
value :ROW_ERROR_CODE_UNSPECIFIED, 0
|
136
|
+
value :FIELDS_ERROR, 1
|
124
137
|
end
|
125
138
|
end
|
126
139
|
end
|
@@ -152,6 +165,8 @@ module Google
|
|
152
165
|
FlushRowsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FlushRowsResponse").msgclass
|
153
166
|
StorageError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError").msgclass
|
154
167
|
StorageError::StorageErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError.StorageErrorCode").enummodule
|
168
|
+
RowError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.RowError").msgclass
|
169
|
+
RowError::RowErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.RowError.RowErrorCode").enummodule
|
155
170
|
end
|
156
171
|
end
|
157
172
|
end
|
@@ -134,6 +134,13 @@ module Google
|
|
134
134
|
# * For PENDING streams, data is not made visible until the stream itself is
|
135
135
|
# finalized (via the `FinalizeWriteStream` rpc), and the stream is explicitly
|
136
136
|
# committed via the `BatchCommitWriteStreams` rpc.
|
137
|
+
#
|
138
|
+
# Note: For users coding against the gRPC api directly, it may be
|
139
|
+
# necessary to supply the x-goog-request-params system parameter
|
140
|
+
# with `write_stream=<full_write_stream_name>`.
|
141
|
+
#
|
142
|
+
# More information about system parameters:
|
143
|
+
# https://cloud.google.com/apis/docs/system-parameters
|
137
144
|
rpc :AppendRows, stream(::Google::Cloud::Bigquery::Storage::V1::AppendRowsRequest), stream(::Google::Cloud::Bigquery::Storage::V1::AppendRowsResponse)
|
138
145
|
# Gets information about a write stream.
|
139
146
|
rpc :GetWriteStream, ::Google::Cloud::Bigquery::Storage::V1::GetWriteStreamRequest, ::Google::Cloud::Bigquery::Storage::V1::WriteStream
|
@@ -1,13 +1,14 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/stream.proto
|
3
3
|
|
4
|
+
require 'google/protobuf'
|
5
|
+
|
4
6
|
require 'google/api/field_behavior_pb'
|
5
7
|
require 'google/api/resource_pb'
|
6
8
|
require 'google/cloud/bigquery/storage/v1/arrow_pb'
|
7
9
|
require 'google/cloud/bigquery/storage/v1/avro_pb'
|
8
10
|
require 'google/cloud/bigquery/storage/v1/table_pb'
|
9
11
|
require 'google/protobuf/timestamp_pb'
|
10
|
-
require 'google/protobuf'
|
11
12
|
|
12
13
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
13
14
|
add_file("google/cloud/bigquery/storage/v1/stream.proto", :syntax => :proto3) do
|
@@ -20,6 +21,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
20
21
|
optional :read_options, :message, 8, "google.cloud.bigquery.storage.v1.ReadSession.TableReadOptions"
|
21
22
|
repeated :streams, :message, 10, "google.cloud.bigquery.storage.v1.ReadStream"
|
22
23
|
optional :estimated_total_bytes_scanned, :int64, 12
|
24
|
+
optional :trace_id, :string, 13
|
23
25
|
oneof :schema do
|
24
26
|
optional :avro_schema, :message, 4, "google.cloud.bigquery.storage.v1.AvroSchema"
|
25
27
|
optional :arrow_schema, :message, 5, "google.cloud.bigquery.storage.v1.ArrowSchema"
|
@@ -1,9 +1,10 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/table.proto
|
3
3
|
|
4
|
-
require 'google/api/field_behavior_pb'
|
5
4
|
require 'google/protobuf'
|
6
5
|
|
6
|
+
require 'google/api/field_behavior_pb'
|
7
|
+
|
7
8
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
8
9
|
add_file("google/cloud/bigquery/storage/v1/table.proto", :syntax => :proto3) do
|
9
10
|
add_message "google.cloud.bigquery.storage.v1.TableSchema" do
|
@@ -42,7 +42,8 @@ module Google
|
|
42
42
|
# IPC-serialized Arrow RecordBatch.
|
43
43
|
# @!attribute [rw] row_count
|
44
44
|
# @return [::Integer]
|
45
|
-
# The count of rows in `serialized_record_batch`.
|
45
|
+
# [Deprecated] The count of rows in `serialized_record_batch`.
|
46
|
+
# Please use the format-independent ReadRowsResponse.row_count instead.
|
46
47
|
class ArrowRecordBatch
|
47
48
|
include ::Google::Protobuf::MessageExts
|
48
49
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -38,7 +38,8 @@ module Google
|
|
38
38
|
# Binary serialized rows in a block.
|
39
39
|
# @!attribute [rw] row_count
|
40
40
|
# @return [::Integer]
|
41
|
-
# The count of rows in the returning block.
|
41
|
+
# [Deprecated] The count of rows in the returning block.
|
42
|
+
# Please use the format-independent ReadRowsResponse.row_count instead.
|
42
43
|
class AvroRows
|
43
44
|
include ::Google::Protobuf::MessageExts
|
44
45
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -35,11 +35,13 @@ module Google
|
|
35
35
|
# Max initial number of streams. If unset or zero, the server will
|
36
36
|
# provide a value of streams so as to produce reasonable throughput. Must be
|
37
37
|
# non-negative. The number of streams may be lower than the requested number,
|
38
|
-
# depending on the amount parallelism that is reasonable for the table.
|
39
|
-
#
|
40
|
-
# max limit of 1,000.
|
38
|
+
# depending on the amount parallelism that is reasonable for the table.
|
39
|
+
# There is a default system max limit of 1,000.
|
41
40
|
#
|
42
|
-
#
|
41
|
+
# This must be greater than or equal to preferred_min_stream_count.
|
42
|
+
# Typically, clients should either leave this unset to let the system to
|
43
|
+
# determine an upper bound OR set this a size for the maximum "units of work"
|
44
|
+
# it can gracefully handle.
|
43
45
|
class CreateReadSessionRequest
|
44
46
|
include ::Google::Protobuf::MessageExts
|
45
47
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -189,10 +191,12 @@ module Google
|
|
189
191
|
# request.
|
190
192
|
#
|
191
193
|
# For explicitly created write streams, the format is:
|
192
|
-
#
|
194
|
+
#
|
195
|
+
# * `projects/{project}/datasets/{dataset}/tables/{table}/streams/{id}`
|
193
196
|
#
|
194
197
|
# For the special default stream, the format is:
|
195
|
-
#
|
198
|
+
#
|
199
|
+
# * `projects/{project}/datasets/{dataset}/tables/{table}/streams/_default`.
|
196
200
|
# @!attribute [rw] offset
|
197
201
|
# @return [::Google::Protobuf::Int64Value]
|
198
202
|
# If present, the write is only performed if the next append offset is same
|
@@ -260,6 +264,11 @@ module Google
|
|
260
264
|
# If backend detects a schema update, pass it to user so that user can
|
261
265
|
# use it to input new type of message. It will be empty when no schema
|
262
266
|
# updates have occurred.
|
267
|
+
# @!attribute [rw] row_errors
|
268
|
+
# @return [::Array<::Google::Cloud::Bigquery::Storage::V1::RowError>]
|
269
|
+
# If a request failed due to corrupted rows, no rows in the batch will be
|
270
|
+
# appended. The API will return row level error info, so that the caller can
|
271
|
+
# remove the bad rows and retry the request.
|
263
272
|
class AppendRowsResponse
|
264
273
|
include ::Google::Protobuf::MessageExts
|
265
274
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -403,6 +412,36 @@ module Google
|
|
403
412
|
# There is a schema mismatch and it is caused by user schema has extra
|
404
413
|
# field than bigquery schema.
|
405
414
|
SCHEMA_MISMATCH_EXTRA_FIELDS = 7
|
415
|
+
|
416
|
+
# Offset already exists.
|
417
|
+
OFFSET_ALREADY_EXISTS = 8
|
418
|
+
|
419
|
+
# Offset out of range.
|
420
|
+
OFFSET_OUT_OF_RANGE = 9
|
421
|
+
end
|
422
|
+
end
|
423
|
+
|
424
|
+
# The message that presents row level error info in a request.
|
425
|
+
# @!attribute [rw] index
|
426
|
+
# @return [::Integer]
|
427
|
+
# Index of the malformed row in the request.
|
428
|
+
# @!attribute [rw] code
|
429
|
+
# @return [::Google::Cloud::Bigquery::Storage::V1::RowError::RowErrorCode]
|
430
|
+
# Structured error reason for a row error.
|
431
|
+
# @!attribute [rw] message
|
432
|
+
# @return [::String]
|
433
|
+
# Description of the issue encountered when processing the row.
|
434
|
+
class RowError
|
435
|
+
include ::Google::Protobuf::MessageExts
|
436
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
437
|
+
|
438
|
+
# Error code for `RowError`.
|
439
|
+
module RowErrorCode
|
440
|
+
# Default error.
|
441
|
+
ROW_ERROR_CODE_UNSPECIFIED = 0
|
442
|
+
|
443
|
+
# One or more fields in the row has errors.
|
444
|
+
FIELDS_ERROR = 1
|
406
445
|
end
|
407
446
|
end
|
408
447
|
end
|
@@ -34,7 +34,7 @@ module Google
|
|
34
34
|
# automatically assigned and currently cannot be specified or updated.
|
35
35
|
# @!attribute [rw] data_format
|
36
36
|
# @return [::Google::Cloud::Bigquery::Storage::V1::DataFormat]
|
37
|
-
# Immutable. Data format of the output data.
|
37
|
+
# Immutable. Data format of the output data. DATA_FORMAT_UNSPECIFIED not supported.
|
38
38
|
# @!attribute [r] avro_schema
|
39
39
|
# @return [::Google::Cloud::Bigquery::Storage::V1::AvroSchema]
|
40
40
|
# Output only. Avro schema.
|
@@ -64,6 +64,14 @@ module Google
|
|
64
64
|
# Output only. An estimate on the number of bytes this session will scan when
|
65
65
|
# all streams are completely consumed. This estimate is based on
|
66
66
|
# metadata from the table which might be incomplete or stale.
|
67
|
+
# @!attribute [rw] trace_id
|
68
|
+
# @return [::String]
|
69
|
+
# Optional. ID set by client to annotate a session identity. This does not need
|
70
|
+
# to be strictly unique, but instead the same ID should be used to group
|
71
|
+
# logically connected sessions (e.g. All using the same ID for all sessions
|
72
|
+
# needed to complete a Spark SQL query is reasonable).
|
73
|
+
#
|
74
|
+
# Maximum length is 256 bytes.
|
67
75
|
class ReadSession
|
68
76
|
include ::Google::Protobuf::MessageExts
|
69
77
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -177,6 +185,7 @@ module Google
|
|
177
185
|
|
178
186
|
# Data format for input or output data.
|
179
187
|
module DataFormat
|
188
|
+
# Data format is unspecified.
|
180
189
|
DATA_FORMAT_UNSPECIFIED = 0
|
181
190
|
|
182
191
|
# Avro is a standard open source row based file format.
|
@@ -22,7 +22,9 @@ module Google
|
|
22
22
|
module Bigquery
|
23
23
|
module Storage
|
24
24
|
module V1
|
25
|
-
# Schema of a table.
|
25
|
+
# Schema of a table. This schema is a subset of
|
26
|
+
# google.cloud.bigquery.v2.TableSchema containing information necessary to
|
27
|
+
# generate valid message to write to BigQuery.
|
26
28
|
# @!attribute [rw] fields
|
27
29
|
# @return [::Array<::Google::Cloud::Bigquery::Storage::V1::TableFieldSchema>]
|
28
30
|
# Describes the fields in a table.
|
@@ -44,7 +44,7 @@ module Google
|
|
44
44
|
# foo = any.unpack(Foo.class);
|
45
45
|
# }
|
46
46
|
#
|
47
|
-
#
|
47
|
+
# Example 3: Pack and unpack a message in Python.
|
48
48
|
#
|
49
49
|
# foo = Foo(...)
|
50
50
|
# any = Any()
|
@@ -54,7 +54,7 @@ module Google
|
|
54
54
|
# any.Unpack(foo)
|
55
55
|
# ...
|
56
56
|
#
|
57
|
-
#
|
57
|
+
# Example 4: Pack and unpack a message in Go
|
58
58
|
#
|
59
59
|
# foo := &pb.Foo{...}
|
60
60
|
# any, err := anypb.New(foo)
|
@@ -75,7 +75,7 @@ module Google
|
|
75
75
|
#
|
76
76
|
#
|
77
77
|
# JSON
|
78
|
-
#
|
78
|
+
#
|
79
79
|
# The JSON representation of an `Any` value uses the regular
|
80
80
|
# representation of the deserialized, embedded message, with an
|
81
81
|
# additional field `@type` which contains the type URL. Example:
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-cloud-bigquery-storage-v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.11.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2022-
|
11
|
+
date: 2022-06-22 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: gapic-common
|
@@ -219,7 +219,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
219
219
|
- !ruby/object:Gem::Version
|
220
220
|
version: '0'
|
221
221
|
requirements: []
|
222
|
-
rubygems_version: 3.3.
|
222
|
+
rubygems_version: 3.3.14
|
223
223
|
signing_key:
|
224
224
|
specification_version: 4
|
225
225
|
summary: API Client library for the BigQuery Storage V1 API
|