google-cloud-bigquery-storage-v1 0.10.1 → 0.12.0

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: dff85b10ef593a6d705ab82f98458dbf22c2bce8b84bb1612d17da8f867ea767
4
- data.tar.gz: 0e19138911fed5e2c44118090a1fecbab1b9586bed1569ec35a636e8c51f395c
3
+ metadata.gz: 97e012ed12b5d17636615b23153be73fd4006ccc47ba0955a722139b09c2f53d
4
+ data.tar.gz: c2dd5d3b6a99d21fe96e99c2348e578307fbe67c5cd21c829385fdfa89e372df
5
5
  SHA512:
6
- metadata.gz: 0f754922de4957aebbb9f24d63743f525de91878889535f79e6821188924d3139400d4f5038ef215efe53e18602eacf0474faf922f93f58efdc2f45ef351b3b9
7
- data.tar.gz: a31410349d6e7457a2fb7ad7666400faa09027d2c5d4a9728060754b921bb7512536fe7e020e7cebb7052e9af04105ad894d93ace5930f63a5a11f715147f6f5
6
+ metadata.gz: 40865f1fe1a363798ce1a66aa55f10b450ce09e9aaa6768b0d7c5723901face80a5cec1a4fd09e742e2b8a51f2638fd94bcdb25f650e65758f46612e14ea9fcc
7
+ data.tar.gz: 9d151f5e6c122112eafca91af8c33d21c01ce03d10895fddca917318300df57c00f2a9c86f9260a10ad5fb37ae02542c3e48d5e73681f1435b59b97b403c247e
data/README.md CHANGED
@@ -69,16 +69,21 @@ module GRPC
69
69
  end
70
70
  ```
71
71
 
72
+
73
+ ## Google Cloud Samples
74
+
75
+ To browse ready to use code samples check [Google Cloud Samples](https://cloud.google.com/docs/samples).
76
+
72
77
  ## Supported Ruby Versions
73
78
 
74
- This library is supported on Ruby 2.5+.
79
+ This library is supported on Ruby 2.6+.
75
80
 
76
81
  Google provides official support for Ruby versions that are actively supported
77
82
  by Ruby Core—that is, Ruby versions that are either in normal maintenance or
78
- in security maintenance, and not end of life. Currently, this means Ruby 2.5
79
- and later. Older versions of Ruby _may_ still work, but are unsupported and not
80
- recommended. See https://www.ruby-lang.org/en/downloads/branches/ for details
81
- about the Ruby support schedule.
83
+ in security maintenance, and not end of life. Older versions of Ruby _may_
84
+ still work, but are unsupported and not recommended. See
85
+ https://www.ruby-lang.org/en/downloads/branches/ for details about the Ruby
86
+ support schedule.
82
87
 
83
88
  ## Which client should I use?
84
89
 
@@ -207,11 +207,13 @@ module Google
207
207
  # Max initial number of streams. If unset or zero, the server will
208
208
  # provide a value of streams so as to produce reasonable throughput. Must be
209
209
  # non-negative. The number of streams may be lower than the requested number,
210
- # depending on the amount parallelism that is reasonable for the table. Error
211
- # will be returned if the max count is greater than the current system
212
- # max limit of 1,000.
210
+ # depending on the amount parallelism that is reasonable for the table.
211
+ # There is a default system max limit of 1,000.
213
212
  #
214
- # Streams must be read starting from offset 0.
213
+ # This must be greater than or equal to preferred_min_stream_count.
214
+ # Typically, clients should either leave this unset to let the system to
215
+ # determine an upper bound OR set this a size for the maximum "units of work"
216
+ # it can gracefully handle.
215
217
  #
216
218
  # @yield [response, operation] Access the result along with the RPC operation
217
219
  # @yieldparam response [::Google::Cloud::Bigquery::Storage::V1::ReadSession]
@@ -70,9 +70,9 @@ module Google
70
70
  end
71
71
  default_config = Client::Configuration.new parent_config
72
72
 
73
- default_config.rpcs.create_write_stream.timeout = 600.0
73
+ default_config.rpcs.create_write_stream.timeout = 1200.0
74
74
  default_config.rpcs.create_write_stream.retry_policy = {
75
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14]
75
+ initial_delay: 10.0, max_delay: 120.0, multiplier: 1.3, retry_codes: [4, 14, 8]
76
76
  }
77
77
 
78
78
  default_config.rpcs.append_rows.timeout = 86_400.0
@@ -1,9 +1,10 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/storage/v1/protobuf.proto
3
3
 
4
- require 'google/protobuf/descriptor_pb'
5
4
  require 'google/protobuf'
6
5
 
6
+ require 'google/protobuf/descriptor_pb'
7
+
7
8
  Google::Protobuf::DescriptorPool.generated_pool.build do
8
9
  add_file("google/cloud/bigquery/storage/v1/protobuf.proto", :syntax => :proto3) do
9
10
  add_message "google.cloud.bigquery.storage.v1.ProtoSchema" do
@@ -1,6 +1,8 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/storage/v1/storage.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/annotations_pb'
5
7
  require 'google/api/client_pb'
6
8
  require 'google/api/field_behavior_pb'
@@ -13,7 +15,6 @@ require 'google/cloud/bigquery/storage/v1/table_pb'
13
15
  require 'google/protobuf/timestamp_pb'
14
16
  require 'google/protobuf/wrappers_pb'
15
17
  require 'google/rpc/status_pb'
16
- require 'google/protobuf'
17
18
 
18
19
  Google::Protobuf::DescriptorPool.generated_pool.build do
19
20
  add_file("google/cloud/bigquery/storage/v1/storage.proto", :syntax => :proto3) do
@@ -75,6 +76,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
75
76
  end
76
77
  add_message "google.cloud.bigquery.storage.v1.AppendRowsResponse" do
77
78
  optional :updated_schema, :message, 3, "google.cloud.bigquery.storage.v1.TableSchema"
79
+ repeated :row_errors, :message, 4, "google.cloud.bigquery.storage.v1.RowError"
78
80
  oneof :response do
79
81
  optional :append_result, :message, 1, "google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult"
80
82
  optional :error, :message, 2, "google.rpc.Status"
@@ -124,6 +126,15 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
124
126
  value :OFFSET_ALREADY_EXISTS, 8
125
127
  value :OFFSET_OUT_OF_RANGE, 9
126
128
  end
129
+ add_message "google.cloud.bigquery.storage.v1.RowError" do
130
+ optional :index, :int64, 1
131
+ optional :code, :enum, 2, "google.cloud.bigquery.storage.v1.RowError.RowErrorCode"
132
+ optional :message, :string, 3
133
+ end
134
+ add_enum "google.cloud.bigquery.storage.v1.RowError.RowErrorCode" do
135
+ value :ROW_ERROR_CODE_UNSPECIFIED, 0
136
+ value :FIELDS_ERROR, 1
137
+ end
127
138
  end
128
139
  end
129
140
 
@@ -154,6 +165,8 @@ module Google
154
165
  FlushRowsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FlushRowsResponse").msgclass
155
166
  StorageError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError").msgclass
156
167
  StorageError::StorageErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError.StorageErrorCode").enummodule
168
+ RowError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.RowError").msgclass
169
+ RowError::RowErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.RowError.RowErrorCode").enummodule
157
170
  end
158
171
  end
159
172
  end
@@ -1,13 +1,14 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/storage/v1/stream.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/field_behavior_pb'
5
7
  require 'google/api/resource_pb'
6
8
  require 'google/cloud/bigquery/storage/v1/arrow_pb'
7
9
  require 'google/cloud/bigquery/storage/v1/avro_pb'
8
10
  require 'google/cloud/bigquery/storage/v1/table_pb'
9
11
  require 'google/protobuf/timestamp_pb'
10
- require 'google/protobuf'
11
12
 
12
13
  Google::Protobuf::DescriptorPool.generated_pool.build do
13
14
  add_file("google/cloud/bigquery/storage/v1/stream.proto", :syntax => :proto3) do
@@ -1,9 +1,10 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/storage/v1/table.proto
3
3
 
4
- require 'google/api/field_behavior_pb'
5
4
  require 'google/protobuf'
6
5
 
6
+ require 'google/api/field_behavior_pb'
7
+
7
8
  Google::Protobuf::DescriptorPool.generated_pool.build do
8
9
  add_file("google/cloud/bigquery/storage/v1/table.proto", :syntax => :proto3) do
9
10
  add_message "google.cloud.bigquery.storage.v1.TableSchema" do
@@ -22,7 +22,7 @@ module Google
22
22
  module Bigquery
23
23
  module Storage
24
24
  module V1
25
- VERSION = "0.10.1"
25
+ VERSION = "0.12.0"
26
26
  end
27
27
  end
28
28
  end
@@ -35,11 +35,13 @@ module Google
35
35
  # Max initial number of streams. If unset or zero, the server will
36
36
  # provide a value of streams so as to produce reasonable throughput. Must be
37
37
  # non-negative. The number of streams may be lower than the requested number,
38
- # depending on the amount parallelism that is reasonable for the table. Error
39
- # will be returned if the max count is greater than the current system
40
- # max limit of 1,000.
38
+ # depending on the amount parallelism that is reasonable for the table.
39
+ # There is a default system max limit of 1,000.
41
40
  #
42
- # Streams must be read starting from offset 0.
41
+ # This must be greater than or equal to preferred_min_stream_count.
42
+ # Typically, clients should either leave this unset to let the system to
43
+ # determine an upper bound OR set this a size for the maximum "units of work"
44
+ # it can gracefully handle.
43
45
  class CreateReadSessionRequest
44
46
  include ::Google::Protobuf::MessageExts
45
47
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -262,6 +264,11 @@ module Google
262
264
  # If backend detects a schema update, pass it to user so that user can
263
265
  # use it to input new type of message. It will be empty when no schema
264
266
  # updates have occurred.
267
+ # @!attribute [rw] row_errors
268
+ # @return [::Array<::Google::Cloud::Bigquery::Storage::V1::RowError>]
269
+ # If a request failed due to corrupted rows, no rows in the batch will be
270
+ # appended. The API will return row level error info, so that the caller can
271
+ # remove the bad rows and retry the request.
265
272
  class AppendRowsResponse
266
273
  include ::Google::Protobuf::MessageExts
267
274
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -413,6 +420,30 @@ module Google
413
420
  OFFSET_OUT_OF_RANGE = 9
414
421
  end
415
422
  end
423
+
424
+ # The message that presents row level error info in a request.
425
+ # @!attribute [rw] index
426
+ # @return [::Integer]
427
+ # Index of the malformed row in the request.
428
+ # @!attribute [rw] code
429
+ # @return [::Google::Cloud::Bigquery::Storage::V1::RowError::RowErrorCode]
430
+ # Structured error reason for a row error.
431
+ # @!attribute [rw] message
432
+ # @return [::String]
433
+ # Description of the issue encountered when processing the row.
434
+ class RowError
435
+ include ::Google::Protobuf::MessageExts
436
+ extend ::Google::Protobuf::MessageExts::ClassMethods
437
+
438
+ # Error code for `RowError`.
439
+ module RowErrorCode
440
+ # Default error.
441
+ ROW_ERROR_CODE_UNSPECIFIED = 0
442
+
443
+ # One or more fields in the row has errors.
444
+ FIELDS_ERROR = 1
445
+ end
446
+ end
416
447
  end
417
448
  end
418
449
  end
@@ -34,7 +34,7 @@ module Google
34
34
  # automatically assigned and currently cannot be specified or updated.
35
35
  # @!attribute [rw] data_format
36
36
  # @return [::Google::Cloud::Bigquery::Storage::V1::DataFormat]
37
- # Immutable. Data format of the output data.
37
+ # Immutable. Data format of the output data. DATA_FORMAT_UNSPECIFIED not supported.
38
38
  # @!attribute [r] avro_schema
39
39
  # @return [::Google::Cloud::Bigquery::Storage::V1::AvroSchema]
40
40
  # Output only. Avro schema.
@@ -185,6 +185,7 @@ module Google
185
185
 
186
186
  # Data format for input or output data.
187
187
  module DataFormat
188
+ # Data format is unspecified.
188
189
  DATA_FORMAT_UNSPECIFIED = 0
189
190
 
190
191
  # Avro is a standard open source row based file format.
@@ -22,7 +22,9 @@ module Google
22
22
  module Bigquery
23
23
  module Storage
24
24
  module V1
25
- # Schema of a table.
25
+ # Schema of a table. This schema is a subset of
26
+ # google.cloud.bigquery.v2.TableSchema containing information necessary to
27
+ # generate valid message to write to BigQuery.
26
28
  # @!attribute [rw] fields
27
29
  # @return [::Array<::Google::Cloud::Bigquery::Storage::V1::TableFieldSchema>]
28
30
  # Describes the fields in a table.
@@ -44,7 +44,7 @@ module Google
44
44
  # foo = any.unpack(Foo.class);
45
45
  # }
46
46
  #
47
- # Example 3: Pack and unpack a message in Python.
47
+ # Example 3: Pack and unpack a message in Python.
48
48
  #
49
49
  # foo = Foo(...)
50
50
  # any = Any()
@@ -54,7 +54,7 @@ module Google
54
54
  # any.Unpack(foo)
55
55
  # ...
56
56
  #
57
- # Example 4: Pack and unpack a message in Go
57
+ # Example 4: Pack and unpack a message in Go
58
58
  #
59
59
  # foo := &pb.Foo{...}
60
60
  # any, err := anypb.New(foo)
@@ -75,7 +75,7 @@ module Google
75
75
  #
76
76
  #
77
77
  # JSON
78
- # ====
78
+ #
79
79
  # The JSON representation of an `Any` value uses the regular
80
80
  # representation of the deserialized, embedded message, with an
81
81
  # additional field `@type` which contains the type URL. Example:
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-bigquery-storage-v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.10.1
4
+ version: 0.12.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2022-04-01 00:00:00.000000000 Z
11
+ date: 2022-07-01 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: gapic-common
@@ -16,7 +16,7 @@ dependencies:
16
16
  requirements:
17
17
  - - ">="
18
18
  - !ruby/object:Gem::Version
19
- version: '0.7'
19
+ version: '0.10'
20
20
  - - "<"
21
21
  - !ruby/object:Gem::Version
22
22
  version: 2.a
@@ -26,7 +26,7 @@ dependencies:
26
26
  requirements:
27
27
  - - ">="
28
28
  - !ruby/object:Gem::Version
29
- version: '0.7'
29
+ version: '0.10'
30
30
  - - "<"
31
31
  - !ruby/object:Gem::Version
32
32
  version: 2.a
@@ -50,28 +50,28 @@ dependencies:
50
50
  requirements:
51
51
  - - "~>"
52
52
  - !ruby/object:Gem::Version
53
- version: 1.25.1
53
+ version: 1.26.1
54
54
  type: :development
55
55
  prerelease: false
56
56
  version_requirements: !ruby/object:Gem::Requirement
57
57
  requirements:
58
58
  - - "~>"
59
59
  - !ruby/object:Gem::Version
60
- version: 1.25.1
60
+ version: 1.26.1
61
61
  - !ruby/object:Gem::Dependency
62
62
  name: minitest
63
63
  requirement: !ruby/object:Gem::Requirement
64
64
  requirements:
65
65
  - - "~>"
66
66
  - !ruby/object:Gem::Version
67
- version: '5.14'
67
+ version: '5.16'
68
68
  type: :development
69
69
  prerelease: false
70
70
  version_requirements: !ruby/object:Gem::Requirement
71
71
  requirements:
72
72
  - - "~>"
73
73
  - !ruby/object:Gem::Version
74
- version: '5.14'
74
+ version: '5.16'
75
75
  - !ruby/object:Gem::Dependency
76
76
  name: minitest-focus
77
77
  requirement: !ruby/object:Gem::Requirement
@@ -106,14 +106,14 @@ dependencies:
106
106
  requirements:
107
107
  - - ">="
108
108
  - !ruby/object:Gem::Version
109
- version: '12.0'
109
+ version: '13.0'
110
110
  type: :development
111
111
  prerelease: false
112
112
  version_requirements: !ruby/object:Gem::Requirement
113
113
  requirements:
114
114
  - - ">="
115
115
  - !ruby/object:Gem::Version
116
- version: '12.0'
116
+ version: '13.0'
117
117
  - !ruby/object:Gem::Dependency
118
118
  name: redcarpet
119
119
  requirement: !ruby/object:Gem::Requirement
@@ -212,14 +212,14 @@ required_ruby_version: !ruby/object:Gem::Requirement
212
212
  requirements:
213
213
  - - ">="
214
214
  - !ruby/object:Gem::Version
215
- version: '2.5'
215
+ version: '2.6'
216
216
  required_rubygems_version: !ruby/object:Gem::Requirement
217
217
  requirements:
218
218
  - - ">="
219
219
  - !ruby/object:Gem::Version
220
220
  version: '0'
221
221
  requirements: []
222
- rubygems_version: 3.3.5
222
+ rubygems_version: 3.3.14
223
223
  signing_key:
224
224
  specification_version: 4
225
225
  summary: API Client library for the BigQuery Storage V1 API