google-cloud-bigquery-storage-v1 0.20.0 → 0.21.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/client.rb +4 -4
- data/lib/google/cloud/bigquery/storage/v1/version.rb +1 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/storage.rb +39 -14
- data/proto_docs/google/cloud/bigquery/storage/v1/stream.rb +4 -4
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 703332e4ce288caa3087714d2396c597a59be06c8419d4d4ebb181128eb7e699
|
4
|
+
data.tar.gz: 52241ca1467778c2ebbe3bd278093a5bc0421cf540aa1f3efdc57b48b9b7073a
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: eff8e669467b02d0cc0e1bf2871728890becb60548b7ce3e8c055130a58e703c412578f5a8d1fa769f0995ed00efa4fd25a4c2e8baa99ac0fbaf77b425360436
|
7
|
+
data.tar.gz: 1acc21d67a2ee9cb1bbc481c0ed85768a5ec42eec2bb31587373089e102f88df914a5e0d7cd15b010ebc79208de2ac596e65bbd556c02a064c1b1f33469df4c9
|
@@ -82,22 +82,22 @@ module Google
|
|
82
82
|
|
83
83
|
default_config.rpcs.get_write_stream.timeout = 600.0
|
84
84
|
default_config.rpcs.get_write_stream.retry_policy = {
|
85
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14]
|
85
|
+
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14, 8]
|
86
86
|
}
|
87
87
|
|
88
88
|
default_config.rpcs.finalize_write_stream.timeout = 600.0
|
89
89
|
default_config.rpcs.finalize_write_stream.retry_policy = {
|
90
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14]
|
90
|
+
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14, 8]
|
91
91
|
}
|
92
92
|
|
93
93
|
default_config.rpcs.batch_commit_write_streams.timeout = 600.0
|
94
94
|
default_config.rpcs.batch_commit_write_streams.retry_policy = {
|
95
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14]
|
95
|
+
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14, 8]
|
96
96
|
}
|
97
97
|
|
98
98
|
default_config.rpcs.flush_rows.timeout = 600.0
|
99
99
|
default_config.rpcs.flush_rows.retry_policy = {
|
100
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14]
|
100
|
+
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [4, 14, 8]
|
101
101
|
}
|
102
102
|
|
103
103
|
default_config
|
@@ -190,18 +190,23 @@ module Google
|
|
190
190
|
|
191
191
|
# Request message for `AppendRows`.
|
192
192
|
#
|
193
|
-
#
|
194
|
-
#
|
195
|
-
#
|
193
|
+
# Because AppendRows is a bidirectional streaming RPC, certain parts of the
|
194
|
+
# AppendRowsRequest need only be specified for the first request before
|
195
|
+
# switching table destinations. You can also switch table destinations within
|
196
|
+
# the same connection for the default stream.
|
196
197
|
#
|
197
198
|
# The size of a single AppendRowsRequest must be less than 10 MB in size.
|
198
199
|
# Requests larger than this return an error, typically `INVALID_ARGUMENT`.
|
199
200
|
# @!attribute [rw] write_stream
|
200
201
|
# @return [::String]
|
201
|
-
# Required. The write_stream identifies the
|
202
|
-
#
|
203
|
-
#
|
204
|
-
# the first request.
|
202
|
+
# Required. The write_stream identifies the append operation. It must be
|
203
|
+
# provided in the following scenarios:
|
204
|
+
#
|
205
|
+
# * In the first request to an AppendRows connection.
|
206
|
+
#
|
207
|
+
# * In all subsequent requests to an AppendRows connection, if you use the
|
208
|
+
# same connection to write to multiple tables or change the input schema for
|
209
|
+
# default streams.
|
205
210
|
#
|
206
211
|
# For explicitly created write streams, the format is:
|
207
212
|
#
|
@@ -210,6 +215,22 @@ module Google
|
|
210
215
|
# For the special default stream, the format is:
|
211
216
|
#
|
212
217
|
# * `projects/{project}/datasets/{dataset}/tables/{table}/streams/_default`.
|
218
|
+
#
|
219
|
+
# An example of a possible sequence of requests with write_stream fields
|
220
|
+
# within a single connection:
|
221
|
+
#
|
222
|
+
# * r1: \\{write_stream: stream_name_1}
|
223
|
+
#
|
224
|
+
# * r2: \\{write_stream: /*omit*/}
|
225
|
+
#
|
226
|
+
# * r3: \\{write_stream: /*omit*/}
|
227
|
+
#
|
228
|
+
# * r4: \\{write_stream: stream_name_2}
|
229
|
+
#
|
230
|
+
# * r5: \\{write_stream: stream_name_2}
|
231
|
+
#
|
232
|
+
# The destination changed in request_4, so the write_stream field must be
|
233
|
+
# populated in all subsequent requests in this stream.
|
213
234
|
# @!attribute [rw] offset
|
214
235
|
# @return [::Google::Protobuf::Int64Value]
|
215
236
|
# If present, the write is only performed if the next append offset is same
|
@@ -251,9 +272,14 @@ module Google
|
|
251
272
|
# requests.
|
252
273
|
# @!attribute [rw] writer_schema
|
253
274
|
# @return [::Google::Cloud::Bigquery::Storage::V1::ProtoSchema]
|
254
|
-
#
|
255
|
-
#
|
256
|
-
#
|
275
|
+
# The protocol buffer schema used to serialize the data. Provide this value
|
276
|
+
# whenever:
|
277
|
+
#
|
278
|
+
# * You send the first request of an RPC connection.
|
279
|
+
#
|
280
|
+
# * You change the input schema.
|
281
|
+
#
|
282
|
+
# * You specify a new destination table.
|
257
283
|
# @!attribute [rw] rows
|
258
284
|
# @return [::Google::Cloud::Bigquery::Storage::V1::ProtoRows]
|
259
285
|
# Serialized row data in protobuf message format.
|
@@ -274,10 +300,9 @@ module Google
|
|
274
300
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
275
301
|
end
|
276
302
|
|
277
|
-
# An enum to indicate how to interpret missing values
|
278
|
-
#
|
279
|
-
#
|
280
|
-
# schema.
|
303
|
+
# An enum to indicate how to interpret missing values of fields that are
|
304
|
+
# present in user schema but missing in rows. A missing value can represent a
|
305
|
+
# NULL or a column default value defined in BigQuery table schema.
|
281
306
|
module MissingValueInterpretation
|
282
307
|
# Invalid missing value interpretation. Requests with this value will be
|
283
308
|
# rejected.
|
@@ -69,10 +69,10 @@ module Google
|
|
69
69
|
# metadata from the table which might be incomplete or stale.
|
70
70
|
# @!attribute [r] estimated_total_physical_file_size
|
71
71
|
# @return [::Integer]
|
72
|
-
# Output only. A pre-projected estimate of the total physical size
|
73
|
-
#
|
74
|
-
#
|
75
|
-
# metadata from the table
|
72
|
+
# Output only. A pre-projected estimate of the total physical size of files
|
73
|
+
# (in bytes) that this session will scan when all streams are consumed. This
|
74
|
+
# estimate is independent of the selected columns and can be based on
|
75
|
+
# incomplete or stale metadata from the table. This field is only set for
|
76
76
|
# BigLake tables.
|
77
77
|
# @!attribute [r] estimated_row_count
|
78
78
|
# @return [::Integer]
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-cloud-bigquery-storage-v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.21.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2023-
|
11
|
+
date: 2023-07-07 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: gapic-common
|