google-cloud-bigquery-storage-v1 0.8.0 → 0.9.2
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/.yardopts +1 -1
- data/AUTHENTICATION.md +7 -25
- data/README.md +1 -1
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/client.rb +63 -9
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/credentials.rb +0 -1
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/client.rb +124 -15
- data/lib/google/cloud/bigquery/storage/v1/stream_pb.rb +6 -0
- data/lib/google/cloud/bigquery/storage/v1/version.rb +1 -1
- data/proto_docs/google/api/resource.rb +10 -71
- data/proto_docs/google/cloud/bigquery/storage/v1/stream.rb +13 -0
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: dd6247e85a5ffc569f96771fba5c9a3662c25fbd53e6f8e2e9adb7aee9480376
|
4
|
+
data.tar.gz: 0b501022ceac5a5bf209143d5639eb229adaa604d2aeb7b591910642c20e3dd5
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: b0018a10cb61fd33add4559fc6cd7c2323d79dd70244ffdaa0ee45d9ab79ab4139f0563739868c8fcf8be9ee5810a3ecb1378955e992f024a784ca8a0ef7f64c
|
7
|
+
data.tar.gz: 6f45f00046cb5b0723f7bfb214b5e7918ef5ecb151e0f3cc7768a2eeeb9fc1e0f1a4753cd89eb38208dc10552dbeb34bcaae3f49c774b1115c5ad91146935010
|
data/.yardopts
CHANGED
data/AUTHENTICATION.md
CHANGED
@@ -120,15 +120,6 @@ To configure your system for this, simply:
|
|
120
120
|
**NOTE:** This is _not_ recommended for running in production. The Cloud SDK
|
121
121
|
*should* only be used during development.
|
122
122
|
|
123
|
-
[gce-how-to]: https://cloud.google.com/compute/docs/authentication#using
|
124
|
-
[dev-console]: https://console.cloud.google.com/project
|
125
|
-
|
126
|
-
[enable-apis]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/enable-apis.png
|
127
|
-
|
128
|
-
[create-new-service-account]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/create-new-service-account.png
|
129
|
-
[create-new-service-account-existing-keys]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/create-new-service-account-existing-keys.png
|
130
|
-
[reuse-service-account]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/reuse-service-account.png
|
131
|
-
|
132
123
|
## Creating a Service Account
|
133
124
|
|
134
125
|
Google Cloud requires **Service Account Credentials** to
|
@@ -139,31 +130,22 @@ If you are not running this client within
|
|
139
130
|
[Google Cloud Platform environments](#google-cloud-platform-environments), you
|
140
131
|
need a Google Developers service account.
|
141
132
|
|
142
|
-
1. Visit the [Google
|
133
|
+
1. Visit the [Google Cloud Console](https://console.cloud.google.com/project).
|
143
134
|
2. Create a new project or click on an existing project.
|
144
|
-
3. Activate the
|
135
|
+
3. Activate the menu in the upper left and select **APIs & Services**. From
|
145
136
|
here, you will enable the APIs that your application requires.
|
146
137
|
|
147
|
-
![Enable the APIs that your application requires][enable-apis]
|
148
|
-
|
149
138
|
*Note: You may need to enable billing in order to use these services.*
|
150
139
|
|
151
140
|
4. Select **Credentials** from the side navigation.
|
152
141
|
|
153
|
-
|
154
|
-
|
155
|
-
![Create a new service account][create-new-service-account]
|
156
|
-
|
157
|
-
![Create a new service account With Existing Keys][create-new-service-account-existing-keys]
|
158
|
-
|
159
|
-
Find the "Add credentials" drop down and select "Service account" to be
|
160
|
-
guided through downloading a new JSON key file.
|
142
|
+
Find the "Create credentials" drop down near the top of the page, and select
|
143
|
+
"Service account" to be guided through downloading a new JSON key file.
|
161
144
|
|
162
145
|
If you want to re-use an existing service account, you can easily generate a
|
163
|
-
new key file. Just select the account you wish to re-use,
|
164
|
-
|
165
|
-
|
166
|
-
![Re-use an existing service account][reuse-service-account]
|
146
|
+
new key file. Just select the account you wish to re-use, click the pencil
|
147
|
+
tool on the right side to edit the service account, select the **Keys** tab,
|
148
|
+
and then select **Add Key**.
|
167
149
|
|
168
150
|
The key file you download will be used by this library to authenticate API
|
169
151
|
requests and should be stored in a secure location.
|
data/README.md
CHANGED
@@ -37,7 +37,7 @@ request = ::Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest.new #
|
|
37
37
|
response = client.create_read_session request
|
38
38
|
```
|
39
39
|
|
40
|
-
View the [Client Library Documentation](https://
|
40
|
+
View the [Client Library Documentation](https://cloud.google.com/ruby/docs/reference/google-cloud-bigquery-storage-v1/latest)
|
41
41
|
for class and method documentation.
|
42
42
|
|
43
43
|
See also the [Product Documentation](https://cloud.google.com/bigquery/docs/reference/storage)
|
@@ -221,6 +221,21 @@ module Google
|
|
221
221
|
#
|
222
222
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
223
223
|
#
|
224
|
+
# @example Basic example
|
225
|
+
# require "google/cloud/bigquery/storage/v1"
|
226
|
+
#
|
227
|
+
# # Create a client object. The client can be reused for multiple calls.
|
228
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
|
229
|
+
#
|
230
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
231
|
+
# request = Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest.new
|
232
|
+
#
|
233
|
+
# # Call the create_read_session method.
|
234
|
+
# result = client.create_read_session request
|
235
|
+
#
|
236
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::ReadSession.
|
237
|
+
# p result
|
238
|
+
#
|
224
239
|
def create_read_session request, options = nil
|
225
240
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
226
241
|
|
@@ -238,9 +253,11 @@ module Google
|
|
238
253
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
239
254
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
240
255
|
|
241
|
-
header_params = {
|
242
|
-
|
243
|
-
|
256
|
+
header_params = {}
|
257
|
+
if request.read_session&.table
|
258
|
+
header_params["read_session.table"] = request.read_session.table
|
259
|
+
end
|
260
|
+
|
244
261
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
245
262
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
246
263
|
|
@@ -299,6 +316,24 @@ module Google
|
|
299
316
|
#
|
300
317
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
301
318
|
#
|
319
|
+
# @example Basic example
|
320
|
+
# require "google/cloud/bigquery/storage/v1"
|
321
|
+
#
|
322
|
+
# # Create a client object. The client can be reused for multiple calls.
|
323
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
|
324
|
+
#
|
325
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
326
|
+
# request = Google::Cloud::Bigquery::Storage::V1::ReadRowsRequest.new
|
327
|
+
#
|
328
|
+
# # Call the read_rows method.
|
329
|
+
# result = client.read_rows request
|
330
|
+
#
|
331
|
+
# # The returned object is a streamed enumerable yielding elements of
|
332
|
+
# # type ::Google::Cloud::Bigquery::Storage::V1::ReadRowsResponse.
|
333
|
+
# result.each do |response|
|
334
|
+
# p response
|
335
|
+
# end
|
336
|
+
#
|
302
337
|
def read_rows request, options = nil
|
303
338
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
304
339
|
|
@@ -316,9 +351,11 @@ module Google
|
|
316
351
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
317
352
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
318
353
|
|
319
|
-
header_params = {
|
320
|
-
|
321
|
-
|
354
|
+
header_params = {}
|
355
|
+
if request.read_stream
|
356
|
+
header_params["read_stream"] = request.read_stream
|
357
|
+
end
|
358
|
+
|
322
359
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
323
360
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
324
361
|
|
@@ -386,6 +423,21 @@ module Google
|
|
386
423
|
#
|
387
424
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
388
425
|
#
|
426
|
+
# @example Basic example
|
427
|
+
# require "google/cloud/bigquery/storage/v1"
|
428
|
+
#
|
429
|
+
# # Create a client object. The client can be reused for multiple calls.
|
430
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryRead::Client.new
|
431
|
+
#
|
432
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
433
|
+
# request = Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest.new
|
434
|
+
#
|
435
|
+
# # Call the split_read_stream method.
|
436
|
+
# result = client.split_read_stream request
|
437
|
+
#
|
438
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse.
|
439
|
+
# p result
|
440
|
+
#
|
389
441
|
def split_read_stream request, options = nil
|
390
442
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
391
443
|
|
@@ -403,9 +455,11 @@ module Google
|
|
403
455
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
404
456
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
405
457
|
|
406
|
-
header_params = {
|
407
|
-
|
408
|
-
|
458
|
+
header_params = {}
|
459
|
+
if request.name
|
460
|
+
header_params["name"] = request.name
|
461
|
+
end
|
462
|
+
|
409
463
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
410
464
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
411
465
|
|
@@ -28,7 +28,6 @@ module Google
|
|
28
28
|
class Credentials < ::Google::Auth::Credentials
|
29
29
|
self.scope = [
|
30
30
|
"https://www.googleapis.com/auth/bigquery",
|
31
|
-
"https://www.googleapis.com/auth/bigquery.readonly",
|
32
31
|
"https://www.googleapis.com/auth/cloud-platform"
|
33
32
|
]
|
34
33
|
self.env_vars = [
|
@@ -217,6 +217,21 @@ module Google
|
|
217
217
|
#
|
218
218
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
219
219
|
#
|
220
|
+
# @example Basic example
|
221
|
+
# require "google/cloud/bigquery/storage/v1"
|
222
|
+
#
|
223
|
+
# # Create a client object. The client can be reused for multiple calls.
|
224
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
225
|
+
#
|
226
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
227
|
+
# request = Google::Cloud::Bigquery::Storage::V1::CreateWriteStreamRequest.new
|
228
|
+
#
|
229
|
+
# # Call the create_write_stream method.
|
230
|
+
# result = client.create_write_stream request
|
231
|
+
#
|
232
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::WriteStream.
|
233
|
+
# p result
|
234
|
+
#
|
220
235
|
def create_write_stream request, options = nil
|
221
236
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
222
237
|
|
@@ -234,9 +249,11 @@ module Google
|
|
234
249
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
235
250
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
236
251
|
|
237
|
-
header_params = {
|
238
|
-
|
239
|
-
|
252
|
+
header_params = {}
|
253
|
+
if request.parent
|
254
|
+
header_params["parent"] = request.parent
|
255
|
+
end
|
256
|
+
|
240
257
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
241
258
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
242
259
|
|
@@ -302,6 +319,30 @@ module Google
|
|
302
319
|
#
|
303
320
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
304
321
|
#
|
322
|
+
# @example Basic example
|
323
|
+
# require "google/cloud/bigquery/storage/v1"
|
324
|
+
#
|
325
|
+
# # Create a client object. The client can be reused for multiple calls.
|
326
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
327
|
+
#
|
328
|
+
# # Create an input stream
|
329
|
+
# input = Gapic::StreamInput.new
|
330
|
+
#
|
331
|
+
# # Call the append_rows method to start streaming.
|
332
|
+
# output = client.append_rows input
|
333
|
+
#
|
334
|
+
# # Send requests on the stream. For each request, pass in keyword
|
335
|
+
# # arguments to set fields. Be sure to close the stream when done.
|
336
|
+
# input << Google::Cloud::Bigquery::Storage::V1::AppendRowsRequest.new
|
337
|
+
# input << Google::Cloud::Bigquery::Storage::V1::AppendRowsRequest.new
|
338
|
+
# input.close
|
339
|
+
#
|
340
|
+
# # Handle streamed responses. These may be interleaved with inputs.
|
341
|
+
# # Each response is of type ::Google::Cloud::Bigquery::Storage::V1::AppendRowsResponse.
|
342
|
+
# output.each do |response|
|
343
|
+
# p response
|
344
|
+
# end
|
345
|
+
#
|
305
346
|
def append_rows request, options = nil
|
306
347
|
unless request.is_a? ::Enumerable
|
307
348
|
raise ::ArgumentError, "request must be an Enumerable" unless request.respond_to? :to_enum
|
@@ -370,6 +411,21 @@ module Google
|
|
370
411
|
#
|
371
412
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
372
413
|
#
|
414
|
+
# @example Basic example
|
415
|
+
# require "google/cloud/bigquery/storage/v1"
|
416
|
+
#
|
417
|
+
# # Create a client object. The client can be reused for multiple calls.
|
418
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
419
|
+
#
|
420
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
421
|
+
# request = Google::Cloud::Bigquery::Storage::V1::GetWriteStreamRequest.new
|
422
|
+
#
|
423
|
+
# # Call the get_write_stream method.
|
424
|
+
# result = client.get_write_stream request
|
425
|
+
#
|
426
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::WriteStream.
|
427
|
+
# p result
|
428
|
+
#
|
373
429
|
def get_write_stream request, options = nil
|
374
430
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
375
431
|
|
@@ -387,9 +443,11 @@ module Google
|
|
387
443
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
388
444
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
389
445
|
|
390
|
-
header_params = {
|
391
|
-
|
392
|
-
|
446
|
+
header_params = {}
|
447
|
+
if request.name
|
448
|
+
header_params["name"] = request.name
|
449
|
+
end
|
450
|
+
|
393
451
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
394
452
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
395
453
|
|
@@ -440,6 +498,21 @@ module Google
|
|
440
498
|
#
|
441
499
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
442
500
|
#
|
501
|
+
# @example Basic example
|
502
|
+
# require "google/cloud/bigquery/storage/v1"
|
503
|
+
#
|
504
|
+
# # Create a client object. The client can be reused for multiple calls.
|
505
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
506
|
+
#
|
507
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
508
|
+
# request = Google::Cloud::Bigquery::Storage::V1::FinalizeWriteStreamRequest.new
|
509
|
+
#
|
510
|
+
# # Call the finalize_write_stream method.
|
511
|
+
# result = client.finalize_write_stream request
|
512
|
+
#
|
513
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::FinalizeWriteStreamResponse.
|
514
|
+
# p result
|
515
|
+
#
|
443
516
|
def finalize_write_stream request, options = nil
|
444
517
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
445
518
|
|
@@ -457,9 +530,11 @@ module Google
|
|
457
530
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
458
531
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
459
532
|
|
460
|
-
header_params = {
|
461
|
-
|
462
|
-
|
533
|
+
header_params = {}
|
534
|
+
if request.name
|
535
|
+
header_params["name"] = request.name
|
536
|
+
end
|
537
|
+
|
463
538
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
464
539
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
465
540
|
|
@@ -516,6 +591,21 @@ module Google
|
|
516
591
|
#
|
517
592
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
518
593
|
#
|
594
|
+
# @example Basic example
|
595
|
+
# require "google/cloud/bigquery/storage/v1"
|
596
|
+
#
|
597
|
+
# # Create a client object. The client can be reused for multiple calls.
|
598
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
599
|
+
#
|
600
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
601
|
+
# request = Google::Cloud::Bigquery::Storage::V1::BatchCommitWriteStreamsRequest.new
|
602
|
+
#
|
603
|
+
# # Call the batch_commit_write_streams method.
|
604
|
+
# result = client.batch_commit_write_streams request
|
605
|
+
#
|
606
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::BatchCommitWriteStreamsResponse.
|
607
|
+
# p result
|
608
|
+
#
|
519
609
|
def batch_commit_write_streams request, options = nil
|
520
610
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
521
611
|
|
@@ -533,9 +623,11 @@ module Google
|
|
533
623
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
534
624
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
535
625
|
|
536
|
-
header_params = {
|
537
|
-
|
538
|
-
|
626
|
+
header_params = {}
|
627
|
+
if request.parent
|
628
|
+
header_params["parent"] = request.parent
|
629
|
+
end
|
630
|
+
|
539
631
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
540
632
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
541
633
|
|
@@ -594,6 +686,21 @@ module Google
|
|
594
686
|
#
|
595
687
|
# @raise [::Google::Cloud::Error] if the RPC is aborted.
|
596
688
|
#
|
689
|
+
# @example Basic example
|
690
|
+
# require "google/cloud/bigquery/storage/v1"
|
691
|
+
#
|
692
|
+
# # Create a client object. The client can be reused for multiple calls.
|
693
|
+
# client = Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
694
|
+
#
|
695
|
+
# # Create a request. To set request fields, pass in keyword arguments.
|
696
|
+
# request = Google::Cloud::Bigquery::Storage::V1::FlushRowsRequest.new
|
697
|
+
#
|
698
|
+
# # Call the flush_rows method.
|
699
|
+
# result = client.flush_rows request
|
700
|
+
#
|
701
|
+
# # The returned object is of type Google::Cloud::Bigquery::Storage::V1::FlushRowsResponse.
|
702
|
+
# p result
|
703
|
+
#
|
597
704
|
def flush_rows request, options = nil
|
598
705
|
raise ::ArgumentError, "request must be provided" if request.nil?
|
599
706
|
|
@@ -611,9 +718,11 @@ module Google
|
|
611
718
|
gapic_version: ::Google::Cloud::Bigquery::Storage::V1::VERSION
|
612
719
|
metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
|
613
720
|
|
614
|
-
header_params = {
|
615
|
-
|
616
|
-
|
721
|
+
header_params = {}
|
722
|
+
if request.write_stream
|
723
|
+
header_params["write_stream"] = request.write_stream
|
724
|
+
end
|
725
|
+
|
617
726
|
request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
|
618
727
|
metadata[:"x-goog-request-params"] ||= request_params_header
|
619
728
|
|
@@ -44,6 +44,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
44
44
|
optional :create_time, :message, 3, "google.protobuf.Timestamp"
|
45
45
|
optional :commit_time, :message, 4, "google.protobuf.Timestamp"
|
46
46
|
optional :table_schema, :message, 5, "google.cloud.bigquery.storage.v1.TableSchema"
|
47
|
+
optional :write_mode, :enum, 7, "google.cloud.bigquery.storage.v1.WriteStream.WriteMode"
|
47
48
|
end
|
48
49
|
add_enum "google.cloud.bigquery.storage.v1.WriteStream.Type" do
|
49
50
|
value :TYPE_UNSPECIFIED, 0
|
@@ -51,6 +52,10 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
51
52
|
value :PENDING, 2
|
52
53
|
value :BUFFERED, 3
|
53
54
|
end
|
55
|
+
add_enum "google.cloud.bigquery.storage.v1.WriteStream.WriteMode" do
|
56
|
+
value :WRITE_MODE_UNSPECIFIED, 0
|
57
|
+
value :INSERT, 1
|
58
|
+
end
|
54
59
|
add_enum "google.cloud.bigquery.storage.v1.DataFormat" do
|
55
60
|
value :DATA_FORMAT_UNSPECIFIED, 0
|
56
61
|
value :AVRO, 1
|
@@ -70,6 +75,7 @@ module Google
|
|
70
75
|
ReadStream = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ReadStream").msgclass
|
71
76
|
WriteStream = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.WriteStream").msgclass
|
72
77
|
WriteStream::Type = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.WriteStream.Type").enummodule
|
78
|
+
WriteStream::WriteMode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.WriteStream.WriteMode").enummodule
|
73
79
|
DataFormat = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.DataFormat").enummodule
|
74
80
|
end
|
75
81
|
end
|
@@ -33,11 +33,7 @@ module Google
|
|
33
33
|
# // For Kubernetes resources, the format is {api group}/{kind}.
|
34
34
|
# option (google.api.resource) = {
|
35
35
|
# type: "pubsub.googleapis.com/Topic"
|
36
|
-
#
|
37
|
-
# pattern: "projects/{project}/topics/{topic}"
|
38
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Project"
|
39
|
-
# parent_name_extractor: "projects/{project}"
|
40
|
-
# }
|
36
|
+
# pattern: "projects/{project}/topics/{topic}"
|
41
37
|
# };
|
42
38
|
# }
|
43
39
|
#
|
@@ -45,10 +41,7 @@ module Google
|
|
45
41
|
#
|
46
42
|
# resources:
|
47
43
|
# - type: "pubsub.googleapis.com/Topic"
|
48
|
-
#
|
49
|
-
# - pattern: "projects/{project}/topics/{topic}"
|
50
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Project"
|
51
|
-
# parent_name_extractor: "projects/{project}"
|
44
|
+
# pattern: "projects/{project}/topics/{topic}"
|
52
45
|
#
|
53
46
|
# Sometimes, resources have multiple patterns, typically because they can
|
54
47
|
# live under multiple parents.
|
@@ -58,26 +51,10 @@ module Google
|
|
58
51
|
# message LogEntry {
|
59
52
|
# option (google.api.resource) = {
|
60
53
|
# type: "logging.googleapis.com/LogEntry"
|
61
|
-
#
|
62
|
-
#
|
63
|
-
#
|
64
|
-
#
|
65
|
-
# }
|
66
|
-
# name_descriptor: {
|
67
|
-
# pattern: "folders/{folder}/logs/{log}"
|
68
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Folder"
|
69
|
-
# parent_name_extractor: "folders/{folder}"
|
70
|
-
# }
|
71
|
-
# name_descriptor: {
|
72
|
-
# pattern: "organizations/{organization}/logs/{log}"
|
73
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Organization"
|
74
|
-
# parent_name_extractor: "organizations/{organization}"
|
75
|
-
# }
|
76
|
-
# name_descriptor: {
|
77
|
-
# pattern: "billingAccounts/{billing_account}/logs/{log}"
|
78
|
-
# parent_type: "billing.googleapis.com/BillingAccount"
|
79
|
-
# parent_name_extractor: "billingAccounts/{billing_account}"
|
80
|
-
# }
|
54
|
+
# pattern: "projects/{project}/logs/{log}"
|
55
|
+
# pattern: "folders/{folder}/logs/{log}"
|
56
|
+
# pattern: "organizations/{organization}/logs/{log}"
|
57
|
+
# pattern: "billingAccounts/{billing_account}/logs/{log}"
|
81
58
|
# };
|
82
59
|
# }
|
83
60
|
#
|
@@ -85,48 +62,10 @@ module Google
|
|
85
62
|
#
|
86
63
|
# resources:
|
87
64
|
# - type: 'logging.googleapis.com/LogEntry'
|
88
|
-
#
|
89
|
-
#
|
90
|
-
#
|
91
|
-
#
|
92
|
-
# - pattern: "folders/{folder}/logs/{log}"
|
93
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Folder"
|
94
|
-
# parent_name_extractor: "folders/{folder}"
|
95
|
-
# - pattern: "organizations/{organization}/logs/{log}"
|
96
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Organization"
|
97
|
-
# parent_name_extractor: "organizations/{organization}"
|
98
|
-
# - pattern: "billingAccounts/{billing_account}/logs/{log}"
|
99
|
-
# parent_type: "billing.googleapis.com/BillingAccount"
|
100
|
-
# parent_name_extractor: "billingAccounts/{billing_account}"
|
101
|
-
#
|
102
|
-
# For flexible resources, the resource name doesn't contain parent names, but
|
103
|
-
# the resource itself has parents for policy evaluation.
|
104
|
-
#
|
105
|
-
# Example:
|
106
|
-
#
|
107
|
-
# message Shelf {
|
108
|
-
# option (google.api.resource) = {
|
109
|
-
# type: "library.googleapis.com/Shelf"
|
110
|
-
# name_descriptor: {
|
111
|
-
# pattern: "shelves/{shelf}"
|
112
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Project"
|
113
|
-
# }
|
114
|
-
# name_descriptor: {
|
115
|
-
# pattern: "shelves/{shelf}"
|
116
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Folder"
|
117
|
-
# }
|
118
|
-
# };
|
119
|
-
# }
|
120
|
-
#
|
121
|
-
# The ResourceDescriptor Yaml config will look like:
|
122
|
-
#
|
123
|
-
# resources:
|
124
|
-
# - type: 'library.googleapis.com/Shelf'
|
125
|
-
# name_descriptor:
|
126
|
-
# - pattern: "shelves/{shelf}"
|
127
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Project"
|
128
|
-
# - pattern: "shelves/{shelf}"
|
129
|
-
# parent_type: "cloudresourcemanager.googleapis.com/Folder"
|
65
|
+
# pattern: "projects/{project}/logs/{log}"
|
66
|
+
# pattern: "folders/{folder}/logs/{log}"
|
67
|
+
# pattern: "organizations/{organization}/logs/{log}"
|
68
|
+
# pattern: "billingAccounts/{billing_account}/logs/{log}"
|
130
69
|
# @!attribute [rw] type
|
131
70
|
# @return [::String]
|
132
71
|
# The resource type. It must be in the format of
|
@@ -141,6 +141,9 @@ module Google
|
|
141
141
|
# `CreateWriteStream` response. Caller should generate data that's
|
142
142
|
# compatible with this schema to send in initial `AppendRowsRequest`.
|
143
143
|
# The table schema could go out of date during the life time of the stream.
|
144
|
+
# @!attribute [rw] write_mode
|
145
|
+
# @return [::Google::Cloud::Bigquery::Storage::V1::WriteStream::WriteMode]
|
146
|
+
# Immutable. Mode of the stream.
|
144
147
|
class WriteStream
|
145
148
|
include ::Google::Protobuf::MessageExts
|
146
149
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -160,6 +163,16 @@ module Google
|
|
160
163
|
# Data is only visible up to the offset to which it was flushed.
|
161
164
|
BUFFERED = 3
|
162
165
|
end
|
166
|
+
|
167
|
+
# Mode enum of the stream.
|
168
|
+
module WriteMode
|
169
|
+
# Unknown type.
|
170
|
+
WRITE_MODE_UNSPECIFIED = 0
|
171
|
+
|
172
|
+
# Insert new records into the table.
|
173
|
+
# It is the default value if customers do not specify it.
|
174
|
+
INSERT = 1
|
175
|
+
end
|
163
176
|
end
|
164
177
|
|
165
178
|
# Data format for input or output data.
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-cloud-bigquery-storage-v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.9.2
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2022-02-18 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: gapic-common
|
@@ -219,7 +219,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
219
219
|
- !ruby/object:Gem::Version
|
220
220
|
version: '0'
|
221
221
|
requirements: []
|
222
|
-
rubygems_version: 3.
|
222
|
+
rubygems_version: 3.3.5
|
223
223
|
signing_key:
|
224
224
|
specification_version: 4
|
225
225
|
summary: API Client library for the BigQuery Storage V1 API
|