google-cloud-bigquery-storage-v1 0.6.2 → 0.8.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/google/cloud/bigquery/storage/v1/big_query_read/client.rb +102 -49
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/client.rb +936 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/credentials.rb +55 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_write/paths.rb +75 -0
- data/lib/google/cloud/bigquery/storage/v1/big_query_write.rb +56 -0
- data/lib/google/cloud/bigquery/storage/v1/protobuf_pb.rb +29 -0
- data/lib/google/cloud/bigquery/storage/v1/storage_pb.rb +86 -2
- data/lib/google/cloud/bigquery/storage/v1/storage_services_pb.rb +81 -2
- data/lib/google/cloud/bigquery/storage/v1/stream_pb.rb +19 -2
- data/lib/google/cloud/bigquery/storage/v1/table_pb.rb +62 -0
- data/lib/google/cloud/bigquery/storage/v1/version.rb +1 -1
- data/lib/google/cloud/bigquery/storage/v1.rb +1 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/protobuf.rb +56 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/storage.rb +244 -1
- data/proto_docs/google/cloud/bigquery/storage/v1/stream.rb +50 -0
- data/proto_docs/google/cloud/bigquery/storage/v1/table.rb +172 -0
- data/proto_docs/google/protobuf/any.rb +141 -0
- data/proto_docs/google/protobuf/wrappers.rb +121 -0
- data/proto_docs/google/rpc/status.rb +46 -0
- metadata +15 -4
@@ -0,0 +1,55 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2021 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
require "googleauth"
|
20
|
+
|
21
|
+
module Google
|
22
|
+
module Cloud
|
23
|
+
module Bigquery
|
24
|
+
module Storage
|
25
|
+
module V1
|
26
|
+
module BigQueryWrite
|
27
|
+
# Credentials for the BigQueryWrite API.
|
28
|
+
class Credentials < ::Google::Auth::Credentials
|
29
|
+
self.scope = [
|
30
|
+
"https://www.googleapis.com/auth/bigquery",
|
31
|
+
"https://www.googleapis.com/auth/bigquery.insertdata",
|
32
|
+
"https://www.googleapis.com/auth/cloud-platform"
|
33
|
+
]
|
34
|
+
self.env_vars = [
|
35
|
+
"BIGQUERY_STORAGE_CREDENTIALS",
|
36
|
+
"BIGQUERY_STORAGE_KEYFILE",
|
37
|
+
"GOOGLE_CLOUD_CREDENTIALS",
|
38
|
+
"GOOGLE_CLOUD_KEYFILE",
|
39
|
+
"GCLOUD_KEYFILE",
|
40
|
+
"BIGQUERY_STORAGE_CREDENTIALS_JSON",
|
41
|
+
"BIGQUERY_STORAGE_KEYFILE_JSON",
|
42
|
+
"GOOGLE_CLOUD_CREDENTIALS_JSON",
|
43
|
+
"GOOGLE_CLOUD_KEYFILE_JSON",
|
44
|
+
"GCLOUD_KEYFILE_JSON"
|
45
|
+
]
|
46
|
+
self.paths = [
|
47
|
+
"~/.config/google_cloud/application_default_credentials.json"
|
48
|
+
]
|
49
|
+
end
|
50
|
+
end
|
51
|
+
end
|
52
|
+
end
|
53
|
+
end
|
54
|
+
end
|
55
|
+
end
|
@@ -0,0 +1,75 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2021 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module Bigquery
|
23
|
+
module Storage
|
24
|
+
module V1
|
25
|
+
module BigQueryWrite
|
26
|
+
# Path helper methods for the BigQueryWrite API.
|
27
|
+
module Paths
|
28
|
+
##
|
29
|
+
# Create a fully-qualified Table resource string.
|
30
|
+
#
|
31
|
+
# The resource will be in the following format:
|
32
|
+
#
|
33
|
+
# `projects/{project}/datasets/{dataset}/tables/{table}`
|
34
|
+
#
|
35
|
+
# @param project [String]
|
36
|
+
# @param dataset [String]
|
37
|
+
# @param table [String]
|
38
|
+
#
|
39
|
+
# @return [::String]
|
40
|
+
def table_path project:, dataset:, table:
|
41
|
+
raise ::ArgumentError, "project cannot contain /" if project.to_s.include? "/"
|
42
|
+
raise ::ArgumentError, "dataset cannot contain /" if dataset.to_s.include? "/"
|
43
|
+
|
44
|
+
"projects/#{project}/datasets/#{dataset}/tables/#{table}"
|
45
|
+
end
|
46
|
+
|
47
|
+
##
|
48
|
+
# Create a fully-qualified WriteStream resource string.
|
49
|
+
#
|
50
|
+
# The resource will be in the following format:
|
51
|
+
#
|
52
|
+
# `projects/{project}/datasets/{dataset}/tables/{table}/streams/{stream}`
|
53
|
+
#
|
54
|
+
# @param project [String]
|
55
|
+
# @param dataset [String]
|
56
|
+
# @param table [String]
|
57
|
+
# @param stream [String]
|
58
|
+
#
|
59
|
+
# @return [::String]
|
60
|
+
def write_stream_path project:, dataset:, table:, stream:
|
61
|
+
raise ::ArgumentError, "project cannot contain /" if project.to_s.include? "/"
|
62
|
+
raise ::ArgumentError, "dataset cannot contain /" if dataset.to_s.include? "/"
|
63
|
+
raise ::ArgumentError, "table cannot contain /" if table.to_s.include? "/"
|
64
|
+
|
65
|
+
"projects/#{project}/datasets/#{dataset}/tables/#{table}/streams/#{stream}"
|
66
|
+
end
|
67
|
+
|
68
|
+
extend self
|
69
|
+
end
|
70
|
+
end
|
71
|
+
end
|
72
|
+
end
|
73
|
+
end
|
74
|
+
end
|
75
|
+
end
|
@@ -0,0 +1,56 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2021 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
require "gapic/common"
|
20
|
+
require "gapic/config"
|
21
|
+
require "gapic/config/method"
|
22
|
+
|
23
|
+
require "google/cloud/bigquery/storage/v1/version"
|
24
|
+
|
25
|
+
require "google/cloud/bigquery/storage/v1/big_query_write/credentials"
|
26
|
+
require "google/cloud/bigquery/storage/v1/big_query_write/paths"
|
27
|
+
require "google/cloud/bigquery/storage/v1/big_query_write/client"
|
28
|
+
|
29
|
+
module Google
|
30
|
+
module Cloud
|
31
|
+
module Bigquery
|
32
|
+
module Storage
|
33
|
+
module V1
|
34
|
+
##
|
35
|
+
# BigQuery Write API.
|
36
|
+
#
|
37
|
+
# The Write API can be used to write data to BigQuery.
|
38
|
+
#
|
39
|
+
# For supplementary information about the Write API, see:
|
40
|
+
# https://cloud.google.com/bigquery/docs/write-api
|
41
|
+
#
|
42
|
+
# To load this service and instantiate a client:
|
43
|
+
#
|
44
|
+
# require "google/cloud/bigquery/storage/v1/big_query_write"
|
45
|
+
# client = ::Google::Cloud::Bigquery::Storage::V1::BigQueryWrite::Client.new
|
46
|
+
#
|
47
|
+
module BigQueryWrite
|
48
|
+
end
|
49
|
+
end
|
50
|
+
end
|
51
|
+
end
|
52
|
+
end
|
53
|
+
end
|
54
|
+
|
55
|
+
helper_path = ::File.join __dir__, "big_query_write", "helpers.rb"
|
56
|
+
require "google/cloud/bigquery/storage/v1/big_query_write/helpers" if ::File.file? helper_path
|
@@ -0,0 +1,29 @@
|
|
1
|
+
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
|
+
# source: google/cloud/bigquery/storage/v1/protobuf.proto
|
3
|
+
|
4
|
+
require 'google/protobuf/descriptor_pb'
|
5
|
+
require 'google/protobuf'
|
6
|
+
|
7
|
+
Google::Protobuf::DescriptorPool.generated_pool.build do
|
8
|
+
add_file("google/cloud/bigquery/storage/v1/protobuf.proto", :syntax => :proto3) do
|
9
|
+
add_message "google.cloud.bigquery.storage.v1.ProtoSchema" do
|
10
|
+
optional :proto_descriptor, :message, 1, "google.protobuf.DescriptorProto"
|
11
|
+
end
|
12
|
+
add_message "google.cloud.bigquery.storage.v1.ProtoRows" do
|
13
|
+
repeated :serialized_rows, :bytes, 1
|
14
|
+
end
|
15
|
+
end
|
16
|
+
end
|
17
|
+
|
18
|
+
module Google
|
19
|
+
module Cloud
|
20
|
+
module Bigquery
|
21
|
+
module Storage
|
22
|
+
module V1
|
23
|
+
ProtoSchema = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ProtoSchema").msgclass
|
24
|
+
ProtoRows = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ProtoRows").msgclass
|
25
|
+
end
|
26
|
+
end
|
27
|
+
end
|
28
|
+
end
|
29
|
+
end
|
@@ -1,15 +1,20 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/storage.proto
|
3
3
|
|
4
|
-
require 'google/protobuf'
|
5
|
-
|
6
4
|
require 'google/api/annotations_pb'
|
7
5
|
require 'google/api/client_pb'
|
8
6
|
require 'google/api/field_behavior_pb'
|
9
7
|
require 'google/api/resource_pb'
|
10
8
|
require 'google/cloud/bigquery/storage/v1/arrow_pb'
|
11
9
|
require 'google/cloud/bigquery/storage/v1/avro_pb'
|
10
|
+
require 'google/cloud/bigquery/storage/v1/protobuf_pb'
|
12
11
|
require 'google/cloud/bigquery/storage/v1/stream_pb'
|
12
|
+
require 'google/cloud/bigquery/storage/v1/table_pb'
|
13
|
+
require 'google/protobuf/timestamp_pb'
|
14
|
+
require 'google/protobuf/wrappers_pb'
|
15
|
+
require 'google/rpc/status_pb'
|
16
|
+
require 'google/protobuf'
|
17
|
+
|
13
18
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
14
19
|
add_file("google/cloud/bigquery/storage/v1/storage.proto", :syntax => :proto3) do
|
15
20
|
add_message "google.cloud.bigquery.storage.v1.CreateReadSessionRequest" do
|
@@ -52,6 +57,71 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
52
57
|
optional :primary_stream, :message, 1, "google.cloud.bigquery.storage.v1.ReadStream"
|
53
58
|
optional :remainder_stream, :message, 2, "google.cloud.bigquery.storage.v1.ReadStream"
|
54
59
|
end
|
60
|
+
add_message "google.cloud.bigquery.storage.v1.CreateWriteStreamRequest" do
|
61
|
+
optional :parent, :string, 1
|
62
|
+
optional :write_stream, :message, 2, "google.cloud.bigquery.storage.v1.WriteStream"
|
63
|
+
end
|
64
|
+
add_message "google.cloud.bigquery.storage.v1.AppendRowsRequest" do
|
65
|
+
optional :write_stream, :string, 1
|
66
|
+
optional :offset, :message, 2, "google.protobuf.Int64Value"
|
67
|
+
optional :trace_id, :string, 6
|
68
|
+
oneof :rows do
|
69
|
+
optional :proto_rows, :message, 4, "google.cloud.bigquery.storage.v1.AppendRowsRequest.ProtoData"
|
70
|
+
end
|
71
|
+
end
|
72
|
+
add_message "google.cloud.bigquery.storage.v1.AppendRowsRequest.ProtoData" do
|
73
|
+
optional :writer_schema, :message, 1, "google.cloud.bigquery.storage.v1.ProtoSchema"
|
74
|
+
optional :rows, :message, 2, "google.cloud.bigquery.storage.v1.ProtoRows"
|
75
|
+
end
|
76
|
+
add_message "google.cloud.bigquery.storage.v1.AppendRowsResponse" do
|
77
|
+
optional :updated_schema, :message, 3, "google.cloud.bigquery.storage.v1.TableSchema"
|
78
|
+
oneof :response do
|
79
|
+
optional :append_result, :message, 1, "google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult"
|
80
|
+
optional :error, :message, 2, "google.rpc.Status"
|
81
|
+
end
|
82
|
+
end
|
83
|
+
add_message "google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult" do
|
84
|
+
optional :offset, :message, 1, "google.protobuf.Int64Value"
|
85
|
+
end
|
86
|
+
add_message "google.cloud.bigquery.storage.v1.GetWriteStreamRequest" do
|
87
|
+
optional :name, :string, 1
|
88
|
+
end
|
89
|
+
add_message "google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsRequest" do
|
90
|
+
optional :parent, :string, 1
|
91
|
+
repeated :write_streams, :string, 2
|
92
|
+
end
|
93
|
+
add_message "google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsResponse" do
|
94
|
+
optional :commit_time, :message, 1, "google.protobuf.Timestamp"
|
95
|
+
repeated :stream_errors, :message, 2, "google.cloud.bigquery.storage.v1.StorageError"
|
96
|
+
end
|
97
|
+
add_message "google.cloud.bigquery.storage.v1.FinalizeWriteStreamRequest" do
|
98
|
+
optional :name, :string, 1
|
99
|
+
end
|
100
|
+
add_message "google.cloud.bigquery.storage.v1.FinalizeWriteStreamResponse" do
|
101
|
+
optional :row_count, :int64, 1
|
102
|
+
end
|
103
|
+
add_message "google.cloud.bigquery.storage.v1.FlushRowsRequest" do
|
104
|
+
optional :write_stream, :string, 1
|
105
|
+
optional :offset, :message, 2, "google.protobuf.Int64Value"
|
106
|
+
end
|
107
|
+
add_message "google.cloud.bigquery.storage.v1.FlushRowsResponse" do
|
108
|
+
optional :offset, :int64, 1
|
109
|
+
end
|
110
|
+
add_message "google.cloud.bigquery.storage.v1.StorageError" do
|
111
|
+
optional :code, :enum, 1, "google.cloud.bigquery.storage.v1.StorageError.StorageErrorCode"
|
112
|
+
optional :entity, :string, 2
|
113
|
+
optional :error_message, :string, 3
|
114
|
+
end
|
115
|
+
add_enum "google.cloud.bigquery.storage.v1.StorageError.StorageErrorCode" do
|
116
|
+
value :STORAGE_ERROR_CODE_UNSPECIFIED, 0
|
117
|
+
value :TABLE_NOT_FOUND, 1
|
118
|
+
value :STREAM_ALREADY_COMMITTED, 2
|
119
|
+
value :STREAM_NOT_FOUND, 3
|
120
|
+
value :INVALID_STREAM_TYPE, 4
|
121
|
+
value :INVALID_STREAM_STATE, 5
|
122
|
+
value :STREAM_FINALIZED, 6
|
123
|
+
value :SCHEMA_MISMATCH_EXTRA_FIELDS, 7
|
124
|
+
end
|
55
125
|
end
|
56
126
|
end
|
57
127
|
|
@@ -68,6 +138,20 @@ module Google
|
|
68
138
|
ReadRowsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ReadRowsResponse").msgclass
|
69
139
|
SplitReadStreamRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.SplitReadStreamRequest").msgclass
|
70
140
|
SplitReadStreamResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.SplitReadStreamResponse").msgclass
|
141
|
+
CreateWriteStreamRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.CreateWriteStreamRequest").msgclass
|
142
|
+
AppendRowsRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.AppendRowsRequest").msgclass
|
143
|
+
AppendRowsRequest::ProtoData = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.AppendRowsRequest.ProtoData").msgclass
|
144
|
+
AppendRowsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.AppendRowsResponse").msgclass
|
145
|
+
AppendRowsResponse::AppendResult = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult").msgclass
|
146
|
+
GetWriteStreamRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.GetWriteStreamRequest").msgclass
|
147
|
+
BatchCommitWriteStreamsRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsRequest").msgclass
|
148
|
+
BatchCommitWriteStreamsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsResponse").msgclass
|
149
|
+
FinalizeWriteStreamRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FinalizeWriteStreamRequest").msgclass
|
150
|
+
FinalizeWriteStreamResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FinalizeWriteStreamResponse").msgclass
|
151
|
+
FlushRowsRequest = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FlushRowsRequest").msgclass
|
152
|
+
FlushRowsResponse = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.FlushRowsResponse").msgclass
|
153
|
+
StorageError = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError").msgclass
|
154
|
+
StorageError::StorageErrorCode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.StorageError.StorageErrorCode").enummodule
|
71
155
|
end
|
72
156
|
end
|
73
157
|
end
|
@@ -30,7 +30,7 @@ module Google
|
|
30
30
|
# The Read API can be used to read data from BigQuery.
|
31
31
|
class Service
|
32
32
|
|
33
|
-
include GRPC::GenericService
|
33
|
+
include ::GRPC::GenericService
|
34
34
|
|
35
35
|
self.marshal_class_method = :encode
|
36
36
|
self.unmarshal_class_method = :decode
|
@@ -53,7 +53,7 @@ module Google
|
|
53
53
|
# limits are enforced based on the number of pre-filtered rows, so some
|
54
54
|
# filters can lead to lopsided assignments.
|
55
55
|
#
|
56
|
-
# Read sessions automatically expire
|
56
|
+
# Read sessions automatically expire 6 hours after they are created and do
|
57
57
|
# not require manual clean-up by the caller.
|
58
58
|
rpc :CreateReadSession, ::Google::Cloud::Bigquery::Storage::V1::CreateReadSessionRequest, ::Google::Cloud::Bigquery::Storage::V1::ReadSession
|
59
59
|
# Reads rows from the stream in the format prescribed by the ReadSession.
|
@@ -79,6 +79,85 @@ module Google
|
|
79
79
|
rpc :SplitReadStream, ::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamRequest, ::Google::Cloud::Bigquery::Storage::V1::SplitReadStreamResponse
|
80
80
|
end
|
81
81
|
|
82
|
+
Stub = Service.rpc_stub_class
|
83
|
+
end
|
84
|
+
module BigQueryWrite
|
85
|
+
# BigQuery Write API.
|
86
|
+
#
|
87
|
+
# The Write API can be used to write data to BigQuery.
|
88
|
+
#
|
89
|
+
# For supplementary information about the Write API, see:
|
90
|
+
# https://cloud.google.com/bigquery/docs/write-api
|
91
|
+
class Service
|
92
|
+
|
93
|
+
include ::GRPC::GenericService
|
94
|
+
|
95
|
+
self.marshal_class_method = :encode
|
96
|
+
self.unmarshal_class_method = :decode
|
97
|
+
self.service_name = 'google.cloud.bigquery.storage.v1.BigQueryWrite'
|
98
|
+
|
99
|
+
# Creates a write stream to the given table.
|
100
|
+
# Additionally, every table has a special stream named '_default'
|
101
|
+
# to which data can be written. This stream doesn't need to be created using
|
102
|
+
# CreateWriteStream. It is a stream that can be used simultaneously by any
|
103
|
+
# number of clients. Data written to this stream is considered committed as
|
104
|
+
# soon as an acknowledgement is received.
|
105
|
+
rpc :CreateWriteStream, ::Google::Cloud::Bigquery::Storage::V1::CreateWriteStreamRequest, ::Google::Cloud::Bigquery::Storage::V1::WriteStream
|
106
|
+
# Appends data to the given stream.
|
107
|
+
#
|
108
|
+
# If `offset` is specified, the `offset` is checked against the end of
|
109
|
+
# stream. The server returns `OUT_OF_RANGE` in `AppendRowsResponse` if an
|
110
|
+
# attempt is made to append to an offset beyond the current end of the stream
|
111
|
+
# or `ALREADY_EXISTS` if user provides an `offset` that has already been
|
112
|
+
# written to. User can retry with adjusted offset within the same RPC
|
113
|
+
# connection. If `offset` is not specified, append happens at the end of the
|
114
|
+
# stream.
|
115
|
+
#
|
116
|
+
# The response contains an optional offset at which the append
|
117
|
+
# happened. No offset information will be returned for appends to a
|
118
|
+
# default stream.
|
119
|
+
#
|
120
|
+
# Responses are received in the same order in which requests are sent.
|
121
|
+
# There will be one response for each successful inserted request. Responses
|
122
|
+
# may optionally embed error information if the originating AppendRequest was
|
123
|
+
# not successfully processed.
|
124
|
+
#
|
125
|
+
# The specifics of when successfully appended data is made visible to the
|
126
|
+
# table are governed by the type of stream:
|
127
|
+
#
|
128
|
+
# * For COMMITTED streams (which includes the default stream), data is
|
129
|
+
# visible immediately upon successful append.
|
130
|
+
#
|
131
|
+
# * For BUFFERED streams, data is made visible via a subsequent `FlushRows`
|
132
|
+
# rpc which advances a cursor to a newer offset in the stream.
|
133
|
+
#
|
134
|
+
# * For PENDING streams, data is not made visible until the stream itself is
|
135
|
+
# finalized (via the `FinalizeWriteStream` rpc), and the stream is explicitly
|
136
|
+
# committed via the `BatchCommitWriteStreams` rpc.
|
137
|
+
rpc :AppendRows, stream(::Google::Cloud::Bigquery::Storage::V1::AppendRowsRequest), stream(::Google::Cloud::Bigquery::Storage::V1::AppendRowsResponse)
|
138
|
+
# Gets information about a write stream.
|
139
|
+
rpc :GetWriteStream, ::Google::Cloud::Bigquery::Storage::V1::GetWriteStreamRequest, ::Google::Cloud::Bigquery::Storage::V1::WriteStream
|
140
|
+
# Finalize a write stream so that no new data can be appended to the
|
141
|
+
# stream. Finalize is not supported on the '_default' stream.
|
142
|
+
rpc :FinalizeWriteStream, ::Google::Cloud::Bigquery::Storage::V1::FinalizeWriteStreamRequest, ::Google::Cloud::Bigquery::Storage::V1::FinalizeWriteStreamResponse
|
143
|
+
# Atomically commits a group of `PENDING` streams that belong to the same
|
144
|
+
# `parent` table.
|
145
|
+
#
|
146
|
+
# Streams must be finalized before commit and cannot be committed multiple
|
147
|
+
# times. Once a stream is committed, data in the stream becomes available
|
148
|
+
# for read operations.
|
149
|
+
rpc :BatchCommitWriteStreams, ::Google::Cloud::Bigquery::Storage::V1::BatchCommitWriteStreamsRequest, ::Google::Cloud::Bigquery::Storage::V1::BatchCommitWriteStreamsResponse
|
150
|
+
# Flushes rows to a BUFFERED stream.
|
151
|
+
#
|
152
|
+
# If users are appending rows to BUFFERED stream, flush operation is
|
153
|
+
# required in order for the rows to become available for reading. A
|
154
|
+
# Flush operation flushes up to any previously flushed offset in a BUFFERED
|
155
|
+
# stream, to the offset specified in the request.
|
156
|
+
#
|
157
|
+
# Flush is not supported on the _default stream, since it is not BUFFERED.
|
158
|
+
rpc :FlushRows, ::Google::Cloud::Bigquery::Storage::V1::FlushRowsRequest, ::Google::Cloud::Bigquery::Storage::V1::FlushRowsResponse
|
159
|
+
end
|
160
|
+
|
82
161
|
Stub = Service.rpc_stub_class
|
83
162
|
end
|
84
163
|
end
|
@@ -1,13 +1,14 @@
|
|
1
1
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
2
|
# source: google/cloud/bigquery/storage/v1/stream.proto
|
3
3
|
|
4
|
-
require 'google/protobuf'
|
5
|
-
|
6
4
|
require 'google/api/field_behavior_pb'
|
7
5
|
require 'google/api/resource_pb'
|
8
6
|
require 'google/cloud/bigquery/storage/v1/arrow_pb'
|
9
7
|
require 'google/cloud/bigquery/storage/v1/avro_pb'
|
8
|
+
require 'google/cloud/bigquery/storage/v1/table_pb'
|
10
9
|
require 'google/protobuf/timestamp_pb'
|
10
|
+
require 'google/protobuf'
|
11
|
+
|
11
12
|
Google::Protobuf::DescriptorPool.generated_pool.build do
|
12
13
|
add_file("google/cloud/bigquery/storage/v1/stream.proto", :syntax => :proto3) do
|
13
14
|
add_message "google.cloud.bigquery.storage.v1.ReadSession" do
|
@@ -18,6 +19,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
18
19
|
optional :table_modifiers, :message, 7, "google.cloud.bigquery.storage.v1.ReadSession.TableModifiers"
|
19
20
|
optional :read_options, :message, 8, "google.cloud.bigquery.storage.v1.ReadSession.TableReadOptions"
|
20
21
|
repeated :streams, :message, 10, "google.cloud.bigquery.storage.v1.ReadStream"
|
22
|
+
optional :estimated_total_bytes_scanned, :int64, 12
|
21
23
|
oneof :schema do
|
22
24
|
optional :avro_schema, :message, 4, "google.cloud.bigquery.storage.v1.AvroSchema"
|
23
25
|
optional :arrow_schema, :message, 5, "google.cloud.bigquery.storage.v1.ArrowSchema"
|
@@ -36,6 +38,19 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
36
38
|
add_message "google.cloud.bigquery.storage.v1.ReadStream" do
|
37
39
|
optional :name, :string, 1
|
38
40
|
end
|
41
|
+
add_message "google.cloud.bigquery.storage.v1.WriteStream" do
|
42
|
+
optional :name, :string, 1
|
43
|
+
optional :type, :enum, 2, "google.cloud.bigquery.storage.v1.WriteStream.Type"
|
44
|
+
optional :create_time, :message, 3, "google.protobuf.Timestamp"
|
45
|
+
optional :commit_time, :message, 4, "google.protobuf.Timestamp"
|
46
|
+
optional :table_schema, :message, 5, "google.cloud.bigquery.storage.v1.TableSchema"
|
47
|
+
end
|
48
|
+
add_enum "google.cloud.bigquery.storage.v1.WriteStream.Type" do
|
49
|
+
value :TYPE_UNSPECIFIED, 0
|
50
|
+
value :COMMITTED, 1
|
51
|
+
value :PENDING, 2
|
52
|
+
value :BUFFERED, 3
|
53
|
+
end
|
39
54
|
add_enum "google.cloud.bigquery.storage.v1.DataFormat" do
|
40
55
|
value :DATA_FORMAT_UNSPECIFIED, 0
|
41
56
|
value :AVRO, 1
|
@@ -53,6 +68,8 @@ module Google
|
|
53
68
|
ReadSession::TableModifiers = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ReadSession.TableModifiers").msgclass
|
54
69
|
ReadSession::TableReadOptions = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ReadSession.TableReadOptions").msgclass
|
55
70
|
ReadStream = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.ReadStream").msgclass
|
71
|
+
WriteStream = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.WriteStream").msgclass
|
72
|
+
WriteStream::Type = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.WriteStream.Type").enummodule
|
56
73
|
DataFormat = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.DataFormat").enummodule
|
57
74
|
end
|
58
75
|
end
|
@@ -0,0 +1,62 @@
|
|
1
|
+
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
|
+
# source: google/cloud/bigquery/storage/v1/table.proto
|
3
|
+
|
4
|
+
require 'google/api/field_behavior_pb'
|
5
|
+
require 'google/protobuf'
|
6
|
+
|
7
|
+
Google::Protobuf::DescriptorPool.generated_pool.build do
|
8
|
+
add_file("google/cloud/bigquery/storage/v1/table.proto", :syntax => :proto3) do
|
9
|
+
add_message "google.cloud.bigquery.storage.v1.TableSchema" do
|
10
|
+
repeated :fields, :message, 1, "google.cloud.bigquery.storage.v1.TableFieldSchema"
|
11
|
+
end
|
12
|
+
add_message "google.cloud.bigquery.storage.v1.TableFieldSchema" do
|
13
|
+
optional :name, :string, 1
|
14
|
+
optional :type, :enum, 2, "google.cloud.bigquery.storage.v1.TableFieldSchema.Type"
|
15
|
+
optional :mode, :enum, 3, "google.cloud.bigquery.storage.v1.TableFieldSchema.Mode"
|
16
|
+
repeated :fields, :message, 4, "google.cloud.bigquery.storage.v1.TableFieldSchema"
|
17
|
+
optional :description, :string, 6
|
18
|
+
optional :max_length, :int64, 7
|
19
|
+
optional :precision, :int64, 8
|
20
|
+
optional :scale, :int64, 9
|
21
|
+
end
|
22
|
+
add_enum "google.cloud.bigquery.storage.v1.TableFieldSchema.Type" do
|
23
|
+
value :TYPE_UNSPECIFIED, 0
|
24
|
+
value :STRING, 1
|
25
|
+
value :INT64, 2
|
26
|
+
value :DOUBLE, 3
|
27
|
+
value :STRUCT, 4
|
28
|
+
value :BYTES, 5
|
29
|
+
value :BOOL, 6
|
30
|
+
value :TIMESTAMP, 7
|
31
|
+
value :DATE, 8
|
32
|
+
value :TIME, 9
|
33
|
+
value :DATETIME, 10
|
34
|
+
value :GEOGRAPHY, 11
|
35
|
+
value :NUMERIC, 12
|
36
|
+
value :BIGNUMERIC, 13
|
37
|
+
value :INTERVAL, 14
|
38
|
+
value :JSON, 15
|
39
|
+
end
|
40
|
+
add_enum "google.cloud.bigquery.storage.v1.TableFieldSchema.Mode" do
|
41
|
+
value :MODE_UNSPECIFIED, 0
|
42
|
+
value :NULLABLE, 1
|
43
|
+
value :REQUIRED, 2
|
44
|
+
value :REPEATED, 3
|
45
|
+
end
|
46
|
+
end
|
47
|
+
end
|
48
|
+
|
49
|
+
module Google
|
50
|
+
module Cloud
|
51
|
+
module Bigquery
|
52
|
+
module Storage
|
53
|
+
module V1
|
54
|
+
TableSchema = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.TableSchema").msgclass
|
55
|
+
TableFieldSchema = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.TableFieldSchema").msgclass
|
56
|
+
TableFieldSchema::Type = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.TableFieldSchema.Type").enummodule
|
57
|
+
TableFieldSchema::Mode = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.storage.v1.TableFieldSchema.Mode").enummodule
|
58
|
+
end
|
59
|
+
end
|
60
|
+
end
|
61
|
+
end
|
62
|
+
end
|
@@ -0,0 +1,56 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2021 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module Bigquery
|
23
|
+
module Storage
|
24
|
+
module V1
|
25
|
+
# ProtoSchema describes the schema of the serialized protocol buffer data rows.
|
26
|
+
# @!attribute [rw] proto_descriptor
|
27
|
+
# @return [::Google::Protobuf::DescriptorProto]
|
28
|
+
# Descriptor for input message. The provided descriptor must be self
|
29
|
+
# contained, such that data rows sent can be fully decoded using only the
|
30
|
+
# single descriptor. For data rows that are compositions of multiple
|
31
|
+
# independent messages, this means the descriptor may need to be transformed
|
32
|
+
# to only use nested types:
|
33
|
+
# https://developers.google.com/protocol-buffers/docs/proto#nested
|
34
|
+
#
|
35
|
+
# For additional information for how proto types and values map onto BigQuery
|
36
|
+
# see: https://cloud.google.com/bigquery/docs/write-api#data_type_conversions
|
37
|
+
class ProtoSchema
|
38
|
+
include ::Google::Protobuf::MessageExts
|
39
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
40
|
+
end
|
41
|
+
|
42
|
+
# @!attribute [rw] serialized_rows
|
43
|
+
# @return [::Array<::String>]
|
44
|
+
# A sequence of rows serialized as a Protocol Buffer.
|
45
|
+
#
|
46
|
+
# See https://developers.google.com/protocol-buffers/docs/overview for more
|
47
|
+
# information on deserializing this field.
|
48
|
+
class ProtoRows
|
49
|
+
include ::Google::Protobuf::MessageExts
|
50
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
51
|
+
end
|
52
|
+
end
|
53
|
+
end
|
54
|
+
end
|
55
|
+
end
|
56
|
+
end
|