google-cloud-bigquery-data_transfer 0.2.3 → 0.2.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.yardopts +2 -0
- data/AUTHENTICATION.md +199 -0
- data/README.md +3 -3
- data/lib/google/cloud/bigquery/data_transfer.rb +7 -7
- data/lib/google/cloud/bigquery/data_transfer/v1.rb +7 -7
- data/lib/google/cloud/bigquery/data_transfer/v1/credentials.rb +1 -1
- data/lib/google/cloud/bigquery/data_transfer/v1/data_transfer_service_client.rb +58 -58
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/cloud/bigquery/datatransfer/v1/datatransfer.rb +1 -1
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/cloud/bigquery/datatransfer/v1/transfer.rb +14 -14
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/any.rb +3 -2
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/duration.rb +1 -1
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/empty.rb +1 -1
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/field_mask.rb +19 -27
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/struct.rb +1 -1
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/timestamp.rb +16 -14
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/rpc/status.rb +18 -15
- data/lib/google/cloud/bigquery/datatransfer/v1/datatransfer_services_pb.rb +3 -3
- metadata +6 -6
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: dc8f0001f8d494c4c0caa52faa1365faf4eac9f526284a1ce5dad1d371714cbb
|
4
|
+
data.tar.gz: 65389fd9f60a871f1f1e89d703d425d3d8ed732bd2d561be2da56cf803d98389
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 0147e39b896cd0d646b661480303fbae70d2b6de5b39b1509a167712155b139cff4b44cf339d11a0c242bc5d8f625b219dedeae78d3c03b33822e26fa1a49d7f
|
7
|
+
data.tar.gz: 8061c0b0292be433bc6455327c0c9c9fec477dc9f21c850ad1c50586c25de1a2b3efbbaed7642fb434937fc9d002c0c34a48f125c65cfbff95b24b2c7f15a074
|
data/.yardopts
CHANGED
data/AUTHENTICATION.md
ADDED
@@ -0,0 +1,199 @@
|
|
1
|
+
# Authentication
|
2
|
+
|
3
|
+
In general, the google-cloud-bigquery-data_transfer library uses [Service
|
4
|
+
Account](https://cloud.google.com/iam/docs/creating-managing-service-accounts)
|
5
|
+
credentials to connect to Google Cloud services. When running within [Google
|
6
|
+
Cloud Platform environments](#google-cloud-platform-environments)
|
7
|
+
the credentials will be discovered automatically. When running on other
|
8
|
+
environments, the Service Account credentials can be specified by providing the
|
9
|
+
path to the [JSON
|
10
|
+
keyfile](https://cloud.google.com/iam/docs/managing-service-account-keys) for
|
11
|
+
the account (or the JSON itself) in [environment
|
12
|
+
variables](#environment-variables). Additionally, Cloud SDK credentials can also
|
13
|
+
be discovered automatically, but this is only recommended during development.
|
14
|
+
|
15
|
+
## Quickstart
|
16
|
+
|
17
|
+
1. [Create a service account and credentials](#creating-a-service-account).
|
18
|
+
2. Set the [environment variable](#environment-variables).
|
19
|
+
|
20
|
+
```sh
|
21
|
+
export DATA_TRANSFER_CREDENTIALS=/path/to/json`
|
22
|
+
```
|
23
|
+
|
24
|
+
3. Initialize the client.
|
25
|
+
|
26
|
+
```ruby
|
27
|
+
require "google/cloud/bigquery/data_transfer"
|
28
|
+
|
29
|
+
client = Google::Cloud::Bigquery::DataTransfer.new
|
30
|
+
```
|
31
|
+
|
32
|
+
## Project and Credential Lookup
|
33
|
+
|
34
|
+
The google-cloud-bigquery-data_transfer library aims to make authentication
|
35
|
+
as simple as possible, and provides several mechanisms to configure your system
|
36
|
+
without providing **Project ID** and **Service Account Credentials** directly in
|
37
|
+
code.
|
38
|
+
|
39
|
+
**Project ID** is discovered in the following order:
|
40
|
+
|
41
|
+
1. Specify project ID in method arguments
|
42
|
+
2. Specify project ID in configuration
|
43
|
+
3. Discover project ID in environment variables
|
44
|
+
4. Discover GCE project ID
|
45
|
+
5. Discover project ID in credentials JSON
|
46
|
+
|
47
|
+
**Credentials** are discovered in the following order:
|
48
|
+
|
49
|
+
1. Specify credentials in method arguments
|
50
|
+
2. Specify credentials in configuration
|
51
|
+
3. Discover credentials path in environment variables
|
52
|
+
4. Discover credentials JSON in environment variables
|
53
|
+
5. Discover credentials file in the Cloud SDK's path
|
54
|
+
6. Discover GCE credentials
|
55
|
+
|
56
|
+
### Google Cloud Platform environments
|
57
|
+
|
58
|
+
While running on Google Cloud Platform environments such as Google Compute
|
59
|
+
Engine, Google App Engine and Google Kubernetes Engine, no extra work is needed.
|
60
|
+
The **Project ID** and **Credentials** and are discovered automatically. Code
|
61
|
+
should be written as if already authenticated. Just be sure when you [set up the
|
62
|
+
GCE instance][gce-how-to], you add the correct scopes for the APIs you want to
|
63
|
+
access. For example:
|
64
|
+
|
65
|
+
* **All APIs**
|
66
|
+
* `https://www.googleapis.com/auth/cloud-platform`
|
67
|
+
* `https://www.googleapis.com/auth/cloud-platform.read-only`
|
68
|
+
* **BigQuery**
|
69
|
+
* `https://www.googleapis.com/auth/bigquery`
|
70
|
+
* `https://www.googleapis.com/auth/bigquery.insertdata`
|
71
|
+
* **Compute Engine**
|
72
|
+
* `https://www.googleapis.com/auth/compute`
|
73
|
+
* **Datastore**
|
74
|
+
* `https://www.googleapis.com/auth/datastore`
|
75
|
+
* `https://www.googleapis.com/auth/userinfo.email`
|
76
|
+
* **DNS**
|
77
|
+
* `https://www.googleapis.com/auth/ndev.clouddns.readwrite`
|
78
|
+
* **Pub/Sub**
|
79
|
+
* `https://www.googleapis.com/auth/pubsub`
|
80
|
+
* **Storage**
|
81
|
+
* `https://www.googleapis.com/auth/devstorage.full_control`
|
82
|
+
* `https://www.googleapis.com/auth/devstorage.read_only`
|
83
|
+
* `https://www.googleapis.com/auth/devstorage.read_write`
|
84
|
+
|
85
|
+
### Environment Variables
|
86
|
+
|
87
|
+
The **Project ID** and **Credentials JSON** can be placed in environment
|
88
|
+
variables instead of declaring them directly in code. Each service has its own
|
89
|
+
environment variable, allowing for different service accounts to be used for
|
90
|
+
different services. (See the READMEs for the individual service gems for
|
91
|
+
details.) The path to the **Credentials JSON** file can be stored in the
|
92
|
+
environment variable, or the **Credentials JSON** itself can be stored for
|
93
|
+
environments such as Docker containers where writing files is difficult or not
|
94
|
+
encouraged.
|
95
|
+
|
96
|
+
The environment variables that google-cloud-bigquery-data_transfer checks for project ID are:
|
97
|
+
|
98
|
+
1. `DATA_TRANSFER_PROJECT`
|
99
|
+
2. `GOOGLE_CLOUD_PROJECT`
|
100
|
+
|
101
|
+
The environment variables that google-cloud-bigquery-data_transfer checks for credentials are configured on {Google::Cloud::Bigquery::DataTransfer::V1::Credentials}:
|
102
|
+
|
103
|
+
1. `DATA_TRANSFER_CREDENTIALS` - Path to JSON file, or JSON contents
|
104
|
+
2. `DATA_TRANSFER_KEYFILE` - Path to JSON file, or JSON contents
|
105
|
+
3. `GOOGLE_CLOUD_CREDENTIALS` - Path to JSON file, or JSON contents
|
106
|
+
4. `GOOGLE_CLOUD_KEYFILE` - Path to JSON file, or JSON contents
|
107
|
+
5. `GOOGLE_APPLICATION_CREDENTIALS` - Path to JSON file
|
108
|
+
|
109
|
+
```ruby
|
110
|
+
require "google/cloud/bigquery/data_transfer"
|
111
|
+
|
112
|
+
ENV["DATA_TRANSFER_PROJECT"] = "my-project-id"
|
113
|
+
ENV["DATA_TRANSFER_CREDENTIALS"] = "path/to/keyfile.json"
|
114
|
+
|
115
|
+
client = Google::Cloud::Bigquery::DataTransfer.new
|
116
|
+
```
|
117
|
+
|
118
|
+
### Configuration
|
119
|
+
|
120
|
+
The **Project ID** and **Credentials JSON** can be configured instead of placing them in environment variables or providing them as arguments.
|
121
|
+
|
122
|
+
```ruby
|
123
|
+
require "google/cloud/bigquery/data_transfer"
|
124
|
+
|
125
|
+
Google::Cloud::Bigquery::DataTransfer.configure do |config|
|
126
|
+
config.project_id = "my-project-id"
|
127
|
+
config.credentials = "path/to/keyfile.json"
|
128
|
+
end
|
129
|
+
|
130
|
+
client = Google::Cloud::Bigquery::DataTransfer.new
|
131
|
+
```
|
132
|
+
|
133
|
+
### Cloud SDK
|
134
|
+
|
135
|
+
This option allows for an easy way to authenticate during development. If
|
136
|
+
credentials are not provided in code or in environment variables, then Cloud SDK
|
137
|
+
credentials are discovered.
|
138
|
+
|
139
|
+
To configure your system for this, simply:
|
140
|
+
|
141
|
+
1. [Download and install the Cloud SDK](https://cloud.google.com/sdk)
|
142
|
+
2. Authenticate using OAuth 2.0 `$ gcloud auth login`
|
143
|
+
3. Write code as if already authenticated.
|
144
|
+
|
145
|
+
**NOTE:** This is _not_ recommended for running in production. The Cloud SDK
|
146
|
+
*should* only be used during development.
|
147
|
+
|
148
|
+
[gce-how-to]: https://cloud.google.com/compute/docs/authentication#using
|
149
|
+
[dev-console]: https://console.cloud.google.com/project
|
150
|
+
|
151
|
+
[enable-apis]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/enable-apis.png
|
152
|
+
|
153
|
+
[create-new-service-account]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/create-new-service-account.png
|
154
|
+
[create-new-service-account-existing-keys]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/create-new-service-account-existing-keys.png
|
155
|
+
[reuse-service-account]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/reuse-service-account.png
|
156
|
+
|
157
|
+
## Creating a Service Account
|
158
|
+
|
159
|
+
Google Cloud requires a **Project ID** and **Service Account Credentials** to
|
160
|
+
connect to the APIs. You will use the **Project ID** and **JSON key file** to
|
161
|
+
connect to most services with google-cloud-bigquery-data_transfer.
|
162
|
+
|
163
|
+
If you are not running this client within [Google Cloud Platform
|
164
|
+
environments](#google-cloud-platform-environments), you need a Google
|
165
|
+
Developers service account.
|
166
|
+
|
167
|
+
1. Visit the [Google Developers Console][dev-console].
|
168
|
+
1. Create a new project or click on an existing project.
|
169
|
+
1. Activate the slide-out navigation tray and select **API Manager**. From
|
170
|
+
here, you will enable the APIs that your application requires.
|
171
|
+
|
172
|
+
![Enable the APIs that your application requires][enable-apis]
|
173
|
+
|
174
|
+
*Note: You may need to enable billing in order to use these services.*
|
175
|
+
|
176
|
+
1. Select **Credentials** from the side navigation.
|
177
|
+
|
178
|
+
You should see a screen like one of the following.
|
179
|
+
|
180
|
+
![Create a new service account][create-new-service-account]
|
181
|
+
|
182
|
+
![Create a new service account With Existing Keys][create-new-service-account-existing-keys]
|
183
|
+
|
184
|
+
Find the "Add credentials" drop down and select "Service account" to be
|
185
|
+
guided through downloading a new JSON key file.
|
186
|
+
|
187
|
+
If you want to re-use an existing service account, you can easily generate a
|
188
|
+
new key file. Just select the account you wish to re-use, and click "Generate
|
189
|
+
new JSON key":
|
190
|
+
|
191
|
+
![Re-use an existing service account][reuse-service-account]
|
192
|
+
|
193
|
+
The key file you download will be used by this library to authenticate API
|
194
|
+
requests and should be stored in a secure location.
|
195
|
+
|
196
|
+
## Troubleshooting
|
197
|
+
|
198
|
+
If you're having trouble authenticating you can ask for help by following the
|
199
|
+
{file:TROUBLESHOOTING.md Troubleshooting Guide}.
|
data/README.md
CHANGED
@@ -25,16 +25,16 @@ $ gem install google-cloud-bigquery-data_transfer
|
|
25
25
|
```rb
|
26
26
|
require "google/cloud/bigquery/data_transfer"
|
27
27
|
|
28
|
-
|
28
|
+
data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new
|
29
29
|
formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_path(project_id)
|
30
30
|
|
31
31
|
# Iterate over all results.
|
32
|
-
|
32
|
+
data_transfer_client.list_data_sources(formatted_parent).each do |element|
|
33
33
|
# Process element.
|
34
34
|
end
|
35
35
|
|
36
36
|
# Or iterate over results one page at a time.
|
37
|
-
|
37
|
+
data_transfer_client.list_data_sources(formatted_parent).each_page do |page|
|
38
38
|
# Process each page at a time.
|
39
39
|
page.each do |element|
|
40
40
|
# Process element.
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -48,16 +48,16 @@ module Google
|
|
48
48
|
# ```rb
|
49
49
|
# require "google/cloud/bigquery/data_transfer"
|
50
50
|
#
|
51
|
-
#
|
51
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new
|
52
52
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_path(project_id)
|
53
53
|
#
|
54
54
|
# # Iterate over all results.
|
55
|
-
#
|
55
|
+
# data_transfer_client.list_data_sources(formatted_parent).each do |element|
|
56
56
|
# # Process element.
|
57
57
|
# end
|
58
58
|
#
|
59
59
|
# # Or iterate over results one page at a time.
|
60
|
-
#
|
60
|
+
# data_transfer_client.list_data_sources(formatted_parent).each_page do |page|
|
61
61
|
# # Process each page at a time.
|
62
62
|
# page.each do |element|
|
63
63
|
# # Process element.
|
@@ -112,9 +112,9 @@ module Google
|
|
112
112
|
|
113
113
|
##
|
114
114
|
# The Google BigQuery Data Transfer Service API enables BigQuery users to
|
115
|
-
# configure the transfer of their data from other Google Products into
|
116
|
-
# This service contains methods that are end user exposed. It backs
|
117
|
-
# frontend.
|
115
|
+
# configure the transfer of their data from other Google Products into
|
116
|
+
# BigQuery. This service contains methods that are end user exposed. It backs
|
117
|
+
# up the frontend.
|
118
118
|
#
|
119
119
|
# @param version [Symbol, String]
|
120
120
|
# The major version of the service to be used. By default :v1
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -48,16 +48,16 @@ module Google
|
|
48
48
|
# ```rb
|
49
49
|
# require "google/cloud/bigquery/data_transfer"
|
50
50
|
#
|
51
|
-
#
|
51
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
52
52
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_path(project_id)
|
53
53
|
#
|
54
54
|
# # Iterate over all results.
|
55
|
-
#
|
55
|
+
# data_transfer_client.list_data_sources(formatted_parent).each do |element|
|
56
56
|
# # Process element.
|
57
57
|
# end
|
58
58
|
#
|
59
59
|
# # Or iterate over results one page at a time.
|
60
|
-
#
|
60
|
+
# data_transfer_client.list_data_sources(formatted_parent).each_page do |page|
|
61
61
|
# # Process each page at a time.
|
62
62
|
# page.each do |element|
|
63
63
|
# # Process element.
|
@@ -104,9 +104,9 @@ module Google
|
|
104
104
|
|
105
105
|
##
|
106
106
|
# The Google BigQuery Data Transfer Service API enables BigQuery users to
|
107
|
-
# configure the transfer of their data from other Google Products into
|
108
|
-
# This service contains methods that are end user exposed. It backs
|
109
|
-
# frontend.
|
107
|
+
# configure the transfer of their data from other Google Products into
|
108
|
+
# BigQuery. This service contains methods that are end user exposed. It backs
|
109
|
+
# up the frontend.
|
110
110
|
#
|
111
111
|
# @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
|
112
112
|
# Provides the means for authenticating requests made by the client. This parameter can
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -34,9 +34,9 @@ module Google
|
|
34
34
|
module DataTransfer
|
35
35
|
module V1
|
36
36
|
# The Google BigQuery Data Transfer Service API enables BigQuery users to
|
37
|
-
# configure the transfer of their data from other Google Products into
|
38
|
-
# This service contains methods that are end user exposed. It backs
|
39
|
-
# frontend.
|
37
|
+
# configure the transfer of their data from other Google Products into
|
38
|
+
# BigQuery. This service contains methods that are end user exposed. It backs
|
39
|
+
# up the frontend.
|
40
40
|
#
|
41
41
|
# @!attribute [r] data_transfer_service_stub
|
42
42
|
# @return [Google::Cloud::Bigquery::Datatransfer::V1::DataTransferService::Stub]
|
@@ -83,23 +83,17 @@ module Google
|
|
83
83
|
].freeze
|
84
84
|
|
85
85
|
|
86
|
-
PROJECT_DATA_SOURCE_PATH_TEMPLATE = Google::Gax::PathTemplate.new(
|
87
|
-
"projects/{project}/dataSources/{data_source}"
|
88
|
-
)
|
89
|
-
|
90
|
-
private_constant :PROJECT_DATA_SOURCE_PATH_TEMPLATE
|
91
|
-
|
92
86
|
PROJECT_PATH_TEMPLATE = Google::Gax::PathTemplate.new(
|
93
87
|
"projects/{project}"
|
94
88
|
)
|
95
89
|
|
96
90
|
private_constant :PROJECT_PATH_TEMPLATE
|
97
91
|
|
98
|
-
|
99
|
-
"projects/{project}/
|
92
|
+
PROJECT_DATA_SOURCE_PATH_TEMPLATE = Google::Gax::PathTemplate.new(
|
93
|
+
"projects/{project}/dataSources/{data_source}"
|
100
94
|
)
|
101
95
|
|
102
|
-
private_constant :
|
96
|
+
private_constant :PROJECT_DATA_SOURCE_PATH_TEMPLATE
|
103
97
|
|
104
98
|
PROJECT_RUN_PATH_TEMPLATE = Google::Gax::PathTemplate.new(
|
105
99
|
"projects/{project}/transferConfigs/{transfer_config}/runs/{run}"
|
@@ -107,16 +101,11 @@ module Google
|
|
107
101
|
|
108
102
|
private_constant :PROJECT_RUN_PATH_TEMPLATE
|
109
103
|
|
110
|
-
|
111
|
-
|
112
|
-
|
113
|
-
|
114
|
-
|
115
|
-
PROJECT_DATA_SOURCE_PATH_TEMPLATE.render(
|
116
|
-
:"project" => project,
|
117
|
-
:"data_source" => data_source
|
118
|
-
)
|
119
|
-
end
|
104
|
+
PROJECT_TRANSFER_CONFIG_PATH_TEMPLATE = Google::Gax::PathTemplate.new(
|
105
|
+
"projects/{project}/transferConfigs/{transfer_config}"
|
106
|
+
)
|
107
|
+
|
108
|
+
private_constant :PROJECT_TRANSFER_CONFIG_PATH_TEMPLATE
|
120
109
|
|
121
110
|
# Returns a fully-qualified project resource name string.
|
122
111
|
# @param project [String]
|
@@ -127,14 +116,14 @@ module Google
|
|
127
116
|
)
|
128
117
|
end
|
129
118
|
|
130
|
-
# Returns a fully-qualified
|
119
|
+
# Returns a fully-qualified project_data_source resource name string.
|
131
120
|
# @param project [String]
|
132
|
-
# @param
|
121
|
+
# @param data_source [String]
|
133
122
|
# @return [String]
|
134
|
-
def self.
|
135
|
-
|
123
|
+
def self.project_data_source_path project, data_source
|
124
|
+
PROJECT_DATA_SOURCE_PATH_TEMPLATE.render(
|
136
125
|
:"project" => project,
|
137
|
-
:"
|
126
|
+
:"data_source" => data_source
|
138
127
|
)
|
139
128
|
end
|
140
129
|
|
@@ -151,6 +140,17 @@ module Google
|
|
151
140
|
)
|
152
141
|
end
|
153
142
|
|
143
|
+
# Returns a fully-qualified project_transfer_config resource name string.
|
144
|
+
# @param project [String]
|
145
|
+
# @param transfer_config [String]
|
146
|
+
# @return [String]
|
147
|
+
def self.project_transfer_config_path project, transfer_config
|
148
|
+
PROJECT_TRANSFER_CONFIG_PATH_TEMPLATE.render(
|
149
|
+
:"project" => project,
|
150
|
+
:"transfer_config" => transfer_config
|
151
|
+
)
|
152
|
+
end
|
153
|
+
|
154
154
|
# @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
|
155
155
|
# Provides the means for authenticating requests made by the client. This parameter can
|
156
156
|
# be many types.
|
@@ -379,9 +379,9 @@ module Google
|
|
379
379
|
# @example
|
380
380
|
# require "google/cloud/bigquery/data_transfer"
|
381
381
|
#
|
382
|
-
#
|
382
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
383
383
|
# formatted_name = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_data_source_path("[PROJECT]", "[DATA_SOURCE]")
|
384
|
-
# response =
|
384
|
+
# response = data_transfer_client.get_data_source(formatted_name)
|
385
385
|
|
386
386
|
def get_data_source \
|
387
387
|
name,
|
@@ -421,16 +421,16 @@ module Google
|
|
421
421
|
# @example
|
422
422
|
# require "google/cloud/bigquery/data_transfer"
|
423
423
|
#
|
424
|
-
#
|
424
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
425
425
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_path("[PROJECT]")
|
426
426
|
#
|
427
427
|
# # Iterate over all results.
|
428
|
-
#
|
428
|
+
# data_transfer_client.list_data_sources(formatted_parent).each do |element|
|
429
429
|
# # Process element.
|
430
430
|
# end
|
431
431
|
#
|
432
432
|
# # Or iterate over results one page at a time.
|
433
|
-
#
|
433
|
+
# data_transfer_client.list_data_sources(formatted_parent).each_page do |page|
|
434
434
|
# # Process each page at a time.
|
435
435
|
# page.each do |element|
|
436
436
|
# # Process element.
|
@@ -489,12 +489,12 @@ module Google
|
|
489
489
|
# @example
|
490
490
|
# require "google/cloud/bigquery/data_transfer"
|
491
491
|
#
|
492
|
-
#
|
492
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
493
493
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_path("[PROJECT]")
|
494
494
|
#
|
495
495
|
# # TODO: Initialize `transfer_config`:
|
496
496
|
# transfer_config = {}
|
497
|
-
# response =
|
497
|
+
# response = data_transfer_client.create_transfer_config(formatted_parent, transfer_config)
|
498
498
|
|
499
499
|
def create_transfer_config \
|
500
500
|
parent,
|
@@ -550,14 +550,14 @@ module Google
|
|
550
550
|
# @example
|
551
551
|
# require "google/cloud/bigquery/data_transfer"
|
552
552
|
#
|
553
|
-
#
|
553
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
554
554
|
#
|
555
555
|
# # TODO: Initialize `transfer_config`:
|
556
556
|
# transfer_config = {}
|
557
557
|
#
|
558
558
|
# # TODO: Initialize `update_mask`:
|
559
559
|
# update_mask = {}
|
560
|
-
# response =
|
560
|
+
# response = data_transfer_client.update_transfer_config(transfer_config, update_mask)
|
561
561
|
|
562
562
|
def update_transfer_config \
|
563
563
|
transfer_config,
|
@@ -590,9 +590,9 @@ module Google
|
|
590
590
|
# @example
|
591
591
|
# require "google/cloud/bigquery/data_transfer"
|
592
592
|
#
|
593
|
-
#
|
593
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
594
594
|
# formatted_name = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_transfer_config_path("[PROJECT]", "[TRANSFER_CONFIG]")
|
595
|
-
#
|
595
|
+
# data_transfer_client.delete_transfer_config(formatted_name)
|
596
596
|
|
597
597
|
def delete_transfer_config \
|
598
598
|
name,
|
@@ -622,9 +622,9 @@ module Google
|
|
622
622
|
# @example
|
623
623
|
# require "google/cloud/bigquery/data_transfer"
|
624
624
|
#
|
625
|
-
#
|
625
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
626
626
|
# formatted_name = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_transfer_config_path("[PROJECT]", "[TRANSFER_CONFIG]")
|
627
|
-
# response =
|
627
|
+
# response = data_transfer_client.get_transfer_config(formatted_name)
|
628
628
|
|
629
629
|
def get_transfer_config \
|
630
630
|
name,
|
@@ -665,16 +665,16 @@ module Google
|
|
665
665
|
# @example
|
666
666
|
# require "google/cloud/bigquery/data_transfer"
|
667
667
|
#
|
668
|
-
#
|
668
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
669
669
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_path("[PROJECT]")
|
670
670
|
#
|
671
671
|
# # Iterate over all results.
|
672
|
-
#
|
672
|
+
# data_transfer_client.list_transfer_configs(formatted_parent).each do |element|
|
673
673
|
# # Process element.
|
674
674
|
# end
|
675
675
|
#
|
676
676
|
# # Or iterate over results one page at a time.
|
677
|
-
#
|
677
|
+
# data_transfer_client.list_transfer_configs(formatted_parent).each_page do |page|
|
678
678
|
# # Process each page at a time.
|
679
679
|
# page.each do |element|
|
680
680
|
# # Process element.
|
@@ -725,7 +725,7 @@ module Google
|
|
725
725
|
# @example
|
726
726
|
# require "google/cloud/bigquery/data_transfer"
|
727
727
|
#
|
728
|
-
#
|
728
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
729
729
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_transfer_config_path("[PROJECT]", "[TRANSFER_CONFIG]")
|
730
730
|
#
|
731
731
|
# # TODO: Initialize `start_time`:
|
@@ -733,7 +733,7 @@ module Google
|
|
733
733
|
#
|
734
734
|
# # TODO: Initialize `end_time`:
|
735
735
|
# end_time = {}
|
736
|
-
# response =
|
736
|
+
# response = data_transfer_client.schedule_transfer_runs(formatted_parent, start_time, end_time)
|
737
737
|
|
738
738
|
def schedule_transfer_runs \
|
739
739
|
parent,
|
@@ -766,9 +766,9 @@ module Google
|
|
766
766
|
# @example
|
767
767
|
# require "google/cloud/bigquery/data_transfer"
|
768
768
|
#
|
769
|
-
#
|
769
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
770
770
|
# formatted_name = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_run_path("[PROJECT]", "[TRANSFER_CONFIG]", "[RUN]")
|
771
|
-
# response =
|
771
|
+
# response = data_transfer_client.get_transfer_run(formatted_name)
|
772
772
|
|
773
773
|
def get_transfer_run \
|
774
774
|
name,
|
@@ -796,9 +796,9 @@ module Google
|
|
796
796
|
# @example
|
797
797
|
# require "google/cloud/bigquery/data_transfer"
|
798
798
|
#
|
799
|
-
#
|
799
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
800
800
|
# formatted_name = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_run_path("[PROJECT]", "[TRANSFER_CONFIG]", "[RUN]")
|
801
|
-
#
|
801
|
+
# data_transfer_client.delete_transfer_run(formatted_name)
|
802
802
|
|
803
803
|
def delete_transfer_run \
|
804
804
|
name,
|
@@ -843,16 +843,16 @@ module Google
|
|
843
843
|
# @example
|
844
844
|
# require "google/cloud/bigquery/data_transfer"
|
845
845
|
#
|
846
|
-
#
|
846
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
847
847
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_transfer_config_path("[PROJECT]", "[TRANSFER_CONFIG]")
|
848
848
|
#
|
849
849
|
# # Iterate over all results.
|
850
|
-
#
|
850
|
+
# data_transfer_client.list_transfer_runs(formatted_parent).each do |element|
|
851
851
|
# # Process element.
|
852
852
|
# end
|
853
853
|
#
|
854
854
|
# # Or iterate over results one page at a time.
|
855
|
-
#
|
855
|
+
# data_transfer_client.list_transfer_runs(formatted_parent).each_page do |page|
|
856
856
|
# # Process each page at a time.
|
857
857
|
# page.each do |element|
|
858
858
|
# # Process element.
|
@@ -905,16 +905,16 @@ module Google
|
|
905
905
|
# @example
|
906
906
|
# require "google/cloud/bigquery/data_transfer"
|
907
907
|
#
|
908
|
-
#
|
908
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
909
909
|
# formatted_parent = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_run_path("[PROJECT]", "[TRANSFER_CONFIG]", "[RUN]")
|
910
910
|
#
|
911
911
|
# # Iterate over all results.
|
912
|
-
#
|
912
|
+
# data_transfer_client.list_transfer_logs(formatted_parent).each do |element|
|
913
913
|
# # Process element.
|
914
914
|
# end
|
915
915
|
#
|
916
916
|
# # Or iterate over results one page at a time.
|
917
|
-
#
|
917
|
+
# data_transfer_client.list_transfer_logs(formatted_parent).each_page do |page|
|
918
918
|
# # Process each page at a time.
|
919
919
|
# page.each do |element|
|
920
920
|
# # Process element.
|
@@ -957,9 +957,9 @@ module Google
|
|
957
957
|
# @example
|
958
958
|
# require "google/cloud/bigquery/data_transfer"
|
959
959
|
#
|
960
|
-
#
|
960
|
+
# data_transfer_client = Google::Cloud::Bigquery::DataTransfer.new(version: :v1)
|
961
961
|
# formatted_name = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferServiceClient.project_data_source_path("[PROJECT]", "[DATA_SOURCE]")
|
962
|
-
# response =
|
962
|
+
# response = data_transfer_client.check_valid_creds(formatted_name)
|
963
963
|
|
964
964
|
def check_valid_creds \
|
965
965
|
name,
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -174,19 +174,6 @@ module Google
|
|
174
174
|
end
|
175
175
|
end
|
176
176
|
|
177
|
-
# DEPRECATED. Represents data transfer type.
|
178
|
-
module TransferType
|
179
|
-
# Invalid or Unknown transfer type placeholder.
|
180
|
-
TRANSFER_TYPE_UNSPECIFIED = 0
|
181
|
-
|
182
|
-
# Batch data transfer.
|
183
|
-
BATCH = 1
|
184
|
-
|
185
|
-
# Streaming data transfer. Streaming data source currently doesn't
|
186
|
-
# support multiple transfer configs per project.
|
187
|
-
STREAMING = 2
|
188
|
-
end
|
189
|
-
|
190
177
|
# Represents data transfer run state.
|
191
178
|
module TransferState
|
192
179
|
# State placeholder.
|
@@ -208,6 +195,19 @@ module Google
|
|
208
195
|
# Data transfer is cancelled.
|
209
196
|
CANCELLED = 6
|
210
197
|
end
|
198
|
+
|
199
|
+
# DEPRECATED. Represents data transfer type.
|
200
|
+
module TransferType
|
201
|
+
# Invalid or Unknown transfer type placeholder.
|
202
|
+
TRANSFER_TYPE_UNSPECIFIED = 0
|
203
|
+
|
204
|
+
# Batch data transfer.
|
205
|
+
BATCH = 1
|
206
|
+
|
207
|
+
# Streaming data transfer. Streaming data source currently doesn't
|
208
|
+
# support multiple transfer configs per project.
|
209
|
+
STREAMING = 2
|
210
|
+
end
|
211
211
|
end
|
212
212
|
end
|
213
213
|
end
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -97,7 +97,8 @@ module Google
|
|
97
97
|
# @!attribute [rw] type_url
|
98
98
|
# @return [String]
|
99
99
|
# A URL/resource name that uniquely identifies the type of the serialized
|
100
|
-
# protocol buffer message.
|
100
|
+
# protocol buffer message. This string must contain at least
|
101
|
+
# one "/" character. The last segment of the URL's path must represent
|
101
102
|
# the fully qualified name of the type (as in
|
102
103
|
# `path/google.protobuf.Duration`). The name should be in a canonical form
|
103
104
|
# (e.g., leading "." is not accepted).
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -83,57 +83,49 @@ module Google
|
|
83
83
|
# describe the updated values, the API ignores the values of all
|
84
84
|
# fields not covered by the mask.
|
85
85
|
#
|
86
|
-
# If a repeated field is specified for an update operation,
|
87
|
-
#
|
88
|
-
#
|
89
|
-
# string.
|
86
|
+
# If a repeated field is specified for an update operation, new values will
|
87
|
+
# be appended to the existing repeated field in the target resource. Note that
|
88
|
+
# a repeated field is only allowed in the last position of a `paths` string.
|
90
89
|
#
|
91
90
|
# If a sub-message is specified in the last position of the field mask for an
|
92
|
-
# update operation, then the existing sub-message
|
93
|
-
#
|
91
|
+
# update operation, then new value will be merged into the existing sub-message
|
92
|
+
# in the target resource.
|
93
|
+
#
|
94
|
+
# For example, given the target message:
|
94
95
|
#
|
95
96
|
# f {
|
96
97
|
# b {
|
97
|
-
# d
|
98
|
-
# x
|
98
|
+
# d: 1
|
99
|
+
# x: 2
|
99
100
|
# }
|
100
|
-
# c
|
101
|
+
# c: [1]
|
101
102
|
# }
|
102
103
|
#
|
103
104
|
# And an update message:
|
104
105
|
#
|
105
106
|
# f {
|
106
107
|
# b {
|
107
|
-
# d
|
108
|
+
# d: 10
|
108
109
|
# }
|
110
|
+
# c: [2]
|
109
111
|
# }
|
110
112
|
#
|
111
113
|
# then if the field mask is:
|
112
114
|
#
|
113
|
-
# paths: "f.b"
|
115
|
+
# paths: ["f.b", "f.c"]
|
114
116
|
#
|
115
117
|
# then the result will be:
|
116
118
|
#
|
117
119
|
# f {
|
118
120
|
# b {
|
119
|
-
# d
|
121
|
+
# d: 10
|
122
|
+
# x: 2
|
120
123
|
# }
|
121
|
-
# c
|
124
|
+
# c: [1, 2]
|
122
125
|
# }
|
123
126
|
#
|
124
|
-
#
|
125
|
-
#
|
126
|
-
# paths: "f.b.d"
|
127
|
-
#
|
128
|
-
# then the result would be:
|
129
|
-
#
|
130
|
-
# f {
|
131
|
-
# b {
|
132
|
-
# d : 10
|
133
|
-
# x : 2
|
134
|
-
# }
|
135
|
-
# c : 1
|
136
|
-
# }
|
127
|
+
# An implementation may provide options to override this default behavior for
|
128
|
+
# repeated and message fields.
|
137
129
|
#
|
138
130
|
# In order to reset a field's value to the default, the field must
|
139
131
|
# be in the mask and set to the default value in the provided resource.
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -15,17 +15,19 @@
|
|
15
15
|
|
16
16
|
module Google
|
17
17
|
module Protobuf
|
18
|
-
# A Timestamp represents a point in time independent of any time zone
|
19
|
-
#
|
20
|
-
# nanosecond resolution
|
21
|
-
#
|
22
|
-
# backwards to year one.
|
23
|
-
#
|
24
|
-
#
|
25
|
-
#
|
26
|
-
#
|
27
|
-
#
|
28
|
-
#
|
18
|
+
# A Timestamp represents a point in time independent of any time zone or local
|
19
|
+
# calendar, encoded as a count of seconds and fractions of seconds at
|
20
|
+
# nanosecond resolution. The count is relative to an epoch at UTC midnight on
|
21
|
+
# January 1, 1970, in the proleptic Gregorian calendar which extends the
|
22
|
+
# Gregorian calendar backwards to year one.
|
23
|
+
#
|
24
|
+
# All minutes are 60 seconds long. Leap seconds are "smeared" so that no leap
|
25
|
+
# second table is needed for interpretation, using a [24-hour linear
|
26
|
+
# smear](https://developers.google.com/time/smear).
|
27
|
+
#
|
28
|
+
# The range is from 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z. By
|
29
|
+
# restricting to that range, we ensure that we can convert to and from [RFC
|
30
|
+
# 3339](https://www.ietf.org/rfc/rfc3339.txt) date strings.
|
29
31
|
#
|
30
32
|
# = Examples
|
31
33
|
#
|
@@ -86,12 +88,12 @@ module Google
|
|
86
88
|
# 01:30 UTC on January 15, 2017.
|
87
89
|
#
|
88
90
|
# In JavaScript, one can convert a Date object to this format using the
|
89
|
-
# standard [toISOString()](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Date/toISOString
|
91
|
+
# standard [toISOString()](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Date/toISOString)
|
90
92
|
# method. In Python, a standard `datetime.datetime` object can be converted
|
91
93
|
# to this format using [`strftime`](https://docs.python.org/2/library/time.html#time.strftime)
|
92
94
|
# with the time format spec '%Y-%m-%dT%H:%M:%S.%fZ'. Likewise, in Java, one
|
93
95
|
# can use the Joda Time's [`ISODateTimeFormat.dateTime()`](
|
94
|
-
# http://www.joda.org/joda-time/apidocs/org/joda/time/format/ISODateTimeFormat.html#dateTime
|
96
|
+
# http://www.joda.org/joda-time/apidocs/org/joda/time/format/ISODateTimeFormat.html#dateTime%2D%2D
|
95
97
|
# ) to obtain a formatter capable of generating timestamps in this format.
|
96
98
|
# @!attribute [rw] seconds
|
97
99
|
# @return [Integer]
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -15,24 +15,25 @@
|
|
15
15
|
|
16
16
|
module Google
|
17
17
|
module Rpc
|
18
|
-
# The `Status` type defines a logical error model that is suitable for
|
19
|
-
# programming environments, including REST APIs and RPC APIs. It is
|
20
|
-
# [gRPC](https://github.com/grpc). The error model is designed to be:
|
18
|
+
# The `Status` type defines a logical error model that is suitable for
|
19
|
+
# different programming environments, including REST APIs and RPC APIs. It is
|
20
|
+
# used by [gRPC](https://github.com/grpc). The error model is designed to be:
|
21
21
|
#
|
22
22
|
# * Simple to use and understand for most users
|
23
23
|
# * Flexible enough to meet unexpected needs
|
24
24
|
#
|
25
25
|
# = Overview
|
26
26
|
#
|
27
|
-
# The `Status` message contains three pieces of data: error code, error
|
28
|
-
# and error details. The error code should be an enum value of
|
29
|
-
# {Google::Rpc::Code}, but it may accept additional error codes
|
30
|
-
# error message should be a developer-facing English message
|
31
|
-
# developers *understand* and *resolve* the error. If a localized
|
32
|
-
# error message is needed, put the localized message in the error
|
33
|
-
# localize it in the client. The optional error details may contain
|
34
|
-
# information about the error. There is a predefined set of error
|
35
|
-
# in the package `google.rpc` that can be used for common error
|
27
|
+
# The `Status` message contains three pieces of data: error code, error
|
28
|
+
# message, and error details. The error code should be an enum value of
|
29
|
+
# {Google::Rpc::Code}, but it may accept additional error codes
|
30
|
+
# if needed. The error message should be a developer-facing English message
|
31
|
+
# that helps developers *understand* and *resolve* the error. If a localized
|
32
|
+
# user-facing error message is needed, put the localized message in the error
|
33
|
+
# details or localize it in the client. The optional error details may contain
|
34
|
+
# arbitrary information about the error. There is a predefined set of error
|
35
|
+
# detail types in the package `google.rpc` that can be used for common error
|
36
|
+
# conditions.
|
36
37
|
#
|
37
38
|
# = Language mapping
|
38
39
|
#
|
@@ -69,12 +70,14 @@ module Google
|
|
69
70
|
# be used directly after any stripping needed for security/privacy reasons.
|
70
71
|
# @!attribute [rw] code
|
71
72
|
# @return [Integer]
|
72
|
-
# The status code, which should be an enum value of
|
73
|
+
# The status code, which should be an enum value of
|
74
|
+
# {Google::Rpc::Code}.
|
73
75
|
# @!attribute [rw] message
|
74
76
|
# @return [String]
|
75
77
|
# A developer-facing error message, which should be in English. Any
|
76
78
|
# user-facing error message should be localized and sent in the
|
77
|
-
# {Google::Rpc::Status#details} field, or localized
|
79
|
+
# {Google::Rpc::Status#details} field, or localized
|
80
|
+
# by the client.
|
78
81
|
# @!attribute [rw] details
|
79
82
|
# @return [Array<Google::Protobuf::Any>]
|
80
83
|
# A list of messages that carry the error details. There is a common set of
|
@@ -27,9 +27,9 @@ module Google
|
|
27
27
|
module V1
|
28
28
|
module DataTransferService
|
29
29
|
# The Google BigQuery Data Transfer Service API enables BigQuery users to
|
30
|
-
# configure the transfer of their data from other Google Products into
|
31
|
-
# This service contains methods that are end user exposed. It backs
|
32
|
-
# frontend.
|
30
|
+
# configure the transfer of their data from other Google Products into
|
31
|
+
# BigQuery. This service contains methods that are end user exposed. It backs
|
32
|
+
# up the frontend.
|
33
33
|
class Service
|
34
34
|
|
35
35
|
include GRPC::GenericService
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-cloud-bigquery-data_transfer
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.2.
|
4
|
+
version: 0.2.4
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2019-04-29 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: google-gax
|
@@ -58,14 +58,14 @@ dependencies:
|
|
58
58
|
requirements:
|
59
59
|
- - "~>"
|
60
60
|
- !ruby/object:Gem::Version
|
61
|
-
version: 0.
|
61
|
+
version: 0.64.0
|
62
62
|
type: :development
|
63
63
|
prerelease: false
|
64
64
|
version_requirements: !ruby/object:Gem::Requirement
|
65
65
|
requirements:
|
66
66
|
- - "~>"
|
67
67
|
- !ruby/object:Gem::Version
|
68
|
-
version: 0.
|
68
|
+
version: 0.64.0
|
69
69
|
- !ruby/object:Gem::Dependency
|
70
70
|
name: simplecov
|
71
71
|
requirement: !ruby/object:Gem::Requirement
|
@@ -102,6 +102,7 @@ extensions: []
|
|
102
102
|
extra_rdoc_files: []
|
103
103
|
files:
|
104
104
|
- ".yardopts"
|
105
|
+
- AUTHENTICATION.md
|
105
106
|
- LICENSE
|
106
107
|
- README.md
|
107
108
|
- lib/google/cloud/bigquery/data_transfer.rb
|
@@ -147,8 +148,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
147
148
|
- !ruby/object:Gem::Version
|
148
149
|
version: '0'
|
149
150
|
requirements: []
|
150
|
-
|
151
|
-
rubygems_version: 2.7.7
|
151
|
+
rubygems_version: 3.0.3
|
152
152
|
signing_key:
|
153
153
|
specification_version: 4
|
154
154
|
summary: API Client library for BigQuery Data Transfer API
|