google-cloud-bigquery-data_transfer 0.2.5 → 0.3.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/README.md +4 -4
- data/lib/google/cloud/bigquery/data_transfer.rb +7 -3
- data/lib/google/cloud/bigquery/data_transfer/v1.rb +11 -3
- data/lib/google/cloud/bigquery/data_transfer/v1/data_transfer_service_client.rb +160 -5
- data/lib/google/cloud/bigquery/data_transfer/v1/data_transfer_service_client_config.json +5 -0
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/cloud/bigquery/data_transfer/v1/data_transfer.rb +1 -2
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/cloud/bigquery/datatransfer/v1/datasource.rb +25 -0
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/cloud/bigquery/datatransfer/v1/datatransfer.rb +66 -14
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/cloud/bigquery/datatransfer/v1/transfer.rb +37 -19
- data/lib/google/cloud/bigquery/data_transfer/v1/doc/google/protobuf/wrappers.rb +3 -67
- data/lib/google/cloud/bigquery/data_transfer/version.rb +1 -1
- data/lib/google/cloud/bigquery/datatransfer/v1/datasource_pb.rb +170 -0
- data/lib/google/cloud/bigquery/datatransfer/v1/datasource_services_pb.rb +103 -0
- data/lib/google/cloud/bigquery/datatransfer/v1/datatransfer_pb.rb +21 -0
- data/lib/google/cloud/bigquery/datatransfer/v1/datatransfer_services_pb.rb +8 -1
- data/lib/google/cloud/bigquery/datatransfer/v1/transfer_pb.rb +7 -0
- metadata +8 -5
@@ -41,7 +41,7 @@ module Google
|
|
41
41
|
# Is parameter required.
|
42
42
|
# @!attribute [rw] repeated
|
43
43
|
# @return [true, false]
|
44
|
-
#
|
44
|
+
# Deprecated. This field has no effect.
|
45
45
|
# @!attribute [rw] validation_regex
|
46
46
|
# @return [String]
|
47
47
|
# Regular expression which can be used for parameter validation.
|
@@ -56,7 +56,7 @@ module Google
|
|
56
56
|
# For integer and double values specifies maxminum allowed value.
|
57
57
|
# @!attribute [rw] fields
|
58
58
|
# @return [Array<Google::Cloud::Bigquery::Datatransfer::V1::DataSourceParameter>]
|
59
|
-
#
|
59
|
+
# Deprecated. This field has no effect.
|
60
60
|
# @!attribute [rw] validation_description
|
61
61
|
# @return [String]
|
62
62
|
# Description of the requirements for this field, in case the user input does
|
@@ -69,8 +69,11 @@ module Google
|
|
69
69
|
# Cannot be changed after initial creation.
|
70
70
|
# @!attribute [rw] recurse
|
71
71
|
# @return [true, false]
|
72
|
-
#
|
73
|
-
#
|
72
|
+
# Deprecated. This field has no effect.
|
73
|
+
# @!attribute [rw] deprecated
|
74
|
+
# @return [true, false]
|
75
|
+
# If true, it should not be used in new transfers, and it should not be
|
76
|
+
# visible to users.
|
74
77
|
class DataSourceParameter
|
75
78
|
# Parameter type.
|
76
79
|
module Type
|
@@ -90,7 +93,7 @@ module Google
|
|
90
93
|
# Boolean parameter.
|
91
94
|
BOOLEAN = 4
|
92
95
|
|
93
|
-
#
|
96
|
+
# Deprecated. This field has no effect.
|
94
97
|
RECORD = 5
|
95
98
|
|
96
99
|
# Page ID for a Google+ Page.
|
@@ -115,24 +118,21 @@ module Google
|
|
115
118
|
# @!attribute [rw] client_id
|
116
119
|
# @return [String]
|
117
120
|
# Data source client id which should be used to receive refresh token.
|
118
|
-
# When not supplied, no offline credentials are populated for data transfer.
|
119
121
|
# @!attribute [rw] scopes
|
120
122
|
# @return [Array<String>]
|
121
|
-
# Api auth scopes for which refresh token needs to be obtained.
|
122
|
-
#
|
123
|
-
#
|
124
|
-
# e.g., https://www.googleapis.com/auth/bigquery
|
123
|
+
# Api auth scopes for which refresh token needs to be obtained. These are
|
124
|
+
# scopes needed by a data source to prepare data and ingest them into
|
125
|
+
# BigQuery, e.g., https://www.googleapis.com/auth/bigquery
|
125
126
|
# @!attribute [rw] transfer_type
|
126
127
|
# @return [Google::Cloud::Bigquery::Datatransfer::V1::TransferType]
|
127
128
|
# Deprecated. This field has no effect.
|
128
129
|
# @!attribute [rw] supports_multiple_transfers
|
129
130
|
# @return [true, false]
|
130
|
-
#
|
131
|
-
# to different BigQuery targets.
|
131
|
+
# Deprecated. This field has no effect.
|
132
132
|
# @!attribute [rw] update_deadline_seconds
|
133
133
|
# @return [Integer]
|
134
134
|
# The number of seconds to wait for an update from the data source
|
135
|
-
# before
|
135
|
+
# before the Data Transfer Service marks the transfer as FAILED.
|
136
136
|
# @!attribute [rw] default_schedule
|
137
137
|
# @return [String]
|
138
138
|
# Default data transfer schedule.
|
@@ -248,7 +248,7 @@ module Google
|
|
248
248
|
# @!attribute [rw] parent
|
249
249
|
# @return [String]
|
250
250
|
# The BigQuery project id where the transfer configuration should be created.
|
251
|
-
# Must be in the format
|
251
|
+
# Must be in the format projects/\\{project_id}/locations/\\{location_id}
|
252
252
|
# If specified location and location of the destination bigquery dataset
|
253
253
|
# do not match - the request will fail.
|
254
254
|
# @!attribute [rw] transfer_config
|
@@ -272,6 +272,14 @@ module Google
|
|
272
272
|
# urn:ietf:wg:oauth:2.0:oob means that authorization code should be
|
273
273
|
# returned in the title bar of the browser, with the page text prompting
|
274
274
|
# the user to copy the code and paste it in the application.
|
275
|
+
# @!attribute [rw] version_info
|
276
|
+
# @return [String]
|
277
|
+
# Optional version info. If users want to find a very recent access token,
|
278
|
+
# that is, immediately after approving access, users have to set the
|
279
|
+
# version_info claim in the token request. To obtain the version_info, users
|
280
|
+
# must use the "none+gsession" response type. which be return a
|
281
|
+
# version_info back in the authorization response which be be put in a JWT
|
282
|
+
# claim in the token request.
|
275
283
|
class CreateTransferConfigRequest; end
|
276
284
|
|
277
285
|
# A request to update a transfer configuration. To update the user id of the
|
@@ -300,6 +308,14 @@ module Google
|
|
300
308
|
# @!attribute [rw] update_mask
|
301
309
|
# @return [Google::Protobuf::FieldMask]
|
302
310
|
# Required list of fields to be updated in this request.
|
311
|
+
# @!attribute [rw] version_info
|
312
|
+
# @return [String]
|
313
|
+
# Optional version info. If users want to find a very recent access token,
|
314
|
+
# that is, immediately after approving access, users have to set the
|
315
|
+
# version_info claim in the token request. To obtain the version_info, users
|
316
|
+
# must use the "none+gsession" response type. which be return a
|
317
|
+
# version_info back in the authorization response which be be put in a JWT
|
318
|
+
# claim in the token request.
|
303
319
|
class UpdateTransferConfigRequest; end
|
304
320
|
|
305
321
|
# A request to get data transfer information.
|
@@ -481,6 +497,42 @@ module Google
|
|
481
497
|
# @return [Array<Google::Cloud::Bigquery::Datatransfer::V1::TransferRun>]
|
482
498
|
# The transfer runs that were scheduled.
|
483
499
|
class ScheduleTransferRunsResponse; end
|
500
|
+
|
501
|
+
# A request to start manual transfer runs.
|
502
|
+
# @!attribute [rw] parent
|
503
|
+
# @return [String]
|
504
|
+
# Transfer configuration name in the form:
|
505
|
+
# `projects/{project_id}/transferConfigs/{config_id}`.
|
506
|
+
# @!attribute [rw] requested_time_range
|
507
|
+
# @return [Google::Cloud::Bigquery::Datatransfer::V1::StartManualTransferRunsRequest::TimeRange]
|
508
|
+
# Time range for the transfer runs that should be started.
|
509
|
+
# @!attribute [rw] requested_run_time
|
510
|
+
# @return [Google::Protobuf::Timestamp]
|
511
|
+
# Specific run_time for a transfer run to be started. The
|
512
|
+
# requested_run_time must not be in the future.
|
513
|
+
class StartManualTransferRunsRequest
|
514
|
+
# A specification for a time range, this will request transfer runs with
|
515
|
+
# run_time between start_time (inclusive) and end_time (exclusive).
|
516
|
+
# @!attribute [rw] start_time
|
517
|
+
# @return [Google::Protobuf::Timestamp]
|
518
|
+
# Start time of the range of transfer runs. For example,
|
519
|
+
# `"2017-05-25T00:00:00+00:00"`. The start_time must be strictly less than
|
520
|
+
# the end_time. Creates transfer runs where run_time is in the range betwen
|
521
|
+
# start_time (inclusive) and end_time (exlusive).
|
522
|
+
# @!attribute [rw] end_time
|
523
|
+
# @return [Google::Protobuf::Timestamp]
|
524
|
+
# End time of the range of transfer runs. For example,
|
525
|
+
# `"2017-05-30T00:00:00+00:00"`. The end_time must not be in the future.
|
526
|
+
# Creates transfer runs where run_time is in the range betwen start_time
|
527
|
+
# (inclusive) and end_time (exlusive).
|
528
|
+
class TimeRange; end
|
529
|
+
end
|
530
|
+
|
531
|
+
# A response to start manual transfer runs.
|
532
|
+
# @!attribute [rw] runs
|
533
|
+
# @return [Array<Google::Cloud::Bigquery::Datatransfer::V1::TransferRun>]
|
534
|
+
# The transfer runs that were created.
|
535
|
+
class StartManualTransferRunsResponse; end
|
484
536
|
end
|
485
537
|
end
|
486
538
|
end
|
@@ -18,6 +18,28 @@ module Google
|
|
18
18
|
module Bigquery
|
19
19
|
module Datatransfer
|
20
20
|
module V1
|
21
|
+
# Options customizing the data transfer schedule.
|
22
|
+
# @!attribute [rw] disable_auto_scheduling
|
23
|
+
# @return [true, false]
|
24
|
+
# If true, automatic scheduling of data transfer runs for this configuration
|
25
|
+
# will be disabled. The runs can be started on ad-hoc basis using
|
26
|
+
# StartManualTransferRuns API. When automatic scheduling is disabled, the
|
27
|
+
# TransferConfig.schedule field will be ignored.
|
28
|
+
# @!attribute [rw] start_time
|
29
|
+
# @return [Google::Protobuf::Timestamp]
|
30
|
+
# Specifies time to start scheduling transfer runs. The first run will be
|
31
|
+
# scheduled at or after the start time according to a recurrence pattern
|
32
|
+
# defined in the schedule string. The start time can be changed at any
|
33
|
+
# moment. The time when a data transfer can be trigerred manually is not
|
34
|
+
# limited by this option.
|
35
|
+
# @!attribute [rw] end_time
|
36
|
+
# @return [Google::Protobuf::Timestamp]
|
37
|
+
# Defines time to stop scheduling transfer runs. A transfer run cannot be
|
38
|
+
# scheduled at or after the end time. The end time can be changed at any
|
39
|
+
# moment. The time when a data transfer can be trigerred manually is not
|
40
|
+
# limited by this option.
|
41
|
+
class ScheduleOptions; end
|
42
|
+
|
21
43
|
# Represents a data transfer configuration. A transfer configuration
|
22
44
|
# contains all metadata needed to perform a data transfer. For example,
|
23
45
|
# `destination_dataset_id` specifies where data should be stored.
|
@@ -27,11 +49,12 @@ module Google
|
|
27
49
|
# @!attribute [rw] name
|
28
50
|
# @return [String]
|
29
51
|
# The resource name of the transfer config.
|
30
|
-
# Transfer config names have the form
|
31
|
-
# `projects/{project_id}/transferConfigs/{config_id}`.
|
32
|
-
#
|
33
|
-
#
|
34
|
-
#
|
52
|
+
# Transfer config names have the form of
|
53
|
+
# `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
|
54
|
+
# The name is automatically generated based on the config_id specified in
|
55
|
+
# CreateTransferConfigRequest along with project_id and region. If config_id
|
56
|
+
# is not provided, usually a uuid, even though it is not guaranteed or
|
57
|
+
# required, will be generated for config_id.
|
35
58
|
# @!attribute [rw] destination_dataset_id
|
36
59
|
# @return [String]
|
37
60
|
# The BigQuery target dataset id.
|
@@ -58,6 +81,9 @@ module Google
|
|
58
81
|
# See more explanation about the format here:
|
59
82
|
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
|
60
83
|
# NOTE: the granularity should be at least 8 hours, or less frequent.
|
84
|
+
# @!attribute [rw] schedule_options
|
85
|
+
# @return [Google::Cloud::Bigquery::Datatransfer::V1::ScheduleOptions]
|
86
|
+
# Options customizing the data transfer schedule.
|
61
87
|
# @!attribute [rw] data_refresh_window_days
|
62
88
|
# @return [Integer]
|
63
89
|
# The number of days to look back to automatically refresh the data.
|
@@ -81,11 +107,7 @@ module Google
|
|
81
107
|
# Output only. State of the most recently updated transfer run.
|
82
108
|
# @!attribute [rw] user_id
|
83
109
|
# @return [Integer]
|
84
|
-
#
|
85
|
-
# Applicable only to data sources that do not support service accounts.
|
86
|
-
# When set to 0, the data source service account credentials are used.
|
87
|
-
# May be negative. Note, that this identifier is not stable.
|
88
|
-
# It may change over time even for the same user.
|
110
|
+
# Deprecated. Unique ID of the user on whose behalf transfer is done.
|
89
111
|
# @!attribute [rw] dataset_region
|
90
112
|
# @return [String]
|
91
113
|
# Output only. Region in which BigQuery dataset is located.
|
@@ -103,8 +125,8 @@ module Google
|
|
103
125
|
# Minimum time after which a transfer run can be started.
|
104
126
|
# @!attribute [rw] run_time
|
105
127
|
# @return [Google::Protobuf::Timestamp]
|
106
|
-
# For batch transfer runs, specifies the date and time
|
107
|
-
#
|
128
|
+
# For batch transfer runs, specifies the date and time of the data should be
|
129
|
+
# ingested.
|
108
130
|
# @!attribute [rw] error_status
|
109
131
|
# @return [Google::Rpc::Status]
|
110
132
|
# Status of the transfer run.
|
@@ -133,18 +155,14 @@ module Google
|
|
133
155
|
# Data transfer run state. Ignored for input requests.
|
134
156
|
# @!attribute [rw] user_id
|
135
157
|
# @return [Integer]
|
136
|
-
#
|
137
|
-
# Applicable only to data sources that do not support service accounts.
|
138
|
-
# When set to 0, the data source service account credentials are used.
|
139
|
-
# May be negative. Note, that this identifier is not stable.
|
140
|
-
# It may change over time even for the same user.
|
158
|
+
# Deprecated. Unique ID of the user on whose behalf transfer is done.
|
141
159
|
# @!attribute [rw] schedule
|
142
160
|
# @return [String]
|
143
161
|
# Output only. Describes the schedule of this transfer run if it was
|
144
162
|
# created as part of a regular schedule. For batch transfer runs that are
|
145
163
|
# scheduled manually, this is empty.
|
146
164
|
# NOTE: the system might choose to delay the schedule depending on the
|
147
|
-
# current load, so `schedule_time` doesn't always
|
165
|
+
# current load, so `schedule_time` doesn't always match this.
|
148
166
|
class TransferRun; end
|
149
167
|
|
150
168
|
# Represents a user facing message for a particular data transfer run.
|
@@ -186,7 +204,7 @@ module Google
|
|
186
204
|
# Data transfer is in progress.
|
187
205
|
RUNNING = 3
|
188
206
|
|
189
|
-
# Data transfer completed
|
207
|
+
# Data transfer completed successfully.
|
190
208
|
SUCCEEDED = 4
|
191
209
|
|
192
210
|
# Data transfer failed.
|
@@ -1,4 +1,4 @@
|
|
1
|
-
# Copyright
|
1
|
+
# Copyright 2019 Google LLC
|
2
2
|
#
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
4
|
# you may not use this file except in compliance with the License.
|
@@ -15,76 +15,12 @@
|
|
15
15
|
|
16
16
|
module Google
|
17
17
|
module Protobuf
|
18
|
-
# Wrapper message for
|
18
|
+
# Wrapper message for `double`.
|
19
19
|
#
|
20
|
-
# The JSON representation for
|
20
|
+
# The JSON representation for `DoubleValue` is JSON number.
|
21
21
|
# @!attribute [rw] value
|
22
22
|
# @return [Float]
|
23
23
|
# The double value.
|
24
24
|
class DoubleValue; end
|
25
|
-
|
26
|
-
# Wrapper message for +float+.
|
27
|
-
#
|
28
|
-
# The JSON representation for +FloatValue+ is JSON number.
|
29
|
-
# @!attribute [rw] value
|
30
|
-
# @return [Float]
|
31
|
-
# The float value.
|
32
|
-
class FloatValue; end
|
33
|
-
|
34
|
-
# Wrapper message for +int64+.
|
35
|
-
#
|
36
|
-
# The JSON representation for +Int64Value+ is JSON string.
|
37
|
-
# @!attribute [rw] value
|
38
|
-
# @return [Integer]
|
39
|
-
# The int64 value.
|
40
|
-
class Int64Value; end
|
41
|
-
|
42
|
-
# Wrapper message for +uint64+.
|
43
|
-
#
|
44
|
-
# The JSON representation for +UInt64Value+ is JSON string.
|
45
|
-
# @!attribute [rw] value
|
46
|
-
# @return [Integer]
|
47
|
-
# The uint64 value.
|
48
|
-
class UInt64Value; end
|
49
|
-
|
50
|
-
# Wrapper message for +int32+.
|
51
|
-
#
|
52
|
-
# The JSON representation for +Int32Value+ is JSON number.
|
53
|
-
# @!attribute [rw] value
|
54
|
-
# @return [Integer]
|
55
|
-
# The int32 value.
|
56
|
-
class Int32Value; end
|
57
|
-
|
58
|
-
# Wrapper message for +uint32+.
|
59
|
-
#
|
60
|
-
# The JSON representation for +UInt32Value+ is JSON number.
|
61
|
-
# @!attribute [rw] value
|
62
|
-
# @return [Integer]
|
63
|
-
# The uint32 value.
|
64
|
-
class UInt32Value; end
|
65
|
-
|
66
|
-
# Wrapper message for +bool+.
|
67
|
-
#
|
68
|
-
# The JSON representation for +BoolValue+ is JSON +true+ and +false+.
|
69
|
-
# @!attribute [rw] value
|
70
|
-
# @return [true, false]
|
71
|
-
# The bool value.
|
72
|
-
class BoolValue; end
|
73
|
-
|
74
|
-
# Wrapper message for +string+.
|
75
|
-
#
|
76
|
-
# The JSON representation for +StringValue+ is JSON string.
|
77
|
-
# @!attribute [rw] value
|
78
|
-
# @return [String]
|
79
|
-
# The string value.
|
80
|
-
class StringValue; end
|
81
|
-
|
82
|
-
# Wrapper message for +bytes+.
|
83
|
-
#
|
84
|
-
# The JSON representation for +BytesValue+ is JSON string.
|
85
|
-
# @!attribute [rw] value
|
86
|
-
# @return [String]
|
87
|
-
# The bytes value.
|
88
|
-
class BytesValue; end
|
89
25
|
end
|
90
26
|
end
|
@@ -0,0 +1,170 @@
|
|
1
|
+
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
|
+
# source: google/cloud/bigquery/datatransfer/v1/datasource.proto
|
3
|
+
|
4
|
+
|
5
|
+
require 'google/protobuf'
|
6
|
+
|
7
|
+
require 'google/api/annotations_pb'
|
8
|
+
require 'google/cloud/bigquery/datatransfer/v1/datatransfer_pb'
|
9
|
+
require 'google/cloud/bigquery/datatransfer/v1/transfer_pb'
|
10
|
+
require 'google/protobuf/duration_pb'
|
11
|
+
require 'google/protobuf/empty_pb'
|
12
|
+
require 'google/protobuf/field_mask_pb'
|
13
|
+
require 'google/protobuf/timestamp_pb'
|
14
|
+
require 'google/protobuf/wrappers_pb'
|
15
|
+
require 'google/api/client_pb'
|
16
|
+
Google::Protobuf::DescriptorPool.generated_pool.build do
|
17
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo" do
|
18
|
+
optional :sql, :string, 1
|
19
|
+
optional :destination_table_id, :string, 2
|
20
|
+
optional :destination_table_description, :string, 10
|
21
|
+
repeated :table_defs, :message, 3, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.TableDefinition"
|
22
|
+
repeated :user_defined_functions, :string, 4
|
23
|
+
optional :write_disposition, :enum, 6, "google.cloud.bigquery.datatransfer.v1.WriteDisposition"
|
24
|
+
end
|
25
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.FieldSchema" do
|
26
|
+
optional :field_name, :string, 1
|
27
|
+
optional :type, :enum, 2, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.FieldSchema.Type"
|
28
|
+
optional :is_repeated, :bool, 3
|
29
|
+
optional :description, :string, 4
|
30
|
+
optional :schema, :message, 5, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.RecordSchema"
|
31
|
+
end
|
32
|
+
add_enum "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.FieldSchema.Type" do
|
33
|
+
value :TYPE_UNSPECIFIED, 0
|
34
|
+
value :STRING, 1
|
35
|
+
value :INTEGER, 2
|
36
|
+
value :FLOAT, 3
|
37
|
+
value :RECORD, 4
|
38
|
+
value :BYTES, 5
|
39
|
+
value :BOOLEAN, 6
|
40
|
+
value :TIMESTAMP, 7
|
41
|
+
value :DATE, 8
|
42
|
+
value :TIME, 9
|
43
|
+
value :DATETIME, 10
|
44
|
+
value :NUMERIC, 11
|
45
|
+
value :GEOGRAPHY, 12
|
46
|
+
end
|
47
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.RecordSchema" do
|
48
|
+
repeated :fields, :message, 1, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.FieldSchema"
|
49
|
+
end
|
50
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.TableDefinition" do
|
51
|
+
optional :table_id, :string, 1
|
52
|
+
repeated :source_uris, :string, 2
|
53
|
+
optional :format, :enum, 3, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.Format"
|
54
|
+
optional :max_bad_records, :int32, 4
|
55
|
+
optional :encoding, :enum, 5, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.Encoding"
|
56
|
+
optional :csv_options, :message, 6, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.TableDefinition.CsvOptions"
|
57
|
+
optional :schema, :message, 7, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.RecordSchema"
|
58
|
+
optional :ignore_unknown_values, :message, 10, "google.protobuf.BoolValue"
|
59
|
+
end
|
60
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.TableDefinition.CsvOptions" do
|
61
|
+
optional :field_delimiter, :message, 1, "google.protobuf.StringValue"
|
62
|
+
optional :allow_quoted_newlines, :message, 2, "google.protobuf.BoolValue"
|
63
|
+
optional :quote_char, :message, 3, "google.protobuf.StringValue"
|
64
|
+
optional :skip_leading_rows, :message, 4, "google.protobuf.Int64Value"
|
65
|
+
optional :allow_jagged_rows, :message, 5, "google.protobuf.BoolValue"
|
66
|
+
end
|
67
|
+
add_enum "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.Format" do
|
68
|
+
value :FORMAT_UNSPECIFIED, 0
|
69
|
+
value :CSV, 1
|
70
|
+
value :JSON, 2
|
71
|
+
value :AVRO, 3
|
72
|
+
value :RECORDIO, 4
|
73
|
+
value :COLUMNIO, 5
|
74
|
+
value :CAPACITOR, 6
|
75
|
+
value :PARQUET, 7
|
76
|
+
value :ORC, 8
|
77
|
+
end
|
78
|
+
add_enum "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.Encoding" do
|
79
|
+
value :ENCODING_UNSPECIFIED, 0
|
80
|
+
value :ISO_8859_1, 1
|
81
|
+
value :UTF8, 2
|
82
|
+
end
|
83
|
+
add_message "google.cloud.bigquery.datatransfer.v1.UpdateTransferRunRequest" do
|
84
|
+
optional :transfer_run, :message, 1, "google.cloud.bigquery.datatransfer.v1.TransferRun"
|
85
|
+
optional :update_mask, :message, 2, "google.protobuf.FieldMask"
|
86
|
+
end
|
87
|
+
add_message "google.cloud.bigquery.datatransfer.v1.LogTransferRunMessagesRequest" do
|
88
|
+
optional :name, :string, 1
|
89
|
+
repeated :transfer_messages, :message, 2, "google.cloud.bigquery.datatransfer.v1.TransferMessage"
|
90
|
+
end
|
91
|
+
add_message "google.cloud.bigquery.datatransfer.v1.StartBigQueryJobsRequest" do
|
92
|
+
optional :name, :string, 1
|
93
|
+
repeated :imported_data, :message, 2, "google.cloud.bigquery.datatransfer.v1.ImportedDataInfo"
|
94
|
+
optional :user_credentials, :bytes, 3
|
95
|
+
optional :max_parallelism, :int32, 8
|
96
|
+
end
|
97
|
+
add_message "google.cloud.bigquery.datatransfer.v1.FinishRunRequest" do
|
98
|
+
optional :name, :string, 1
|
99
|
+
end
|
100
|
+
add_message "google.cloud.bigquery.datatransfer.v1.CreateDataSourceDefinitionRequest" do
|
101
|
+
optional :parent, :string, 1
|
102
|
+
optional :data_source_definition, :message, 2, "google.cloud.bigquery.datatransfer.v1.DataSourceDefinition"
|
103
|
+
end
|
104
|
+
add_message "google.cloud.bigquery.datatransfer.v1.UpdateDataSourceDefinitionRequest" do
|
105
|
+
optional :data_source_definition, :message, 1, "google.cloud.bigquery.datatransfer.v1.DataSourceDefinition"
|
106
|
+
optional :update_mask, :message, 2, "google.protobuf.FieldMask"
|
107
|
+
end
|
108
|
+
add_message "google.cloud.bigquery.datatransfer.v1.DeleteDataSourceDefinitionRequest" do
|
109
|
+
optional :name, :string, 1
|
110
|
+
end
|
111
|
+
add_message "google.cloud.bigquery.datatransfer.v1.GetDataSourceDefinitionRequest" do
|
112
|
+
optional :name, :string, 1
|
113
|
+
end
|
114
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ListDataSourceDefinitionsRequest" do
|
115
|
+
optional :parent, :string, 1
|
116
|
+
optional :page_token, :string, 2
|
117
|
+
optional :page_size, :int32, 3
|
118
|
+
end
|
119
|
+
add_message "google.cloud.bigquery.datatransfer.v1.ListDataSourceDefinitionsResponse" do
|
120
|
+
repeated :data_source_definitions, :message, 1, "google.cloud.bigquery.datatransfer.v1.DataSourceDefinition"
|
121
|
+
optional :next_page_token, :string, 2
|
122
|
+
end
|
123
|
+
add_message "google.cloud.bigquery.datatransfer.v1.DataSourceDefinition" do
|
124
|
+
optional :name, :string, 21
|
125
|
+
optional :data_source, :message, 1, "google.cloud.bigquery.datatransfer.v1.DataSource"
|
126
|
+
optional :transfer_run_pubsub_topic, :string, 13
|
127
|
+
optional :run_time_offset, :message, 16, "google.protobuf.Duration"
|
128
|
+
optional :support_email, :string, 22
|
129
|
+
optional :service_account, :string, 2
|
130
|
+
optional :disabled, :bool, 5
|
131
|
+
optional :transfer_config_pubsub_topic, :string, 12
|
132
|
+
repeated :supported_location_ids, :string, 23
|
133
|
+
end
|
134
|
+
add_enum "google.cloud.bigquery.datatransfer.v1.WriteDisposition" do
|
135
|
+
value :WRITE_DISPOSITION_UNSPECIFIED, 0
|
136
|
+
value :WRITE_TRUNCATE, 1
|
137
|
+
value :WRITE_APPEND, 2
|
138
|
+
end
|
139
|
+
end
|
140
|
+
|
141
|
+
module Google
|
142
|
+
module Cloud
|
143
|
+
module Bigquery
|
144
|
+
module Datatransfer
|
145
|
+
module V1
|
146
|
+
ImportedDataInfo = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo").msgclass
|
147
|
+
ImportedDataInfo::FieldSchema = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.FieldSchema").msgclass
|
148
|
+
ImportedDataInfo::FieldSchema::Type = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.FieldSchema.Type").enummodule
|
149
|
+
ImportedDataInfo::RecordSchema = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.RecordSchema").msgclass
|
150
|
+
ImportedDataInfo::TableDefinition = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.TableDefinition").msgclass
|
151
|
+
ImportedDataInfo::TableDefinition::CsvOptions = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.TableDefinition.CsvOptions").msgclass
|
152
|
+
ImportedDataInfo::Format = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.Format").enummodule
|
153
|
+
ImportedDataInfo::Encoding = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ImportedDataInfo.Encoding").enummodule
|
154
|
+
UpdateTransferRunRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.UpdateTransferRunRequest").msgclass
|
155
|
+
LogTransferRunMessagesRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.LogTransferRunMessagesRequest").msgclass
|
156
|
+
StartBigQueryJobsRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.StartBigQueryJobsRequest").msgclass
|
157
|
+
FinishRunRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.FinishRunRequest").msgclass
|
158
|
+
CreateDataSourceDefinitionRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.CreateDataSourceDefinitionRequest").msgclass
|
159
|
+
UpdateDataSourceDefinitionRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.UpdateDataSourceDefinitionRequest").msgclass
|
160
|
+
DeleteDataSourceDefinitionRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.DeleteDataSourceDefinitionRequest").msgclass
|
161
|
+
GetDataSourceDefinitionRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.GetDataSourceDefinitionRequest").msgclass
|
162
|
+
ListDataSourceDefinitionsRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ListDataSourceDefinitionsRequest").msgclass
|
163
|
+
ListDataSourceDefinitionsResponse = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.ListDataSourceDefinitionsResponse").msgclass
|
164
|
+
DataSourceDefinition = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.DataSourceDefinition").msgclass
|
165
|
+
WriteDisposition = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.bigquery.datatransfer.v1.WriteDisposition").enummodule
|
166
|
+
end
|
167
|
+
end
|
168
|
+
end
|
169
|
+
end
|
170
|
+
end
|