google-cloud-bigquery-data_transfer-v1 0.4.1 → 0.4.5

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: fcc4faa89caae40371c403488bc3c44a952fa421b1737b03e8af332016b0a17e
4
- data.tar.gz: acf1ca1a755615520405b789399235ff88d4eea1ff8b3df41b6f455eec9b4abf
3
+ metadata.gz: 858ac9046282b3f9f1ad16896571a7ec0ff7167e6a41f4a677e3623b4e08eaaa
4
+ data.tar.gz: caefa5a2b16a67126c8915ddcb5bc9d85cf62a06bce500d35fb6c7c5630b2b75
5
5
  SHA512:
6
- metadata.gz: f25a01f491c142e65f7340c437d7121f2c500179049914b39a38121d2c449ada8d1343b23ec2fc4827eb2a9cb680cd25ce7dd3081930764f8225585a787dbb80
7
- data.tar.gz: e15f2a12b29b8685fcfce8a33cce18d568859268acbb60b4694bb821578870e4a50563ac8c3213ffeac8d2a6f1365c9a2ec467b81ca74bb0df07bda20d9a2179
6
+ metadata.gz: cbe0687f2bb91583fec9192380f098421d1bec4f6726aa4155cba8ba855e21dc551d063eec692974aa3955003bb8fb8bd91e3605bcb9e0f9110b97a04544e361
7
+ data.tar.gz: 1587893e7feeb969dcc82f8973facf0e835c19cfcbb46b80c538c4ef91d7edf091c26b7ab37d4eeabb769c89b160bcdb3d4a1b370809626a3fbadd19ca74872e
data/AUTHENTICATION.md CHANGED
@@ -66,11 +66,11 @@ The environment variables that google-cloud-bigquery-data_transfer-v1
66
66
  checks for credentials are configured on the service Credentials class (such as
67
67
  {::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Credentials}):
68
68
 
69
- 1. `DATA_TRANSFER_CREDENTIALS` - Path to JSON file, or JSON contents
70
- 2. `DATA_TRANSFER_KEYFILE` - Path to JSON file, or JSON contents
71
- 3. `GOOGLE_CLOUD_CREDENTIALS` - Path to JSON file, or JSON contents
72
- 4. `GOOGLE_CLOUD_KEYFILE` - Path to JSON file, or JSON contents
73
- 5. `GOOGLE_APPLICATION_CREDENTIALS` - Path to JSON file
69
+ * `DATA_TRANSFER_CREDENTIALS` - Path to JSON file, or JSON contents
70
+ * `DATA_TRANSFER_KEYFILE` - Path to JSON file, or JSON contents
71
+ * `GOOGLE_CLOUD_CREDENTIALS` - Path to JSON file, or JSON contents
72
+ * `GOOGLE_CLOUD_KEYFILE` - Path to JSON file, or JSON contents
73
+ * `GOOGLE_APPLICATION_CREDENTIALS` - Path to JSON file
74
74
 
75
75
  ```ruby
76
76
  require "google/cloud/bigquery/data_transfer/v1"
@@ -82,8 +82,8 @@ client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Clien
82
82
 
83
83
  ### Configuration
84
84
 
85
- The **Credentials JSON** can be configured instead of placing them in
86
- environment variables. Either on an individual client initialization:
85
+ The path to the **Credentials JSON** file can be configured instead of storing
86
+ it in an environment variable. Either on an individual client initialization:
87
87
 
88
88
  ```ruby
89
89
  require "google/cloud/bigquery/data_transfer/v1"
@@ -93,7 +93,7 @@ client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Clien
93
93
  end
94
94
  ```
95
95
 
96
- Or configured globally for all clients:
96
+ Or globally for all clients:
97
97
 
98
98
  ```ruby
99
99
  require "google/cloud/bigquery/data_transfer/v1"
data/README.md CHANGED
@@ -33,7 +33,7 @@ In order to use this library, you first need to go through the following steps:
33
33
  require "google/cloud/bigquery/data_transfer/v1"
34
34
 
35
35
  client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
36
- request = my_create_request
36
+ request = ::Google::Cloud::Bigquery::DataTransfer::V1::GetDataSourceRequest.new # (request fields as keyword arguments...)
37
37
  response = client.get_data_source request
38
38
  ```
39
39
 
@@ -45,13 +45,12 @@ module Google
45
45
  # See {::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client::Configuration}
46
46
  # for a description of the configuration fields.
47
47
  #
48
- # ## Example
48
+ # @example
49
49
  #
50
- # To modify the configuration for all DataTransferService clients:
51
- #
52
- # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
53
- # config.timeout = 10.0
54
- # end
50
+ # # Modify the configuration for all DataTransferService clients
51
+ # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
52
+ # config.timeout = 10.0
53
+ # end
55
54
  #
56
55
  # @yield [config] Configure the Client client.
57
56
  # @yieldparam config [Client::Configuration]
@@ -71,18 +70,12 @@ module Google
71
70
 
72
71
  default_config.rpcs.get_data_source.timeout = 20.0
73
72
  default_config.rpcs.get_data_source.retry_policy = {
74
- initial_delay: 0.1,
75
- max_delay: 60.0,
76
- multiplier: 1.3,
77
- retry_codes: [14, 4]
73
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
78
74
  }
79
75
 
80
76
  default_config.rpcs.list_data_sources.timeout = 20.0
81
77
  default_config.rpcs.list_data_sources.retry_policy = {
82
- initial_delay: 0.1,
83
- max_delay: 60.0,
84
- multiplier: 1.3,
85
- retry_codes: [14, 4]
78
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
86
79
  }
87
80
 
88
81
  default_config.rpcs.create_transfer_config.timeout = 30.0
@@ -91,68 +84,44 @@ module Google
91
84
 
92
85
  default_config.rpcs.delete_transfer_config.timeout = 20.0
93
86
  default_config.rpcs.delete_transfer_config.retry_policy = {
94
- initial_delay: 0.1,
95
- max_delay: 60.0,
96
- multiplier: 1.3,
97
- retry_codes: [14, 4]
87
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
98
88
  }
99
89
 
100
90
  default_config.rpcs.get_transfer_config.timeout = 20.0
101
91
  default_config.rpcs.get_transfer_config.retry_policy = {
102
- initial_delay: 0.1,
103
- max_delay: 60.0,
104
- multiplier: 1.3,
105
- retry_codes: [14, 4]
92
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
106
93
  }
107
94
 
108
95
  default_config.rpcs.list_transfer_configs.timeout = 20.0
109
96
  default_config.rpcs.list_transfer_configs.retry_policy = {
110
- initial_delay: 0.1,
111
- max_delay: 60.0,
112
- multiplier: 1.3,
113
- retry_codes: [14, 4]
97
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
114
98
  }
115
99
 
116
100
  default_config.rpcs.schedule_transfer_runs.timeout = 30.0
117
101
 
118
102
  default_config.rpcs.get_transfer_run.timeout = 20.0
119
103
  default_config.rpcs.get_transfer_run.retry_policy = {
120
- initial_delay: 0.1,
121
- max_delay: 60.0,
122
- multiplier: 1.3,
123
- retry_codes: [14, 4]
104
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
124
105
  }
125
106
 
126
107
  default_config.rpcs.delete_transfer_run.timeout = 20.0
127
108
  default_config.rpcs.delete_transfer_run.retry_policy = {
128
- initial_delay: 0.1,
129
- max_delay: 60.0,
130
- multiplier: 1.3,
131
- retry_codes: [14, 4]
109
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
132
110
  }
133
111
 
134
112
  default_config.rpcs.list_transfer_runs.timeout = 20.0
135
113
  default_config.rpcs.list_transfer_runs.retry_policy = {
136
- initial_delay: 0.1,
137
- max_delay: 60.0,
138
- multiplier: 1.3,
139
- retry_codes: [14, 4]
114
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
140
115
  }
141
116
 
142
117
  default_config.rpcs.list_transfer_logs.timeout = 20.0
143
118
  default_config.rpcs.list_transfer_logs.retry_policy = {
144
- initial_delay: 0.1,
145
- max_delay: 60.0,
146
- multiplier: 1.3,
147
- retry_codes: [14, 4]
119
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
148
120
  }
149
121
 
150
122
  default_config.rpcs.check_valid_creds.timeout = 20.0
151
123
  default_config.rpcs.check_valid_creds.retry_policy = {
152
- initial_delay: 0.1,
153
- max_delay: 60.0,
154
- multiplier: 1.3,
155
- retry_codes: [14, 4]
124
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
156
125
  }
157
126
 
158
127
  default_config
@@ -184,19 +153,15 @@ module Google
184
153
  ##
185
154
  # Create a new DataTransferService client object.
186
155
  #
187
- # ## Examples
188
- #
189
- # To create a new DataTransferService client with the default
190
- # configuration:
156
+ # @example
191
157
  #
192
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
158
+ # # Create a client using the default configuration
159
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
193
160
  #
194
- # To create a new DataTransferService client with a custom
195
- # configuration:
196
- #
197
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
198
- # config.timeout = 10.0
199
- # end
161
+ # # Create a client using a custom configuration
162
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
163
+ # config.timeout = 10.0
164
+ # end
200
165
  #
201
166
  # @yield [config] Configure the DataTransferService client.
202
167
  # @yieldparam config [Client::Configuration]
@@ -216,14 +181,13 @@ module Google
216
181
 
217
182
  # Create credentials
218
183
  credentials = @config.credentials
219
- # Use self-signed JWT if the scope and endpoint are unchanged from default,
184
+ # Use self-signed JWT if the endpoint is unchanged from default,
220
185
  # but only if the default endpoint does not have a region prefix.
221
- enable_self_signed_jwt = @config.scope == Client.configure.scope &&
222
- @config.endpoint == Client.configure.endpoint &&
186
+ enable_self_signed_jwt = @config.endpoint == Client.configure.endpoint &&
223
187
  !@config.endpoint.split(".").first.include?("-")
224
188
  credentials ||= Credentials.default scope: @config.scope,
225
189
  enable_self_signed_jwt: enable_self_signed_jwt
226
- if credentials.is_a?(String) || credentials.is_a?(Hash)
190
+ if credentials.is_a?(::String) || credentials.is_a?(::Hash)
227
191
  credentials = Credentials.new credentials, scope: @config.scope
228
192
  end
229
193
  @quota_project_id = @config.quota_project
@@ -272,6 +236,21 @@ module Google
272
236
  #
273
237
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
274
238
  #
239
+ # @example Basic example
240
+ # require "google/cloud/bigquery/data_transfer/v1"
241
+ #
242
+ # # Create a client object. The client can be reused for multiple calls.
243
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
244
+ #
245
+ # # Create a request. To set request fields, pass in keyword arguments.
246
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetDataSourceRequest.new
247
+ #
248
+ # # Call the get_data_source method.
249
+ # result = client.get_data_source request
250
+ #
251
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::DataSource.
252
+ # p result
253
+ #
275
254
  def get_data_source request, options = nil
276
255
  raise ::ArgumentError, "request must be provided" if request.nil?
277
256
 
@@ -289,16 +268,20 @@ module Google
289
268
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
290
269
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
291
270
 
292
- header_params = {
293
- "name" => request.name
294
- }
271
+ header_params = {}
272
+ if request.name
273
+ header_params["name"] = request.name
274
+ end
275
+
295
276
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
296
277
  metadata[:"x-goog-request-params"] ||= request_params_header
297
278
 
298
279
  options.apply_defaults timeout: @config.rpcs.get_data_source.timeout,
299
280
  metadata: metadata,
300
281
  retry_policy: @config.rpcs.get_data_source.retry_policy
301
- options.apply_defaults metadata: @config.metadata,
282
+
283
+ options.apply_defaults timeout: @config.timeout,
284
+ metadata: @config.metadata,
302
285
  retry_policy: @config.retry_policy
303
286
 
304
287
  @data_transfer_service_stub.call_rpc :get_data_source, request, options: options do |response, operation|
@@ -349,6 +332,27 @@ module Google
349
332
  #
350
333
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
351
334
  #
335
+ # @example Basic example
336
+ # require "google/cloud/bigquery/data_transfer/v1"
337
+ #
338
+ # # Create a client object. The client can be reused for multiple calls.
339
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
340
+ #
341
+ # # Create a request. To set request fields, pass in keyword arguments.
342
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListDataSourcesRequest.new
343
+ #
344
+ # # Call the list_data_sources method.
345
+ # result = client.list_data_sources request
346
+ #
347
+ # # The returned object is of type Gapic::PagedEnumerable. You can
348
+ # # iterate over all elements by calling #each, and the enumerable
349
+ # # will lazily make API calls to fetch subsequent pages. Other
350
+ # # methods are also available for managing paging directly.
351
+ # result.each do |response|
352
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::DataSource.
353
+ # p response
354
+ # end
355
+ #
352
356
  def list_data_sources request, options = nil
353
357
  raise ::ArgumentError, "request must be provided" if request.nil?
354
358
 
@@ -366,16 +370,20 @@ module Google
366
370
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
367
371
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
368
372
 
369
- header_params = {
370
- "parent" => request.parent
371
- }
373
+ header_params = {}
374
+ if request.parent
375
+ header_params["parent"] = request.parent
376
+ end
377
+
372
378
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
373
379
  metadata[:"x-goog-request-params"] ||= request_params_header
374
380
 
375
381
  options.apply_defaults timeout: @config.rpcs.list_data_sources.timeout,
376
382
  metadata: metadata,
377
383
  retry_policy: @config.rpcs.list_data_sources.retry_policy
378
- options.apply_defaults metadata: @config.metadata,
384
+
385
+ options.apply_defaults timeout: @config.timeout,
386
+ metadata: @config.metadata,
379
387
  retry_policy: @config.retry_policy
380
388
 
381
389
  @data_transfer_service_stub.call_rpc :list_data_sources, request, options: options do |response, operation|
@@ -450,6 +458,21 @@ module Google
450
458
  #
451
459
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
452
460
  #
461
+ # @example Basic example
462
+ # require "google/cloud/bigquery/data_transfer/v1"
463
+ #
464
+ # # Create a client object. The client can be reused for multiple calls.
465
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
466
+ #
467
+ # # Create a request. To set request fields, pass in keyword arguments.
468
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::CreateTransferConfigRequest.new
469
+ #
470
+ # # Call the create_transfer_config method.
471
+ # result = client.create_transfer_config request
472
+ #
473
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
474
+ # p result
475
+ #
453
476
  def create_transfer_config request, options = nil
454
477
  raise ::ArgumentError, "request must be provided" if request.nil?
455
478
 
@@ -467,16 +490,20 @@ module Google
467
490
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
468
491
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
469
492
 
470
- header_params = {
471
- "parent" => request.parent
472
- }
493
+ header_params = {}
494
+ if request.parent
495
+ header_params["parent"] = request.parent
496
+ end
497
+
473
498
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
474
499
  metadata[:"x-goog-request-params"] ||= request_params_header
475
500
 
476
501
  options.apply_defaults timeout: @config.rpcs.create_transfer_config.timeout,
477
502
  metadata: metadata,
478
503
  retry_policy: @config.rpcs.create_transfer_config.retry_policy
479
- options.apply_defaults metadata: @config.metadata,
504
+
505
+ options.apply_defaults timeout: @config.timeout,
506
+ metadata: @config.metadata,
480
507
  retry_policy: @config.retry_policy
481
508
 
482
509
  @data_transfer_service_stub.call_rpc :create_transfer_config, request, options: options do |response, operation|
@@ -549,6 +576,21 @@ module Google
549
576
  #
550
577
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
551
578
  #
579
+ # @example Basic example
580
+ # require "google/cloud/bigquery/data_transfer/v1"
581
+ #
582
+ # # Create a client object. The client can be reused for multiple calls.
583
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
584
+ #
585
+ # # Create a request. To set request fields, pass in keyword arguments.
586
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::UpdateTransferConfigRequest.new
587
+ #
588
+ # # Call the update_transfer_config method.
589
+ # result = client.update_transfer_config request
590
+ #
591
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
592
+ # p result
593
+ #
552
594
  def update_transfer_config request, options = nil
553
595
  raise ::ArgumentError, "request must be provided" if request.nil?
554
596
 
@@ -566,16 +608,20 @@ module Google
566
608
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
567
609
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
568
610
 
569
- header_params = {
570
- "transfer_config.name" => request.transfer_config.name
571
- }
611
+ header_params = {}
612
+ if request.transfer_config&.name
613
+ header_params["transfer_config.name"] = request.transfer_config.name
614
+ end
615
+
572
616
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
573
617
  metadata[:"x-goog-request-params"] ||= request_params_header
574
618
 
575
619
  options.apply_defaults timeout: @config.rpcs.update_transfer_config.timeout,
576
620
  metadata: metadata,
577
621
  retry_policy: @config.rpcs.update_transfer_config.retry_policy
578
- options.apply_defaults metadata: @config.metadata,
622
+
623
+ options.apply_defaults timeout: @config.timeout,
624
+ metadata: @config.metadata,
579
625
  retry_policy: @config.retry_policy
580
626
 
581
627
  @data_transfer_service_stub.call_rpc :update_transfer_config, request, options: options do |response, operation|
@@ -618,6 +664,21 @@ module Google
618
664
  #
619
665
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
620
666
  #
667
+ # @example Basic example
668
+ # require "google/cloud/bigquery/data_transfer/v1"
669
+ #
670
+ # # Create a client object. The client can be reused for multiple calls.
671
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
672
+ #
673
+ # # Create a request. To set request fields, pass in keyword arguments.
674
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferConfigRequest.new
675
+ #
676
+ # # Call the delete_transfer_config method.
677
+ # result = client.delete_transfer_config request
678
+ #
679
+ # # The returned object is of type Google::Protobuf::Empty.
680
+ # p result
681
+ #
621
682
  def delete_transfer_config request, options = nil
622
683
  raise ::ArgumentError, "request must be provided" if request.nil?
623
684
 
@@ -635,16 +696,20 @@ module Google
635
696
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
636
697
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
637
698
 
638
- header_params = {
639
- "name" => request.name
640
- }
699
+ header_params = {}
700
+ if request.name
701
+ header_params["name"] = request.name
702
+ end
703
+
641
704
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
642
705
  metadata[:"x-goog-request-params"] ||= request_params_header
643
706
 
644
707
  options.apply_defaults timeout: @config.rpcs.delete_transfer_config.timeout,
645
708
  metadata: metadata,
646
709
  retry_policy: @config.rpcs.delete_transfer_config.retry_policy
647
- options.apply_defaults metadata: @config.metadata,
710
+
711
+ options.apply_defaults timeout: @config.timeout,
712
+ metadata: @config.metadata,
648
713
  retry_policy: @config.retry_policy
649
714
 
650
715
  @data_transfer_service_stub.call_rpc :delete_transfer_config, request, options: options do |response, operation|
@@ -686,6 +751,21 @@ module Google
686
751
  #
687
752
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
688
753
  #
754
+ # @example Basic example
755
+ # require "google/cloud/bigquery/data_transfer/v1"
756
+ #
757
+ # # Create a client object. The client can be reused for multiple calls.
758
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
759
+ #
760
+ # # Create a request. To set request fields, pass in keyword arguments.
761
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetTransferConfigRequest.new
762
+ #
763
+ # # Call the get_transfer_config method.
764
+ # result = client.get_transfer_config request
765
+ #
766
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
767
+ # p result
768
+ #
689
769
  def get_transfer_config request, options = nil
690
770
  raise ::ArgumentError, "request must be provided" if request.nil?
691
771
 
@@ -703,16 +783,20 @@ module Google
703
783
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
704
784
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
705
785
 
706
- header_params = {
707
- "name" => request.name
708
- }
786
+ header_params = {}
787
+ if request.name
788
+ header_params["name"] = request.name
789
+ end
790
+
709
791
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
710
792
  metadata[:"x-goog-request-params"] ||= request_params_header
711
793
 
712
794
  options.apply_defaults timeout: @config.rpcs.get_transfer_config.timeout,
713
795
  metadata: metadata,
714
796
  retry_policy: @config.rpcs.get_transfer_config.retry_policy
715
- options.apply_defaults metadata: @config.metadata,
797
+
798
+ options.apply_defaults timeout: @config.timeout,
799
+ metadata: @config.metadata,
716
800
  retry_policy: @config.retry_policy
717
801
 
718
802
  @data_transfer_service_stub.call_rpc :get_transfer_config, request, options: options do |response, operation|
@@ -724,7 +808,8 @@ module Google
724
808
  end
725
809
 
726
810
  ##
727
- # Returns information about all data transfers in the project.
811
+ # Returns information about all transfer configs owned by a project in the
812
+ # specified location.
728
813
  #
729
814
  # @overload list_transfer_configs(request, options = nil)
730
815
  # Pass arguments to `list_transfer_configs` via a request object, either of type
@@ -764,6 +849,27 @@ module Google
764
849
  #
765
850
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
766
851
  #
852
+ # @example Basic example
853
+ # require "google/cloud/bigquery/data_transfer/v1"
854
+ #
855
+ # # Create a client object. The client can be reused for multiple calls.
856
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
857
+ #
858
+ # # Create a request. To set request fields, pass in keyword arguments.
859
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsRequest.new
860
+ #
861
+ # # Call the list_transfer_configs method.
862
+ # result = client.list_transfer_configs request
863
+ #
864
+ # # The returned object is of type Gapic::PagedEnumerable. You can
865
+ # # iterate over all elements by calling #each, and the enumerable
866
+ # # will lazily make API calls to fetch subsequent pages. Other
867
+ # # methods are also available for managing paging directly.
868
+ # result.each do |response|
869
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
870
+ # p response
871
+ # end
872
+ #
767
873
  def list_transfer_configs request, options = nil
768
874
  raise ::ArgumentError, "request must be provided" if request.nil?
769
875
 
@@ -781,16 +887,20 @@ module Google
781
887
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
782
888
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
783
889
 
784
- header_params = {
785
- "parent" => request.parent
786
- }
890
+ header_params = {}
891
+ if request.parent
892
+ header_params["parent"] = request.parent
893
+ end
894
+
787
895
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
788
896
  metadata[:"x-goog-request-params"] ||= request_params_header
789
897
 
790
898
  options.apply_defaults timeout: @config.rpcs.list_transfer_configs.timeout,
791
899
  metadata: metadata,
792
900
  retry_policy: @config.rpcs.list_transfer_configs.retry_policy
793
- options.apply_defaults metadata: @config.metadata,
901
+
902
+ options.apply_defaults timeout: @config.timeout,
903
+ metadata: @config.metadata,
794
904
  retry_policy: @config.retry_policy
795
905
 
796
906
  @data_transfer_service_stub.call_rpc :list_transfer_configs, request, options: options do |response, operation|
@@ -809,6 +919,8 @@ module Google
809
919
  # Note that runs are created per UTC time in the time range.
810
920
  # DEPRECATED: use StartManualTransferRuns instead.
811
921
  #
922
+ # @deprecated This method is deprecated and may be removed in the next major version update.
923
+ #
812
924
  # @overload schedule_transfer_runs(request, options = nil)
813
925
  # Pass arguments to `schedule_transfer_runs` via a request object, either of type
814
926
  # {::Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsRequest} or an equivalent Hash.
@@ -843,6 +955,21 @@ module Google
843
955
  #
844
956
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
845
957
  #
958
+ # @example Basic example
959
+ # require "google/cloud/bigquery/data_transfer/v1"
960
+ #
961
+ # # Create a client object. The client can be reused for multiple calls.
962
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
963
+ #
964
+ # # Create a request. To set request fields, pass in keyword arguments.
965
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsRequest.new
966
+ #
967
+ # # Call the schedule_transfer_runs method.
968
+ # result = client.schedule_transfer_runs request
969
+ #
970
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsResponse.
971
+ # p result
972
+ #
846
973
  def schedule_transfer_runs request, options = nil
847
974
  raise ::ArgumentError, "request must be provided" if request.nil?
848
975
 
@@ -860,16 +987,20 @@ module Google
860
987
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
861
988
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
862
989
 
863
- header_params = {
864
- "parent" => request.parent
865
- }
990
+ header_params = {}
991
+ if request.parent
992
+ header_params["parent"] = request.parent
993
+ end
994
+
866
995
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
867
996
  metadata[:"x-goog-request-params"] ||= request_params_header
868
997
 
869
998
  options.apply_defaults timeout: @config.rpcs.schedule_transfer_runs.timeout,
870
999
  metadata: metadata,
871
1000
  retry_policy: @config.rpcs.schedule_transfer_runs.retry_policy
872
- options.apply_defaults metadata: @config.metadata,
1001
+
1002
+ options.apply_defaults timeout: @config.timeout,
1003
+ metadata: @config.metadata,
873
1004
  retry_policy: @config.retry_policy
874
1005
 
875
1006
  @data_transfer_service_stub.call_rpc :schedule_transfer_runs, request, options: options do |response, operation|
@@ -919,6 +1050,21 @@ module Google
919
1050
  #
920
1051
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
921
1052
  #
1053
+ # @example Basic example
1054
+ # require "google/cloud/bigquery/data_transfer/v1"
1055
+ #
1056
+ # # Create a client object. The client can be reused for multiple calls.
1057
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1058
+ #
1059
+ # # Create a request. To set request fields, pass in keyword arguments.
1060
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::StartManualTransferRunsRequest.new
1061
+ #
1062
+ # # Call the start_manual_transfer_runs method.
1063
+ # result = client.start_manual_transfer_runs request
1064
+ #
1065
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::StartManualTransferRunsResponse.
1066
+ # p result
1067
+ #
922
1068
  def start_manual_transfer_runs request, options = nil
923
1069
  raise ::ArgumentError, "request must be provided" if request.nil?
924
1070
 
@@ -936,16 +1082,20 @@ module Google
936
1082
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
937
1083
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
938
1084
 
939
- header_params = {
940
- "parent" => request.parent
941
- }
1085
+ header_params = {}
1086
+ if request.parent
1087
+ header_params["parent"] = request.parent
1088
+ end
1089
+
942
1090
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
943
1091
  metadata[:"x-goog-request-params"] ||= request_params_header
944
1092
 
945
1093
  options.apply_defaults timeout: @config.rpcs.start_manual_transfer_runs.timeout,
946
1094
  metadata: metadata,
947
1095
  retry_policy: @config.rpcs.start_manual_transfer_runs.retry_policy
948
- options.apply_defaults metadata: @config.metadata,
1096
+
1097
+ options.apply_defaults timeout: @config.timeout,
1098
+ metadata: @config.metadata,
949
1099
  retry_policy: @config.retry_policy
950
1100
 
951
1101
  @data_transfer_service_stub.call_rpc :start_manual_transfer_runs, request, options: options do |response, operation|
@@ -987,6 +1137,21 @@ module Google
987
1137
  #
988
1138
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
989
1139
  #
1140
+ # @example Basic example
1141
+ # require "google/cloud/bigquery/data_transfer/v1"
1142
+ #
1143
+ # # Create a client object. The client can be reused for multiple calls.
1144
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1145
+ #
1146
+ # # Create a request. To set request fields, pass in keyword arguments.
1147
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetTransferRunRequest.new
1148
+ #
1149
+ # # Call the get_transfer_run method.
1150
+ # result = client.get_transfer_run request
1151
+ #
1152
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferRun.
1153
+ # p result
1154
+ #
990
1155
  def get_transfer_run request, options = nil
991
1156
  raise ::ArgumentError, "request must be provided" if request.nil?
992
1157
 
@@ -1004,16 +1169,20 @@ module Google
1004
1169
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1005
1170
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1006
1171
 
1007
- header_params = {
1008
- "name" => request.name
1009
- }
1172
+ header_params = {}
1173
+ if request.name
1174
+ header_params["name"] = request.name
1175
+ end
1176
+
1010
1177
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1011
1178
  metadata[:"x-goog-request-params"] ||= request_params_header
1012
1179
 
1013
1180
  options.apply_defaults timeout: @config.rpcs.get_transfer_run.timeout,
1014
1181
  metadata: metadata,
1015
1182
  retry_policy: @config.rpcs.get_transfer_run.retry_policy
1016
- options.apply_defaults metadata: @config.metadata,
1183
+
1184
+ options.apply_defaults timeout: @config.timeout,
1185
+ metadata: @config.metadata,
1017
1186
  retry_policy: @config.retry_policy
1018
1187
 
1019
1188
  @data_transfer_service_stub.call_rpc :get_transfer_run, request, options: options do |response, operation|
@@ -1055,6 +1224,21 @@ module Google
1055
1224
  #
1056
1225
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1057
1226
  #
1227
+ # @example Basic example
1228
+ # require "google/cloud/bigquery/data_transfer/v1"
1229
+ #
1230
+ # # Create a client object. The client can be reused for multiple calls.
1231
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1232
+ #
1233
+ # # Create a request. To set request fields, pass in keyword arguments.
1234
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferRunRequest.new
1235
+ #
1236
+ # # Call the delete_transfer_run method.
1237
+ # result = client.delete_transfer_run request
1238
+ #
1239
+ # # The returned object is of type Google::Protobuf::Empty.
1240
+ # p result
1241
+ #
1058
1242
  def delete_transfer_run request, options = nil
1059
1243
  raise ::ArgumentError, "request must be provided" if request.nil?
1060
1244
 
@@ -1072,16 +1256,20 @@ module Google
1072
1256
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1073
1257
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1074
1258
 
1075
- header_params = {
1076
- "name" => request.name
1077
- }
1259
+ header_params = {}
1260
+ if request.name
1261
+ header_params["name"] = request.name
1262
+ end
1263
+
1078
1264
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1079
1265
  metadata[:"x-goog-request-params"] ||= request_params_header
1080
1266
 
1081
1267
  options.apply_defaults timeout: @config.rpcs.delete_transfer_run.timeout,
1082
1268
  metadata: metadata,
1083
1269
  retry_policy: @config.rpcs.delete_transfer_run.retry_policy
1084
- options.apply_defaults metadata: @config.metadata,
1270
+
1271
+ options.apply_defaults timeout: @config.timeout,
1272
+ metadata: @config.metadata,
1085
1273
  retry_policy: @config.retry_policy
1086
1274
 
1087
1275
  @data_transfer_service_stub.call_rpc :delete_transfer_run, request, options: options do |response, operation|
@@ -1136,6 +1324,27 @@ module Google
1136
1324
  #
1137
1325
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1138
1326
  #
1327
+ # @example Basic example
1328
+ # require "google/cloud/bigquery/data_transfer/v1"
1329
+ #
1330
+ # # Create a client object. The client can be reused for multiple calls.
1331
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1332
+ #
1333
+ # # Create a request. To set request fields, pass in keyword arguments.
1334
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferRunsRequest.new
1335
+ #
1336
+ # # Call the list_transfer_runs method.
1337
+ # result = client.list_transfer_runs request
1338
+ #
1339
+ # # The returned object is of type Gapic::PagedEnumerable. You can
1340
+ # # iterate over all elements by calling #each, and the enumerable
1341
+ # # will lazily make API calls to fetch subsequent pages. Other
1342
+ # # methods are also available for managing paging directly.
1343
+ # result.each do |response|
1344
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferRun.
1345
+ # p response
1346
+ # end
1347
+ #
1139
1348
  def list_transfer_runs request, options = nil
1140
1349
  raise ::ArgumentError, "request must be provided" if request.nil?
1141
1350
 
@@ -1153,16 +1362,20 @@ module Google
1153
1362
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1154
1363
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1155
1364
 
1156
- header_params = {
1157
- "parent" => request.parent
1158
- }
1365
+ header_params = {}
1366
+ if request.parent
1367
+ header_params["parent"] = request.parent
1368
+ end
1369
+
1159
1370
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1160
1371
  metadata[:"x-goog-request-params"] ||= request_params_header
1161
1372
 
1162
1373
  options.apply_defaults timeout: @config.rpcs.list_transfer_runs.timeout,
1163
1374
  metadata: metadata,
1164
1375
  retry_policy: @config.rpcs.list_transfer_runs.retry_policy
1165
- options.apply_defaults metadata: @config.metadata,
1376
+
1377
+ options.apply_defaults timeout: @config.timeout,
1378
+ metadata: @config.metadata,
1166
1379
  retry_policy: @config.retry_policy
1167
1380
 
1168
1381
  @data_transfer_service_stub.call_rpc :list_transfer_runs, request, options: options do |response, operation|
@@ -1216,6 +1429,27 @@ module Google
1216
1429
  #
1217
1430
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1218
1431
  #
1432
+ # @example Basic example
1433
+ # require "google/cloud/bigquery/data_transfer/v1"
1434
+ #
1435
+ # # Create a client object. The client can be reused for multiple calls.
1436
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1437
+ #
1438
+ # # Create a request. To set request fields, pass in keyword arguments.
1439
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferLogsRequest.new
1440
+ #
1441
+ # # Call the list_transfer_logs method.
1442
+ # result = client.list_transfer_logs request
1443
+ #
1444
+ # # The returned object is of type Gapic::PagedEnumerable. You can
1445
+ # # iterate over all elements by calling #each, and the enumerable
1446
+ # # will lazily make API calls to fetch subsequent pages. Other
1447
+ # # methods are also available for managing paging directly.
1448
+ # result.each do |response|
1449
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferMessage.
1450
+ # p response
1451
+ # end
1452
+ #
1219
1453
  def list_transfer_logs request, options = nil
1220
1454
  raise ::ArgumentError, "request must be provided" if request.nil?
1221
1455
 
@@ -1233,16 +1467,20 @@ module Google
1233
1467
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1234
1468
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1235
1469
 
1236
- header_params = {
1237
- "parent" => request.parent
1238
- }
1470
+ header_params = {}
1471
+ if request.parent
1472
+ header_params["parent"] = request.parent
1473
+ end
1474
+
1239
1475
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1240
1476
  metadata[:"x-goog-request-params"] ||= request_params_header
1241
1477
 
1242
1478
  options.apply_defaults timeout: @config.rpcs.list_transfer_logs.timeout,
1243
1479
  metadata: metadata,
1244
1480
  retry_policy: @config.rpcs.list_transfer_logs.retry_policy
1245
- options.apply_defaults metadata: @config.metadata,
1481
+
1482
+ options.apply_defaults timeout: @config.timeout,
1483
+ metadata: @config.metadata,
1246
1484
  retry_policy: @config.retry_policy
1247
1485
 
1248
1486
  @data_transfer_service_stub.call_rpc :list_transfer_logs, request, options: options do |response, operation|
@@ -1290,6 +1528,21 @@ module Google
1290
1528
  #
1291
1529
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1292
1530
  #
1531
+ # @example Basic example
1532
+ # require "google/cloud/bigquery/data_transfer/v1"
1533
+ #
1534
+ # # Create a client object. The client can be reused for multiple calls.
1535
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1536
+ #
1537
+ # # Create a request. To set request fields, pass in keyword arguments.
1538
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::CheckValidCredsRequest.new
1539
+ #
1540
+ # # Call the check_valid_creds method.
1541
+ # result = client.check_valid_creds request
1542
+ #
1543
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::CheckValidCredsResponse.
1544
+ # p result
1545
+ #
1293
1546
  def check_valid_creds request, options = nil
1294
1547
  raise ::ArgumentError, "request must be provided" if request.nil?
1295
1548
 
@@ -1307,16 +1560,20 @@ module Google
1307
1560
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1308
1561
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1309
1562
 
1310
- header_params = {
1311
- "name" => request.name
1312
- }
1563
+ header_params = {}
1564
+ if request.name
1565
+ header_params["name"] = request.name
1566
+ end
1567
+
1313
1568
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1314
1569
  metadata[:"x-goog-request-params"] ||= request_params_header
1315
1570
 
1316
1571
  options.apply_defaults timeout: @config.rpcs.check_valid_creds.timeout,
1317
1572
  metadata: metadata,
1318
1573
  retry_policy: @config.rpcs.check_valid_creds.retry_policy
1319
- options.apply_defaults metadata: @config.metadata,
1574
+
1575
+ options.apply_defaults timeout: @config.timeout,
1576
+ metadata: @config.metadata,
1320
1577
  retry_policy: @config.retry_policy
1321
1578
 
1322
1579
  @data_transfer_service_stub.call_rpc :check_valid_creds, request, options: options do |response, operation|
@@ -1340,22 +1597,21 @@ module Google
1340
1597
  # Configuration can be applied globally to all clients, or to a single client
1341
1598
  # on construction.
1342
1599
  #
1343
- # # Examples
1344
- #
1345
- # To modify the global config, setting the timeout for get_data_source
1346
- # to 20 seconds, and all remaining timeouts to 10 seconds:
1347
- #
1348
- # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
1349
- # config.timeout = 10.0
1350
- # config.rpcs.get_data_source.timeout = 20.0
1351
- # end
1352
- #
1353
- # To apply the above configuration only to a new client:
1354
- #
1355
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
1356
- # config.timeout = 10.0
1357
- # config.rpcs.get_data_source.timeout = 20.0
1358
- # end
1600
+ # @example
1601
+ #
1602
+ # # Modify the global config, setting the timeout for
1603
+ # # get_data_source to 20 seconds,
1604
+ # # and all remaining timeouts to 10 seconds.
1605
+ # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
1606
+ # config.timeout = 10.0
1607
+ # config.rpcs.get_data_source.timeout = 20.0
1608
+ # end
1609
+ #
1610
+ # # Apply the above configuration only to a new client.
1611
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
1612
+ # config.timeout = 10.0
1613
+ # config.rpcs.get_data_source.timeout = 20.0
1614
+ # end
1359
1615
  #
1360
1616
  # @!attribute [rw] endpoint
1361
1617
  # The hostname or hostname:port of the service endpoint.
@@ -22,7 +22,7 @@ module Google
22
22
  module Bigquery
23
23
  module DataTransfer
24
24
  module V1
25
- VERSION = "0.4.1"
25
+ VERSION = "0.4.5"
26
26
  end
27
27
  end
28
28
  end
@@ -1,8 +1,6 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/datatransfer/v1/datatransfer.proto
3
3
 
4
- require 'google/protobuf'
5
-
6
4
  require 'google/api/annotations_pb'
7
5
  require 'google/api/client_pb'
8
6
  require 'google/api/field_behavior_pb'
@@ -13,6 +11,8 @@ require 'google/protobuf/empty_pb'
13
11
  require 'google/protobuf/field_mask_pb'
14
12
  require 'google/protobuf/timestamp_pb'
15
13
  require 'google/protobuf/wrappers_pb'
14
+ require 'google/protobuf'
15
+
16
16
  Google::Protobuf::DescriptorPool.generated_pool.build do
17
17
  add_file("google/cloud/bigquery/datatransfer/v1/datatransfer.proto", :syntax => :proto3) do
18
18
  add_message "google.cloud.bigquery.datatransfer.v1.DataSourceParameter" do
@@ -53,7 +53,8 @@ module Google
53
53
  rpc :DeleteTransferConfig, ::Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferConfigRequest, ::Google::Protobuf::Empty
54
54
  # Returns information about a data transfer config.
55
55
  rpc :GetTransferConfig, ::Google::Cloud::Bigquery::DataTransfer::V1::GetTransferConfigRequest, ::Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig
56
- # Returns information about all data transfers in the project.
56
+ # Returns information about all transfer configs owned by a project in the
57
+ # specified location.
57
58
  rpc :ListTransferConfigs, ::Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsRequest, ::Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsResponse
58
59
  # Creates transfer runs for a time range [start_time, end_time].
59
60
  # For each date - or whatever granularity the data source supports - in the
@@ -1,13 +1,14 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/datatransfer/v1/transfer.proto
3
3
 
4
- require 'google/protobuf'
5
-
6
4
  require 'google/api/field_behavior_pb'
7
5
  require 'google/api/resource_pb'
6
+ require 'google/protobuf/duration_pb'
8
7
  require 'google/protobuf/struct_pb'
9
8
  require 'google/protobuf/timestamp_pb'
10
9
  require 'google/rpc/status_pb'
10
+ require 'google/protobuf'
11
+
11
12
  Google::Protobuf::DescriptorPool.generated_pool.build do
12
13
  add_file("google/cloud/bigquery/datatransfer/v1/transfer.proto", :syntax => :proto3) do
13
14
  add_message "google.cloud.bigquery.datatransfer.v1.EmailPreferences" do
@@ -57,9 +57,15 @@ module Google
57
57
 
58
58
  # Denotes that a (repeated) field is an unordered list.
59
59
  # This indicates that the service may provide the elements of the list
60
- # in any arbitrary order, rather than the order the user originally
60
+ # in any arbitrary order, rather than the order the user originally
61
61
  # provided. Additionally, the list's order may or may not be stable.
62
62
  UNORDERED_LIST = 6
63
+
64
+ # Denotes that this field returns a non-empty default value if not set.
65
+ # This indicates that if the user provides the empty value in a request,
66
+ # a non-empty value will be returned. The user will not be aware of what
67
+ # non-empty value to expect.
68
+ NON_EMPTY_DEFAULT = 7
63
69
  end
64
70
  end
65
71
  end
@@ -195,9 +195,7 @@ module Google
195
195
  # exchanged for a refresh token on the backend.
196
196
  GOOGLE_PLUS_AUTHORIZATION_CODE = 2
197
197
 
198
- # Use First Party Client OAuth. First Party Client OAuth doesn't require a
199
- # refresh token to get an offline access token. Instead, it uses a
200
- # client-signed JWT assertion to retrieve an access token.
198
+ # Use First Party OAuth.
201
199
  FIRST_PARTY_OAUTH = 3
202
200
  end
203
201
 
@@ -617,14 +615,14 @@ module Google
617
615
  # @return [::Google::Protobuf::Timestamp]
618
616
  # Start time of the range of transfer runs. For example,
619
617
  # `"2017-05-25T00:00:00+00:00"`. The start_time must be strictly less than
620
- # the end_time. Creates transfer runs where run_time is in the range betwen
621
- # start_time (inclusive) and end_time (exlusive).
618
+ # the end_time. Creates transfer runs where run_time is in the range
619
+ # between start_time (inclusive) and end_time (exclusive).
622
620
  # @!attribute [rw] end_time
623
621
  # @return [::Google::Protobuf::Timestamp]
624
622
  # End time of the range of transfer runs. For example,
625
623
  # `"2017-05-30T00:00:00+00:00"`. The end_time must not be in the future.
626
- # Creates transfer runs where run_time is in the range betwen start_time
627
- # (inclusive) and end_time (exlusive).
624
+ # Creates transfer runs where run_time is in the range between start_time
625
+ # (inclusive) and end_time (exclusive).
628
626
  class TimeRange
629
627
  include ::Google::Protobuf::MessageExts
630
628
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -66,12 +66,11 @@ module Google
66
66
  # @!attribute [rw] name
67
67
  # @return [::String]
68
68
  # The resource name of the transfer config.
69
- # Transfer config names have the form of
69
+ # Transfer config names have the form
70
70
  # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
71
- # The name is automatically generated based on the config_id specified in
72
- # CreateTransferConfigRequest along with project_id and region. If config_id
73
- # is not provided, usually a uuid, even though it is not guaranteed or
74
- # required, will be generated for config_id.
71
+ # Where `config_id` is usually a uuid, even though it is not
72
+ # guaranteed or required. The name is ignored when creating a transfer
73
+ # config.
75
74
  # @!attribute [rw] destination_dataset_id
76
75
  # @return [::String]
77
76
  # The BigQuery target dataset id.
@@ -83,7 +82,10 @@ module Google
83
82
  # Data source id. Cannot be changed once data transfer is created.
84
83
  # @!attribute [rw] params
85
84
  # @return [::Google::Protobuf::Struct]
86
- # Data transfer specific parameters.
85
+ # Parameters specific to each data source. For more information see the
86
+ # bq tab in the 'Setting up a data transfer' section for each data source.
87
+ # For example the parameters for Cloud Storage transfers are listed here:
88
+ # https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
87
89
  # @!attribute [rw] schedule
88
90
  # @return [::String]
89
91
  # Data transfer schedule.
@@ -132,6 +134,9 @@ module Google
132
134
  # @return [::String]
133
135
  # Pub/Sub topic where notifications will be sent after transfer runs
134
136
  # associated with this transfer config finish.
137
+ #
138
+ # The format for specifying a pubsub topic is:
139
+ # `projects/{project}/topics/{topic}`
135
140
  # @!attribute [rw] email_preferences
136
141
  # @return [::Google::Cloud::Bigquery::DataTransfer::V1::EmailPreferences]
137
142
  # Email notifications will be sent according to these preferences
@@ -171,7 +176,10 @@ module Google
171
176
  # Output only. Last time the data transfer run state was updated.
172
177
  # @!attribute [r] params
173
178
  # @return [::Google::Protobuf::Struct]
174
- # Output only. Data transfer specific parameters.
179
+ # Output only. Parameters specific to each data source. For more information see the
180
+ # bq tab in the 'Setting up a data transfer' section for each data source.
181
+ # For example the parameters for Cloud Storage transfers are listed here:
182
+ # https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
175
183
  # @!attribute [r] destination_dataset_id
176
184
  # @return [::String]
177
185
  # Output only. The BigQuery target dataset id.
@@ -194,7 +202,10 @@ module Google
194
202
  # @!attribute [r] notification_pubsub_topic
195
203
  # @return [::String]
196
204
  # Output only. Pub/Sub topic where a notification will be sent after this
197
- # transfer run finishes
205
+ # transfer run finishes.
206
+ #
207
+ # The format for specifying a pubsub topic is:
208
+ # `projects/{project}/topics/{topic}`
198
209
  # @!attribute [r] email_preferences
199
210
  # @return [::Google::Cloud::Bigquery::DataTransfer::V1::EmailPreferences]
200
211
  # Output only. Email notifications will be sent according to these
@@ -250,23 +261,23 @@ module Google
250
261
 
251
262
  # Represents data transfer run state.
252
263
  module TransferState
253
- # State placeholder.
264
+ # State placeholder (0).
254
265
  TRANSFER_STATE_UNSPECIFIED = 0
255
266
 
256
267
  # Data transfer is scheduled and is waiting to be picked up by
257
- # data transfer backend.
268
+ # data transfer backend (2).
258
269
  PENDING = 2
259
270
 
260
- # Data transfer is in progress.
271
+ # Data transfer is in progress (3).
261
272
  RUNNING = 3
262
273
 
263
- # Data transfer completed successfully.
274
+ # Data transfer completed successfully (4).
264
275
  SUCCEEDED = 4
265
276
 
266
- # Data transfer failed.
277
+ # Data transfer failed (5).
267
278
  FAILED = 5
268
279
 
269
- # Data transfer is cancelled.
280
+ # Data transfer is cancelled (6).
270
281
  CANCELLED = 6
271
282
  end
272
283
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-bigquery-data_transfer-v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.4.1
4
+ version: 0.4.5
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2021-06-17 00:00:00.000000000 Z
11
+ date: 2021-11-08 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: gapic-common
@@ -16,7 +16,7 @@ dependencies:
16
16
  requirements:
17
17
  - - ">="
18
18
  - !ruby/object:Gem::Version
19
- version: '0.5'
19
+ version: '0.7'
20
20
  - - "<"
21
21
  - !ruby/object:Gem::Version
22
22
  version: 2.a
@@ -26,7 +26,7 @@ dependencies:
26
26
  requirements:
27
27
  - - ">="
28
28
  - !ruby/object:Gem::Version
29
- version: '0.5'
29
+ version: '0.7'
30
30
  - - "<"
31
31
  - !ruby/object:Gem::Version
32
32
  version: 2.a