google-cloud-bigquery-data_transfer-v1 0.4.1 → 0.4.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: fcc4faa89caae40371c403488bc3c44a952fa421b1737b03e8af332016b0a17e
4
- data.tar.gz: acf1ca1a755615520405b789399235ff88d4eea1ff8b3df41b6f455eec9b4abf
3
+ metadata.gz: 858ac9046282b3f9f1ad16896571a7ec0ff7167e6a41f4a677e3623b4e08eaaa
4
+ data.tar.gz: caefa5a2b16a67126c8915ddcb5bc9d85cf62a06bce500d35fb6c7c5630b2b75
5
5
  SHA512:
6
- metadata.gz: f25a01f491c142e65f7340c437d7121f2c500179049914b39a38121d2c449ada8d1343b23ec2fc4827eb2a9cb680cd25ce7dd3081930764f8225585a787dbb80
7
- data.tar.gz: e15f2a12b29b8685fcfce8a33cce18d568859268acbb60b4694bb821578870e4a50563ac8c3213ffeac8d2a6f1365c9a2ec467b81ca74bb0df07bda20d9a2179
6
+ metadata.gz: cbe0687f2bb91583fec9192380f098421d1bec4f6726aa4155cba8ba855e21dc551d063eec692974aa3955003bb8fb8bd91e3605bcb9e0f9110b97a04544e361
7
+ data.tar.gz: 1587893e7feeb969dcc82f8973facf0e835c19cfcbb46b80c538c4ef91d7edf091c26b7ab37d4eeabb769c89b160bcdb3d4a1b370809626a3fbadd19ca74872e
data/AUTHENTICATION.md CHANGED
@@ -66,11 +66,11 @@ The environment variables that google-cloud-bigquery-data_transfer-v1
66
66
  checks for credentials are configured on the service Credentials class (such as
67
67
  {::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Credentials}):
68
68
 
69
- 1. `DATA_TRANSFER_CREDENTIALS` - Path to JSON file, or JSON contents
70
- 2. `DATA_TRANSFER_KEYFILE` - Path to JSON file, or JSON contents
71
- 3. `GOOGLE_CLOUD_CREDENTIALS` - Path to JSON file, or JSON contents
72
- 4. `GOOGLE_CLOUD_KEYFILE` - Path to JSON file, or JSON contents
73
- 5. `GOOGLE_APPLICATION_CREDENTIALS` - Path to JSON file
69
+ * `DATA_TRANSFER_CREDENTIALS` - Path to JSON file, or JSON contents
70
+ * `DATA_TRANSFER_KEYFILE` - Path to JSON file, or JSON contents
71
+ * `GOOGLE_CLOUD_CREDENTIALS` - Path to JSON file, or JSON contents
72
+ * `GOOGLE_CLOUD_KEYFILE` - Path to JSON file, or JSON contents
73
+ * `GOOGLE_APPLICATION_CREDENTIALS` - Path to JSON file
74
74
 
75
75
  ```ruby
76
76
  require "google/cloud/bigquery/data_transfer/v1"
@@ -82,8 +82,8 @@ client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Clien
82
82
 
83
83
  ### Configuration
84
84
 
85
- The **Credentials JSON** can be configured instead of placing them in
86
- environment variables. Either on an individual client initialization:
85
+ The path to the **Credentials JSON** file can be configured instead of storing
86
+ it in an environment variable. Either on an individual client initialization:
87
87
 
88
88
  ```ruby
89
89
  require "google/cloud/bigquery/data_transfer/v1"
@@ -93,7 +93,7 @@ client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Clien
93
93
  end
94
94
  ```
95
95
 
96
- Or configured globally for all clients:
96
+ Or globally for all clients:
97
97
 
98
98
  ```ruby
99
99
  require "google/cloud/bigquery/data_transfer/v1"
data/README.md CHANGED
@@ -33,7 +33,7 @@ In order to use this library, you first need to go through the following steps:
33
33
  require "google/cloud/bigquery/data_transfer/v1"
34
34
 
35
35
  client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
36
- request = my_create_request
36
+ request = ::Google::Cloud::Bigquery::DataTransfer::V1::GetDataSourceRequest.new # (request fields as keyword arguments...)
37
37
  response = client.get_data_source request
38
38
  ```
39
39
 
@@ -45,13 +45,12 @@ module Google
45
45
  # See {::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client::Configuration}
46
46
  # for a description of the configuration fields.
47
47
  #
48
- # ## Example
48
+ # @example
49
49
  #
50
- # To modify the configuration for all DataTransferService clients:
51
- #
52
- # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
53
- # config.timeout = 10.0
54
- # end
50
+ # # Modify the configuration for all DataTransferService clients
51
+ # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
52
+ # config.timeout = 10.0
53
+ # end
55
54
  #
56
55
  # @yield [config] Configure the Client client.
57
56
  # @yieldparam config [Client::Configuration]
@@ -71,18 +70,12 @@ module Google
71
70
 
72
71
  default_config.rpcs.get_data_source.timeout = 20.0
73
72
  default_config.rpcs.get_data_source.retry_policy = {
74
- initial_delay: 0.1,
75
- max_delay: 60.0,
76
- multiplier: 1.3,
77
- retry_codes: [14, 4]
73
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
78
74
  }
79
75
 
80
76
  default_config.rpcs.list_data_sources.timeout = 20.0
81
77
  default_config.rpcs.list_data_sources.retry_policy = {
82
- initial_delay: 0.1,
83
- max_delay: 60.0,
84
- multiplier: 1.3,
85
- retry_codes: [14, 4]
78
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
86
79
  }
87
80
 
88
81
  default_config.rpcs.create_transfer_config.timeout = 30.0
@@ -91,68 +84,44 @@ module Google
91
84
 
92
85
  default_config.rpcs.delete_transfer_config.timeout = 20.0
93
86
  default_config.rpcs.delete_transfer_config.retry_policy = {
94
- initial_delay: 0.1,
95
- max_delay: 60.0,
96
- multiplier: 1.3,
97
- retry_codes: [14, 4]
87
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
98
88
  }
99
89
 
100
90
  default_config.rpcs.get_transfer_config.timeout = 20.0
101
91
  default_config.rpcs.get_transfer_config.retry_policy = {
102
- initial_delay: 0.1,
103
- max_delay: 60.0,
104
- multiplier: 1.3,
105
- retry_codes: [14, 4]
92
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
106
93
  }
107
94
 
108
95
  default_config.rpcs.list_transfer_configs.timeout = 20.0
109
96
  default_config.rpcs.list_transfer_configs.retry_policy = {
110
- initial_delay: 0.1,
111
- max_delay: 60.0,
112
- multiplier: 1.3,
113
- retry_codes: [14, 4]
97
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
114
98
  }
115
99
 
116
100
  default_config.rpcs.schedule_transfer_runs.timeout = 30.0
117
101
 
118
102
  default_config.rpcs.get_transfer_run.timeout = 20.0
119
103
  default_config.rpcs.get_transfer_run.retry_policy = {
120
- initial_delay: 0.1,
121
- max_delay: 60.0,
122
- multiplier: 1.3,
123
- retry_codes: [14, 4]
104
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
124
105
  }
125
106
 
126
107
  default_config.rpcs.delete_transfer_run.timeout = 20.0
127
108
  default_config.rpcs.delete_transfer_run.retry_policy = {
128
- initial_delay: 0.1,
129
- max_delay: 60.0,
130
- multiplier: 1.3,
131
- retry_codes: [14, 4]
109
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
132
110
  }
133
111
 
134
112
  default_config.rpcs.list_transfer_runs.timeout = 20.0
135
113
  default_config.rpcs.list_transfer_runs.retry_policy = {
136
- initial_delay: 0.1,
137
- max_delay: 60.0,
138
- multiplier: 1.3,
139
- retry_codes: [14, 4]
114
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
140
115
  }
141
116
 
142
117
  default_config.rpcs.list_transfer_logs.timeout = 20.0
143
118
  default_config.rpcs.list_transfer_logs.retry_policy = {
144
- initial_delay: 0.1,
145
- max_delay: 60.0,
146
- multiplier: 1.3,
147
- retry_codes: [14, 4]
119
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
148
120
  }
149
121
 
150
122
  default_config.rpcs.check_valid_creds.timeout = 20.0
151
123
  default_config.rpcs.check_valid_creds.retry_policy = {
152
- initial_delay: 0.1,
153
- max_delay: 60.0,
154
- multiplier: 1.3,
155
- retry_codes: [14, 4]
124
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 4]
156
125
  }
157
126
 
158
127
  default_config
@@ -184,19 +153,15 @@ module Google
184
153
  ##
185
154
  # Create a new DataTransferService client object.
186
155
  #
187
- # ## Examples
188
- #
189
- # To create a new DataTransferService client with the default
190
- # configuration:
156
+ # @example
191
157
  #
192
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
158
+ # # Create a client using the default configuration
159
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
193
160
  #
194
- # To create a new DataTransferService client with a custom
195
- # configuration:
196
- #
197
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
198
- # config.timeout = 10.0
199
- # end
161
+ # # Create a client using a custom configuration
162
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
163
+ # config.timeout = 10.0
164
+ # end
200
165
  #
201
166
  # @yield [config] Configure the DataTransferService client.
202
167
  # @yieldparam config [Client::Configuration]
@@ -216,14 +181,13 @@ module Google
216
181
 
217
182
  # Create credentials
218
183
  credentials = @config.credentials
219
- # Use self-signed JWT if the scope and endpoint are unchanged from default,
184
+ # Use self-signed JWT if the endpoint is unchanged from default,
220
185
  # but only if the default endpoint does not have a region prefix.
221
- enable_self_signed_jwt = @config.scope == Client.configure.scope &&
222
- @config.endpoint == Client.configure.endpoint &&
186
+ enable_self_signed_jwt = @config.endpoint == Client.configure.endpoint &&
223
187
  !@config.endpoint.split(".").first.include?("-")
224
188
  credentials ||= Credentials.default scope: @config.scope,
225
189
  enable_self_signed_jwt: enable_self_signed_jwt
226
- if credentials.is_a?(String) || credentials.is_a?(Hash)
190
+ if credentials.is_a?(::String) || credentials.is_a?(::Hash)
227
191
  credentials = Credentials.new credentials, scope: @config.scope
228
192
  end
229
193
  @quota_project_id = @config.quota_project
@@ -272,6 +236,21 @@ module Google
272
236
  #
273
237
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
274
238
  #
239
+ # @example Basic example
240
+ # require "google/cloud/bigquery/data_transfer/v1"
241
+ #
242
+ # # Create a client object. The client can be reused for multiple calls.
243
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
244
+ #
245
+ # # Create a request. To set request fields, pass in keyword arguments.
246
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetDataSourceRequest.new
247
+ #
248
+ # # Call the get_data_source method.
249
+ # result = client.get_data_source request
250
+ #
251
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::DataSource.
252
+ # p result
253
+ #
275
254
  def get_data_source request, options = nil
276
255
  raise ::ArgumentError, "request must be provided" if request.nil?
277
256
 
@@ -289,16 +268,20 @@ module Google
289
268
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
290
269
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
291
270
 
292
- header_params = {
293
- "name" => request.name
294
- }
271
+ header_params = {}
272
+ if request.name
273
+ header_params["name"] = request.name
274
+ end
275
+
295
276
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
296
277
  metadata[:"x-goog-request-params"] ||= request_params_header
297
278
 
298
279
  options.apply_defaults timeout: @config.rpcs.get_data_source.timeout,
299
280
  metadata: metadata,
300
281
  retry_policy: @config.rpcs.get_data_source.retry_policy
301
- options.apply_defaults metadata: @config.metadata,
282
+
283
+ options.apply_defaults timeout: @config.timeout,
284
+ metadata: @config.metadata,
302
285
  retry_policy: @config.retry_policy
303
286
 
304
287
  @data_transfer_service_stub.call_rpc :get_data_source, request, options: options do |response, operation|
@@ -349,6 +332,27 @@ module Google
349
332
  #
350
333
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
351
334
  #
335
+ # @example Basic example
336
+ # require "google/cloud/bigquery/data_transfer/v1"
337
+ #
338
+ # # Create a client object. The client can be reused for multiple calls.
339
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
340
+ #
341
+ # # Create a request. To set request fields, pass in keyword arguments.
342
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListDataSourcesRequest.new
343
+ #
344
+ # # Call the list_data_sources method.
345
+ # result = client.list_data_sources request
346
+ #
347
+ # # The returned object is of type Gapic::PagedEnumerable. You can
348
+ # # iterate over all elements by calling #each, and the enumerable
349
+ # # will lazily make API calls to fetch subsequent pages. Other
350
+ # # methods are also available for managing paging directly.
351
+ # result.each do |response|
352
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::DataSource.
353
+ # p response
354
+ # end
355
+ #
352
356
  def list_data_sources request, options = nil
353
357
  raise ::ArgumentError, "request must be provided" if request.nil?
354
358
 
@@ -366,16 +370,20 @@ module Google
366
370
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
367
371
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
368
372
 
369
- header_params = {
370
- "parent" => request.parent
371
- }
373
+ header_params = {}
374
+ if request.parent
375
+ header_params["parent"] = request.parent
376
+ end
377
+
372
378
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
373
379
  metadata[:"x-goog-request-params"] ||= request_params_header
374
380
 
375
381
  options.apply_defaults timeout: @config.rpcs.list_data_sources.timeout,
376
382
  metadata: metadata,
377
383
  retry_policy: @config.rpcs.list_data_sources.retry_policy
378
- options.apply_defaults metadata: @config.metadata,
384
+
385
+ options.apply_defaults timeout: @config.timeout,
386
+ metadata: @config.metadata,
379
387
  retry_policy: @config.retry_policy
380
388
 
381
389
  @data_transfer_service_stub.call_rpc :list_data_sources, request, options: options do |response, operation|
@@ -450,6 +458,21 @@ module Google
450
458
  #
451
459
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
452
460
  #
461
+ # @example Basic example
462
+ # require "google/cloud/bigquery/data_transfer/v1"
463
+ #
464
+ # # Create a client object. The client can be reused for multiple calls.
465
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
466
+ #
467
+ # # Create a request. To set request fields, pass in keyword arguments.
468
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::CreateTransferConfigRequest.new
469
+ #
470
+ # # Call the create_transfer_config method.
471
+ # result = client.create_transfer_config request
472
+ #
473
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
474
+ # p result
475
+ #
453
476
  def create_transfer_config request, options = nil
454
477
  raise ::ArgumentError, "request must be provided" if request.nil?
455
478
 
@@ -467,16 +490,20 @@ module Google
467
490
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
468
491
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
469
492
 
470
- header_params = {
471
- "parent" => request.parent
472
- }
493
+ header_params = {}
494
+ if request.parent
495
+ header_params["parent"] = request.parent
496
+ end
497
+
473
498
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
474
499
  metadata[:"x-goog-request-params"] ||= request_params_header
475
500
 
476
501
  options.apply_defaults timeout: @config.rpcs.create_transfer_config.timeout,
477
502
  metadata: metadata,
478
503
  retry_policy: @config.rpcs.create_transfer_config.retry_policy
479
- options.apply_defaults metadata: @config.metadata,
504
+
505
+ options.apply_defaults timeout: @config.timeout,
506
+ metadata: @config.metadata,
480
507
  retry_policy: @config.retry_policy
481
508
 
482
509
  @data_transfer_service_stub.call_rpc :create_transfer_config, request, options: options do |response, operation|
@@ -549,6 +576,21 @@ module Google
549
576
  #
550
577
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
551
578
  #
579
+ # @example Basic example
580
+ # require "google/cloud/bigquery/data_transfer/v1"
581
+ #
582
+ # # Create a client object. The client can be reused for multiple calls.
583
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
584
+ #
585
+ # # Create a request. To set request fields, pass in keyword arguments.
586
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::UpdateTransferConfigRequest.new
587
+ #
588
+ # # Call the update_transfer_config method.
589
+ # result = client.update_transfer_config request
590
+ #
591
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
592
+ # p result
593
+ #
552
594
  def update_transfer_config request, options = nil
553
595
  raise ::ArgumentError, "request must be provided" if request.nil?
554
596
 
@@ -566,16 +608,20 @@ module Google
566
608
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
567
609
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
568
610
 
569
- header_params = {
570
- "transfer_config.name" => request.transfer_config.name
571
- }
611
+ header_params = {}
612
+ if request.transfer_config&.name
613
+ header_params["transfer_config.name"] = request.transfer_config.name
614
+ end
615
+
572
616
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
573
617
  metadata[:"x-goog-request-params"] ||= request_params_header
574
618
 
575
619
  options.apply_defaults timeout: @config.rpcs.update_transfer_config.timeout,
576
620
  metadata: metadata,
577
621
  retry_policy: @config.rpcs.update_transfer_config.retry_policy
578
- options.apply_defaults metadata: @config.metadata,
622
+
623
+ options.apply_defaults timeout: @config.timeout,
624
+ metadata: @config.metadata,
579
625
  retry_policy: @config.retry_policy
580
626
 
581
627
  @data_transfer_service_stub.call_rpc :update_transfer_config, request, options: options do |response, operation|
@@ -618,6 +664,21 @@ module Google
618
664
  #
619
665
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
620
666
  #
667
+ # @example Basic example
668
+ # require "google/cloud/bigquery/data_transfer/v1"
669
+ #
670
+ # # Create a client object. The client can be reused for multiple calls.
671
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
672
+ #
673
+ # # Create a request. To set request fields, pass in keyword arguments.
674
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferConfigRequest.new
675
+ #
676
+ # # Call the delete_transfer_config method.
677
+ # result = client.delete_transfer_config request
678
+ #
679
+ # # The returned object is of type Google::Protobuf::Empty.
680
+ # p result
681
+ #
621
682
  def delete_transfer_config request, options = nil
622
683
  raise ::ArgumentError, "request must be provided" if request.nil?
623
684
 
@@ -635,16 +696,20 @@ module Google
635
696
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
636
697
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
637
698
 
638
- header_params = {
639
- "name" => request.name
640
- }
699
+ header_params = {}
700
+ if request.name
701
+ header_params["name"] = request.name
702
+ end
703
+
641
704
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
642
705
  metadata[:"x-goog-request-params"] ||= request_params_header
643
706
 
644
707
  options.apply_defaults timeout: @config.rpcs.delete_transfer_config.timeout,
645
708
  metadata: metadata,
646
709
  retry_policy: @config.rpcs.delete_transfer_config.retry_policy
647
- options.apply_defaults metadata: @config.metadata,
710
+
711
+ options.apply_defaults timeout: @config.timeout,
712
+ metadata: @config.metadata,
648
713
  retry_policy: @config.retry_policy
649
714
 
650
715
  @data_transfer_service_stub.call_rpc :delete_transfer_config, request, options: options do |response, operation|
@@ -686,6 +751,21 @@ module Google
686
751
  #
687
752
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
688
753
  #
754
+ # @example Basic example
755
+ # require "google/cloud/bigquery/data_transfer/v1"
756
+ #
757
+ # # Create a client object. The client can be reused for multiple calls.
758
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
759
+ #
760
+ # # Create a request. To set request fields, pass in keyword arguments.
761
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetTransferConfigRequest.new
762
+ #
763
+ # # Call the get_transfer_config method.
764
+ # result = client.get_transfer_config request
765
+ #
766
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
767
+ # p result
768
+ #
689
769
  def get_transfer_config request, options = nil
690
770
  raise ::ArgumentError, "request must be provided" if request.nil?
691
771
 
@@ -703,16 +783,20 @@ module Google
703
783
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
704
784
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
705
785
 
706
- header_params = {
707
- "name" => request.name
708
- }
786
+ header_params = {}
787
+ if request.name
788
+ header_params["name"] = request.name
789
+ end
790
+
709
791
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
710
792
  metadata[:"x-goog-request-params"] ||= request_params_header
711
793
 
712
794
  options.apply_defaults timeout: @config.rpcs.get_transfer_config.timeout,
713
795
  metadata: metadata,
714
796
  retry_policy: @config.rpcs.get_transfer_config.retry_policy
715
- options.apply_defaults metadata: @config.metadata,
797
+
798
+ options.apply_defaults timeout: @config.timeout,
799
+ metadata: @config.metadata,
716
800
  retry_policy: @config.retry_policy
717
801
 
718
802
  @data_transfer_service_stub.call_rpc :get_transfer_config, request, options: options do |response, operation|
@@ -724,7 +808,8 @@ module Google
724
808
  end
725
809
 
726
810
  ##
727
- # Returns information about all data transfers in the project.
811
+ # Returns information about all transfer configs owned by a project in the
812
+ # specified location.
728
813
  #
729
814
  # @overload list_transfer_configs(request, options = nil)
730
815
  # Pass arguments to `list_transfer_configs` via a request object, either of type
@@ -764,6 +849,27 @@ module Google
764
849
  #
765
850
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
766
851
  #
852
+ # @example Basic example
853
+ # require "google/cloud/bigquery/data_transfer/v1"
854
+ #
855
+ # # Create a client object. The client can be reused for multiple calls.
856
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
857
+ #
858
+ # # Create a request. To set request fields, pass in keyword arguments.
859
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsRequest.new
860
+ #
861
+ # # Call the list_transfer_configs method.
862
+ # result = client.list_transfer_configs request
863
+ #
864
+ # # The returned object is of type Gapic::PagedEnumerable. You can
865
+ # # iterate over all elements by calling #each, and the enumerable
866
+ # # will lazily make API calls to fetch subsequent pages. Other
867
+ # # methods are also available for managing paging directly.
868
+ # result.each do |response|
869
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
870
+ # p response
871
+ # end
872
+ #
767
873
  def list_transfer_configs request, options = nil
768
874
  raise ::ArgumentError, "request must be provided" if request.nil?
769
875
 
@@ -781,16 +887,20 @@ module Google
781
887
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
782
888
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
783
889
 
784
- header_params = {
785
- "parent" => request.parent
786
- }
890
+ header_params = {}
891
+ if request.parent
892
+ header_params["parent"] = request.parent
893
+ end
894
+
787
895
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
788
896
  metadata[:"x-goog-request-params"] ||= request_params_header
789
897
 
790
898
  options.apply_defaults timeout: @config.rpcs.list_transfer_configs.timeout,
791
899
  metadata: metadata,
792
900
  retry_policy: @config.rpcs.list_transfer_configs.retry_policy
793
- options.apply_defaults metadata: @config.metadata,
901
+
902
+ options.apply_defaults timeout: @config.timeout,
903
+ metadata: @config.metadata,
794
904
  retry_policy: @config.retry_policy
795
905
 
796
906
  @data_transfer_service_stub.call_rpc :list_transfer_configs, request, options: options do |response, operation|
@@ -809,6 +919,8 @@ module Google
809
919
  # Note that runs are created per UTC time in the time range.
810
920
  # DEPRECATED: use StartManualTransferRuns instead.
811
921
  #
922
+ # @deprecated This method is deprecated and may be removed in the next major version update.
923
+ #
812
924
  # @overload schedule_transfer_runs(request, options = nil)
813
925
  # Pass arguments to `schedule_transfer_runs` via a request object, either of type
814
926
  # {::Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsRequest} or an equivalent Hash.
@@ -843,6 +955,21 @@ module Google
843
955
  #
844
956
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
845
957
  #
958
+ # @example Basic example
959
+ # require "google/cloud/bigquery/data_transfer/v1"
960
+ #
961
+ # # Create a client object. The client can be reused for multiple calls.
962
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
963
+ #
964
+ # # Create a request. To set request fields, pass in keyword arguments.
965
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsRequest.new
966
+ #
967
+ # # Call the schedule_transfer_runs method.
968
+ # result = client.schedule_transfer_runs request
969
+ #
970
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsResponse.
971
+ # p result
972
+ #
846
973
  def schedule_transfer_runs request, options = nil
847
974
  raise ::ArgumentError, "request must be provided" if request.nil?
848
975
 
@@ -860,16 +987,20 @@ module Google
860
987
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
861
988
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
862
989
 
863
- header_params = {
864
- "parent" => request.parent
865
- }
990
+ header_params = {}
991
+ if request.parent
992
+ header_params["parent"] = request.parent
993
+ end
994
+
866
995
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
867
996
  metadata[:"x-goog-request-params"] ||= request_params_header
868
997
 
869
998
  options.apply_defaults timeout: @config.rpcs.schedule_transfer_runs.timeout,
870
999
  metadata: metadata,
871
1000
  retry_policy: @config.rpcs.schedule_transfer_runs.retry_policy
872
- options.apply_defaults metadata: @config.metadata,
1001
+
1002
+ options.apply_defaults timeout: @config.timeout,
1003
+ metadata: @config.metadata,
873
1004
  retry_policy: @config.retry_policy
874
1005
 
875
1006
  @data_transfer_service_stub.call_rpc :schedule_transfer_runs, request, options: options do |response, operation|
@@ -919,6 +1050,21 @@ module Google
919
1050
  #
920
1051
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
921
1052
  #
1053
+ # @example Basic example
1054
+ # require "google/cloud/bigquery/data_transfer/v1"
1055
+ #
1056
+ # # Create a client object. The client can be reused for multiple calls.
1057
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1058
+ #
1059
+ # # Create a request. To set request fields, pass in keyword arguments.
1060
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::StartManualTransferRunsRequest.new
1061
+ #
1062
+ # # Call the start_manual_transfer_runs method.
1063
+ # result = client.start_manual_transfer_runs request
1064
+ #
1065
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::StartManualTransferRunsResponse.
1066
+ # p result
1067
+ #
922
1068
  def start_manual_transfer_runs request, options = nil
923
1069
  raise ::ArgumentError, "request must be provided" if request.nil?
924
1070
 
@@ -936,16 +1082,20 @@ module Google
936
1082
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
937
1083
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
938
1084
 
939
- header_params = {
940
- "parent" => request.parent
941
- }
1085
+ header_params = {}
1086
+ if request.parent
1087
+ header_params["parent"] = request.parent
1088
+ end
1089
+
942
1090
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
943
1091
  metadata[:"x-goog-request-params"] ||= request_params_header
944
1092
 
945
1093
  options.apply_defaults timeout: @config.rpcs.start_manual_transfer_runs.timeout,
946
1094
  metadata: metadata,
947
1095
  retry_policy: @config.rpcs.start_manual_transfer_runs.retry_policy
948
- options.apply_defaults metadata: @config.metadata,
1096
+
1097
+ options.apply_defaults timeout: @config.timeout,
1098
+ metadata: @config.metadata,
949
1099
  retry_policy: @config.retry_policy
950
1100
 
951
1101
  @data_transfer_service_stub.call_rpc :start_manual_transfer_runs, request, options: options do |response, operation|
@@ -987,6 +1137,21 @@ module Google
987
1137
  #
988
1138
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
989
1139
  #
1140
+ # @example Basic example
1141
+ # require "google/cloud/bigquery/data_transfer/v1"
1142
+ #
1143
+ # # Create a client object. The client can be reused for multiple calls.
1144
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1145
+ #
1146
+ # # Create a request. To set request fields, pass in keyword arguments.
1147
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetTransferRunRequest.new
1148
+ #
1149
+ # # Call the get_transfer_run method.
1150
+ # result = client.get_transfer_run request
1151
+ #
1152
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferRun.
1153
+ # p result
1154
+ #
990
1155
  def get_transfer_run request, options = nil
991
1156
  raise ::ArgumentError, "request must be provided" if request.nil?
992
1157
 
@@ -1004,16 +1169,20 @@ module Google
1004
1169
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1005
1170
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1006
1171
 
1007
- header_params = {
1008
- "name" => request.name
1009
- }
1172
+ header_params = {}
1173
+ if request.name
1174
+ header_params["name"] = request.name
1175
+ end
1176
+
1010
1177
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1011
1178
  metadata[:"x-goog-request-params"] ||= request_params_header
1012
1179
 
1013
1180
  options.apply_defaults timeout: @config.rpcs.get_transfer_run.timeout,
1014
1181
  metadata: metadata,
1015
1182
  retry_policy: @config.rpcs.get_transfer_run.retry_policy
1016
- options.apply_defaults metadata: @config.metadata,
1183
+
1184
+ options.apply_defaults timeout: @config.timeout,
1185
+ metadata: @config.metadata,
1017
1186
  retry_policy: @config.retry_policy
1018
1187
 
1019
1188
  @data_transfer_service_stub.call_rpc :get_transfer_run, request, options: options do |response, operation|
@@ -1055,6 +1224,21 @@ module Google
1055
1224
  #
1056
1225
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1057
1226
  #
1227
+ # @example Basic example
1228
+ # require "google/cloud/bigquery/data_transfer/v1"
1229
+ #
1230
+ # # Create a client object. The client can be reused for multiple calls.
1231
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1232
+ #
1233
+ # # Create a request. To set request fields, pass in keyword arguments.
1234
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferRunRequest.new
1235
+ #
1236
+ # # Call the delete_transfer_run method.
1237
+ # result = client.delete_transfer_run request
1238
+ #
1239
+ # # The returned object is of type Google::Protobuf::Empty.
1240
+ # p result
1241
+ #
1058
1242
  def delete_transfer_run request, options = nil
1059
1243
  raise ::ArgumentError, "request must be provided" if request.nil?
1060
1244
 
@@ -1072,16 +1256,20 @@ module Google
1072
1256
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1073
1257
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1074
1258
 
1075
- header_params = {
1076
- "name" => request.name
1077
- }
1259
+ header_params = {}
1260
+ if request.name
1261
+ header_params["name"] = request.name
1262
+ end
1263
+
1078
1264
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1079
1265
  metadata[:"x-goog-request-params"] ||= request_params_header
1080
1266
 
1081
1267
  options.apply_defaults timeout: @config.rpcs.delete_transfer_run.timeout,
1082
1268
  metadata: metadata,
1083
1269
  retry_policy: @config.rpcs.delete_transfer_run.retry_policy
1084
- options.apply_defaults metadata: @config.metadata,
1270
+
1271
+ options.apply_defaults timeout: @config.timeout,
1272
+ metadata: @config.metadata,
1085
1273
  retry_policy: @config.retry_policy
1086
1274
 
1087
1275
  @data_transfer_service_stub.call_rpc :delete_transfer_run, request, options: options do |response, operation|
@@ -1136,6 +1324,27 @@ module Google
1136
1324
  #
1137
1325
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1138
1326
  #
1327
+ # @example Basic example
1328
+ # require "google/cloud/bigquery/data_transfer/v1"
1329
+ #
1330
+ # # Create a client object. The client can be reused for multiple calls.
1331
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1332
+ #
1333
+ # # Create a request. To set request fields, pass in keyword arguments.
1334
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferRunsRequest.new
1335
+ #
1336
+ # # Call the list_transfer_runs method.
1337
+ # result = client.list_transfer_runs request
1338
+ #
1339
+ # # The returned object is of type Gapic::PagedEnumerable. You can
1340
+ # # iterate over all elements by calling #each, and the enumerable
1341
+ # # will lazily make API calls to fetch subsequent pages. Other
1342
+ # # methods are also available for managing paging directly.
1343
+ # result.each do |response|
1344
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferRun.
1345
+ # p response
1346
+ # end
1347
+ #
1139
1348
  def list_transfer_runs request, options = nil
1140
1349
  raise ::ArgumentError, "request must be provided" if request.nil?
1141
1350
 
@@ -1153,16 +1362,20 @@ module Google
1153
1362
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1154
1363
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1155
1364
 
1156
- header_params = {
1157
- "parent" => request.parent
1158
- }
1365
+ header_params = {}
1366
+ if request.parent
1367
+ header_params["parent"] = request.parent
1368
+ end
1369
+
1159
1370
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1160
1371
  metadata[:"x-goog-request-params"] ||= request_params_header
1161
1372
 
1162
1373
  options.apply_defaults timeout: @config.rpcs.list_transfer_runs.timeout,
1163
1374
  metadata: metadata,
1164
1375
  retry_policy: @config.rpcs.list_transfer_runs.retry_policy
1165
- options.apply_defaults metadata: @config.metadata,
1376
+
1377
+ options.apply_defaults timeout: @config.timeout,
1378
+ metadata: @config.metadata,
1166
1379
  retry_policy: @config.retry_policy
1167
1380
 
1168
1381
  @data_transfer_service_stub.call_rpc :list_transfer_runs, request, options: options do |response, operation|
@@ -1216,6 +1429,27 @@ module Google
1216
1429
  #
1217
1430
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1218
1431
  #
1432
+ # @example Basic example
1433
+ # require "google/cloud/bigquery/data_transfer/v1"
1434
+ #
1435
+ # # Create a client object. The client can be reused for multiple calls.
1436
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1437
+ #
1438
+ # # Create a request. To set request fields, pass in keyword arguments.
1439
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferLogsRequest.new
1440
+ #
1441
+ # # Call the list_transfer_logs method.
1442
+ # result = client.list_transfer_logs request
1443
+ #
1444
+ # # The returned object is of type Gapic::PagedEnumerable. You can
1445
+ # # iterate over all elements by calling #each, and the enumerable
1446
+ # # will lazily make API calls to fetch subsequent pages. Other
1447
+ # # methods are also available for managing paging directly.
1448
+ # result.each do |response|
1449
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferMessage.
1450
+ # p response
1451
+ # end
1452
+ #
1219
1453
  def list_transfer_logs request, options = nil
1220
1454
  raise ::ArgumentError, "request must be provided" if request.nil?
1221
1455
 
@@ -1233,16 +1467,20 @@ module Google
1233
1467
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1234
1468
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1235
1469
 
1236
- header_params = {
1237
- "parent" => request.parent
1238
- }
1470
+ header_params = {}
1471
+ if request.parent
1472
+ header_params["parent"] = request.parent
1473
+ end
1474
+
1239
1475
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1240
1476
  metadata[:"x-goog-request-params"] ||= request_params_header
1241
1477
 
1242
1478
  options.apply_defaults timeout: @config.rpcs.list_transfer_logs.timeout,
1243
1479
  metadata: metadata,
1244
1480
  retry_policy: @config.rpcs.list_transfer_logs.retry_policy
1245
- options.apply_defaults metadata: @config.metadata,
1481
+
1482
+ options.apply_defaults timeout: @config.timeout,
1483
+ metadata: @config.metadata,
1246
1484
  retry_policy: @config.retry_policy
1247
1485
 
1248
1486
  @data_transfer_service_stub.call_rpc :list_transfer_logs, request, options: options do |response, operation|
@@ -1290,6 +1528,21 @@ module Google
1290
1528
  #
1291
1529
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1292
1530
  #
1531
+ # @example Basic example
1532
+ # require "google/cloud/bigquery/data_transfer/v1"
1533
+ #
1534
+ # # Create a client object. The client can be reused for multiple calls.
1535
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1536
+ #
1537
+ # # Create a request. To set request fields, pass in keyword arguments.
1538
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::CheckValidCredsRequest.new
1539
+ #
1540
+ # # Call the check_valid_creds method.
1541
+ # result = client.check_valid_creds request
1542
+ #
1543
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::CheckValidCredsResponse.
1544
+ # p result
1545
+ #
1293
1546
  def check_valid_creds request, options = nil
1294
1547
  raise ::ArgumentError, "request must be provided" if request.nil?
1295
1548
 
@@ -1307,16 +1560,20 @@ module Google
1307
1560
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1308
1561
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1309
1562
 
1310
- header_params = {
1311
- "name" => request.name
1312
- }
1563
+ header_params = {}
1564
+ if request.name
1565
+ header_params["name"] = request.name
1566
+ end
1567
+
1313
1568
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1314
1569
  metadata[:"x-goog-request-params"] ||= request_params_header
1315
1570
 
1316
1571
  options.apply_defaults timeout: @config.rpcs.check_valid_creds.timeout,
1317
1572
  metadata: metadata,
1318
1573
  retry_policy: @config.rpcs.check_valid_creds.retry_policy
1319
- options.apply_defaults metadata: @config.metadata,
1574
+
1575
+ options.apply_defaults timeout: @config.timeout,
1576
+ metadata: @config.metadata,
1320
1577
  retry_policy: @config.retry_policy
1321
1578
 
1322
1579
  @data_transfer_service_stub.call_rpc :check_valid_creds, request, options: options do |response, operation|
@@ -1340,22 +1597,21 @@ module Google
1340
1597
  # Configuration can be applied globally to all clients, or to a single client
1341
1598
  # on construction.
1342
1599
  #
1343
- # # Examples
1344
- #
1345
- # To modify the global config, setting the timeout for get_data_source
1346
- # to 20 seconds, and all remaining timeouts to 10 seconds:
1347
- #
1348
- # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
1349
- # config.timeout = 10.0
1350
- # config.rpcs.get_data_source.timeout = 20.0
1351
- # end
1352
- #
1353
- # To apply the above configuration only to a new client:
1354
- #
1355
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
1356
- # config.timeout = 10.0
1357
- # config.rpcs.get_data_source.timeout = 20.0
1358
- # end
1600
+ # @example
1601
+ #
1602
+ # # Modify the global config, setting the timeout for
1603
+ # # get_data_source to 20 seconds,
1604
+ # # and all remaining timeouts to 10 seconds.
1605
+ # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
1606
+ # config.timeout = 10.0
1607
+ # config.rpcs.get_data_source.timeout = 20.0
1608
+ # end
1609
+ #
1610
+ # # Apply the above configuration only to a new client.
1611
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
1612
+ # config.timeout = 10.0
1613
+ # config.rpcs.get_data_source.timeout = 20.0
1614
+ # end
1359
1615
  #
1360
1616
  # @!attribute [rw] endpoint
1361
1617
  # The hostname or hostname:port of the service endpoint.
@@ -22,7 +22,7 @@ module Google
22
22
  module Bigquery
23
23
  module DataTransfer
24
24
  module V1
25
- VERSION = "0.4.1"
25
+ VERSION = "0.4.5"
26
26
  end
27
27
  end
28
28
  end
@@ -1,8 +1,6 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/datatransfer/v1/datatransfer.proto
3
3
 
4
- require 'google/protobuf'
5
-
6
4
  require 'google/api/annotations_pb'
7
5
  require 'google/api/client_pb'
8
6
  require 'google/api/field_behavior_pb'
@@ -13,6 +11,8 @@ require 'google/protobuf/empty_pb'
13
11
  require 'google/protobuf/field_mask_pb'
14
12
  require 'google/protobuf/timestamp_pb'
15
13
  require 'google/protobuf/wrappers_pb'
14
+ require 'google/protobuf'
15
+
16
16
  Google::Protobuf::DescriptorPool.generated_pool.build do
17
17
  add_file("google/cloud/bigquery/datatransfer/v1/datatransfer.proto", :syntax => :proto3) do
18
18
  add_message "google.cloud.bigquery.datatransfer.v1.DataSourceParameter" do
@@ -53,7 +53,8 @@ module Google
53
53
  rpc :DeleteTransferConfig, ::Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferConfigRequest, ::Google::Protobuf::Empty
54
54
  # Returns information about a data transfer config.
55
55
  rpc :GetTransferConfig, ::Google::Cloud::Bigquery::DataTransfer::V1::GetTransferConfigRequest, ::Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig
56
- # Returns information about all data transfers in the project.
56
+ # Returns information about all transfer configs owned by a project in the
57
+ # specified location.
57
58
  rpc :ListTransferConfigs, ::Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsRequest, ::Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsResponse
58
59
  # Creates transfer runs for a time range [start_time, end_time].
59
60
  # For each date - or whatever granularity the data source supports - in the
@@ -1,13 +1,14 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/datatransfer/v1/transfer.proto
3
3
 
4
- require 'google/protobuf'
5
-
6
4
  require 'google/api/field_behavior_pb'
7
5
  require 'google/api/resource_pb'
6
+ require 'google/protobuf/duration_pb'
8
7
  require 'google/protobuf/struct_pb'
9
8
  require 'google/protobuf/timestamp_pb'
10
9
  require 'google/rpc/status_pb'
10
+ require 'google/protobuf'
11
+
11
12
  Google::Protobuf::DescriptorPool.generated_pool.build do
12
13
  add_file("google/cloud/bigquery/datatransfer/v1/transfer.proto", :syntax => :proto3) do
13
14
  add_message "google.cloud.bigquery.datatransfer.v1.EmailPreferences" do
@@ -57,9 +57,15 @@ module Google
57
57
 
58
58
  # Denotes that a (repeated) field is an unordered list.
59
59
  # This indicates that the service may provide the elements of the list
60
- # in any arbitrary order, rather than the order the user originally
60
+ # in any arbitrary order, rather than the order the user originally
61
61
  # provided. Additionally, the list's order may or may not be stable.
62
62
  UNORDERED_LIST = 6
63
+
64
+ # Denotes that this field returns a non-empty default value if not set.
65
+ # This indicates that if the user provides the empty value in a request,
66
+ # a non-empty value will be returned. The user will not be aware of what
67
+ # non-empty value to expect.
68
+ NON_EMPTY_DEFAULT = 7
63
69
  end
64
70
  end
65
71
  end
@@ -195,9 +195,7 @@ module Google
195
195
  # exchanged for a refresh token on the backend.
196
196
  GOOGLE_PLUS_AUTHORIZATION_CODE = 2
197
197
 
198
- # Use First Party Client OAuth. First Party Client OAuth doesn't require a
199
- # refresh token to get an offline access token. Instead, it uses a
200
- # client-signed JWT assertion to retrieve an access token.
198
+ # Use First Party OAuth.
201
199
  FIRST_PARTY_OAUTH = 3
202
200
  end
203
201
 
@@ -617,14 +615,14 @@ module Google
617
615
  # @return [::Google::Protobuf::Timestamp]
618
616
  # Start time of the range of transfer runs. For example,
619
617
  # `"2017-05-25T00:00:00+00:00"`. The start_time must be strictly less than
620
- # the end_time. Creates transfer runs where run_time is in the range betwen
621
- # start_time (inclusive) and end_time (exlusive).
618
+ # the end_time. Creates transfer runs where run_time is in the range
619
+ # between start_time (inclusive) and end_time (exclusive).
622
620
  # @!attribute [rw] end_time
623
621
  # @return [::Google::Protobuf::Timestamp]
624
622
  # End time of the range of transfer runs. For example,
625
623
  # `"2017-05-30T00:00:00+00:00"`. The end_time must not be in the future.
626
- # Creates transfer runs where run_time is in the range betwen start_time
627
- # (inclusive) and end_time (exlusive).
624
+ # Creates transfer runs where run_time is in the range between start_time
625
+ # (inclusive) and end_time (exclusive).
628
626
  class TimeRange
629
627
  include ::Google::Protobuf::MessageExts
630
628
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -66,12 +66,11 @@ module Google
66
66
  # @!attribute [rw] name
67
67
  # @return [::String]
68
68
  # The resource name of the transfer config.
69
- # Transfer config names have the form of
69
+ # Transfer config names have the form
70
70
  # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
71
- # The name is automatically generated based on the config_id specified in
72
- # CreateTransferConfigRequest along with project_id and region. If config_id
73
- # is not provided, usually a uuid, even though it is not guaranteed or
74
- # required, will be generated for config_id.
71
+ # Where `config_id` is usually a uuid, even though it is not
72
+ # guaranteed or required. The name is ignored when creating a transfer
73
+ # config.
75
74
  # @!attribute [rw] destination_dataset_id
76
75
  # @return [::String]
77
76
  # The BigQuery target dataset id.
@@ -83,7 +82,10 @@ module Google
83
82
  # Data source id. Cannot be changed once data transfer is created.
84
83
  # @!attribute [rw] params
85
84
  # @return [::Google::Protobuf::Struct]
86
- # Data transfer specific parameters.
85
+ # Parameters specific to each data source. For more information see the
86
+ # bq tab in the 'Setting up a data transfer' section for each data source.
87
+ # For example the parameters for Cloud Storage transfers are listed here:
88
+ # https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
87
89
  # @!attribute [rw] schedule
88
90
  # @return [::String]
89
91
  # Data transfer schedule.
@@ -132,6 +134,9 @@ module Google
132
134
  # @return [::String]
133
135
  # Pub/Sub topic where notifications will be sent after transfer runs
134
136
  # associated with this transfer config finish.
137
+ #
138
+ # The format for specifying a pubsub topic is:
139
+ # `projects/{project}/topics/{topic}`
135
140
  # @!attribute [rw] email_preferences
136
141
  # @return [::Google::Cloud::Bigquery::DataTransfer::V1::EmailPreferences]
137
142
  # Email notifications will be sent according to these preferences
@@ -171,7 +176,10 @@ module Google
171
176
  # Output only. Last time the data transfer run state was updated.
172
177
  # @!attribute [r] params
173
178
  # @return [::Google::Protobuf::Struct]
174
- # Output only. Data transfer specific parameters.
179
+ # Output only. Parameters specific to each data source. For more information see the
180
+ # bq tab in the 'Setting up a data transfer' section for each data source.
181
+ # For example the parameters for Cloud Storage transfers are listed here:
182
+ # https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
175
183
  # @!attribute [r] destination_dataset_id
176
184
  # @return [::String]
177
185
  # Output only. The BigQuery target dataset id.
@@ -194,7 +202,10 @@ module Google
194
202
  # @!attribute [r] notification_pubsub_topic
195
203
  # @return [::String]
196
204
  # Output only. Pub/Sub topic where a notification will be sent after this
197
- # transfer run finishes
205
+ # transfer run finishes.
206
+ #
207
+ # The format for specifying a pubsub topic is:
208
+ # `projects/{project}/topics/{topic}`
198
209
  # @!attribute [r] email_preferences
199
210
  # @return [::Google::Cloud::Bigquery::DataTransfer::V1::EmailPreferences]
200
211
  # Output only. Email notifications will be sent according to these
@@ -250,23 +261,23 @@ module Google
250
261
 
251
262
  # Represents data transfer run state.
252
263
  module TransferState
253
- # State placeholder.
264
+ # State placeholder (0).
254
265
  TRANSFER_STATE_UNSPECIFIED = 0
255
266
 
256
267
  # Data transfer is scheduled and is waiting to be picked up by
257
- # data transfer backend.
268
+ # data transfer backend (2).
258
269
  PENDING = 2
259
270
 
260
- # Data transfer is in progress.
271
+ # Data transfer is in progress (3).
261
272
  RUNNING = 3
262
273
 
263
- # Data transfer completed successfully.
274
+ # Data transfer completed successfully (4).
264
275
  SUCCEEDED = 4
265
276
 
266
- # Data transfer failed.
277
+ # Data transfer failed (5).
267
278
  FAILED = 5
268
279
 
269
- # Data transfer is cancelled.
280
+ # Data transfer is cancelled (6).
270
281
  CANCELLED = 6
271
282
  end
272
283
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-bigquery-data_transfer-v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.4.1
4
+ version: 0.4.5
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2021-06-17 00:00:00.000000000 Z
11
+ date: 2021-11-08 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: gapic-common
@@ -16,7 +16,7 @@ dependencies:
16
16
  requirements:
17
17
  - - ">="
18
18
  - !ruby/object:Gem::Version
19
- version: '0.5'
19
+ version: '0.7'
20
20
  - - "<"
21
21
  - !ruby/object:Gem::Version
22
22
  version: 2.a
@@ -26,7 +26,7 @@ dependencies:
26
26
  requirements:
27
27
  - - ">="
28
28
  - !ruby/object:Gem::Version
29
- version: '0.5'
29
+ version: '0.7'
30
30
  - - "<"
31
31
  - !ruby/object:Gem::Version
32
32
  version: 2.a