google-cloud-bigquery-data_transfer-v1 0.4.2 → 0.4.6

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 24ae4b3454274e215f40d7158f366a38da00f1d48379e8a6a959ce951761d120
4
- data.tar.gz: e509dd369ee6ab15cc5aac337737b690f8d951f4b9814125992dfdf9ea1cd9df
3
+ metadata.gz: 0dc98a0216f9b79b237f5c2df3b6f0de5ede88ca709f8cd4e994327a227d29c8
4
+ data.tar.gz: 634ac33b77c2d07a4d2064c7fff482d488e596bd087ca6855ed1fae442416f83
5
5
  SHA512:
6
- metadata.gz: 83ae129b251781deb0c22a0cbcdd9f01c2958781c56e349fd343ec830b7d13367d8da1eb5a200a9e08c5b9c3fd4c4c9db9d02717a3609c87191ecdd1ffd0bda1
7
- data.tar.gz: 19c757bfffa87c278b0fc51502ce243ddcaebb265e4f67900498c0c43d4aaede65bc9530ab0e36a036493fd2d32020faf8e51ef5f817e32c6ffc80b7f5a201ef
6
+ metadata.gz: fda0deecc390eb39f5abd52c160dbefb67d23740c3a769650b27b356a665f46e92928ad3ccfe38a1278781ee45b75e1f6ac217d54d3b0973f9b1926f7c4e32a9
7
+ data.tar.gz: 937354981b887fd9649ea6b57ce7936686cbb13a095ef5a7a2e2fa49c106603d99e7cfb804d7858a60b6a414d8b822ece743b4449c6862a97b78d255b292871a
data/.yardopts CHANGED
@@ -1,5 +1,5 @@
1
1
  --no-private
2
- --title=BigQuery Data Transfer Service V1 API
2
+ --title="BigQuery Data Transfer Service V1 API"
3
3
  --exclude _pb\.rb$
4
4
  --markup markdown
5
5
  --markup-provider redcarpet
data/AUTHENTICATION.md CHANGED
@@ -120,15 +120,6 @@ To configure your system for this, simply:
120
120
  **NOTE:** This is _not_ recommended for running in production. The Cloud SDK
121
121
  *should* only be used during development.
122
122
 
123
- [gce-how-to]: https://cloud.google.com/compute/docs/authentication#using
124
- [dev-console]: https://console.cloud.google.com/project
125
-
126
- [enable-apis]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/enable-apis.png
127
-
128
- [create-new-service-account]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/create-new-service-account.png
129
- [create-new-service-account-existing-keys]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/create-new-service-account-existing-keys.png
130
- [reuse-service-account]: https://raw.githubusercontent.com/GoogleCloudPlatform/gcloud-common/master/authentication/reuse-service-account.png
131
-
132
123
  ## Creating a Service Account
133
124
 
134
125
  Google Cloud requires **Service Account Credentials** to
@@ -139,31 +130,22 @@ If you are not running this client within
139
130
  [Google Cloud Platform environments](#google-cloud-platform-environments), you
140
131
  need a Google Developers service account.
141
132
 
142
- 1. Visit the [Google Developers Console][dev-console].
133
+ 1. Visit the [Google Cloud Console](https://console.cloud.google.com/project).
143
134
  2. Create a new project or click on an existing project.
144
- 3. Activate the slide-out navigation tray and select **API Manager**. From
135
+ 3. Activate the menu in the upper left and select **APIs & Services**. From
145
136
  here, you will enable the APIs that your application requires.
146
137
 
147
- ![Enable the APIs that your application requires][enable-apis]
148
-
149
138
  *Note: You may need to enable billing in order to use these services.*
150
139
 
151
140
  4. Select **Credentials** from the side navigation.
152
141
 
153
- You should see a screen like one of the following.
154
-
155
- ![Create a new service account][create-new-service-account]
156
-
157
- ![Create a new service account With Existing Keys][create-new-service-account-existing-keys]
158
-
159
- Find the "Add credentials" drop down and select "Service account" to be
160
- guided through downloading a new JSON key file.
142
+ Find the "Create credentials" drop down near the top of the page, and select
143
+ "Service account" to be guided through downloading a new JSON key file.
161
144
 
162
145
  If you want to re-use an existing service account, you can easily generate a
163
- new key file. Just select the account you wish to re-use, and click "Generate
164
- new JSON key":
165
-
166
- ![Re-use an existing service account][reuse-service-account]
146
+ new key file. Just select the account you wish to re-use, click the pencil
147
+ tool on the right side to edit the service account, select the **Keys** tab,
148
+ and then select **Add Key**.
167
149
 
168
150
  The key file you download will be used by this library to authenticate API
169
151
  requests and should be stored in a secure location.
@@ -45,13 +45,12 @@ module Google
45
45
  # See {::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client::Configuration}
46
46
  # for a description of the configuration fields.
47
47
  #
48
- # ## Example
48
+ # @example
49
49
  #
50
- # To modify the configuration for all DataTransferService clients:
51
- #
52
- # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
53
- # config.timeout = 10.0
54
- # end
50
+ # # Modify the configuration for all DataTransferService clients
51
+ # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
52
+ # config.timeout = 10.0
53
+ # end
55
54
  #
56
55
  # @yield [config] Configure the Client client.
57
56
  # @yieldparam config [Client::Configuration]
@@ -154,19 +153,15 @@ module Google
154
153
  ##
155
154
  # Create a new DataTransferService client object.
156
155
  #
157
- # ## Examples
158
- #
159
- # To create a new DataTransferService client with the default
160
- # configuration:
156
+ # @example
161
157
  #
162
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
158
+ # # Create a client using the default configuration
159
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
163
160
  #
164
- # To create a new DataTransferService client with a custom
165
- # configuration:
166
- #
167
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
168
- # config.timeout = 10.0
169
- # end
161
+ # # Create a client using a custom configuration
162
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
163
+ # config.timeout = 10.0
164
+ # end
170
165
  #
171
166
  # @yield [config] Configure the DataTransferService client.
172
167
  # @yieldparam config [Client::Configuration]
@@ -186,10 +181,9 @@ module Google
186
181
 
187
182
  # Create credentials
188
183
  credentials = @config.credentials
189
- # Use self-signed JWT if the scope and endpoint are unchanged from default,
184
+ # Use self-signed JWT if the endpoint is unchanged from default,
190
185
  # but only if the default endpoint does not have a region prefix.
191
- enable_self_signed_jwt = @config.scope == Client.configure.scope &&
192
- @config.endpoint == Client.configure.endpoint &&
186
+ enable_self_signed_jwt = @config.endpoint == Client.configure.endpoint &&
193
187
  !@config.endpoint.split(".").first.include?("-")
194
188
  credentials ||= Credentials.default scope: @config.scope,
195
189
  enable_self_signed_jwt: enable_self_signed_jwt
@@ -242,6 +236,21 @@ module Google
242
236
  #
243
237
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
244
238
  #
239
+ # @example Basic example
240
+ # require "google/cloud/bigquery/data_transfer/v1"
241
+ #
242
+ # # Create a client object. The client can be reused for multiple calls.
243
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
244
+ #
245
+ # # Create a request. To set request fields, pass in keyword arguments.
246
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetDataSourceRequest.new
247
+ #
248
+ # # Call the get_data_source method.
249
+ # result = client.get_data_source request
250
+ #
251
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::DataSource.
252
+ # p result
253
+ #
245
254
  def get_data_source request, options = nil
246
255
  raise ::ArgumentError, "request must be provided" if request.nil?
247
256
 
@@ -259,16 +268,20 @@ module Google
259
268
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
260
269
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
261
270
 
262
- header_params = {
263
- "name" => request.name
264
- }
271
+ header_params = {}
272
+ if request.name
273
+ header_params["name"] = request.name
274
+ end
275
+
265
276
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
266
277
  metadata[:"x-goog-request-params"] ||= request_params_header
267
278
 
268
279
  options.apply_defaults timeout: @config.rpcs.get_data_source.timeout,
269
280
  metadata: metadata,
270
281
  retry_policy: @config.rpcs.get_data_source.retry_policy
271
- options.apply_defaults metadata: @config.metadata,
282
+
283
+ options.apply_defaults timeout: @config.timeout,
284
+ metadata: @config.metadata,
272
285
  retry_policy: @config.retry_policy
273
286
 
274
287
  @data_transfer_service_stub.call_rpc :get_data_source, request, options: options do |response, operation|
@@ -319,6 +332,27 @@ module Google
319
332
  #
320
333
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
321
334
  #
335
+ # @example Basic example
336
+ # require "google/cloud/bigquery/data_transfer/v1"
337
+ #
338
+ # # Create a client object. The client can be reused for multiple calls.
339
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
340
+ #
341
+ # # Create a request. To set request fields, pass in keyword arguments.
342
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListDataSourcesRequest.new
343
+ #
344
+ # # Call the list_data_sources method.
345
+ # result = client.list_data_sources request
346
+ #
347
+ # # The returned object is of type Gapic::PagedEnumerable. You can
348
+ # # iterate over all elements by calling #each, and the enumerable
349
+ # # will lazily make API calls to fetch subsequent pages. Other
350
+ # # methods are also available for managing paging directly.
351
+ # result.each do |response|
352
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::DataSource.
353
+ # p response
354
+ # end
355
+ #
322
356
  def list_data_sources request, options = nil
323
357
  raise ::ArgumentError, "request must be provided" if request.nil?
324
358
 
@@ -336,16 +370,20 @@ module Google
336
370
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
337
371
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
338
372
 
339
- header_params = {
340
- "parent" => request.parent
341
- }
373
+ header_params = {}
374
+ if request.parent
375
+ header_params["parent"] = request.parent
376
+ end
377
+
342
378
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
343
379
  metadata[:"x-goog-request-params"] ||= request_params_header
344
380
 
345
381
  options.apply_defaults timeout: @config.rpcs.list_data_sources.timeout,
346
382
  metadata: metadata,
347
383
  retry_policy: @config.rpcs.list_data_sources.retry_policy
348
- options.apply_defaults metadata: @config.metadata,
384
+
385
+ options.apply_defaults timeout: @config.timeout,
386
+ metadata: @config.metadata,
349
387
  retry_policy: @config.retry_policy
350
388
 
351
389
  @data_transfer_service_stub.call_rpc :list_data_sources, request, options: options do |response, operation|
@@ -420,6 +458,21 @@ module Google
420
458
  #
421
459
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
422
460
  #
461
+ # @example Basic example
462
+ # require "google/cloud/bigquery/data_transfer/v1"
463
+ #
464
+ # # Create a client object. The client can be reused for multiple calls.
465
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
466
+ #
467
+ # # Create a request. To set request fields, pass in keyword arguments.
468
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::CreateTransferConfigRequest.new
469
+ #
470
+ # # Call the create_transfer_config method.
471
+ # result = client.create_transfer_config request
472
+ #
473
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
474
+ # p result
475
+ #
423
476
  def create_transfer_config request, options = nil
424
477
  raise ::ArgumentError, "request must be provided" if request.nil?
425
478
 
@@ -437,16 +490,20 @@ module Google
437
490
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
438
491
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
439
492
 
440
- header_params = {
441
- "parent" => request.parent
442
- }
493
+ header_params = {}
494
+ if request.parent
495
+ header_params["parent"] = request.parent
496
+ end
497
+
443
498
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
444
499
  metadata[:"x-goog-request-params"] ||= request_params_header
445
500
 
446
501
  options.apply_defaults timeout: @config.rpcs.create_transfer_config.timeout,
447
502
  metadata: metadata,
448
503
  retry_policy: @config.rpcs.create_transfer_config.retry_policy
449
- options.apply_defaults metadata: @config.metadata,
504
+
505
+ options.apply_defaults timeout: @config.timeout,
506
+ metadata: @config.metadata,
450
507
  retry_policy: @config.retry_policy
451
508
 
452
509
  @data_transfer_service_stub.call_rpc :create_transfer_config, request, options: options do |response, operation|
@@ -519,6 +576,21 @@ module Google
519
576
  #
520
577
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
521
578
  #
579
+ # @example Basic example
580
+ # require "google/cloud/bigquery/data_transfer/v1"
581
+ #
582
+ # # Create a client object. The client can be reused for multiple calls.
583
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
584
+ #
585
+ # # Create a request. To set request fields, pass in keyword arguments.
586
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::UpdateTransferConfigRequest.new
587
+ #
588
+ # # Call the update_transfer_config method.
589
+ # result = client.update_transfer_config request
590
+ #
591
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
592
+ # p result
593
+ #
522
594
  def update_transfer_config request, options = nil
523
595
  raise ::ArgumentError, "request must be provided" if request.nil?
524
596
 
@@ -536,16 +608,20 @@ module Google
536
608
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
537
609
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
538
610
 
539
- header_params = {
540
- "transfer_config.name" => request.transfer_config.name
541
- }
611
+ header_params = {}
612
+ if request.transfer_config&.name
613
+ header_params["transfer_config.name"] = request.transfer_config.name
614
+ end
615
+
542
616
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
543
617
  metadata[:"x-goog-request-params"] ||= request_params_header
544
618
 
545
619
  options.apply_defaults timeout: @config.rpcs.update_transfer_config.timeout,
546
620
  metadata: metadata,
547
621
  retry_policy: @config.rpcs.update_transfer_config.retry_policy
548
- options.apply_defaults metadata: @config.metadata,
622
+
623
+ options.apply_defaults timeout: @config.timeout,
624
+ metadata: @config.metadata,
549
625
  retry_policy: @config.retry_policy
550
626
 
551
627
  @data_transfer_service_stub.call_rpc :update_transfer_config, request, options: options do |response, operation|
@@ -588,6 +664,21 @@ module Google
588
664
  #
589
665
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
590
666
  #
667
+ # @example Basic example
668
+ # require "google/cloud/bigquery/data_transfer/v1"
669
+ #
670
+ # # Create a client object. The client can be reused for multiple calls.
671
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
672
+ #
673
+ # # Create a request. To set request fields, pass in keyword arguments.
674
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferConfigRequest.new
675
+ #
676
+ # # Call the delete_transfer_config method.
677
+ # result = client.delete_transfer_config request
678
+ #
679
+ # # The returned object is of type Google::Protobuf::Empty.
680
+ # p result
681
+ #
591
682
  def delete_transfer_config request, options = nil
592
683
  raise ::ArgumentError, "request must be provided" if request.nil?
593
684
 
@@ -605,16 +696,20 @@ module Google
605
696
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
606
697
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
607
698
 
608
- header_params = {
609
- "name" => request.name
610
- }
699
+ header_params = {}
700
+ if request.name
701
+ header_params["name"] = request.name
702
+ end
703
+
611
704
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
612
705
  metadata[:"x-goog-request-params"] ||= request_params_header
613
706
 
614
707
  options.apply_defaults timeout: @config.rpcs.delete_transfer_config.timeout,
615
708
  metadata: metadata,
616
709
  retry_policy: @config.rpcs.delete_transfer_config.retry_policy
617
- options.apply_defaults metadata: @config.metadata,
710
+
711
+ options.apply_defaults timeout: @config.timeout,
712
+ metadata: @config.metadata,
618
713
  retry_policy: @config.retry_policy
619
714
 
620
715
  @data_transfer_service_stub.call_rpc :delete_transfer_config, request, options: options do |response, operation|
@@ -656,6 +751,21 @@ module Google
656
751
  #
657
752
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
658
753
  #
754
+ # @example Basic example
755
+ # require "google/cloud/bigquery/data_transfer/v1"
756
+ #
757
+ # # Create a client object. The client can be reused for multiple calls.
758
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
759
+ #
760
+ # # Create a request. To set request fields, pass in keyword arguments.
761
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetTransferConfigRequest.new
762
+ #
763
+ # # Call the get_transfer_config method.
764
+ # result = client.get_transfer_config request
765
+ #
766
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
767
+ # p result
768
+ #
659
769
  def get_transfer_config request, options = nil
660
770
  raise ::ArgumentError, "request must be provided" if request.nil?
661
771
 
@@ -673,16 +783,20 @@ module Google
673
783
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
674
784
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
675
785
 
676
- header_params = {
677
- "name" => request.name
678
- }
786
+ header_params = {}
787
+ if request.name
788
+ header_params["name"] = request.name
789
+ end
790
+
679
791
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
680
792
  metadata[:"x-goog-request-params"] ||= request_params_header
681
793
 
682
794
  options.apply_defaults timeout: @config.rpcs.get_transfer_config.timeout,
683
795
  metadata: metadata,
684
796
  retry_policy: @config.rpcs.get_transfer_config.retry_policy
685
- options.apply_defaults metadata: @config.metadata,
797
+
798
+ options.apply_defaults timeout: @config.timeout,
799
+ metadata: @config.metadata,
686
800
  retry_policy: @config.retry_policy
687
801
 
688
802
  @data_transfer_service_stub.call_rpc :get_transfer_config, request, options: options do |response, operation|
@@ -694,7 +808,8 @@ module Google
694
808
  end
695
809
 
696
810
  ##
697
- # Returns information about all data transfers in the project.
811
+ # Returns information about all transfer configs owned by a project in the
812
+ # specified location.
698
813
  #
699
814
  # @overload list_transfer_configs(request, options = nil)
700
815
  # Pass arguments to `list_transfer_configs` via a request object, either of type
@@ -734,6 +849,27 @@ module Google
734
849
  #
735
850
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
736
851
  #
852
+ # @example Basic example
853
+ # require "google/cloud/bigquery/data_transfer/v1"
854
+ #
855
+ # # Create a client object. The client can be reused for multiple calls.
856
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
857
+ #
858
+ # # Create a request. To set request fields, pass in keyword arguments.
859
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsRequest.new
860
+ #
861
+ # # Call the list_transfer_configs method.
862
+ # result = client.list_transfer_configs request
863
+ #
864
+ # # The returned object is of type Gapic::PagedEnumerable. You can
865
+ # # iterate over all elements by calling #each, and the enumerable
866
+ # # will lazily make API calls to fetch subsequent pages. Other
867
+ # # methods are also available for managing paging directly.
868
+ # result.each do |response|
869
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig.
870
+ # p response
871
+ # end
872
+ #
737
873
  def list_transfer_configs request, options = nil
738
874
  raise ::ArgumentError, "request must be provided" if request.nil?
739
875
 
@@ -751,16 +887,20 @@ module Google
751
887
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
752
888
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
753
889
 
754
- header_params = {
755
- "parent" => request.parent
756
- }
890
+ header_params = {}
891
+ if request.parent
892
+ header_params["parent"] = request.parent
893
+ end
894
+
757
895
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
758
896
  metadata[:"x-goog-request-params"] ||= request_params_header
759
897
 
760
898
  options.apply_defaults timeout: @config.rpcs.list_transfer_configs.timeout,
761
899
  metadata: metadata,
762
900
  retry_policy: @config.rpcs.list_transfer_configs.retry_policy
763
- options.apply_defaults metadata: @config.metadata,
901
+
902
+ options.apply_defaults timeout: @config.timeout,
903
+ metadata: @config.metadata,
764
904
  retry_policy: @config.retry_policy
765
905
 
766
906
  @data_transfer_service_stub.call_rpc :list_transfer_configs, request, options: options do |response, operation|
@@ -779,6 +919,8 @@ module Google
779
919
  # Note that runs are created per UTC time in the time range.
780
920
  # DEPRECATED: use StartManualTransferRuns instead.
781
921
  #
922
+ # @deprecated This method is deprecated and may be removed in the next major version update.
923
+ #
782
924
  # @overload schedule_transfer_runs(request, options = nil)
783
925
  # Pass arguments to `schedule_transfer_runs` via a request object, either of type
784
926
  # {::Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsRequest} or an equivalent Hash.
@@ -813,6 +955,21 @@ module Google
813
955
  #
814
956
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
815
957
  #
958
+ # @example Basic example
959
+ # require "google/cloud/bigquery/data_transfer/v1"
960
+ #
961
+ # # Create a client object. The client can be reused for multiple calls.
962
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
963
+ #
964
+ # # Create a request. To set request fields, pass in keyword arguments.
965
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsRequest.new
966
+ #
967
+ # # Call the schedule_transfer_runs method.
968
+ # result = client.schedule_transfer_runs request
969
+ #
970
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::ScheduleTransferRunsResponse.
971
+ # p result
972
+ #
816
973
  def schedule_transfer_runs request, options = nil
817
974
  raise ::ArgumentError, "request must be provided" if request.nil?
818
975
 
@@ -830,16 +987,20 @@ module Google
830
987
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
831
988
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
832
989
 
833
- header_params = {
834
- "parent" => request.parent
835
- }
990
+ header_params = {}
991
+ if request.parent
992
+ header_params["parent"] = request.parent
993
+ end
994
+
836
995
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
837
996
  metadata[:"x-goog-request-params"] ||= request_params_header
838
997
 
839
998
  options.apply_defaults timeout: @config.rpcs.schedule_transfer_runs.timeout,
840
999
  metadata: metadata,
841
1000
  retry_policy: @config.rpcs.schedule_transfer_runs.retry_policy
842
- options.apply_defaults metadata: @config.metadata,
1001
+
1002
+ options.apply_defaults timeout: @config.timeout,
1003
+ metadata: @config.metadata,
843
1004
  retry_policy: @config.retry_policy
844
1005
 
845
1006
  @data_transfer_service_stub.call_rpc :schedule_transfer_runs, request, options: options do |response, operation|
@@ -889,6 +1050,21 @@ module Google
889
1050
  #
890
1051
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
891
1052
  #
1053
+ # @example Basic example
1054
+ # require "google/cloud/bigquery/data_transfer/v1"
1055
+ #
1056
+ # # Create a client object. The client can be reused for multiple calls.
1057
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1058
+ #
1059
+ # # Create a request. To set request fields, pass in keyword arguments.
1060
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::StartManualTransferRunsRequest.new
1061
+ #
1062
+ # # Call the start_manual_transfer_runs method.
1063
+ # result = client.start_manual_transfer_runs request
1064
+ #
1065
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::StartManualTransferRunsResponse.
1066
+ # p result
1067
+ #
892
1068
  def start_manual_transfer_runs request, options = nil
893
1069
  raise ::ArgumentError, "request must be provided" if request.nil?
894
1070
 
@@ -906,16 +1082,20 @@ module Google
906
1082
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
907
1083
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
908
1084
 
909
- header_params = {
910
- "parent" => request.parent
911
- }
1085
+ header_params = {}
1086
+ if request.parent
1087
+ header_params["parent"] = request.parent
1088
+ end
1089
+
912
1090
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
913
1091
  metadata[:"x-goog-request-params"] ||= request_params_header
914
1092
 
915
1093
  options.apply_defaults timeout: @config.rpcs.start_manual_transfer_runs.timeout,
916
1094
  metadata: metadata,
917
1095
  retry_policy: @config.rpcs.start_manual_transfer_runs.retry_policy
918
- options.apply_defaults metadata: @config.metadata,
1096
+
1097
+ options.apply_defaults timeout: @config.timeout,
1098
+ metadata: @config.metadata,
919
1099
  retry_policy: @config.retry_policy
920
1100
 
921
1101
  @data_transfer_service_stub.call_rpc :start_manual_transfer_runs, request, options: options do |response, operation|
@@ -957,6 +1137,21 @@ module Google
957
1137
  #
958
1138
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
959
1139
  #
1140
+ # @example Basic example
1141
+ # require "google/cloud/bigquery/data_transfer/v1"
1142
+ #
1143
+ # # Create a client object. The client can be reused for multiple calls.
1144
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1145
+ #
1146
+ # # Create a request. To set request fields, pass in keyword arguments.
1147
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::GetTransferRunRequest.new
1148
+ #
1149
+ # # Call the get_transfer_run method.
1150
+ # result = client.get_transfer_run request
1151
+ #
1152
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::TransferRun.
1153
+ # p result
1154
+ #
960
1155
  def get_transfer_run request, options = nil
961
1156
  raise ::ArgumentError, "request must be provided" if request.nil?
962
1157
 
@@ -974,16 +1169,20 @@ module Google
974
1169
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
975
1170
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
976
1171
 
977
- header_params = {
978
- "name" => request.name
979
- }
1172
+ header_params = {}
1173
+ if request.name
1174
+ header_params["name"] = request.name
1175
+ end
1176
+
980
1177
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
981
1178
  metadata[:"x-goog-request-params"] ||= request_params_header
982
1179
 
983
1180
  options.apply_defaults timeout: @config.rpcs.get_transfer_run.timeout,
984
1181
  metadata: metadata,
985
1182
  retry_policy: @config.rpcs.get_transfer_run.retry_policy
986
- options.apply_defaults metadata: @config.metadata,
1183
+
1184
+ options.apply_defaults timeout: @config.timeout,
1185
+ metadata: @config.metadata,
987
1186
  retry_policy: @config.retry_policy
988
1187
 
989
1188
  @data_transfer_service_stub.call_rpc :get_transfer_run, request, options: options do |response, operation|
@@ -1025,6 +1224,21 @@ module Google
1025
1224
  #
1026
1225
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1027
1226
  #
1227
+ # @example Basic example
1228
+ # require "google/cloud/bigquery/data_transfer/v1"
1229
+ #
1230
+ # # Create a client object. The client can be reused for multiple calls.
1231
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1232
+ #
1233
+ # # Create a request. To set request fields, pass in keyword arguments.
1234
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferRunRequest.new
1235
+ #
1236
+ # # Call the delete_transfer_run method.
1237
+ # result = client.delete_transfer_run request
1238
+ #
1239
+ # # The returned object is of type Google::Protobuf::Empty.
1240
+ # p result
1241
+ #
1028
1242
  def delete_transfer_run request, options = nil
1029
1243
  raise ::ArgumentError, "request must be provided" if request.nil?
1030
1244
 
@@ -1042,16 +1256,20 @@ module Google
1042
1256
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1043
1257
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1044
1258
 
1045
- header_params = {
1046
- "name" => request.name
1047
- }
1259
+ header_params = {}
1260
+ if request.name
1261
+ header_params["name"] = request.name
1262
+ end
1263
+
1048
1264
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1049
1265
  metadata[:"x-goog-request-params"] ||= request_params_header
1050
1266
 
1051
1267
  options.apply_defaults timeout: @config.rpcs.delete_transfer_run.timeout,
1052
1268
  metadata: metadata,
1053
1269
  retry_policy: @config.rpcs.delete_transfer_run.retry_policy
1054
- options.apply_defaults metadata: @config.metadata,
1270
+
1271
+ options.apply_defaults timeout: @config.timeout,
1272
+ metadata: @config.metadata,
1055
1273
  retry_policy: @config.retry_policy
1056
1274
 
1057
1275
  @data_transfer_service_stub.call_rpc :delete_transfer_run, request, options: options do |response, operation|
@@ -1106,6 +1324,27 @@ module Google
1106
1324
  #
1107
1325
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1108
1326
  #
1327
+ # @example Basic example
1328
+ # require "google/cloud/bigquery/data_transfer/v1"
1329
+ #
1330
+ # # Create a client object. The client can be reused for multiple calls.
1331
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1332
+ #
1333
+ # # Create a request. To set request fields, pass in keyword arguments.
1334
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferRunsRequest.new
1335
+ #
1336
+ # # Call the list_transfer_runs method.
1337
+ # result = client.list_transfer_runs request
1338
+ #
1339
+ # # The returned object is of type Gapic::PagedEnumerable. You can
1340
+ # # iterate over all elements by calling #each, and the enumerable
1341
+ # # will lazily make API calls to fetch subsequent pages. Other
1342
+ # # methods are also available for managing paging directly.
1343
+ # result.each do |response|
1344
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferRun.
1345
+ # p response
1346
+ # end
1347
+ #
1109
1348
  def list_transfer_runs request, options = nil
1110
1349
  raise ::ArgumentError, "request must be provided" if request.nil?
1111
1350
 
@@ -1123,16 +1362,20 @@ module Google
1123
1362
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1124
1363
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1125
1364
 
1126
- header_params = {
1127
- "parent" => request.parent
1128
- }
1365
+ header_params = {}
1366
+ if request.parent
1367
+ header_params["parent"] = request.parent
1368
+ end
1369
+
1129
1370
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1130
1371
  metadata[:"x-goog-request-params"] ||= request_params_header
1131
1372
 
1132
1373
  options.apply_defaults timeout: @config.rpcs.list_transfer_runs.timeout,
1133
1374
  metadata: metadata,
1134
1375
  retry_policy: @config.rpcs.list_transfer_runs.retry_policy
1135
- options.apply_defaults metadata: @config.metadata,
1376
+
1377
+ options.apply_defaults timeout: @config.timeout,
1378
+ metadata: @config.metadata,
1136
1379
  retry_policy: @config.retry_policy
1137
1380
 
1138
1381
  @data_transfer_service_stub.call_rpc :list_transfer_runs, request, options: options do |response, operation|
@@ -1186,6 +1429,27 @@ module Google
1186
1429
  #
1187
1430
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1188
1431
  #
1432
+ # @example Basic example
1433
+ # require "google/cloud/bigquery/data_transfer/v1"
1434
+ #
1435
+ # # Create a client object. The client can be reused for multiple calls.
1436
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1437
+ #
1438
+ # # Create a request. To set request fields, pass in keyword arguments.
1439
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::ListTransferLogsRequest.new
1440
+ #
1441
+ # # Call the list_transfer_logs method.
1442
+ # result = client.list_transfer_logs request
1443
+ #
1444
+ # # The returned object is of type Gapic::PagedEnumerable. You can
1445
+ # # iterate over all elements by calling #each, and the enumerable
1446
+ # # will lazily make API calls to fetch subsequent pages. Other
1447
+ # # methods are also available for managing paging directly.
1448
+ # result.each do |response|
1449
+ # # Each element is of type ::Google::Cloud::Bigquery::DataTransfer::V1::TransferMessage.
1450
+ # p response
1451
+ # end
1452
+ #
1189
1453
  def list_transfer_logs request, options = nil
1190
1454
  raise ::ArgumentError, "request must be provided" if request.nil?
1191
1455
 
@@ -1203,16 +1467,20 @@ module Google
1203
1467
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1204
1468
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1205
1469
 
1206
- header_params = {
1207
- "parent" => request.parent
1208
- }
1470
+ header_params = {}
1471
+ if request.parent
1472
+ header_params["parent"] = request.parent
1473
+ end
1474
+
1209
1475
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1210
1476
  metadata[:"x-goog-request-params"] ||= request_params_header
1211
1477
 
1212
1478
  options.apply_defaults timeout: @config.rpcs.list_transfer_logs.timeout,
1213
1479
  metadata: metadata,
1214
1480
  retry_policy: @config.rpcs.list_transfer_logs.retry_policy
1215
- options.apply_defaults metadata: @config.metadata,
1481
+
1482
+ options.apply_defaults timeout: @config.timeout,
1483
+ metadata: @config.metadata,
1216
1484
  retry_policy: @config.retry_policy
1217
1485
 
1218
1486
  @data_transfer_service_stub.call_rpc :list_transfer_logs, request, options: options do |response, operation|
@@ -1260,6 +1528,21 @@ module Google
1260
1528
  #
1261
1529
  # @raise [::Google::Cloud::Error] if the RPC is aborted.
1262
1530
  #
1531
+ # @example Basic example
1532
+ # require "google/cloud/bigquery/data_transfer/v1"
1533
+ #
1534
+ # # Create a client object. The client can be reused for multiple calls.
1535
+ # client = Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new
1536
+ #
1537
+ # # Create a request. To set request fields, pass in keyword arguments.
1538
+ # request = Google::Cloud::Bigquery::DataTransfer::V1::CheckValidCredsRequest.new
1539
+ #
1540
+ # # Call the check_valid_creds method.
1541
+ # result = client.check_valid_creds request
1542
+ #
1543
+ # # The returned object is of type Google::Cloud::Bigquery::DataTransfer::V1::CheckValidCredsResponse.
1544
+ # p result
1545
+ #
1263
1546
  def check_valid_creds request, options = nil
1264
1547
  raise ::ArgumentError, "request must be provided" if request.nil?
1265
1548
 
@@ -1277,16 +1560,20 @@ module Google
1277
1560
  gapic_version: ::Google::Cloud::Bigquery::DataTransfer::V1::VERSION
1278
1561
  metadata[:"x-goog-user-project"] = @quota_project_id if @quota_project_id
1279
1562
 
1280
- header_params = {
1281
- "name" => request.name
1282
- }
1563
+ header_params = {}
1564
+ if request.name
1565
+ header_params["name"] = request.name
1566
+ end
1567
+
1283
1568
  request_params_header = header_params.map { |k, v| "#{k}=#{v}" }.join("&")
1284
1569
  metadata[:"x-goog-request-params"] ||= request_params_header
1285
1570
 
1286
1571
  options.apply_defaults timeout: @config.rpcs.check_valid_creds.timeout,
1287
1572
  metadata: metadata,
1288
1573
  retry_policy: @config.rpcs.check_valid_creds.retry_policy
1289
- options.apply_defaults metadata: @config.metadata,
1574
+
1575
+ options.apply_defaults timeout: @config.timeout,
1576
+ metadata: @config.metadata,
1290
1577
  retry_policy: @config.retry_policy
1291
1578
 
1292
1579
  @data_transfer_service_stub.call_rpc :check_valid_creds, request, options: options do |response, operation|
@@ -1310,22 +1597,21 @@ module Google
1310
1597
  # Configuration can be applied globally to all clients, or to a single client
1311
1598
  # on construction.
1312
1599
  #
1313
- # # Examples
1314
- #
1315
- # To modify the global config, setting the timeout for get_data_source
1316
- # to 20 seconds, and all remaining timeouts to 10 seconds:
1317
- #
1318
- # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
1319
- # config.timeout = 10.0
1320
- # config.rpcs.get_data_source.timeout = 20.0
1321
- # end
1322
- #
1323
- # To apply the above configuration only to a new client:
1324
- #
1325
- # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
1326
- # config.timeout = 10.0
1327
- # config.rpcs.get_data_source.timeout = 20.0
1328
- # end
1600
+ # @example
1601
+ #
1602
+ # # Modify the global config, setting the timeout for
1603
+ # # get_data_source to 20 seconds,
1604
+ # # and all remaining timeouts to 10 seconds.
1605
+ # ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.configure do |config|
1606
+ # config.timeout = 10.0
1607
+ # config.rpcs.get_data_source.timeout = 20.0
1608
+ # end
1609
+ #
1610
+ # # Apply the above configuration only to a new client.
1611
+ # client = ::Google::Cloud::Bigquery::DataTransfer::V1::DataTransferService::Client.new do |config|
1612
+ # config.timeout = 10.0
1613
+ # config.rpcs.get_data_source.timeout = 20.0
1614
+ # end
1329
1615
  #
1330
1616
  # @!attribute [rw] endpoint
1331
1617
  # The hostname or hostname:port of the service endpoint.
@@ -22,7 +22,7 @@ module Google
22
22
  module Bigquery
23
23
  module DataTransfer
24
24
  module V1
25
- VERSION = "0.4.2"
25
+ VERSION = "0.4.6"
26
26
  end
27
27
  end
28
28
  end
@@ -1,8 +1,6 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/datatransfer/v1/datatransfer.proto
3
3
 
4
- require 'google/protobuf'
5
-
6
4
  require 'google/api/annotations_pb'
7
5
  require 'google/api/client_pb'
8
6
  require 'google/api/field_behavior_pb'
@@ -13,6 +11,8 @@ require 'google/protobuf/empty_pb'
13
11
  require 'google/protobuf/field_mask_pb'
14
12
  require 'google/protobuf/timestamp_pb'
15
13
  require 'google/protobuf/wrappers_pb'
14
+ require 'google/protobuf'
15
+
16
16
  Google::Protobuf::DescriptorPool.generated_pool.build do
17
17
  add_file("google/cloud/bigquery/datatransfer/v1/datatransfer.proto", :syntax => :proto3) do
18
18
  add_message "google.cloud.bigquery.datatransfer.v1.DataSourceParameter" do
@@ -31,7 +31,7 @@ module Google
31
31
  # up the frontend.
32
32
  class Service
33
33
 
34
- include GRPC::GenericService
34
+ include ::GRPC::GenericService
35
35
 
36
36
  self.marshal_class_method = :encode
37
37
  self.unmarshal_class_method = :decode
@@ -53,7 +53,8 @@ module Google
53
53
  rpc :DeleteTransferConfig, ::Google::Cloud::Bigquery::DataTransfer::V1::DeleteTransferConfigRequest, ::Google::Protobuf::Empty
54
54
  # Returns information about a data transfer config.
55
55
  rpc :GetTransferConfig, ::Google::Cloud::Bigquery::DataTransfer::V1::GetTransferConfigRequest, ::Google::Cloud::Bigquery::DataTransfer::V1::TransferConfig
56
- # Returns information about all data transfers in the project.
56
+ # Returns information about all transfer configs owned by a project in the
57
+ # specified location.
57
58
  rpc :ListTransferConfigs, ::Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsRequest, ::Google::Cloud::Bigquery::DataTransfer::V1::ListTransferConfigsResponse
58
59
  # Creates transfer runs for a time range [start_time, end_time].
59
60
  # For each date - or whatever granularity the data source supports - in the
@@ -1,13 +1,14 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/bigquery/datatransfer/v1/transfer.proto
3
3
 
4
- require 'google/protobuf'
5
-
6
4
  require 'google/api/field_behavior_pb'
7
5
  require 'google/api/resource_pb'
6
+ require 'google/protobuf/duration_pb'
8
7
  require 'google/protobuf/struct_pb'
9
8
  require 'google/protobuf/timestamp_pb'
10
9
  require 'google/rpc/status_pb'
10
+ require 'google/protobuf'
11
+
11
12
  Google::Protobuf::DescriptorPool.generated_pool.build do
12
13
  add_file("google/cloud/bigquery/datatransfer/v1/transfer.proto", :syntax => :proto3) do
13
14
  add_message "google.cloud.bigquery.datatransfer.v1.EmailPreferences" do
@@ -33,11 +33,7 @@ module Google
33
33
  # // For Kubernetes resources, the format is {api group}/{kind}.
34
34
  # option (google.api.resource) = {
35
35
  # type: "pubsub.googleapis.com/Topic"
36
- # name_descriptor: {
37
- # pattern: "projects/{project}/topics/{topic}"
38
- # parent_type: "cloudresourcemanager.googleapis.com/Project"
39
- # parent_name_extractor: "projects/{project}"
40
- # }
36
+ # pattern: "projects/{project}/topics/{topic}"
41
37
  # };
42
38
  # }
43
39
  #
@@ -45,10 +41,7 @@ module Google
45
41
  #
46
42
  # resources:
47
43
  # - type: "pubsub.googleapis.com/Topic"
48
- # name_descriptor:
49
- # - pattern: "projects/{project}/topics/{topic}"
50
- # parent_type: "cloudresourcemanager.googleapis.com/Project"
51
- # parent_name_extractor: "projects/{project}"
44
+ # pattern: "projects/{project}/topics/{topic}"
52
45
  #
53
46
  # Sometimes, resources have multiple patterns, typically because they can
54
47
  # live under multiple parents.
@@ -58,26 +51,10 @@ module Google
58
51
  # message LogEntry {
59
52
  # option (google.api.resource) = {
60
53
  # type: "logging.googleapis.com/LogEntry"
61
- # name_descriptor: {
62
- # pattern: "projects/{project}/logs/{log}"
63
- # parent_type: "cloudresourcemanager.googleapis.com/Project"
64
- # parent_name_extractor: "projects/{project}"
65
- # }
66
- # name_descriptor: {
67
- # pattern: "folders/{folder}/logs/{log}"
68
- # parent_type: "cloudresourcemanager.googleapis.com/Folder"
69
- # parent_name_extractor: "folders/{folder}"
70
- # }
71
- # name_descriptor: {
72
- # pattern: "organizations/{organization}/logs/{log}"
73
- # parent_type: "cloudresourcemanager.googleapis.com/Organization"
74
- # parent_name_extractor: "organizations/{organization}"
75
- # }
76
- # name_descriptor: {
77
- # pattern: "billingAccounts/{billing_account}/logs/{log}"
78
- # parent_type: "billing.googleapis.com/BillingAccount"
79
- # parent_name_extractor: "billingAccounts/{billing_account}"
80
- # }
54
+ # pattern: "projects/{project}/logs/{log}"
55
+ # pattern: "folders/{folder}/logs/{log}"
56
+ # pattern: "organizations/{organization}/logs/{log}"
57
+ # pattern: "billingAccounts/{billing_account}/logs/{log}"
81
58
  # };
82
59
  # }
83
60
  #
@@ -85,48 +62,10 @@ module Google
85
62
  #
86
63
  # resources:
87
64
  # - type: 'logging.googleapis.com/LogEntry'
88
- # name_descriptor:
89
- # - pattern: "projects/{project}/logs/{log}"
90
- # parent_type: "cloudresourcemanager.googleapis.com/Project"
91
- # parent_name_extractor: "projects/{project}"
92
- # - pattern: "folders/{folder}/logs/{log}"
93
- # parent_type: "cloudresourcemanager.googleapis.com/Folder"
94
- # parent_name_extractor: "folders/{folder}"
95
- # - pattern: "organizations/{organization}/logs/{log}"
96
- # parent_type: "cloudresourcemanager.googleapis.com/Organization"
97
- # parent_name_extractor: "organizations/{organization}"
98
- # - pattern: "billingAccounts/{billing_account}/logs/{log}"
99
- # parent_type: "billing.googleapis.com/BillingAccount"
100
- # parent_name_extractor: "billingAccounts/{billing_account}"
101
- #
102
- # For flexible resources, the resource name doesn't contain parent names, but
103
- # the resource itself has parents for policy evaluation.
104
- #
105
- # Example:
106
- #
107
- # message Shelf {
108
- # option (google.api.resource) = {
109
- # type: "library.googleapis.com/Shelf"
110
- # name_descriptor: {
111
- # pattern: "shelves/{shelf}"
112
- # parent_type: "cloudresourcemanager.googleapis.com/Project"
113
- # }
114
- # name_descriptor: {
115
- # pattern: "shelves/{shelf}"
116
- # parent_type: "cloudresourcemanager.googleapis.com/Folder"
117
- # }
118
- # };
119
- # }
120
- #
121
- # The ResourceDescriptor Yaml config will look like:
122
- #
123
- # resources:
124
- # - type: 'library.googleapis.com/Shelf'
125
- # name_descriptor:
126
- # - pattern: "shelves/{shelf}"
127
- # parent_type: "cloudresourcemanager.googleapis.com/Project"
128
- # - pattern: "shelves/{shelf}"
129
- # parent_type: "cloudresourcemanager.googleapis.com/Folder"
65
+ # pattern: "projects/{project}/logs/{log}"
66
+ # pattern: "folders/{folder}/logs/{log}"
67
+ # pattern: "organizations/{organization}/logs/{log}"
68
+ # pattern: "billingAccounts/{billing_account}/logs/{log}"
130
69
  # @!attribute [rw] type
131
70
  # @return [::String]
132
71
  # The resource type. It must be in the format of
@@ -195,9 +195,7 @@ module Google
195
195
  # exchanged for a refresh token on the backend.
196
196
  GOOGLE_PLUS_AUTHORIZATION_CODE = 2
197
197
 
198
- # Use First Party Client OAuth. First Party Client OAuth doesn't require a
199
- # refresh token to get an offline access token. Instead, it uses a
200
- # client-signed JWT assertion to retrieve an access token.
198
+ # Use First Party OAuth.
201
199
  FIRST_PARTY_OAUTH = 3
202
200
  end
203
201
 
@@ -617,14 +615,14 @@ module Google
617
615
  # @return [::Google::Protobuf::Timestamp]
618
616
  # Start time of the range of transfer runs. For example,
619
617
  # `"2017-05-25T00:00:00+00:00"`. The start_time must be strictly less than
620
- # the end_time. Creates transfer runs where run_time is in the range betwen
621
- # start_time (inclusive) and end_time (exlusive).
618
+ # the end_time. Creates transfer runs where run_time is in the range
619
+ # between start_time (inclusive) and end_time (exclusive).
622
620
  # @!attribute [rw] end_time
623
621
  # @return [::Google::Protobuf::Timestamp]
624
622
  # End time of the range of transfer runs. For example,
625
623
  # `"2017-05-30T00:00:00+00:00"`. The end_time must not be in the future.
626
- # Creates transfer runs where run_time is in the range betwen start_time
627
- # (inclusive) and end_time (exlusive).
624
+ # Creates transfer runs where run_time is in the range between start_time
625
+ # (inclusive) and end_time (exclusive).
628
626
  class TimeRange
629
627
  include ::Google::Protobuf::MessageExts
630
628
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -66,12 +66,11 @@ module Google
66
66
  # @!attribute [rw] name
67
67
  # @return [::String]
68
68
  # The resource name of the transfer config.
69
- # Transfer config names have the form of
69
+ # Transfer config names have the form
70
70
  # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
71
- # The name is automatically generated based on the config_id specified in
72
- # CreateTransferConfigRequest along with project_id and region. If config_id
73
- # is not provided, usually a uuid, even though it is not guaranteed or
74
- # required, will be generated for config_id.
71
+ # Where `config_id` is usually a uuid, even though it is not
72
+ # guaranteed or required. The name is ignored when creating a transfer
73
+ # config.
75
74
  # @!attribute [rw] destination_dataset_id
76
75
  # @return [::String]
77
76
  # The BigQuery target dataset id.
@@ -83,7 +82,10 @@ module Google
83
82
  # Data source id. Cannot be changed once data transfer is created.
84
83
  # @!attribute [rw] params
85
84
  # @return [::Google::Protobuf::Struct]
86
- # Data transfer specific parameters.
85
+ # Parameters specific to each data source. For more information see the
86
+ # bq tab in the 'Setting up a data transfer' section for each data source.
87
+ # For example the parameters for Cloud Storage transfers are listed here:
88
+ # https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
87
89
  # @!attribute [rw] schedule
88
90
  # @return [::String]
89
91
  # Data transfer schedule.
@@ -132,6 +134,9 @@ module Google
132
134
  # @return [::String]
133
135
  # Pub/Sub topic where notifications will be sent after transfer runs
134
136
  # associated with this transfer config finish.
137
+ #
138
+ # The format for specifying a pubsub topic is:
139
+ # `projects/{project}/topics/{topic}`
135
140
  # @!attribute [rw] email_preferences
136
141
  # @return [::Google::Cloud::Bigquery::DataTransfer::V1::EmailPreferences]
137
142
  # Email notifications will be sent according to these preferences
@@ -171,7 +176,10 @@ module Google
171
176
  # Output only. Last time the data transfer run state was updated.
172
177
  # @!attribute [r] params
173
178
  # @return [::Google::Protobuf::Struct]
174
- # Output only. Data transfer specific parameters.
179
+ # Output only. Parameters specific to each data source. For more information see the
180
+ # bq tab in the 'Setting up a data transfer' section for each data source.
181
+ # For example the parameters for Cloud Storage transfers are listed here:
182
+ # https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
175
183
  # @!attribute [r] destination_dataset_id
176
184
  # @return [::String]
177
185
  # Output only. The BigQuery target dataset id.
@@ -194,7 +202,10 @@ module Google
194
202
  # @!attribute [r] notification_pubsub_topic
195
203
  # @return [::String]
196
204
  # Output only. Pub/Sub topic where a notification will be sent after this
197
- # transfer run finishes
205
+ # transfer run finishes.
206
+ #
207
+ # The format for specifying a pubsub topic is:
208
+ # `projects/{project}/topics/{topic}`
198
209
  # @!attribute [r] email_preferences
199
210
  # @return [::Google::Cloud::Bigquery::DataTransfer::V1::EmailPreferences]
200
211
  # Output only. Email notifications will be sent according to these
@@ -250,23 +261,23 @@ module Google
250
261
 
251
262
  # Represents data transfer run state.
252
263
  module TransferState
253
- # State placeholder.
264
+ # State placeholder (0).
254
265
  TRANSFER_STATE_UNSPECIFIED = 0
255
266
 
256
267
  # Data transfer is scheduled and is waiting to be picked up by
257
- # data transfer backend.
268
+ # data transfer backend (2).
258
269
  PENDING = 2
259
270
 
260
- # Data transfer is in progress.
271
+ # Data transfer is in progress (3).
261
272
  RUNNING = 3
262
273
 
263
- # Data transfer completed successfully.
274
+ # Data transfer completed successfully (4).
264
275
  SUCCEEDED = 4
265
276
 
266
- # Data transfer failed.
277
+ # Data transfer failed (5).
267
278
  FAILED = 5
268
279
 
269
- # Data transfer is cancelled.
280
+ # Data transfer is cancelled (6).
270
281
  CANCELLED = 6
271
282
  end
272
283
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-bigquery-data_transfer-v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.4.2
4
+ version: 0.4.6
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2021-07-12 00:00:00.000000000 Z
11
+ date: 2022-01-11 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: gapic-common
@@ -16,7 +16,7 @@ dependencies:
16
16
  requirements:
17
17
  - - ">="
18
18
  - !ruby/object:Gem::Version
19
- version: '0.5'
19
+ version: '0.7'
20
20
  - - "<"
21
21
  - !ruby/object:Gem::Version
22
22
  version: 2.a
@@ -26,7 +26,7 @@ dependencies:
26
26
  requirements:
27
27
  - - ">="
28
28
  - !ruby/object:Gem::Version
29
- version: '0.5'
29
+ version: '0.7'
30
30
  - - "<"
31
31
  - !ruby/object:Gem::Version
32
32
  version: 2.a
@@ -212,7 +212,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
212
212
  - !ruby/object:Gem::Version
213
213
  version: '0'
214
214
  requirements: []
215
- rubygems_version: 3.2.17
215
+ rubygems_version: 3.3.4
216
216
  signing_key:
217
217
  specification_version: 4
218
218
  summary: API Client library for the BigQuery Data Transfer Service V1 API