google-cloud-dataproc 0.9.0 → 0.9.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 73149665558ed4fef80dcabc2baf92983137ac4bde65d74448dbc26865f47107
4
- data.tar.gz: d5b0938c6bdd27df5172550f5e635f58fbed0276ec1d3668f61adce7e01c5767
3
+ metadata.gz: 8f4dc9ab9721db6accf4279d8963bc946bffef23d3480d5b661479df3cf76203
4
+ data.tar.gz: 9919cb5e80780916635366bdabe88f9b4675f5cfaedaf07596633a5c316c8745
5
5
  SHA512:
6
- metadata.gz: bf6e7328361ba5d9e64d47732230f288d451e728328a3e3058e378f53e47f2b796e7f9c75d8212d826f24cc72f117eb5b95119a9ce392e86889e1b056757735f
7
- data.tar.gz: b8404d746ddae8bdbfb87acae3f1109e9813d68a0f4604e85c7f1ee71af38d1bdc26ab4f5345505cdc91858f46db86011de784e503b1faa83c839bd2585f8939
6
+ metadata.gz: 9b59bad93eadf97880799b451199ab06ff8fb4957463cfcc35277176352ad38098916931ee78cabf8e0507505c49f1f3c3636dbeebb8868c1b7ec0937bcbb9f3
7
+ data.tar.gz: 9d60a85d4b95872d332bce6a501be08be0c6d61a4a4f0700392f4f3e2e00ff85b886e2a01b23db2502cfd346fd37269eedb284d54d261466a7354dbbc9437e53
@@ -112,7 +112,7 @@ module Google
112
112
  module AutoscalingPolicyService
113
113
  ##
114
114
  # The API interface for managing autoscaling policies in the
115
- # Google Cloud Dataproc API.
115
+ # Cloud Dataproc API.
116
116
  #
117
117
  # @param version [Symbol, String]
118
118
  # The major version of the service to be used. By default :v1
@@ -282,7 +282,7 @@ module Google
282
282
  module WorkflowTemplateService
283
283
  ##
284
284
  # The API interface for managing Workflow Templates in the
285
- # Cloud Dataproc API.
285
+ # Dataproc API.
286
286
  #
287
287
  # @param version [Symbol, String]
288
288
  # The major version of the service to be used. By default :v1
@@ -110,7 +110,7 @@ module Google
110
110
  module AutoscalingPolicyService
111
111
  ##
112
112
  # The API interface for managing autoscaling policies in the
113
- # Google Cloud Dataproc API.
113
+ # Cloud Dataproc API.
114
114
  #
115
115
  # @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
116
116
  # Provides the means for authenticating requests made by the client. This parameter can
@@ -304,7 +304,7 @@ module Google
304
304
  module WorkflowTemplateService
305
305
  ##
306
306
  # The API interface for managing Workflow Templates in the
307
- # Cloud Dataproc API.
307
+ # Dataproc API.
308
308
  #
309
309
  # @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
310
310
  # Provides the means for authenticating requests made by the client. This parameter can
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/autoscaling_policies.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -27,7 +26,7 @@ module Google
27
26
  module V1beta2
28
27
  module AutoscalingPolicyService
29
28
  # The API interface for managing autoscaling policies in the
30
- # Google Cloud Dataproc API.
29
+ # Cloud Dataproc API.
31
30
  class Service
32
31
 
33
32
  include GRPC::GenericService
@@ -34,7 +34,7 @@ module Google
34
34
  module Dataproc
35
35
  module V1beta2
36
36
  # The API interface for managing autoscaling policies in the
37
- # Google Cloud Dataproc API.
37
+ # Cloud Dataproc API.
38
38
  #
39
39
  # @!attribute [r] autoscaling_policy_service_stub
40
40
  # @return [Google::Cloud::Dataproc::V1beta2::AutoscalingPolicyService::Stub]
@@ -239,7 +239,7 @@ module Google
239
239
  # Required. The ID of the Google Cloud Platform project that the cluster
240
240
  # belongs to.
241
241
  # @param region [String]
242
- # Required. The Cloud Dataproc region in which to handle the request.
242
+ # Required. The Dataproc region in which to handle the request.
243
243
  # @param cluster [Google::Cloud::Dataproc::V1beta2::Cluster | Hash]
244
244
  # Required. The cluster to create.
245
245
  # A hash of the same form as `Google::Cloud::Dataproc::V1beta2::Cluster`
@@ -334,7 +334,7 @@ module Google
334
334
  # Required. The ID of the Google Cloud Platform project the
335
335
  # cluster belongs to.
336
336
  # @param region [String]
337
- # Required. The Cloud Dataproc region in which to handle the request.
337
+ # Required. The Dataproc region in which to handle the request.
338
338
  # @param cluster_name [String]
339
339
  # Required. The cluster name.
340
340
  # @param cluster [Google::Cloud::Dataproc::V1beta2::Cluster | Hash]
@@ -410,7 +410,8 @@ module Google
410
410
  # interrupting jobs in progress. Timeout specifies how long to wait for jobs
411
411
  # in progress to finish before forcefully removing nodes (and potentially
412
412
  # interrupting jobs). Default timeout is 0 (for forceful decommission), and
413
- # the maximum allowed timeout is 1 day.
413
+ # the maximum allowed timeout is 1 day (see JSON representation of
414
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
414
415
  #
415
416
  # Only supported on Dataproc image versions 1.2 and higher.
416
417
  # A hash of the same form as `Google::Protobuf::Duration`
@@ -517,7 +518,7 @@ module Google
517
518
  # Required. The ID of the Google Cloud Platform project that the cluster
518
519
  # belongs to.
519
520
  # @param region [String]
520
- # Required. The Cloud Dataproc region in which to handle the request.
521
+ # Required. The Dataproc region in which to handle the request.
521
522
  # @param cluster_name [String]
522
523
  # Required. The cluster name.
523
524
  # @param cluster_uuid [String]
@@ -613,7 +614,7 @@ module Google
613
614
  # Required. The ID of the Google Cloud Platform project that the cluster
614
615
  # belongs to.
615
616
  # @param region [String]
616
- # Required. The Cloud Dataproc region in which to handle the request.
617
+ # Required. The Dataproc region in which to handle the request.
617
618
  # @param cluster_name [String]
618
619
  # Required. The cluster name.
619
620
  # @param options [Google::Gax::CallOptions]
@@ -660,7 +661,7 @@ module Google
660
661
  # Required. The ID of the Google Cloud Platform project that the cluster
661
662
  # belongs to.
662
663
  # @param region [String]
663
- # Required. The Cloud Dataproc region in which to handle the request.
664
+ # Required. The Dataproc region in which to handle the request.
664
665
  # @param filter [String]
665
666
  # Optional. A filter constraining the clusters to list. Filters are
666
667
  # case-sensitive and have the following syntax:
@@ -746,13 +747,13 @@ module Google
746
747
  # After the operation completes,
747
748
  # {Google::Longrunning::Operation#response Operation#response}
748
749
  # contains
749
- # [Empty](https://cloud.google.comgoogle.protobuf.Empty).
750
+ # {Google::Protobuf::Empty Empty}.
750
751
  #
751
752
  # @param project_id [String]
752
753
  # Required. The ID of the Google Cloud Platform project that the cluster
753
754
  # belongs to.
754
755
  # @param region [String]
755
- # Required. The Cloud Dataproc region in which to handle the request.
756
+ # Required. The Dataproc region in which to handle the request.
756
757
  # @param cluster_name [String]
757
758
  # Required. The cluster name.
758
759
  # @param options [Google::Gax::CallOptions]
@@ -7,7 +7,7 @@ require 'google/protobuf'
7
7
  require 'google/api/annotations_pb'
8
8
  require 'google/api/client_pb'
9
9
  require 'google/api/field_behavior_pb'
10
- require 'google/cloud/dataproc/v1beta2/operations_pb'
10
+ require 'google/api/resource_pb'
11
11
  require 'google/cloud/dataproc/v1beta2/shared_pb'
12
12
  require 'google/longrunning/operations_pb'
13
13
  require 'google/protobuf/duration_pb'
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/clusters.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -38,15 +37,15 @@ module Google
38
37
 
39
38
  # Creates a cluster in a project. The returned
40
39
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
41
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
40
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
42
41
  rpc :CreateCluster, CreateClusterRequest, Google::Longrunning::Operation
43
42
  # Updates a cluster in a project. The returned
44
43
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
45
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
44
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
46
45
  rpc :UpdateCluster, UpdateClusterRequest, Google::Longrunning::Operation
47
46
  # Deletes a cluster in a project. The returned
48
47
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
49
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
48
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
50
49
  rpc :DeleteCluster, DeleteClusterRequest, Google::Longrunning::Operation
51
50
  # Gets the resource representation for a cluster in a project.
52
51
  rpc :GetCluster, GetClusterRequest, Cluster
@@ -54,11 +53,11 @@ module Google
54
53
  rpc :ListClusters, ListClustersRequest, ListClustersResponse
55
54
  # Gets cluster diagnostic information. The returned
56
55
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
57
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
56
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
58
57
  # After the operation completes,
59
58
  # [Operation.response][google.longrunning.Operation.response]
60
59
  # contains
61
- # [Empty](google.protobuf.Empty).
60
+ # [Empty][google.protobuf.Empty].
62
61
  rpc :DiagnoseCluster, DiagnoseClusterRequest, Google::Longrunning::Operation
63
62
  end
64
63
 
@@ -28,7 +28,7 @@ module Google
28
28
  # unique. Names of deleted clusters can be reused.
29
29
  # @!attribute [rw] config
30
30
  # @return [Google::Cloud::Dataproc::V1beta2::ClusterConfig]
31
- # Required. The cluster config. Note that Cloud Dataproc may set
31
+ # Required. The cluster config. Note that Dataproc may set
32
32
  # default values, and values may change when clusters are updated.
33
33
  # @!attribute [rw] labels
34
34
  # @return [Hash{String => String}]
@@ -47,7 +47,7 @@ module Google
47
47
  # Output only. The previous cluster status.
48
48
  # @!attribute [rw] cluster_uuid
49
49
  # @return [String]
50
- # Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
50
+ # Output only. A cluster UUID (Unique Universal Identifier). Dataproc
51
51
  # generates this value when it creates the cluster.
52
52
  # @!attribute [rw] metrics
53
53
  # @return [Google::Cloud::Dataproc::V1beta2::ClusterMetrics]
@@ -60,14 +60,14 @@ module Google
60
60
  # The cluster config.
61
61
  # @!attribute [rw] config_bucket
62
62
  # @return [String]
63
- # Optional. A Google Cloud Storage bucket used to stage job
63
+ # Optional. A Cloud Storage bucket used to stage job
64
64
  # dependencies, config files, and job driver console output.
65
65
  # If you do not specify a staging bucket, Cloud
66
66
  # Dataproc will determine a Cloud Storage location (US,
67
- # ASIA, or EU) for your cluster's staging bucket according to the Google
67
+ # ASIA, or EU) for your cluster's staging bucket according to the
68
68
  # Compute Engine zone where your cluster is deployed, and then create
69
69
  # and manage this project-level, per-location bucket (see
70
- # [Cloud Dataproc staging
70
+ # [Dataproc staging
71
71
  # bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
72
72
  # @!attribute [rw] gce_cluster_config
73
73
  # @return [Google::Cloud::Dataproc::V1beta2::GceClusterConfig]
@@ -143,7 +143,7 @@ module Google
143
143
  # * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
144
144
  # * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
145
145
  #
146
- # Note that the policy must be in the same project and Cloud Dataproc region.
146
+ # Note that the policy must be in the same project and Dataproc region.
147
147
  class AutoscalingConfig; end
148
148
 
149
149
  # Encryption settings for the cluster.
@@ -159,7 +159,7 @@ module Google
159
159
  # @return [String]
160
160
  # Optional. The zone where the Compute Engine cluster will be located.
161
161
  # On a create request, it is required in the "global" region. If omitted
162
- # in a non-global Cloud Dataproc region, the service will pick a zone in the
162
+ # in a non-global Dataproc region, the service will pick a zone in the
163
163
  # corresponding Compute Engine region. On a get request, zone will always be
164
164
  # present.
165
165
  #
@@ -201,17 +201,17 @@ module Google
201
201
  # configured to be accessible without external IP addresses.
202
202
  # @!attribute [rw] service_account
203
203
  # @return [String]
204
- # Optional. The service account of the instances. Defaults to the default
205
- # Compute Engine service account. Custom service accounts need
206
- # permissions equivalent to the following IAM roles:
207
- #
208
- # * roles/logging.logWriter
209
- # * roles/storage.objectAdmin
210
- #
211
- # (see
212
- # https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
213
- # for more information).
214
- # Example: `[account_id]@[project_id].iam.gserviceaccount.com`
204
+ # Optional. The [Dataproc service
205
+ # account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
206
+ # (also see [VM Data Plane
207
+ # identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
208
+ # used by Dataproc cluster VM instances to access Google Cloud Platform
209
+ # services.
210
+ #
211
+ # If not specified, the
212
+ # [Compute Engine default service
213
+ # account](/compute/docs/access/service-accounts#default_service_account)
214
+ # is used.
215
215
  # @!attribute [rw] service_account_scopes
216
216
  # @return [Array<String>]
217
217
  # Optional. The URIs of service account scopes to be included in
@@ -250,7 +250,7 @@ module Google
250
250
  # For master instance groups, must be set to 1.
251
251
  # @!attribute [rw] instance_names
252
252
  # @return [Array<String>]
253
- # Output only. The list of instance names. Cloud Dataproc derives the names
253
+ # Output only. The list of instance names. Dataproc derives the names
254
254
  # from `cluster_name`, `num_instances`, and the instance group.
255
255
  # @!attribute [rw] image_uri
256
256
  # @return [String]
@@ -267,7 +267,7 @@ module Google
267
267
  # * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2`
268
268
  # * `n1-standard-2`
269
269
  #
270
- # **Auto Zone Exception**: If you are using the Cloud Dataproc
270
+ # **Auto Zone Exception**: If you are using the Dataproc
271
271
  # [Auto Zone
272
272
  # Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
273
273
  # feature, you must use the short name of the machine type
@@ -291,7 +291,7 @@ module Google
291
291
  # @!attribute [rw] min_cpu_platform
292
292
  # @return [String]
293
293
  # Specifies the minimum cpu platform for the Instance Group.
294
- # See [Cloud Dataproc&rarr;Minimum CPU Platform]
294
+ # See [Dataproc&rarr;Minimum CPU Platform]
295
295
  # (/dataproc/docs/concepts/compute/dataproc-min-cpu).
296
296
  class InstanceGroupConfig; end
297
297
 
@@ -319,7 +319,7 @@ module Google
319
319
  # * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80`
320
320
  # * `nvidia-tesla-k80`
321
321
  #
322
- # **Auto Zone Exception**: If you are using the Cloud Dataproc
322
+ # **Auto Zone Exception**: If you are using the Dataproc
323
323
  # [Auto Zone
324
324
  # Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
325
325
  # feature, you must use the short name of the accelerator type
@@ -351,25 +351,27 @@ module Google
351
351
  # Specifies the cluster auto-delete schedule configuration.
352
352
  # @!attribute [rw] idle_delete_ttl
353
353
  # @return [Google::Protobuf::Duration]
354
- # Optional. The duration to keep the cluster alive while idling.
355
- # Passing this threshold will cause the cluster to be
356
- # deleted. Valid range: **[10m, 14d]**.
357
- #
358
- # Example: **"10m"**, the minimum value, to delete the
359
- # cluster when it has had no jobs running for 10 minutes.
354
+ # Optional. The duration to keep the cluster alive while idling (when no jobs
355
+ # are running). Passing this threshold will cause the cluster to be
356
+ # deleted. Minimum value is 10 minutes; maximum value is 14 days (see JSON
357
+ # representation of
358
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json).
360
359
  # @!attribute [rw] auto_delete_time
361
360
  # @return [Google::Protobuf::Timestamp]
362
- # Optional. The time when cluster will be auto-deleted.
361
+ # Optional. The time when cluster will be auto-deleted. (see JSON representation of
362
+ # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
363
363
  # @!attribute [rw] auto_delete_ttl
364
364
  # @return [Google::Protobuf::Duration]
365
365
  # Optional. The lifetime duration of cluster. The cluster will be
366
- # auto-deleted at the end of this period. Valid range: **[10m, 14d]**.
367
- #
368
- # Example: **"1d"**, to delete the cluster 1 day after its creation..
366
+ # auto-deleted at the end of this period. Minimum value is 10 minutes;
367
+ # maximum value is 14 days (see JSON representation of
368
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
369
369
  # @!attribute [rw] idle_start_time
370
370
  # @return [Google::Protobuf::Timestamp]
371
371
  # Output only. The time when cluster became idle (most recent job finished)
372
- # and became eligible for deletion due to idleness.
372
+ # and became eligible for deletion due to idleness (see JSON representation
373
+ # of
374
+ # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
373
375
  class LifecycleConfig; end
374
376
 
375
377
  # Security related configuration, including encryption, Kerberos, etc.
@@ -455,7 +457,10 @@ module Google
455
457
  # @!attribute [rw] execution_timeout
456
458
  # @return [Google::Protobuf::Duration]
457
459
  # Optional. Amount of time executable has to complete. Default is
458
- # 10 minutes. Cluster creation fails with an explanatory error message (the
460
+ # 10 minutes (see JSON representation of
461
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
462
+ #
463
+ # Cluster creation fails with an explanatory error message (the
459
464
  # name of the executable that caused the error and the exceeded timeout
460
465
  # period) if the executable is not completed at end of the timeout period.
461
466
  class NodeInitializationAction; end
@@ -469,7 +474,8 @@ module Google
469
474
  # Output only. Optional details of cluster's state.
470
475
  # @!attribute [rw] state_start_time
471
476
  # @return [Google::Protobuf::Timestamp]
472
- # Output only. Time when this state was entered.
477
+ # Output only. Time when this state was entered (see JSON representation of
478
+ # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
473
479
  # @!attribute [rw] substate
474
480
  # @return [Google::Cloud::Dataproc::V1beta2::ClusterStatus::Substate]
475
481
  # Output only. Additional state information that includes
@@ -509,7 +515,7 @@ module Google
509
515
  UNHEALTHY = 1
510
516
 
511
517
  # The agent-reported status is out of date (may occur if
512
- # Cloud Dataproc loses communication with Agent).
518
+ # Dataproc loses communication with Agent).
513
519
  #
514
520
  # Applies to RUNNING state.
515
521
  STALE_STATUS = 2
@@ -520,7 +526,7 @@ module Google
520
526
  # @!attribute [rw] image_version
521
527
  # @return [String]
522
528
  # Optional. The version of software inside the cluster. It must be one of the
523
- # supported [Cloud Dataproc
529
+ # supported [Dataproc
524
530
  # Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
525
531
  # such as "1.2" (including a subminor version, such as "1.2.29"), or the
526
532
  # ["preview"
@@ -570,7 +576,7 @@ module Google
570
576
  # belongs to.
571
577
  # @!attribute [rw] region
572
578
  # @return [String]
573
- # Required. The Cloud Dataproc region in which to handle the request.
579
+ # Required. The Dataproc region in which to handle the request.
574
580
  # @!attribute [rw] cluster
575
581
  # @return [Google::Cloud::Dataproc::V1beta2::Cluster]
576
582
  # Required. The cluster to create.
@@ -596,7 +602,7 @@ module Google
596
602
  # cluster belongs to.
597
603
  # @!attribute [rw] region
598
604
  # @return [String]
599
- # Required. The Cloud Dataproc region in which to handle the request.
605
+ # Required. The Dataproc region in which to handle the request.
600
606
  # @!attribute [rw] cluster_name
601
607
  # @return [String]
602
608
  # Required. The cluster name.
@@ -610,7 +616,8 @@ module Google
610
616
  # interrupting jobs in progress. Timeout specifies how long to wait for jobs
611
617
  # in progress to finish before forcefully removing nodes (and potentially
612
618
  # interrupting jobs). Default timeout is 0 (for forceful decommission), and
613
- # the maximum allowed timeout is 1 day.
619
+ # the maximum allowed timeout is 1 day (see JSON representation of
620
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
614
621
  #
615
622
  # Only supported on Dataproc image versions 1.2 and higher.
616
623
  # @!attribute [rw] update_mask
@@ -697,7 +704,7 @@ module Google
697
704
  # belongs to.
698
705
  # @!attribute [rw] region
699
706
  # @return [String]
700
- # Required. The Cloud Dataproc region in which to handle the request.
707
+ # Required. The Dataproc region in which to handle the request.
701
708
  # @!attribute [rw] cluster_name
702
709
  # @return [String]
703
710
  # Required. The cluster name.
@@ -727,7 +734,7 @@ module Google
727
734
  # belongs to.
728
735
  # @!attribute [rw] region
729
736
  # @return [String]
730
- # Required. The Cloud Dataproc region in which to handle the request.
737
+ # Required. The Dataproc region in which to handle the request.
731
738
  # @!attribute [rw] cluster_name
732
739
  # @return [String]
733
740
  # Required. The cluster name.
@@ -740,7 +747,7 @@ module Google
740
747
  # belongs to.
741
748
  # @!attribute [rw] region
742
749
  # @return [String]
743
- # Required. The Cloud Dataproc region in which to handle the request.
750
+ # Required. The Dataproc region in which to handle the request.
744
751
  # @!attribute [rw] filter
745
752
  # @return [String]
746
753
  # Optional. A filter constraining the clusters to list. Filters are
@@ -788,7 +795,7 @@ module Google
788
795
  # belongs to.
789
796
  # @!attribute [rw] region
790
797
  # @return [String]
791
- # Required. The Cloud Dataproc region in which to handle the request.
798
+ # Required. The Dataproc region in which to handle the request.
792
799
  # @!attribute [rw] cluster_name
793
800
  # @return [String]
794
801
  # Required. The cluster name.
@@ -58,7 +58,7 @@ module Google
58
58
  end
59
59
  end
60
60
 
61
- # A Cloud Dataproc job for running
61
+ # A Dataproc job for running
62
62
  # [Apache Hadoop
63
63
  # MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
64
64
  # jobs on [Apache Hadoop
@@ -97,7 +97,7 @@ module Google
97
97
  # @!attribute [rw] properties
98
98
  # @return [Hash{String => String}]
99
99
  # Optional. A mapping of property names to values, used to configure Hadoop.
100
- # Properties that conflict with values set by the Cloud Dataproc API may be
100
+ # Properties that conflict with values set by the Dataproc API may be
101
101
  # overwritten. Can include properties set in /etc/hadoop/conf/*-site and
102
102
  # classes in user code.
103
103
  # @!attribute [rw] logging_config
@@ -105,7 +105,7 @@ module Google
105
105
  # Optional. The runtime log config for job execution.
106
106
  class HadoopJob; end
107
107
 
108
- # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/)
108
+ # A Dataproc job for running [Apache Spark](http://spark.apache.org/)
109
109
  # applications on YARN.
110
110
  # The specification of the main method to call to drive the job.
111
111
  # Specify either the jar file that contains the main class or the main class
@@ -140,7 +140,7 @@ module Google
140
140
  # @!attribute [rw] properties
141
141
  # @return [Hash{String => String}]
142
142
  # Optional. A mapping of property names to values, used to configure Spark.
143
- # Properties that conflict with values set by the Cloud Dataproc API may be
143
+ # Properties that conflict with values set by the Dataproc API may be
144
144
  # overwritten. Can include properties set in
145
145
  # /etc/spark/conf/spark-defaults.conf and classes in user code.
146
146
  # @!attribute [rw] logging_config
@@ -148,7 +148,7 @@ module Google
148
148
  # Optional. The runtime log config for job execution.
149
149
  class SparkJob; end
150
150
 
151
- # A Cloud Dataproc job for running
151
+ # A Dataproc job for running
152
152
  # [Apache
153
153
  # PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
154
154
  # applications on YARN.
@@ -180,7 +180,7 @@ module Google
180
180
  # @!attribute [rw] properties
181
181
  # @return [Hash{String => String}]
182
182
  # Optional. A mapping of property names to values, used to configure PySpark.
183
- # Properties that conflict with values set by the Cloud Dataproc API may be
183
+ # Properties that conflict with values set by the Dataproc API may be
184
184
  # overwritten. Can include properties set in
185
185
  # /etc/spark/conf/spark-defaults.conf and classes in user code.
186
186
  # @!attribute [rw] logging_config
@@ -207,7 +207,7 @@ module Google
207
207
  # }
208
208
  class QueryList; end
209
209
 
210
- # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/)
210
+ # A Dataproc job for running [Apache Hive](https://hive.apache.org/)
211
211
  # queries on YARN.
212
212
  # @!attribute [rw] query_file_uri
213
213
  # @return [String]
@@ -227,7 +227,7 @@ module Google
227
227
  # @!attribute [rw] properties
228
228
  # @return [Hash{String => String}]
229
229
  # Optional. A mapping of property names and values, used to configure Hive.
230
- # Properties that conflict with values set by the Cloud Dataproc API may be
230
+ # Properties that conflict with values set by the Dataproc API may be
231
231
  # overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
232
232
  # /etc/hive/conf/hive-site.xml, and classes in user code.
233
233
  # @!attribute [rw] jar_file_uris
@@ -237,7 +237,7 @@ module Google
237
237
  # and UDFs.
238
238
  class HiveJob; end
239
239
 
240
- # A Cloud Dataproc job for running [Apache Spark
240
+ # A Dataproc job for running [Apache Spark
241
241
  # SQL](http://spark.apache.org/sql/) queries.
242
242
  # @!attribute [rw] query_file_uri
243
243
  # @return [String]
@@ -253,7 +253,7 @@ module Google
253
253
  # @return [Hash{String => String}]
254
254
  # Optional. A mapping of property names to values, used to configure
255
255
  # Spark SQL's SparkConf. Properties that conflict with values set by the
256
- # Cloud Dataproc API may be overwritten.
256
+ # Dataproc API may be overwritten.
257
257
  # @!attribute [rw] jar_file_uris
258
258
  # @return [Array<String>]
259
259
  # Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
@@ -262,7 +262,7 @@ module Google
262
262
  # Optional. The runtime log config for job execution.
263
263
  class SparkSqlJob; end
264
264
 
265
- # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/)
265
+ # A Dataproc job for running [Apache Pig](https://pig.apache.org/)
266
266
  # queries on YARN.
267
267
  # @!attribute [rw] query_file_uri
268
268
  # @return [String]
@@ -282,7 +282,7 @@ module Google
282
282
  # @!attribute [rw] properties
283
283
  # @return [Hash{String => String}]
284
284
  # Optional. A mapping of property names to values, used to configure Pig.
285
- # Properties that conflict with values set by the Cloud Dataproc API may be
285
+ # Properties that conflict with values set by the Dataproc API may be
286
286
  # overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
287
287
  # /etc/pig/conf/pig.properties, and classes in user code.
288
288
  # @!attribute [rw] jar_file_uris
@@ -294,7 +294,7 @@ module Google
294
294
  # Optional. The runtime log config for job execution.
295
295
  class PigJob; end
296
296
 
297
- # A Cloud Dataproc job for running
297
+ # A Dataproc job for running
298
298
  # [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html)
299
299
  # applications on YARN.
300
300
  # @!attribute [rw] main_r_file_uri
@@ -318,7 +318,7 @@ module Google
318
318
  # @!attribute [rw] properties
319
319
  # @return [Hash{String => String}]
320
320
  # Optional. A mapping of property names to values, used to configure SparkR.
321
- # Properties that conflict with values set by the Cloud Dataproc API may be
321
+ # Properties that conflict with values set by the Dataproc API may be
322
322
  # overwritten. Can include properties set in
323
323
  # /etc/spark/conf/spark-defaults.conf and classes in user code.
324
324
  # @!attribute [rw] logging_config
@@ -326,23 +326,23 @@ module Google
326
326
  # Optional. The runtime log config for job execution.
327
327
  class SparkRJob; end
328
328
 
329
- # Cloud Dataproc job config.
329
+ # Dataproc job config.
330
330
  # @!attribute [rw] cluster_name
331
331
  # @return [String]
332
332
  # Required. The name of the cluster where the job will be submitted.
333
333
  # @!attribute [rw] cluster_uuid
334
334
  # @return [String]
335
- # Output only. A cluster UUID generated by the Cloud Dataproc service when
335
+ # Output only. A cluster UUID generated by the Dataproc service when
336
336
  # the job is submitted.
337
337
  class JobPlacement; end
338
338
 
339
- # Cloud Dataproc job status.
339
+ # Dataproc job status.
340
340
  # @!attribute [rw] state
341
341
  # @return [Google::Cloud::Dataproc::V1beta2::JobStatus::State]
342
342
  # Output only. A state message specifying the overall job state.
343
343
  # @!attribute [rw] details
344
344
  # @return [String]
345
- # Output only. Optional job state details, such as an error
345
+ # Output only. Optional Job state details, such as an error
346
346
  # description if the state is <code>ERROR</code>.
347
347
  # @!attribute [rw] state_start_time
348
348
  # @return [Google::Protobuf::Timestamp]
@@ -408,7 +408,7 @@ module Google
408
408
  QUEUED = 2
409
409
 
410
410
  # The agent-reported status is out of date, which may be caused by a
411
- # loss of communication between the agent and Cloud Dataproc. If the
411
+ # loss of communication between the agent and Dataproc. If the
412
412
  # agent does not send a timely update, the job will fail.
413
413
  #
414
414
  # Applies to RUNNING state.
@@ -424,7 +424,6 @@ module Google
424
424
  # @!attribute [rw] job_id
425
425
  # @return [String]
426
426
  # Optional. The job ID, which must be unique within the project.
427
- #
428
427
  # The ID must contain only letters (a-z, A-Z), numbers (0-9),
429
428
  # underscores (_), or hyphens (-). The maximum length is 100 characters.
430
429
  #
@@ -447,7 +446,7 @@ module Google
447
446
  # Output only. The numerical progress of the application, from 1 to 100.
448
447
  # @!attribute [rw] tracking_url
449
448
  # @return [String]
450
- # Optional. Output only. The HTTP URL of the ApplicationMaster, HistoryServer, or
449
+ # Output only. The HTTP URL of the ApplicationMaster, HistoryServer, or
451
450
  # TimelineServer that provides application-specific information. The URL uses
452
451
  # the internal hostname, and requires a proxy server for resolution and,
453
452
  # possibly, access.
@@ -484,7 +483,7 @@ module Google
484
483
  end
485
484
  end
486
485
 
487
- # A Cloud Dataproc job resource.
486
+ # A Dataproc job resource.
488
487
  # @!attribute [rw] reference
489
488
  # @return [Google::Cloud::Dataproc::V1beta2::JobReference]
490
489
  # Optional. The fully qualified reference to the job, which can be used to
@@ -497,25 +496,18 @@ module Google
497
496
  # run the job.
498
497
  # @!attribute [rw] hadoop_job
499
498
  # @return [Google::Cloud::Dataproc::V1beta2::HadoopJob]
500
- # Job is a Hadoop job.
501
499
  # @!attribute [rw] spark_job
502
500
  # @return [Google::Cloud::Dataproc::V1beta2::SparkJob]
503
- # Job is a Spark job.
504
501
  # @!attribute [rw] pyspark_job
505
502
  # @return [Google::Cloud::Dataproc::V1beta2::PySparkJob]
506
- # Job is a Pyspark job.
507
503
  # @!attribute [rw] hive_job
508
504
  # @return [Google::Cloud::Dataproc::V1beta2::HiveJob]
509
- # Job is a Hive job.
510
505
  # @!attribute [rw] pig_job
511
506
  # @return [Google::Cloud::Dataproc::V1beta2::PigJob]
512
- # Job is a Pig job.
513
507
  # @!attribute [rw] spark_r_job
514
508
  # @return [Google::Cloud::Dataproc::V1beta2::SparkRJob]
515
- # Job is a SparkR job.
516
509
  # @!attribute [rw] spark_sql_job
517
510
  # @return [Google::Cloud::Dataproc::V1beta2::SparkSqlJob]
518
- # Job is a SparkSql job.
519
511
  # @!attribute [rw] status
520
512
  # @return [Google::Cloud::Dataproc::V1beta2::JobStatus]
521
513
  # Output only. The job status. Additional application-specific
@@ -582,7 +574,7 @@ module Google
582
574
  # belongs to.
583
575
  # @!attribute [rw] region
584
576
  # @return [String]
585
- # Required. The Cloud Dataproc region in which to handle the request.
577
+ # Required. The Dataproc region in which to handle the request.
586
578
  # @!attribute [rw] job
587
579
  # @return [Google::Cloud::Dataproc::V1beta2::Job]
588
580
  # Required. The job resource.
@@ -608,7 +600,7 @@ module Google
608
600
  # belongs to.
609
601
  # @!attribute [rw] region
610
602
  # @return [String]
611
- # Required. The Cloud Dataproc region in which to handle the request.
603
+ # Required. The Dataproc region in which to handle the request.
612
604
  # @!attribute [rw] job_id
613
605
  # @return [String]
614
606
  # Required. The job ID.
@@ -621,7 +613,7 @@ module Google
621
613
  # belongs to.
622
614
  # @!attribute [rw] region
623
615
  # @return [String]
624
- # Required. The Cloud Dataproc region in which to handle the request.
616
+ # Required. The Dataproc region in which to handle the request.
625
617
  # @!attribute [rw] page_size
626
618
  # @return [Integer]
627
619
  # Optional. The number of results to return in each response.
@@ -677,7 +669,7 @@ module Google
677
669
  # belongs to.
678
670
  # @!attribute [rw] region
679
671
  # @return [String]
680
- # Required. The Cloud Dataproc region in which to handle the request.
672
+ # Required. The Dataproc region in which to handle the request.
681
673
  # @!attribute [rw] job_id
682
674
  # @return [String]
683
675
  # Required. The job ID.
@@ -712,7 +704,7 @@ module Google
712
704
  # belongs to.
713
705
  # @!attribute [rw] region
714
706
  # @return [String]
715
- # Required. The Cloud Dataproc region in which to handle the request.
707
+ # Required. The Dataproc region in which to handle the request.
716
708
  # @!attribute [rw] job_id
717
709
  # @return [String]
718
710
  # Required. The job ID.
@@ -725,7 +717,7 @@ module Google
725
717
  # belongs to.
726
718
  # @!attribute [rw] region
727
719
  # @return [String]
728
- # Required. The Cloud Dataproc region in which to handle the request.
720
+ # Required. The Dataproc region in which to handle the request.
729
721
  # @!attribute [rw] job_id
730
722
  # @return [String]
731
723
  # Required. The job ID.
@@ -17,7 +17,7 @@ module Google
17
17
  module Cloud
18
18
  module Dataproc
19
19
  module V1beta2
20
- # A Cloud Dataproc workflow template resource.
20
+ # A Dataproc workflow template resource.
21
21
  # @!attribute [rw] id
22
22
  # @return [String]
23
23
  # Required. The template id.
@@ -153,22 +153,16 @@ module Google
153
153
  # or hyphen. Must consist of between 3 and 50 characters.
154
154
  # @!attribute [rw] hadoop_job
155
155
  # @return [Google::Cloud::Dataproc::V1beta2::HadoopJob]
156
- # Job is a Hadoop job.
157
156
  # @!attribute [rw] spark_job
158
157
  # @return [Google::Cloud::Dataproc::V1beta2::SparkJob]
159
- # Job is a Spark job.
160
158
  # @!attribute [rw] pyspark_job
161
159
  # @return [Google::Cloud::Dataproc::V1beta2::PySparkJob]
162
- # Job is a Pyspark job.
163
160
  # @!attribute [rw] hive_job
164
161
  # @return [Google::Cloud::Dataproc::V1beta2::HiveJob]
165
- # Job is a Hive job.
166
162
  # @!attribute [rw] pig_job
167
163
  # @return [Google::Cloud::Dataproc::V1beta2::PigJob]
168
- # Job is a Pig job.
169
164
  # @!attribute [rw] spark_sql_job
170
165
  # @return [Google::Cloud::Dataproc::V1beta2::SparkSqlJob]
171
- # Job is a SparkSql job.
172
166
  # @!attribute [rw] labels
173
167
  # @return [Hash{String => String}]
174
168
  # Optional. The labels to associate with this job.
@@ -286,7 +280,7 @@ module Google
286
280
  # Required. List of allowed values for the parameter.
287
281
  class ValueValidation; end
288
282
 
289
- # A Cloud Dataproc workflow template resource.
283
+ # A Dataproc workflow template resource.
290
284
  # @!attribute [rw] template
291
285
  # @return [String]
292
286
  # Output only. The resource name of the workflow template as described
@@ -217,7 +217,7 @@ module Google
217
217
  # Required. The ID of the Google Cloud Platform project that the job
218
218
  # belongs to.
219
219
  # @param region [String]
220
- # Required. The Cloud Dataproc region in which to handle the request.
220
+ # Required. The Dataproc region in which to handle the request.
221
221
  # @param job [Google::Cloud::Dataproc::V1beta2::Job | Hash]
222
222
  # Required. The job resource.
223
223
  # A hash of the same form as `Google::Cloud::Dataproc::V1beta2::Job`
@@ -280,7 +280,7 @@ module Google
280
280
  # Required. The ID of the Google Cloud Platform project that the job
281
281
  # belongs to.
282
282
  # @param region [String]
283
- # Required. The Cloud Dataproc region in which to handle the request.
283
+ # Required. The Dataproc region in which to handle the request.
284
284
  # @param job_id [String]
285
285
  # Required. The job ID.
286
286
  # @param options [Google::Gax::CallOptions]
@@ -327,7 +327,7 @@ module Google
327
327
  # Required. The ID of the Google Cloud Platform project that the job
328
328
  # belongs to.
329
329
  # @param region [String]
330
- # Required. The Cloud Dataproc region in which to handle the request.
330
+ # Required. The Dataproc region in which to handle the request.
331
331
  # @param page_size [Integer]
332
332
  # The maximum number of resources contained in the underlying API
333
333
  # response. If page streaming is performed per-resource, this
@@ -420,7 +420,7 @@ module Google
420
420
  # Required. The ID of the Google Cloud Platform project that the job
421
421
  # belongs to.
422
422
  # @param region [String]
423
- # Required. The Cloud Dataproc region in which to handle the request.
423
+ # Required. The Dataproc region in which to handle the request.
424
424
  # @param job_id [String]
425
425
  # Required. The job ID.
426
426
  # @param job [Google::Cloud::Dataproc::V1beta2::Job | Hash]
@@ -494,7 +494,7 @@ module Google
494
494
  # Required. The ID of the Google Cloud Platform project that the job
495
495
  # belongs to.
496
496
  # @param region [String]
497
- # Required. The Cloud Dataproc region in which to handle the request.
497
+ # Required. The Dataproc region in which to handle the request.
498
498
  # @param job_id [String]
499
499
  # Required. The job ID.
500
500
  # @param options [Google::Gax::CallOptions]
@@ -542,7 +542,7 @@ module Google
542
542
  # Required. The ID of the Google Cloud Platform project that the job
543
543
  # belongs to.
544
544
  # @param region [String]
545
- # Required. The Cloud Dataproc region in which to handle the request.
545
+ # Required. The Dataproc region in which to handle the request.
546
546
  # @param job_id [String]
547
547
  # Required. The job ID.
548
548
  # @param options [Google::Gax::CallOptions]
@@ -7,6 +7,7 @@ require 'google/protobuf'
7
7
  require 'google/api/annotations_pb'
8
8
  require 'google/api/client_pb'
9
9
  require 'google/api/field_behavior_pb'
10
+ require 'google/longrunning/operations_pb'
10
11
  require 'google/protobuf/empty_pb'
11
12
  require 'google/protobuf/field_mask_pb'
12
13
  require 'google/protobuf/timestamp_pb'
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/jobs.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -45,9 +44,9 @@ module Google
45
44
  rpc :UpdateJob, UpdateJobRequest, Job
46
45
  # Starts a job cancellation request. To access the job resource
47
46
  # after cancellation, call
48
- # [regions/\\{region}/jobs.list](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/list)
47
+ # [regions/\\{region}/jobs.list](https://cloud.google.com/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/list)
49
48
  # or
50
- # [regions/\\{region}/jobs.get](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get).
49
+ # [regions/\\{region}/jobs.get](https://cloud.google.com/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get).
51
50
  rpc :CancelJob, CancelJobRequest, Job
52
51
  # Deletes the job from the project. If the job is active, the delete fails,
53
52
  # and the response returns `FAILED_PRECONDITION`.
@@ -36,7 +36,7 @@ module Google
36
36
  module Dataproc
37
37
  module V1beta2
38
38
  # The API interface for managing Workflow Templates in the
39
- # Cloud Dataproc API.
39
+ # Dataproc API.
40
40
  #
41
41
  # @!attribute [r] workflow_template_service_stub
42
42
  # @return [Google::Cloud::Dataproc::V1beta2::WorkflowTemplateService::Stub]
@@ -404,7 +404,7 @@ module Google
404
404
  # The {Google::Longrunning::Operation#metadata Operation#metadata} will be
405
405
  # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
406
406
  # Also see [Using
407
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
407
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
408
408
  #
409
409
  # On successful completion,
410
410
  # {Google::Longrunning::Operation#response Operation#response} will be
@@ -526,7 +526,7 @@ module Google
526
526
  # The {Google::Longrunning::Operation#metadata Operation#metadata} will be
527
527
  # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
528
528
  # Also see [Using
529
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
529
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
530
530
  #
531
531
  # On successful completion,
532
532
  # {Google::Longrunning::Operation#response Operation#response} will be
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/workflow_templates.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -27,7 +26,7 @@ module Google
27
26
  module V1beta2
28
27
  module WorkflowTemplateService
29
28
  # The API interface for managing Workflow Templates in the
30
- # Cloud Dataproc API.
29
+ # Dataproc API.
31
30
  class Service
32
31
 
33
32
  include GRPC::GenericService
@@ -56,9 +55,9 @@ module Google
56
55
  # clusters to be deleted.
57
56
  #
58
57
  # The [Operation.metadata][google.longrunning.Operation.metadata] will be
59
- # [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
58
+ # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
60
59
  # Also see [Using
61
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
60
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
62
61
  #
63
62
  # On successful completion,
64
63
  # [Operation.response][google.longrunning.Operation.response] will be
@@ -81,9 +80,9 @@ module Google
81
80
  # clusters to be deleted.
82
81
  #
83
82
  # The [Operation.metadata][google.longrunning.Operation.metadata] will be
84
- # [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
83
+ # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
85
84
  # Also see [Using
86
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
85
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
87
86
  #
88
87
  # On successful completion,
89
88
  # [Operation.response][google.longrunning.Operation.response] will be
@@ -16,7 +16,7 @@
16
16
  module Google
17
17
  module Cloud
18
18
  module Dataproc
19
- VERSION = "0.9.0".freeze
19
+ VERSION = "0.9.1".freeze
20
20
  end
21
21
  end
22
22
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-dataproc
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.9.0
4
+ version: 0.9.1
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2020-02-24 00:00:00.000000000 Z
11
+ date: 2020-03-02 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: google-gax