google-cloud-dataproc 0.9.0 → 0.9.1

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 73149665558ed4fef80dcabc2baf92983137ac4bde65d74448dbc26865f47107
4
- data.tar.gz: d5b0938c6bdd27df5172550f5e635f58fbed0276ec1d3668f61adce7e01c5767
3
+ metadata.gz: 8f4dc9ab9721db6accf4279d8963bc946bffef23d3480d5b661479df3cf76203
4
+ data.tar.gz: 9919cb5e80780916635366bdabe88f9b4675f5cfaedaf07596633a5c316c8745
5
5
  SHA512:
6
- metadata.gz: bf6e7328361ba5d9e64d47732230f288d451e728328a3e3058e378f53e47f2b796e7f9c75d8212d826f24cc72f117eb5b95119a9ce392e86889e1b056757735f
7
- data.tar.gz: b8404d746ddae8bdbfb87acae3f1109e9813d68a0f4604e85c7f1ee71af38d1bdc26ab4f5345505cdc91858f46db86011de784e503b1faa83c839bd2585f8939
6
+ metadata.gz: 9b59bad93eadf97880799b451199ab06ff8fb4957463cfcc35277176352ad38098916931ee78cabf8e0507505c49f1f3c3636dbeebb8868c1b7ec0937bcbb9f3
7
+ data.tar.gz: 9d60a85d4b95872d332bce6a501be08be0c6d61a4a4f0700392f4f3e2e00ff85b886e2a01b23db2502cfd346fd37269eedb284d54d261466a7354dbbc9437e53
@@ -112,7 +112,7 @@ module Google
112
112
  module AutoscalingPolicyService
113
113
  ##
114
114
  # The API interface for managing autoscaling policies in the
115
- # Google Cloud Dataproc API.
115
+ # Cloud Dataproc API.
116
116
  #
117
117
  # @param version [Symbol, String]
118
118
  # The major version of the service to be used. By default :v1
@@ -282,7 +282,7 @@ module Google
282
282
  module WorkflowTemplateService
283
283
  ##
284
284
  # The API interface for managing Workflow Templates in the
285
- # Cloud Dataproc API.
285
+ # Dataproc API.
286
286
  #
287
287
  # @param version [Symbol, String]
288
288
  # The major version of the service to be used. By default :v1
@@ -110,7 +110,7 @@ module Google
110
110
  module AutoscalingPolicyService
111
111
  ##
112
112
  # The API interface for managing autoscaling policies in the
113
- # Google Cloud Dataproc API.
113
+ # Cloud Dataproc API.
114
114
  #
115
115
  # @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
116
116
  # Provides the means for authenticating requests made by the client. This parameter can
@@ -304,7 +304,7 @@ module Google
304
304
  module WorkflowTemplateService
305
305
  ##
306
306
  # The API interface for managing Workflow Templates in the
307
- # Cloud Dataproc API.
307
+ # Dataproc API.
308
308
  #
309
309
  # @param credentials [Google::Auth::Credentials, String, Hash, GRPC::Core::Channel, GRPC::Core::ChannelCredentials, Proc]
310
310
  # Provides the means for authenticating requests made by the client. This parameter can
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/autoscaling_policies.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -27,7 +26,7 @@ module Google
27
26
  module V1beta2
28
27
  module AutoscalingPolicyService
29
28
  # The API interface for managing autoscaling policies in the
30
- # Google Cloud Dataproc API.
29
+ # Cloud Dataproc API.
31
30
  class Service
32
31
 
33
32
  include GRPC::GenericService
@@ -34,7 +34,7 @@ module Google
34
34
  module Dataproc
35
35
  module V1beta2
36
36
  # The API interface for managing autoscaling policies in the
37
- # Google Cloud Dataproc API.
37
+ # Cloud Dataproc API.
38
38
  #
39
39
  # @!attribute [r] autoscaling_policy_service_stub
40
40
  # @return [Google::Cloud::Dataproc::V1beta2::AutoscalingPolicyService::Stub]
@@ -239,7 +239,7 @@ module Google
239
239
  # Required. The ID of the Google Cloud Platform project that the cluster
240
240
  # belongs to.
241
241
  # @param region [String]
242
- # Required. The Cloud Dataproc region in which to handle the request.
242
+ # Required. The Dataproc region in which to handle the request.
243
243
  # @param cluster [Google::Cloud::Dataproc::V1beta2::Cluster | Hash]
244
244
  # Required. The cluster to create.
245
245
  # A hash of the same form as `Google::Cloud::Dataproc::V1beta2::Cluster`
@@ -334,7 +334,7 @@ module Google
334
334
  # Required. The ID of the Google Cloud Platform project the
335
335
  # cluster belongs to.
336
336
  # @param region [String]
337
- # Required. The Cloud Dataproc region in which to handle the request.
337
+ # Required. The Dataproc region in which to handle the request.
338
338
  # @param cluster_name [String]
339
339
  # Required. The cluster name.
340
340
  # @param cluster [Google::Cloud::Dataproc::V1beta2::Cluster | Hash]
@@ -410,7 +410,8 @@ module Google
410
410
  # interrupting jobs in progress. Timeout specifies how long to wait for jobs
411
411
  # in progress to finish before forcefully removing nodes (and potentially
412
412
  # interrupting jobs). Default timeout is 0 (for forceful decommission), and
413
- # the maximum allowed timeout is 1 day.
413
+ # the maximum allowed timeout is 1 day (see JSON representation of
414
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
414
415
  #
415
416
  # Only supported on Dataproc image versions 1.2 and higher.
416
417
  # A hash of the same form as `Google::Protobuf::Duration`
@@ -517,7 +518,7 @@ module Google
517
518
  # Required. The ID of the Google Cloud Platform project that the cluster
518
519
  # belongs to.
519
520
  # @param region [String]
520
- # Required. The Cloud Dataproc region in which to handle the request.
521
+ # Required. The Dataproc region in which to handle the request.
521
522
  # @param cluster_name [String]
522
523
  # Required. The cluster name.
523
524
  # @param cluster_uuid [String]
@@ -613,7 +614,7 @@ module Google
613
614
  # Required. The ID of the Google Cloud Platform project that the cluster
614
615
  # belongs to.
615
616
  # @param region [String]
616
- # Required. The Cloud Dataproc region in which to handle the request.
617
+ # Required. The Dataproc region in which to handle the request.
617
618
  # @param cluster_name [String]
618
619
  # Required. The cluster name.
619
620
  # @param options [Google::Gax::CallOptions]
@@ -660,7 +661,7 @@ module Google
660
661
  # Required. The ID of the Google Cloud Platform project that the cluster
661
662
  # belongs to.
662
663
  # @param region [String]
663
- # Required. The Cloud Dataproc region in which to handle the request.
664
+ # Required. The Dataproc region in which to handle the request.
664
665
  # @param filter [String]
665
666
  # Optional. A filter constraining the clusters to list. Filters are
666
667
  # case-sensitive and have the following syntax:
@@ -746,13 +747,13 @@ module Google
746
747
  # After the operation completes,
747
748
  # {Google::Longrunning::Operation#response Operation#response}
748
749
  # contains
749
- # [Empty](https://cloud.google.comgoogle.protobuf.Empty).
750
+ # {Google::Protobuf::Empty Empty}.
750
751
  #
751
752
  # @param project_id [String]
752
753
  # Required. The ID of the Google Cloud Platform project that the cluster
753
754
  # belongs to.
754
755
  # @param region [String]
755
- # Required. The Cloud Dataproc region in which to handle the request.
756
+ # Required. The Dataproc region in which to handle the request.
756
757
  # @param cluster_name [String]
757
758
  # Required. The cluster name.
758
759
  # @param options [Google::Gax::CallOptions]
@@ -7,7 +7,7 @@ require 'google/protobuf'
7
7
  require 'google/api/annotations_pb'
8
8
  require 'google/api/client_pb'
9
9
  require 'google/api/field_behavior_pb'
10
- require 'google/cloud/dataproc/v1beta2/operations_pb'
10
+ require 'google/api/resource_pb'
11
11
  require 'google/cloud/dataproc/v1beta2/shared_pb'
12
12
  require 'google/longrunning/operations_pb'
13
13
  require 'google/protobuf/duration_pb'
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/clusters.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -38,15 +37,15 @@ module Google
38
37
 
39
38
  # Creates a cluster in a project. The returned
40
39
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
41
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
40
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
42
41
  rpc :CreateCluster, CreateClusterRequest, Google::Longrunning::Operation
43
42
  # Updates a cluster in a project. The returned
44
43
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
45
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
44
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
46
45
  rpc :UpdateCluster, UpdateClusterRequest, Google::Longrunning::Operation
47
46
  # Deletes a cluster in a project. The returned
48
47
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
49
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
48
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
50
49
  rpc :DeleteCluster, DeleteClusterRequest, Google::Longrunning::Operation
51
50
  # Gets the resource representation for a cluster in a project.
52
51
  rpc :GetCluster, GetClusterRequest, Cluster
@@ -54,11 +53,11 @@ module Google
54
53
  rpc :ListClusters, ListClustersRequest, ListClustersResponse
55
54
  # Gets cluster diagnostic information. The returned
56
55
  # [Operation.metadata][google.longrunning.Operation.metadata] will be
57
- # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
56
+ # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
58
57
  # After the operation completes,
59
58
  # [Operation.response][google.longrunning.Operation.response]
60
59
  # contains
61
- # [Empty](google.protobuf.Empty).
60
+ # [Empty][google.protobuf.Empty].
62
61
  rpc :DiagnoseCluster, DiagnoseClusterRequest, Google::Longrunning::Operation
63
62
  end
64
63
 
@@ -28,7 +28,7 @@ module Google
28
28
  # unique. Names of deleted clusters can be reused.
29
29
  # @!attribute [rw] config
30
30
  # @return [Google::Cloud::Dataproc::V1beta2::ClusterConfig]
31
- # Required. The cluster config. Note that Cloud Dataproc may set
31
+ # Required. The cluster config. Note that Dataproc may set
32
32
  # default values, and values may change when clusters are updated.
33
33
  # @!attribute [rw] labels
34
34
  # @return [Hash{String => String}]
@@ -47,7 +47,7 @@ module Google
47
47
  # Output only. The previous cluster status.
48
48
  # @!attribute [rw] cluster_uuid
49
49
  # @return [String]
50
- # Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
50
+ # Output only. A cluster UUID (Unique Universal Identifier). Dataproc
51
51
  # generates this value when it creates the cluster.
52
52
  # @!attribute [rw] metrics
53
53
  # @return [Google::Cloud::Dataproc::V1beta2::ClusterMetrics]
@@ -60,14 +60,14 @@ module Google
60
60
  # The cluster config.
61
61
  # @!attribute [rw] config_bucket
62
62
  # @return [String]
63
- # Optional. A Google Cloud Storage bucket used to stage job
63
+ # Optional. A Cloud Storage bucket used to stage job
64
64
  # dependencies, config files, and job driver console output.
65
65
  # If you do not specify a staging bucket, Cloud
66
66
  # Dataproc will determine a Cloud Storage location (US,
67
- # ASIA, or EU) for your cluster's staging bucket according to the Google
67
+ # ASIA, or EU) for your cluster's staging bucket according to the
68
68
  # Compute Engine zone where your cluster is deployed, and then create
69
69
  # and manage this project-level, per-location bucket (see
70
- # [Cloud Dataproc staging
70
+ # [Dataproc staging
71
71
  # bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
72
72
  # @!attribute [rw] gce_cluster_config
73
73
  # @return [Google::Cloud::Dataproc::V1beta2::GceClusterConfig]
@@ -143,7 +143,7 @@ module Google
143
143
  # * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
144
144
  # * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
145
145
  #
146
- # Note that the policy must be in the same project and Cloud Dataproc region.
146
+ # Note that the policy must be in the same project and Dataproc region.
147
147
  class AutoscalingConfig; end
148
148
 
149
149
  # Encryption settings for the cluster.
@@ -159,7 +159,7 @@ module Google
159
159
  # @return [String]
160
160
  # Optional. The zone where the Compute Engine cluster will be located.
161
161
  # On a create request, it is required in the "global" region. If omitted
162
- # in a non-global Cloud Dataproc region, the service will pick a zone in the
162
+ # in a non-global Dataproc region, the service will pick a zone in the
163
163
  # corresponding Compute Engine region. On a get request, zone will always be
164
164
  # present.
165
165
  #
@@ -201,17 +201,17 @@ module Google
201
201
  # configured to be accessible without external IP addresses.
202
202
  # @!attribute [rw] service_account
203
203
  # @return [String]
204
- # Optional. The service account of the instances. Defaults to the default
205
- # Compute Engine service account. Custom service accounts need
206
- # permissions equivalent to the following IAM roles:
207
- #
208
- # * roles/logging.logWriter
209
- # * roles/storage.objectAdmin
210
- #
211
- # (see
212
- # https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
213
- # for more information).
214
- # Example: `[account_id]@[project_id].iam.gserviceaccount.com`
204
+ # Optional. The [Dataproc service
205
+ # account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
206
+ # (also see [VM Data Plane
207
+ # identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
208
+ # used by Dataproc cluster VM instances to access Google Cloud Platform
209
+ # services.
210
+ #
211
+ # If not specified, the
212
+ # [Compute Engine default service
213
+ # account](/compute/docs/access/service-accounts#default_service_account)
214
+ # is used.
215
215
  # @!attribute [rw] service_account_scopes
216
216
  # @return [Array<String>]
217
217
  # Optional. The URIs of service account scopes to be included in
@@ -250,7 +250,7 @@ module Google
250
250
  # For master instance groups, must be set to 1.
251
251
  # @!attribute [rw] instance_names
252
252
  # @return [Array<String>]
253
- # Output only. The list of instance names. Cloud Dataproc derives the names
253
+ # Output only. The list of instance names. Dataproc derives the names
254
254
  # from `cluster_name`, `num_instances`, and the instance group.
255
255
  # @!attribute [rw] image_uri
256
256
  # @return [String]
@@ -267,7 +267,7 @@ module Google
267
267
  # * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2`
268
268
  # * `n1-standard-2`
269
269
  #
270
- # **Auto Zone Exception**: If you are using the Cloud Dataproc
270
+ # **Auto Zone Exception**: If you are using the Dataproc
271
271
  # [Auto Zone
272
272
  # Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
273
273
  # feature, you must use the short name of the machine type
@@ -291,7 +291,7 @@ module Google
291
291
  # @!attribute [rw] min_cpu_platform
292
292
  # @return [String]
293
293
  # Specifies the minimum cpu platform for the Instance Group.
294
- # See [Cloud Dataproc&rarr;Minimum CPU Platform]
294
+ # See [Dataproc&rarr;Minimum CPU Platform]
295
295
  # (/dataproc/docs/concepts/compute/dataproc-min-cpu).
296
296
  class InstanceGroupConfig; end
297
297
 
@@ -319,7 +319,7 @@ module Google
319
319
  # * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80`
320
320
  # * `nvidia-tesla-k80`
321
321
  #
322
- # **Auto Zone Exception**: If you are using the Cloud Dataproc
322
+ # **Auto Zone Exception**: If you are using the Dataproc
323
323
  # [Auto Zone
324
324
  # Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
325
325
  # feature, you must use the short name of the accelerator type
@@ -351,25 +351,27 @@ module Google
351
351
  # Specifies the cluster auto-delete schedule configuration.
352
352
  # @!attribute [rw] idle_delete_ttl
353
353
  # @return [Google::Protobuf::Duration]
354
- # Optional. The duration to keep the cluster alive while idling.
355
- # Passing this threshold will cause the cluster to be
356
- # deleted. Valid range: **[10m, 14d]**.
357
- #
358
- # Example: **"10m"**, the minimum value, to delete the
359
- # cluster when it has had no jobs running for 10 minutes.
354
+ # Optional. The duration to keep the cluster alive while idling (when no jobs
355
+ # are running). Passing this threshold will cause the cluster to be
356
+ # deleted. Minimum value is 10 minutes; maximum value is 14 days (see JSON
357
+ # representation of
358
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json).
360
359
  # @!attribute [rw] auto_delete_time
361
360
  # @return [Google::Protobuf::Timestamp]
362
- # Optional. The time when cluster will be auto-deleted.
361
+ # Optional. The time when cluster will be auto-deleted. (see JSON representation of
362
+ # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
363
363
  # @!attribute [rw] auto_delete_ttl
364
364
  # @return [Google::Protobuf::Duration]
365
365
  # Optional. The lifetime duration of cluster. The cluster will be
366
- # auto-deleted at the end of this period. Valid range: **[10m, 14d]**.
367
- #
368
- # Example: **"1d"**, to delete the cluster 1 day after its creation..
366
+ # auto-deleted at the end of this period. Minimum value is 10 minutes;
367
+ # maximum value is 14 days (see JSON representation of
368
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
369
369
  # @!attribute [rw] idle_start_time
370
370
  # @return [Google::Protobuf::Timestamp]
371
371
  # Output only. The time when cluster became idle (most recent job finished)
372
- # and became eligible for deletion due to idleness.
372
+ # and became eligible for deletion due to idleness (see JSON representation
373
+ # of
374
+ # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
373
375
  class LifecycleConfig; end
374
376
 
375
377
  # Security related configuration, including encryption, Kerberos, etc.
@@ -455,7 +457,10 @@ module Google
455
457
  # @!attribute [rw] execution_timeout
456
458
  # @return [Google::Protobuf::Duration]
457
459
  # Optional. Amount of time executable has to complete. Default is
458
- # 10 minutes. Cluster creation fails with an explanatory error message (the
460
+ # 10 minutes (see JSON representation of
461
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
462
+ #
463
+ # Cluster creation fails with an explanatory error message (the
459
464
  # name of the executable that caused the error and the exceeded timeout
460
465
  # period) if the executable is not completed at end of the timeout period.
461
466
  class NodeInitializationAction; end
@@ -469,7 +474,8 @@ module Google
469
474
  # Output only. Optional details of cluster's state.
470
475
  # @!attribute [rw] state_start_time
471
476
  # @return [Google::Protobuf::Timestamp]
472
- # Output only. Time when this state was entered.
477
+ # Output only. Time when this state was entered (see JSON representation of
478
+ # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
473
479
  # @!attribute [rw] substate
474
480
  # @return [Google::Cloud::Dataproc::V1beta2::ClusterStatus::Substate]
475
481
  # Output only. Additional state information that includes
@@ -509,7 +515,7 @@ module Google
509
515
  UNHEALTHY = 1
510
516
 
511
517
  # The agent-reported status is out of date (may occur if
512
- # Cloud Dataproc loses communication with Agent).
518
+ # Dataproc loses communication with Agent).
513
519
  #
514
520
  # Applies to RUNNING state.
515
521
  STALE_STATUS = 2
@@ -520,7 +526,7 @@ module Google
520
526
  # @!attribute [rw] image_version
521
527
  # @return [String]
522
528
  # Optional. The version of software inside the cluster. It must be one of the
523
- # supported [Cloud Dataproc
529
+ # supported [Dataproc
524
530
  # Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
525
531
  # such as "1.2" (including a subminor version, such as "1.2.29"), or the
526
532
  # ["preview"
@@ -570,7 +576,7 @@ module Google
570
576
  # belongs to.
571
577
  # @!attribute [rw] region
572
578
  # @return [String]
573
- # Required. The Cloud Dataproc region in which to handle the request.
579
+ # Required. The Dataproc region in which to handle the request.
574
580
  # @!attribute [rw] cluster
575
581
  # @return [Google::Cloud::Dataproc::V1beta2::Cluster]
576
582
  # Required. The cluster to create.
@@ -596,7 +602,7 @@ module Google
596
602
  # cluster belongs to.
597
603
  # @!attribute [rw] region
598
604
  # @return [String]
599
- # Required. The Cloud Dataproc region in which to handle the request.
605
+ # Required. The Dataproc region in which to handle the request.
600
606
  # @!attribute [rw] cluster_name
601
607
  # @return [String]
602
608
  # Required. The cluster name.
@@ -610,7 +616,8 @@ module Google
610
616
  # interrupting jobs in progress. Timeout specifies how long to wait for jobs
611
617
  # in progress to finish before forcefully removing nodes (and potentially
612
618
  # interrupting jobs). Default timeout is 0 (for forceful decommission), and
613
- # the maximum allowed timeout is 1 day.
619
+ # the maximum allowed timeout is 1 day (see JSON representation of
620
+ # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
614
621
  #
615
622
  # Only supported on Dataproc image versions 1.2 and higher.
616
623
  # @!attribute [rw] update_mask
@@ -697,7 +704,7 @@ module Google
697
704
  # belongs to.
698
705
  # @!attribute [rw] region
699
706
  # @return [String]
700
- # Required. The Cloud Dataproc region in which to handle the request.
707
+ # Required. The Dataproc region in which to handle the request.
701
708
  # @!attribute [rw] cluster_name
702
709
  # @return [String]
703
710
  # Required. The cluster name.
@@ -727,7 +734,7 @@ module Google
727
734
  # belongs to.
728
735
  # @!attribute [rw] region
729
736
  # @return [String]
730
- # Required. The Cloud Dataproc region in which to handle the request.
737
+ # Required. The Dataproc region in which to handle the request.
731
738
  # @!attribute [rw] cluster_name
732
739
  # @return [String]
733
740
  # Required. The cluster name.
@@ -740,7 +747,7 @@ module Google
740
747
  # belongs to.
741
748
  # @!attribute [rw] region
742
749
  # @return [String]
743
- # Required. The Cloud Dataproc region in which to handle the request.
750
+ # Required. The Dataproc region in which to handle the request.
744
751
  # @!attribute [rw] filter
745
752
  # @return [String]
746
753
  # Optional. A filter constraining the clusters to list. Filters are
@@ -788,7 +795,7 @@ module Google
788
795
  # belongs to.
789
796
  # @!attribute [rw] region
790
797
  # @return [String]
791
- # Required. The Cloud Dataproc region in which to handle the request.
798
+ # Required. The Dataproc region in which to handle the request.
792
799
  # @!attribute [rw] cluster_name
793
800
  # @return [String]
794
801
  # Required. The cluster name.
@@ -58,7 +58,7 @@ module Google
58
58
  end
59
59
  end
60
60
 
61
- # A Cloud Dataproc job for running
61
+ # A Dataproc job for running
62
62
  # [Apache Hadoop
63
63
  # MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
64
64
  # jobs on [Apache Hadoop
@@ -97,7 +97,7 @@ module Google
97
97
  # @!attribute [rw] properties
98
98
  # @return [Hash{String => String}]
99
99
  # Optional. A mapping of property names to values, used to configure Hadoop.
100
- # Properties that conflict with values set by the Cloud Dataproc API may be
100
+ # Properties that conflict with values set by the Dataproc API may be
101
101
  # overwritten. Can include properties set in /etc/hadoop/conf/*-site and
102
102
  # classes in user code.
103
103
  # @!attribute [rw] logging_config
@@ -105,7 +105,7 @@ module Google
105
105
  # Optional. The runtime log config for job execution.
106
106
  class HadoopJob; end
107
107
 
108
- # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/)
108
+ # A Dataproc job for running [Apache Spark](http://spark.apache.org/)
109
109
  # applications on YARN.
110
110
  # The specification of the main method to call to drive the job.
111
111
  # Specify either the jar file that contains the main class or the main class
@@ -140,7 +140,7 @@ module Google
140
140
  # @!attribute [rw] properties
141
141
  # @return [Hash{String => String}]
142
142
  # Optional. A mapping of property names to values, used to configure Spark.
143
- # Properties that conflict with values set by the Cloud Dataproc API may be
143
+ # Properties that conflict with values set by the Dataproc API may be
144
144
  # overwritten. Can include properties set in
145
145
  # /etc/spark/conf/spark-defaults.conf and classes in user code.
146
146
  # @!attribute [rw] logging_config
@@ -148,7 +148,7 @@ module Google
148
148
  # Optional. The runtime log config for job execution.
149
149
  class SparkJob; end
150
150
 
151
- # A Cloud Dataproc job for running
151
+ # A Dataproc job for running
152
152
  # [Apache
153
153
  # PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
154
154
  # applications on YARN.
@@ -180,7 +180,7 @@ module Google
180
180
  # @!attribute [rw] properties
181
181
  # @return [Hash{String => String}]
182
182
  # Optional. A mapping of property names to values, used to configure PySpark.
183
- # Properties that conflict with values set by the Cloud Dataproc API may be
183
+ # Properties that conflict with values set by the Dataproc API may be
184
184
  # overwritten. Can include properties set in
185
185
  # /etc/spark/conf/spark-defaults.conf and classes in user code.
186
186
  # @!attribute [rw] logging_config
@@ -207,7 +207,7 @@ module Google
207
207
  # }
208
208
  class QueryList; end
209
209
 
210
- # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/)
210
+ # A Dataproc job for running [Apache Hive](https://hive.apache.org/)
211
211
  # queries on YARN.
212
212
  # @!attribute [rw] query_file_uri
213
213
  # @return [String]
@@ -227,7 +227,7 @@ module Google
227
227
  # @!attribute [rw] properties
228
228
  # @return [Hash{String => String}]
229
229
  # Optional. A mapping of property names and values, used to configure Hive.
230
- # Properties that conflict with values set by the Cloud Dataproc API may be
230
+ # Properties that conflict with values set by the Dataproc API may be
231
231
  # overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
232
232
  # /etc/hive/conf/hive-site.xml, and classes in user code.
233
233
  # @!attribute [rw] jar_file_uris
@@ -237,7 +237,7 @@ module Google
237
237
  # and UDFs.
238
238
  class HiveJob; end
239
239
 
240
- # A Cloud Dataproc job for running [Apache Spark
240
+ # A Dataproc job for running [Apache Spark
241
241
  # SQL](http://spark.apache.org/sql/) queries.
242
242
  # @!attribute [rw] query_file_uri
243
243
  # @return [String]
@@ -253,7 +253,7 @@ module Google
253
253
  # @return [Hash{String => String}]
254
254
  # Optional. A mapping of property names to values, used to configure
255
255
  # Spark SQL's SparkConf. Properties that conflict with values set by the
256
- # Cloud Dataproc API may be overwritten.
256
+ # Dataproc API may be overwritten.
257
257
  # @!attribute [rw] jar_file_uris
258
258
  # @return [Array<String>]
259
259
  # Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
@@ -262,7 +262,7 @@ module Google
262
262
  # Optional. The runtime log config for job execution.
263
263
  class SparkSqlJob; end
264
264
 
265
- # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/)
265
+ # A Dataproc job for running [Apache Pig](https://pig.apache.org/)
266
266
  # queries on YARN.
267
267
  # @!attribute [rw] query_file_uri
268
268
  # @return [String]
@@ -282,7 +282,7 @@ module Google
282
282
  # @!attribute [rw] properties
283
283
  # @return [Hash{String => String}]
284
284
  # Optional. A mapping of property names to values, used to configure Pig.
285
- # Properties that conflict with values set by the Cloud Dataproc API may be
285
+ # Properties that conflict with values set by the Dataproc API may be
286
286
  # overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
287
287
  # /etc/pig/conf/pig.properties, and classes in user code.
288
288
  # @!attribute [rw] jar_file_uris
@@ -294,7 +294,7 @@ module Google
294
294
  # Optional. The runtime log config for job execution.
295
295
  class PigJob; end
296
296
 
297
- # A Cloud Dataproc job for running
297
+ # A Dataproc job for running
298
298
  # [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html)
299
299
  # applications on YARN.
300
300
  # @!attribute [rw] main_r_file_uri
@@ -318,7 +318,7 @@ module Google
318
318
  # @!attribute [rw] properties
319
319
  # @return [Hash{String => String}]
320
320
  # Optional. A mapping of property names to values, used to configure SparkR.
321
- # Properties that conflict with values set by the Cloud Dataproc API may be
321
+ # Properties that conflict with values set by the Dataproc API may be
322
322
  # overwritten. Can include properties set in
323
323
  # /etc/spark/conf/spark-defaults.conf and classes in user code.
324
324
  # @!attribute [rw] logging_config
@@ -326,23 +326,23 @@ module Google
326
326
  # Optional. The runtime log config for job execution.
327
327
  class SparkRJob; end
328
328
 
329
- # Cloud Dataproc job config.
329
+ # Dataproc job config.
330
330
  # @!attribute [rw] cluster_name
331
331
  # @return [String]
332
332
  # Required. The name of the cluster where the job will be submitted.
333
333
  # @!attribute [rw] cluster_uuid
334
334
  # @return [String]
335
- # Output only. A cluster UUID generated by the Cloud Dataproc service when
335
+ # Output only. A cluster UUID generated by the Dataproc service when
336
336
  # the job is submitted.
337
337
  class JobPlacement; end
338
338
 
339
- # Cloud Dataproc job status.
339
+ # Dataproc job status.
340
340
  # @!attribute [rw] state
341
341
  # @return [Google::Cloud::Dataproc::V1beta2::JobStatus::State]
342
342
  # Output only. A state message specifying the overall job state.
343
343
  # @!attribute [rw] details
344
344
  # @return [String]
345
- # Output only. Optional job state details, such as an error
345
+ # Output only. Optional Job state details, such as an error
346
346
  # description if the state is <code>ERROR</code>.
347
347
  # @!attribute [rw] state_start_time
348
348
  # @return [Google::Protobuf::Timestamp]
@@ -408,7 +408,7 @@ module Google
408
408
  QUEUED = 2
409
409
 
410
410
  # The agent-reported status is out of date, which may be caused by a
411
- # loss of communication between the agent and Cloud Dataproc. If the
411
+ # loss of communication between the agent and Dataproc. If the
412
412
  # agent does not send a timely update, the job will fail.
413
413
  #
414
414
  # Applies to RUNNING state.
@@ -424,7 +424,6 @@ module Google
424
424
  # @!attribute [rw] job_id
425
425
  # @return [String]
426
426
  # Optional. The job ID, which must be unique within the project.
427
- #
428
427
  # The ID must contain only letters (a-z, A-Z), numbers (0-9),
429
428
  # underscores (_), or hyphens (-). The maximum length is 100 characters.
430
429
  #
@@ -447,7 +446,7 @@ module Google
447
446
  # Output only. The numerical progress of the application, from 1 to 100.
448
447
  # @!attribute [rw] tracking_url
449
448
  # @return [String]
450
- # Optional. Output only. The HTTP URL of the ApplicationMaster, HistoryServer, or
449
+ # Output only. The HTTP URL of the ApplicationMaster, HistoryServer, or
451
450
  # TimelineServer that provides application-specific information. The URL uses
452
451
  # the internal hostname, and requires a proxy server for resolution and,
453
452
  # possibly, access.
@@ -484,7 +483,7 @@ module Google
484
483
  end
485
484
  end
486
485
 
487
- # A Cloud Dataproc job resource.
486
+ # A Dataproc job resource.
488
487
  # @!attribute [rw] reference
489
488
  # @return [Google::Cloud::Dataproc::V1beta2::JobReference]
490
489
  # Optional. The fully qualified reference to the job, which can be used to
@@ -497,25 +496,18 @@ module Google
497
496
  # run the job.
498
497
  # @!attribute [rw] hadoop_job
499
498
  # @return [Google::Cloud::Dataproc::V1beta2::HadoopJob]
500
- # Job is a Hadoop job.
501
499
  # @!attribute [rw] spark_job
502
500
  # @return [Google::Cloud::Dataproc::V1beta2::SparkJob]
503
- # Job is a Spark job.
504
501
  # @!attribute [rw] pyspark_job
505
502
  # @return [Google::Cloud::Dataproc::V1beta2::PySparkJob]
506
- # Job is a Pyspark job.
507
503
  # @!attribute [rw] hive_job
508
504
  # @return [Google::Cloud::Dataproc::V1beta2::HiveJob]
509
- # Job is a Hive job.
510
505
  # @!attribute [rw] pig_job
511
506
  # @return [Google::Cloud::Dataproc::V1beta2::PigJob]
512
- # Job is a Pig job.
513
507
  # @!attribute [rw] spark_r_job
514
508
  # @return [Google::Cloud::Dataproc::V1beta2::SparkRJob]
515
- # Job is a SparkR job.
516
509
  # @!attribute [rw] spark_sql_job
517
510
  # @return [Google::Cloud::Dataproc::V1beta2::SparkSqlJob]
518
- # Job is a SparkSql job.
519
511
  # @!attribute [rw] status
520
512
  # @return [Google::Cloud::Dataproc::V1beta2::JobStatus]
521
513
  # Output only. The job status. Additional application-specific
@@ -582,7 +574,7 @@ module Google
582
574
  # belongs to.
583
575
  # @!attribute [rw] region
584
576
  # @return [String]
585
- # Required. The Cloud Dataproc region in which to handle the request.
577
+ # Required. The Dataproc region in which to handle the request.
586
578
  # @!attribute [rw] job
587
579
  # @return [Google::Cloud::Dataproc::V1beta2::Job]
588
580
  # Required. The job resource.
@@ -608,7 +600,7 @@ module Google
608
600
  # belongs to.
609
601
  # @!attribute [rw] region
610
602
  # @return [String]
611
- # Required. The Cloud Dataproc region in which to handle the request.
603
+ # Required. The Dataproc region in which to handle the request.
612
604
  # @!attribute [rw] job_id
613
605
  # @return [String]
614
606
  # Required. The job ID.
@@ -621,7 +613,7 @@ module Google
621
613
  # belongs to.
622
614
  # @!attribute [rw] region
623
615
  # @return [String]
624
- # Required. The Cloud Dataproc region in which to handle the request.
616
+ # Required. The Dataproc region in which to handle the request.
625
617
  # @!attribute [rw] page_size
626
618
  # @return [Integer]
627
619
  # Optional. The number of results to return in each response.
@@ -677,7 +669,7 @@ module Google
677
669
  # belongs to.
678
670
  # @!attribute [rw] region
679
671
  # @return [String]
680
- # Required. The Cloud Dataproc region in which to handle the request.
672
+ # Required. The Dataproc region in which to handle the request.
681
673
  # @!attribute [rw] job_id
682
674
  # @return [String]
683
675
  # Required. The job ID.
@@ -712,7 +704,7 @@ module Google
712
704
  # belongs to.
713
705
  # @!attribute [rw] region
714
706
  # @return [String]
715
- # Required. The Cloud Dataproc region in which to handle the request.
707
+ # Required. The Dataproc region in which to handle the request.
716
708
  # @!attribute [rw] job_id
717
709
  # @return [String]
718
710
  # Required. The job ID.
@@ -725,7 +717,7 @@ module Google
725
717
  # belongs to.
726
718
  # @!attribute [rw] region
727
719
  # @return [String]
728
- # Required. The Cloud Dataproc region in which to handle the request.
720
+ # Required. The Dataproc region in which to handle the request.
729
721
  # @!attribute [rw] job_id
730
722
  # @return [String]
731
723
  # Required. The job ID.
@@ -17,7 +17,7 @@ module Google
17
17
  module Cloud
18
18
  module Dataproc
19
19
  module V1beta2
20
- # A Cloud Dataproc workflow template resource.
20
+ # A Dataproc workflow template resource.
21
21
  # @!attribute [rw] id
22
22
  # @return [String]
23
23
  # Required. The template id.
@@ -153,22 +153,16 @@ module Google
153
153
  # or hyphen. Must consist of between 3 and 50 characters.
154
154
  # @!attribute [rw] hadoop_job
155
155
  # @return [Google::Cloud::Dataproc::V1beta2::HadoopJob]
156
- # Job is a Hadoop job.
157
156
  # @!attribute [rw] spark_job
158
157
  # @return [Google::Cloud::Dataproc::V1beta2::SparkJob]
159
- # Job is a Spark job.
160
158
  # @!attribute [rw] pyspark_job
161
159
  # @return [Google::Cloud::Dataproc::V1beta2::PySparkJob]
162
- # Job is a Pyspark job.
163
160
  # @!attribute [rw] hive_job
164
161
  # @return [Google::Cloud::Dataproc::V1beta2::HiveJob]
165
- # Job is a Hive job.
166
162
  # @!attribute [rw] pig_job
167
163
  # @return [Google::Cloud::Dataproc::V1beta2::PigJob]
168
- # Job is a Pig job.
169
164
  # @!attribute [rw] spark_sql_job
170
165
  # @return [Google::Cloud::Dataproc::V1beta2::SparkSqlJob]
171
- # Job is a SparkSql job.
172
166
  # @!attribute [rw] labels
173
167
  # @return [Hash{String => String}]
174
168
  # Optional. The labels to associate with this job.
@@ -286,7 +280,7 @@ module Google
286
280
  # Required. List of allowed values for the parameter.
287
281
  class ValueValidation; end
288
282
 
289
- # A Cloud Dataproc workflow template resource.
283
+ # A Dataproc workflow template resource.
290
284
  # @!attribute [rw] template
291
285
  # @return [String]
292
286
  # Output only. The resource name of the workflow template as described
@@ -217,7 +217,7 @@ module Google
217
217
  # Required. The ID of the Google Cloud Platform project that the job
218
218
  # belongs to.
219
219
  # @param region [String]
220
- # Required. The Cloud Dataproc region in which to handle the request.
220
+ # Required. The Dataproc region in which to handle the request.
221
221
  # @param job [Google::Cloud::Dataproc::V1beta2::Job | Hash]
222
222
  # Required. The job resource.
223
223
  # A hash of the same form as `Google::Cloud::Dataproc::V1beta2::Job`
@@ -280,7 +280,7 @@ module Google
280
280
  # Required. The ID of the Google Cloud Platform project that the job
281
281
  # belongs to.
282
282
  # @param region [String]
283
- # Required. The Cloud Dataproc region in which to handle the request.
283
+ # Required. The Dataproc region in which to handle the request.
284
284
  # @param job_id [String]
285
285
  # Required. The job ID.
286
286
  # @param options [Google::Gax::CallOptions]
@@ -327,7 +327,7 @@ module Google
327
327
  # Required. The ID of the Google Cloud Platform project that the job
328
328
  # belongs to.
329
329
  # @param region [String]
330
- # Required. The Cloud Dataproc region in which to handle the request.
330
+ # Required. The Dataproc region in which to handle the request.
331
331
  # @param page_size [Integer]
332
332
  # The maximum number of resources contained in the underlying API
333
333
  # response. If page streaming is performed per-resource, this
@@ -420,7 +420,7 @@ module Google
420
420
  # Required. The ID of the Google Cloud Platform project that the job
421
421
  # belongs to.
422
422
  # @param region [String]
423
- # Required. The Cloud Dataproc region in which to handle the request.
423
+ # Required. The Dataproc region in which to handle the request.
424
424
  # @param job_id [String]
425
425
  # Required. The job ID.
426
426
  # @param job [Google::Cloud::Dataproc::V1beta2::Job | Hash]
@@ -494,7 +494,7 @@ module Google
494
494
  # Required. The ID of the Google Cloud Platform project that the job
495
495
  # belongs to.
496
496
  # @param region [String]
497
- # Required. The Cloud Dataproc region in which to handle the request.
497
+ # Required. The Dataproc region in which to handle the request.
498
498
  # @param job_id [String]
499
499
  # Required. The job ID.
500
500
  # @param options [Google::Gax::CallOptions]
@@ -542,7 +542,7 @@ module Google
542
542
  # Required. The ID of the Google Cloud Platform project that the job
543
543
  # belongs to.
544
544
  # @param region [String]
545
- # Required. The Cloud Dataproc region in which to handle the request.
545
+ # Required. The Dataproc region in which to handle the request.
546
546
  # @param job_id [String]
547
547
  # Required. The job ID.
548
548
  # @param options [Google::Gax::CallOptions]
@@ -7,6 +7,7 @@ require 'google/protobuf'
7
7
  require 'google/api/annotations_pb'
8
8
  require 'google/api/client_pb'
9
9
  require 'google/api/field_behavior_pb'
10
+ require 'google/longrunning/operations_pb'
10
11
  require 'google/protobuf/empty_pb'
11
12
  require 'google/protobuf/field_mask_pb'
12
13
  require 'google/protobuf/timestamp_pb'
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/jobs.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -45,9 +44,9 @@ module Google
45
44
  rpc :UpdateJob, UpdateJobRequest, Job
46
45
  # Starts a job cancellation request. To access the job resource
47
46
  # after cancellation, call
48
- # [regions/\\{region}/jobs.list](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/list)
47
+ # [regions/\\{region}/jobs.list](https://cloud.google.com/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/list)
49
48
  # or
50
- # [regions/\\{region}/jobs.get](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get).
49
+ # [regions/\\{region}/jobs.get](https://cloud.google.com/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get).
51
50
  rpc :CancelJob, CancelJobRequest, Job
52
51
  # Deletes the job from the project. If the job is active, the delete fails,
53
52
  # and the response returns `FAILED_PRECONDITION`.
@@ -36,7 +36,7 @@ module Google
36
36
  module Dataproc
37
37
  module V1beta2
38
38
  # The API interface for managing Workflow Templates in the
39
- # Cloud Dataproc API.
39
+ # Dataproc API.
40
40
  #
41
41
  # @!attribute [r] workflow_template_service_stub
42
42
  # @return [Google::Cloud::Dataproc::V1beta2::WorkflowTemplateService::Stub]
@@ -404,7 +404,7 @@ module Google
404
404
  # The {Google::Longrunning::Operation#metadata Operation#metadata} will be
405
405
  # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
406
406
  # Also see [Using
407
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
407
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
408
408
  #
409
409
  # On successful completion,
410
410
  # {Google::Longrunning::Operation#response Operation#response} will be
@@ -526,7 +526,7 @@ module Google
526
526
  # The {Google::Longrunning::Operation#metadata Operation#metadata} will be
527
527
  # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
528
528
  # Also see [Using
529
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
529
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
530
530
  #
531
531
  # On successful completion,
532
532
  # {Google::Longrunning::Operation#response Operation#response} will be
@@ -1,7 +1,7 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # Source: google/cloud/dataproc/v1beta2/workflow_templates.proto for package 'google.cloud.dataproc.v1beta2'
3
3
  # Original file comments:
4
- # Copyright 2019 Google LLC.
4
+ # Copyright 2020 Google LLC
5
5
  #
6
6
  # Licensed under the Apache License, Version 2.0 (the "License");
7
7
  # you may not use this file except in compliance with the License.
@@ -15,7 +15,6 @@
15
15
  # See the License for the specific language governing permissions and
16
16
  # limitations under the License.
17
17
  #
18
- #
19
18
 
20
19
 
21
20
  require 'grpc'
@@ -27,7 +26,7 @@ module Google
27
26
  module V1beta2
28
27
  module WorkflowTemplateService
29
28
  # The API interface for managing Workflow Templates in the
30
- # Cloud Dataproc API.
29
+ # Dataproc API.
31
30
  class Service
32
31
 
33
32
  include GRPC::GenericService
@@ -56,9 +55,9 @@ module Google
56
55
  # clusters to be deleted.
57
56
  #
58
57
  # The [Operation.metadata][google.longrunning.Operation.metadata] will be
59
- # [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
58
+ # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
60
59
  # Also see [Using
61
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
60
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
62
61
  #
63
62
  # On successful completion,
64
63
  # [Operation.response][google.longrunning.Operation.response] will be
@@ -81,9 +80,9 @@ module Google
81
80
  # clusters to be deleted.
82
81
  #
83
82
  # The [Operation.metadata][google.longrunning.Operation.metadata] will be
84
- # [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
83
+ # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
85
84
  # Also see [Using
86
- # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
85
+ # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
87
86
  #
88
87
  # On successful completion,
89
88
  # [Operation.response][google.longrunning.Operation.response] will be
@@ -16,7 +16,7 @@
16
16
  module Google
17
17
  module Cloud
18
18
  module Dataproc
19
- VERSION = "0.9.0".freeze
19
+ VERSION = "0.9.1".freeze
20
20
  end
21
21
  end
22
22
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-dataproc
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.9.0
4
+ version: 0.9.1
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2020-02-24 00:00:00.000000000 Z
11
+ date: 2020-03-02 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: google-gax