google-cloud-dataproc-v1 0.8.0 → 0.11.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 75b10ffcc3efa847d380e239d97ac92298df724a70e6f76e161a4357b31e2464
4
- data.tar.gz: ff2d2916e09c5503d8cb4ee6f74058770b64254ecb962aa4e71f453f4bae6401
3
+ metadata.gz: 8e5fe6da17d8ccadfc7dbd12eaac80d02427f71070bbb1b1225a59dd838d9d7c
4
+ data.tar.gz: d21e87aaa4b34a8f0ed28cd1600c82a06abef2ed9ef44d7f5637372c96c2bbe5
5
5
  SHA512:
6
- metadata.gz: d2f533ae3bc23fa57971f8fe1faf5ba3abf1bec75d25935cb3e3de5229e9b11bcf4e9dd8d2c7c887f6a0fb54ccfdc21764aea94e5f077c6cf5fa5738648a0b6a
7
- data.tar.gz: 73a70eb049fc046aa5ece71bf473adad5df91872e7b83172c1439c1766901b5af6df17fff54dd561fa48e68c70cfb83282cd417a44a3a127058a36adaf547357
6
+ metadata.gz: 61fab34408f18b1a613a1a73cb09ad44de583aead63f99f4e6651853efed34461bef7845942dbae126fea638f852b634d5272e6d56be18cadb87ea2a5f34e809
7
+ data.tar.gz: 552a549a52d4e57e98854b025502ab3e6ed2a7d98c7be9b76773644d33f1ac521172747104da44313f7703e46f2b4ca3cda8dd80649da6ee4f175eec6fe51493
data/README.md CHANGED
@@ -37,7 +37,7 @@ request = ::Google::Cloud::Dataproc::V1::CreateAutoscalingPolicyRequest.new # (r
37
37
  response = client.create_autoscaling_policy request
38
38
  ```
39
39
 
40
- View the [Client Library Documentation](https://googleapis.dev/ruby/google-cloud-dataproc-v1/latest)
40
+ View the [Client Library Documentation](https://cloud.google.com/ruby/docs/reference/google-cloud-dataproc-v1/latest)
41
41
  for class and method documentation.
42
42
 
43
43
  See also the [Product Documentation](https://cloud.google.com/dataproc)
@@ -69,6 +69,11 @@ module GRPC
69
69
  end
70
70
  ```
71
71
 
72
+
73
+ ## Google Cloud Samples
74
+
75
+ To browse ready to use code samples check [Google Cloud Samples](https://cloud.google.com/docs/samples).
76
+
72
77
  ## Supported Ruby Versions
73
78
 
74
79
  This library is supported on Ruby 2.5+.
@@ -1,13 +1,14 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/autoscaling_policies.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/annotations_pb'
5
7
  require 'google/api/client_pb'
6
8
  require 'google/api/field_behavior_pb'
7
9
  require 'google/api/resource_pb'
8
10
  require 'google/protobuf/duration_pb'
9
11
  require 'google/protobuf/empty_pb'
10
- require 'google/protobuf'
11
12
 
12
13
  Google::Protobuf::DescriptorPool.generated_pool.build do
13
14
  add_file("google/cloud/dataproc/v1/autoscaling_policies.proto", :syntax => :proto3) do
@@ -22,8 +23,10 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
22
23
  end
23
24
  end
24
25
  add_message "google.cloud.dataproc.v1.BasicAutoscalingAlgorithm" do
25
- optional :yarn_config, :message, 1, "google.cloud.dataproc.v1.BasicYarnAutoscalingConfig"
26
26
  optional :cooldown_period, :message, 2, "google.protobuf.Duration"
27
+ oneof :config do
28
+ optional :yarn_config, :message, 1, "google.cloud.dataproc.v1.BasicYarnAutoscalingConfig"
29
+ end
27
30
  end
28
31
  add_message "google.cloud.dataproc.v1.BasicYarnAutoscalingConfig" do
29
32
  optional :graceful_decommission_timeout, :message, 5, "google.protobuf.Duration"
@@ -1,6 +1,8 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/batches.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/annotations_pb'
5
7
  require 'google/api/client_pb'
6
8
  require 'google/api/field_behavior_pb'
@@ -9,7 +11,6 @@ require 'google/cloud/dataproc/v1/shared_pb'
9
11
  require 'google/longrunning/operations_pb'
10
12
  require 'google/protobuf/empty_pb'
11
13
  require 'google/protobuf/timestamp_pb'
12
- require 'google/protobuf'
13
14
 
14
15
  Google::Protobuf::DescriptorPool.generated_pool.build do
15
16
  add_file("google/cloud/dataproc/v1/batches.proto", :syntax => :proto3) do
@@ -24,25 +24,6 @@ module Google
24
24
  module ClusterController
25
25
  # Path helper methods for the ClusterController API.
26
26
  module Paths
27
- ##
28
- # Create a fully-qualified Cluster resource string.
29
- #
30
- # The resource will be in the following format:
31
- #
32
- # `projects/{project}/locations/{location}/clusters/{cluster}`
33
- #
34
- # @param project [String]
35
- # @param location [String]
36
- # @param cluster [String]
37
- #
38
- # @return [::String]
39
- def cluster_path project:, location:, cluster:
40
- raise ::ArgumentError, "project cannot contain /" if project.to_s.include? "/"
41
- raise ::ArgumentError, "location cannot contain /" if location.to_s.include? "/"
42
-
43
- "projects/#{project}/locations/#{location}/clusters/#{cluster}"
44
- end
45
-
46
27
  ##
47
28
  # Create a fully-qualified Service resource string.
48
29
  #
@@ -1,6 +1,8 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/clusters.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/annotations_pb'
5
7
  require 'google/api/client_pb'
6
8
  require 'google/api/field_behavior_pb'
@@ -10,7 +12,6 @@ require 'google/longrunning/operations_pb'
10
12
  require 'google/protobuf/duration_pb'
11
13
  require 'google/protobuf/field_mask_pb'
12
14
  require 'google/protobuf/timestamp_pb'
13
- require 'google/protobuf'
14
15
 
15
16
  Google::Protobuf::DescriptorPool.generated_pool.build do
16
17
  add_file("google/cloud/dataproc/v1/clusters.proto", :syntax => :proto3) do
@@ -18,6 +19,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
18
19
  optional :project_id, :string, 1
19
20
  optional :cluster_name, :string, 2
20
21
  optional :config, :message, 3, "google.cloud.dataproc.v1.ClusterConfig"
22
+ optional :virtual_cluster_config, :message, 10, "google.cloud.dataproc.v1.VirtualClusterConfig"
21
23
  map :labels, :string, :string, 8
22
24
  optional :status, :message, 4, "google.cloud.dataproc.v1.ClusterStatus"
23
25
  repeated :status_history, :message, 7, "google.cloud.dataproc.v1.ClusterStatus"
@@ -39,14 +41,17 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
39
41
  optional :lifecycle_config, :message, 17, "google.cloud.dataproc.v1.LifecycleConfig"
40
42
  optional :endpoint_config, :message, 19, "google.cloud.dataproc.v1.EndpointConfig"
41
43
  optional :metastore_config, :message, 20, "google.cloud.dataproc.v1.MetastoreConfig"
42
- optional :gke_cluster_config, :message, 21, "google.cloud.dataproc.v1.GkeClusterConfig"
43
44
  end
44
- add_message "google.cloud.dataproc.v1.GkeClusterConfig" do
45
- optional :namespaced_gke_deployment_target, :message, 1, "google.cloud.dataproc.v1.GkeClusterConfig.NamespacedGkeDeploymentTarget"
45
+ add_message "google.cloud.dataproc.v1.VirtualClusterConfig" do
46
+ optional :staging_bucket, :string, 1
47
+ optional :auxiliary_services_config, :message, 7, "google.cloud.dataproc.v1.AuxiliaryServicesConfig"
48
+ oneof :infrastructure_config do
49
+ optional :kubernetes_cluster_config, :message, 6, "google.cloud.dataproc.v1.KubernetesClusterConfig"
50
+ end
46
51
  end
47
- add_message "google.cloud.dataproc.v1.GkeClusterConfig.NamespacedGkeDeploymentTarget" do
48
- optional :target_gke_cluster, :string, 1
49
- optional :cluster_namespace, :string, 2
52
+ add_message "google.cloud.dataproc.v1.AuxiliaryServicesConfig" do
53
+ optional :metastore_config, :message, 1, "google.cloud.dataproc.v1.MetastoreConfig"
54
+ optional :spark_history_server_config, :message, 2, "google.cloud.dataproc.v1.SparkHistoryServerConfig"
50
55
  end
51
56
  add_message "google.cloud.dataproc.v1.EndpointConfig" do
52
57
  map :http_ports, :string, :string, 1
@@ -273,8 +278,8 @@ module Google
273
278
  module V1
274
279
  Cluster = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.Cluster").msgclass
275
280
  ClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.ClusterConfig").msgclass
276
- GkeClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeClusterConfig").msgclass
277
- GkeClusterConfig::NamespacedGkeDeploymentTarget = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeClusterConfig.NamespacedGkeDeploymentTarget").msgclass
281
+ VirtualClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.VirtualClusterConfig").msgclass
282
+ AuxiliaryServicesConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.AuxiliaryServicesConfig").msgclass
278
283
  EndpointConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.EndpointConfig").msgclass
279
284
  AutoscalingConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.AutoscalingConfig").msgclass
280
285
  EncryptionConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.EncryptionConfig").msgclass
@@ -1,6 +1,8 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/jobs.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/annotations_pb'
5
7
  require 'google/api/client_pb'
6
8
  require 'google/api/field_behavior_pb'
@@ -8,7 +10,6 @@ require 'google/longrunning/operations_pb'
8
10
  require 'google/protobuf/empty_pb'
9
11
  require 'google/protobuf/field_mask_pb'
10
12
  require 'google/protobuf/timestamp_pb'
11
- require 'google/protobuf'
12
13
 
13
14
  Google::Protobuf::DescriptorPool.generated_pool.build do
14
15
  add_file("google/cloud/dataproc/v1/jobs.proto", :syntax => :proto3) do
@@ -1,9 +1,10 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/operations.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/field_behavior_pb'
5
7
  require 'google/protobuf/timestamp_pb'
6
- require 'google/protobuf'
7
8
 
8
9
  Google::Protobuf::DescriptorPool.generated_pool.build do
9
10
  add_file("google/cloud/dataproc/v1/operations.proto", :syntax => :proto3) do
@@ -1,9 +1,10 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/shared.proto
3
3
 
4
- require 'google/api/field_behavior_pb'
5
4
  require 'google/protobuf'
6
5
 
6
+ require 'google/api/field_behavior_pb'
7
+
7
8
  Google::Protobuf::DescriptorPool.generated_pool.build do
8
9
  add_file("google/cloud/dataproc/v1/shared.proto", :syntax => :proto3) do
9
10
  add_message "google.cloud.dataproc.v1.RuntimeConfig" do
@@ -36,6 +37,53 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
36
37
  optional :output_uri, :string, 2
37
38
  optional :diagnostic_output_uri, :string, 3
38
39
  end
40
+ add_message "google.cloud.dataproc.v1.GkeClusterConfig" do
41
+ optional :gke_cluster_target, :string, 2
42
+ repeated :node_pool_target, :message, 3, "google.cloud.dataproc.v1.GkeNodePoolTarget"
43
+ end
44
+ add_message "google.cloud.dataproc.v1.KubernetesClusterConfig" do
45
+ optional :kubernetes_namespace, :string, 1
46
+ optional :kubernetes_software_config, :message, 3, "google.cloud.dataproc.v1.KubernetesSoftwareConfig"
47
+ oneof :config do
48
+ optional :gke_cluster_config, :message, 2, "google.cloud.dataproc.v1.GkeClusterConfig"
49
+ end
50
+ end
51
+ add_message "google.cloud.dataproc.v1.KubernetesSoftwareConfig" do
52
+ map :component_version, :string, :string, 1
53
+ map :properties, :string, :string, 2
54
+ end
55
+ add_message "google.cloud.dataproc.v1.GkeNodePoolTarget" do
56
+ optional :node_pool, :string, 1
57
+ repeated :roles, :enum, 2, "google.cloud.dataproc.v1.GkeNodePoolTarget.Role"
58
+ optional :node_pool_config, :message, 3, "google.cloud.dataproc.v1.GkeNodePoolConfig"
59
+ end
60
+ add_enum "google.cloud.dataproc.v1.GkeNodePoolTarget.Role" do
61
+ value :ROLE_UNSPECIFIED, 0
62
+ value :DEFAULT, 1
63
+ value :CONTROLLER, 2
64
+ value :SPARK_DRIVER, 3
65
+ value :SPARK_EXECUTOR, 4
66
+ end
67
+ add_message "google.cloud.dataproc.v1.GkeNodePoolConfig" do
68
+ optional :config, :message, 2, "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodeConfig"
69
+ repeated :locations, :string, 13
70
+ optional :autoscaling, :message, 4, "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAutoscalingConfig"
71
+ end
72
+ add_message "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodeConfig" do
73
+ optional :machine_type, :string, 1
74
+ optional :preemptible, :bool, 10
75
+ optional :local_ssd_count, :int32, 7
76
+ repeated :accelerators, :message, 11, "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAcceleratorConfig"
77
+ optional :min_cpu_platform, :string, 13
78
+ end
79
+ add_message "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAcceleratorConfig" do
80
+ optional :accelerator_count, :int64, 1
81
+ optional :accelerator_type, :string, 2
82
+ end
83
+ add_message "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAutoscalingConfig" do
84
+ optional :min_node_count, :int32, 2
85
+ optional :max_node_count, :int32, 3
86
+ end
39
87
  add_enum "google.cloud.dataproc.v1.Component" do
40
88
  value :COMPONENT_UNSPECIFIED, 0
41
89
  value :ANACONDA, 5
@@ -69,6 +117,15 @@ module Google
69
117
  SparkHistoryServerConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkHistoryServerConfig").msgclass
70
118
  PeripheralsConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.PeripheralsConfig").msgclass
71
119
  RuntimeInfo = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.RuntimeInfo").msgclass
120
+ GkeClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeClusterConfig").msgclass
121
+ KubernetesClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.KubernetesClusterConfig").msgclass
122
+ KubernetesSoftwareConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.KubernetesSoftwareConfig").msgclass
123
+ GkeNodePoolTarget = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolTarget").msgclass
124
+ GkeNodePoolTarget::Role = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolTarget.Role").enummodule
125
+ GkeNodePoolConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig").msgclass
126
+ GkeNodePoolConfig::GkeNodeConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodeConfig").msgclass
127
+ GkeNodePoolConfig::GkeNodePoolAcceleratorConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAcceleratorConfig").msgclass
128
+ GkeNodePoolConfig::GkeNodePoolAutoscalingConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAutoscalingConfig").msgclass
72
129
  Component = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.Component").enummodule
73
130
  FailureAction = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.FailureAction").enummodule
74
131
  end
@@ -21,7 +21,7 @@ module Google
21
21
  module Cloud
22
22
  module Dataproc
23
23
  module V1
24
- VERSION = "0.8.0"
24
+ VERSION = "0.11.0"
25
25
  end
26
26
  end
27
27
  end
@@ -24,25 +24,6 @@ module Google
24
24
  module WorkflowTemplateService
25
25
  # Path helper methods for the WorkflowTemplateService API.
26
26
  module Paths
27
- ##
28
- # Create a fully-qualified Cluster resource string.
29
- #
30
- # The resource will be in the following format:
31
- #
32
- # `projects/{project}/locations/{location}/clusters/{cluster}`
33
- #
34
- # @param project [String]
35
- # @param location [String]
36
- # @param cluster [String]
37
- #
38
- # @return [::String]
39
- def cluster_path project:, location:, cluster:
40
- raise ::ArgumentError, "project cannot contain /" if project.to_s.include? "/"
41
- raise ::ArgumentError, "location cannot contain /" if location.to_s.include? "/"
42
-
43
- "projects/#{project}/locations/#{location}/clusters/#{cluster}"
44
- end
45
-
46
27
  ##
47
28
  # Create a fully-qualified Location resource string.
48
29
  #
@@ -1,6 +1,8 @@
1
1
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
2
  # source: google/cloud/dataproc/v1/workflow_templates.proto
3
3
 
4
+ require 'google/protobuf'
5
+
4
6
  require 'google/api/annotations_pb'
5
7
  require 'google/api/client_pb'
6
8
  require 'google/api/field_behavior_pb'
@@ -11,7 +13,6 @@ require 'google/longrunning/operations_pb'
11
13
  require 'google/protobuf/duration_pb'
12
14
  require 'google/protobuf/empty_pb'
13
15
  require 'google/protobuf/timestamp_pb'
14
- require 'google/protobuf'
15
16
 
16
17
  Google::Protobuf::DescriptorPool.generated_pool.build do
17
18
  add_file("google/cloud/dataproc/v1/workflow_templates.proto", :syntax => :proto3) do
@@ -29,6 +29,8 @@ module Google
29
29
  ##
30
30
  # To load this package, including all its services, and instantiate a client:
31
31
  #
32
+ # @example
33
+ #
32
34
  # require "google/cloud/dataproc/v1"
33
35
  # client = ::Google::Cloud::Dataproc::V1::AutoscalingPolicyService::Client.new
34
36
  #
@@ -35,6 +35,15 @@ module Google
35
35
  # Optional. The cluster config for a cluster of Compute Engine Instances.
36
36
  # Note that Dataproc may set default values, and values may change
37
37
  # when clusters are updated.
38
+ # @!attribute [rw] virtual_cluster_config
39
+ # @return [::Google::Cloud::Dataproc::V1::VirtualClusterConfig]
40
+ # Optional. The virtual cluster config, used when creating a Dataproc cluster that
41
+ # does not directly control the underlying compute resources, for example,
42
+ # when creating a [Dataproc-on-GKE
43
+ # cluster](https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
44
+ # Note that Dataproc may set default values, and values may change when
45
+ # clusters are updated. Exactly one of config or virtualClusterConfig must be
46
+ # specified.
38
47
  # @!attribute [rw] labels
39
48
  # @return [::Google::Protobuf::Map{::String => ::String}]
40
49
  # Optional. The labels to associate with this cluster.
@@ -156,37 +165,48 @@ module Google
156
165
  # @!attribute [rw] metastore_config
157
166
  # @return [::Google::Cloud::Dataproc::V1::MetastoreConfig]
158
167
  # Optional. Metastore configuration.
159
- # @!attribute [rw] gke_cluster_config
160
- # @return [::Google::Cloud::Dataproc::V1::GkeClusterConfig]
161
- # Optional. BETA. The Kubernetes Engine config for Dataproc clusters deployed to
162
- # Kubernetes. Setting this is considered mutually exclusive with Compute
163
- # Engine-based options such as `gce_cluster_config`, `master_config`,
164
- # `worker_config`, `secondary_worker_config`, and `autoscaling_config`.
165
168
  class ClusterConfig
166
169
  include ::Google::Protobuf::MessageExts
167
170
  extend ::Google::Protobuf::MessageExts::ClassMethods
168
171
  end
169
172
 
170
- # The GKE config for this cluster.
171
- # @!attribute [rw] namespaced_gke_deployment_target
172
- # @return [::Google::Cloud::Dataproc::V1::GkeClusterConfig::NamespacedGkeDeploymentTarget]
173
- # Optional. A target for the deployment.
174
- class GkeClusterConfig
173
+ # Dataproc cluster config for a cluster that does not directly control the
174
+ # underlying compute resources, such as a [Dataproc-on-GKE
175
+ # cluster](https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
176
+ # @!attribute [rw] staging_bucket
177
+ # @return [::String]
178
+ # Optional. A Storage bucket used to stage job
179
+ # dependencies, config files, and job driver console output.
180
+ # If you do not specify a staging bucket, Cloud
181
+ # Dataproc will determine a Cloud Storage location (US,
182
+ # ASIA, or EU) for your cluster's staging bucket according to the
183
+ # Compute Engine zone where your cluster is deployed, and then create
184
+ # and manage this project-level, per-location bucket (see
185
+ # [Dataproc staging and temp
186
+ # buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
187
+ # **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
188
+ # a Cloud Storage bucket.**
189
+ # @!attribute [rw] kubernetes_cluster_config
190
+ # @return [::Google::Cloud::Dataproc::V1::KubernetesClusterConfig]
191
+ # Required. The configuration for running the Dataproc cluster on Kubernetes.
192
+ # @!attribute [rw] auxiliary_services_config
193
+ # @return [::Google::Cloud::Dataproc::V1::AuxiliaryServicesConfig]
194
+ # Optional. Configuration of auxiliary services used by this cluster.
195
+ class VirtualClusterConfig
175
196
  include ::Google::Protobuf::MessageExts
176
197
  extend ::Google::Protobuf::MessageExts::ClassMethods
198
+ end
177
199
 
178
- # A full, namespace-isolated deployment target for an existing GKE cluster.
179
- # @!attribute [rw] target_gke_cluster
180
- # @return [::String]
181
- # Optional. The target GKE cluster to deploy to.
182
- # Format: 'projects/\\{project}/locations/\\{location}/clusters/\\{cluster_id}'
183
- # @!attribute [rw] cluster_namespace
184
- # @return [::String]
185
- # Optional. A namespace within the GKE cluster to deploy into.
186
- class NamespacedGkeDeploymentTarget
187
- include ::Google::Protobuf::MessageExts
188
- extend ::Google::Protobuf::MessageExts::ClassMethods
189
- end
200
+ # Auxiliary services configuration for a Cluster.
201
+ # @!attribute [rw] metastore_config
202
+ # @return [::Google::Cloud::Dataproc::V1::MetastoreConfig]
203
+ # Optional. The Hive Metastore configuration for this workload.
204
+ # @!attribute [rw] spark_history_server_config
205
+ # @return [::Google::Cloud::Dataproc::V1::SparkHistoryServerConfig]
206
+ # Optional. The Spark History Server configuration for the workload.
207
+ class AuxiliaryServicesConfig
208
+ include ::Google::Protobuf::MessageExts
209
+ extend ::Google::Protobuf::MessageExts::ClassMethods
190
210
  end
191
211
 
192
212
  # Endpoint config for this cluster
@@ -594,8 +614,8 @@ module Google
594
614
  # Optional. Interface type of local SSDs (default is "scsi").
595
615
  # Valid values: "scsi" (Small Computer System Interface),
596
616
  # "nvme" (Non-Volatile Memory Express).
597
- # See [SSD Interface
598
- # types](https://cloud.google.com/compute/docs/disks/local-ssd#performance).
617
+ # See [local SSD
618
+ # performance](https://cloud.google.com/compute/docs/disks/local-ssd#performance).
599
619
  class DiskConfig
600
620
  include ::Google::Protobuf::MessageExts
601
621
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -648,6 +668,10 @@ module Google
648
668
  CREATING = 1
649
669
 
650
670
  # The cluster is currently running and healthy. It is ready for use.
671
+ #
672
+ # **Note:** The cluster state changes from "creating" to "running" status
673
+ # after the master node(s), first two primary worker nodes (and the last
674
+ # primary worker node if primary workers > 2) are running.
651
675
  RUNNING = 2
652
676
 
653
677
  # The cluster encountered an error. It is not ready for use.
@@ -135,6 +135,215 @@ module Google
135
135
  end
136
136
  end
137
137
 
138
+ # The cluster's GKE config.
139
+ # @!attribute [rw] gke_cluster_target
140
+ # @return [::String]
141
+ # Optional. A target GKE cluster to deploy to. It must be in the same project and
142
+ # region as the Dataproc cluster (the GKE cluster can be zonal or regional).
143
+ # Format: 'projects/\\{project}/locations/\\{location}/clusters/\\{cluster_id}'
144
+ # @!attribute [rw] node_pool_target
145
+ # @return [::Array<::Google::Cloud::Dataproc::V1::GkeNodePoolTarget>]
146
+ # Optional. GKE NodePools where workloads will be scheduled. At least one node pool
147
+ # must be assigned the 'default' role. Each role can be given to only a
148
+ # single NodePoolTarget. All NodePools must have the same location settings.
149
+ # If a nodePoolTarget is not specified, Dataproc constructs a default
150
+ # nodePoolTarget.
151
+ class GkeClusterConfig
152
+ include ::Google::Protobuf::MessageExts
153
+ extend ::Google::Protobuf::MessageExts::ClassMethods
154
+ end
155
+
156
+ # The configuration for running the Dataproc cluster on Kubernetes.
157
+ # @!attribute [rw] kubernetes_namespace
158
+ # @return [::String]
159
+ # Optional. A namespace within the Kubernetes cluster to deploy into. If this namespace
160
+ # does not exist, it is created. If it exists, Dataproc
161
+ # verifies that another Dataproc VirtualCluster is not installed
162
+ # into it. If not specified, the name of the Dataproc Cluster is used.
163
+ # @!attribute [rw] gke_cluster_config
164
+ # @return [::Google::Cloud::Dataproc::V1::GkeClusterConfig]
165
+ # Required. The configuration for running the Dataproc cluster on GKE.
166
+ # @!attribute [rw] kubernetes_software_config
167
+ # @return [::Google::Cloud::Dataproc::V1::KubernetesSoftwareConfig]
168
+ # Optional. The software configuration for this Dataproc cluster running on Kubernetes.
169
+ class KubernetesClusterConfig
170
+ include ::Google::Protobuf::MessageExts
171
+ extend ::Google::Protobuf::MessageExts::ClassMethods
172
+ end
173
+
174
+ # The software configuration for this Dataproc cluster running on Kubernetes.
175
+ # @!attribute [rw] component_version
176
+ # @return [::Google::Protobuf::Map{::String => ::String}]
177
+ # The components that should be installed in this Dataproc cluster. The key
178
+ # must be a string from the KubernetesComponent enumeration. The value is
179
+ # the version of the software to be installed.
180
+ # At least one entry must be specified.
181
+ # @!attribute [rw] properties
182
+ # @return [::Google::Protobuf::Map{::String => ::String}]
183
+ # The properties to set on daemon config files.
184
+ #
185
+ # Property keys are specified in `prefix:property` format, for example
186
+ # `spark:spark.kubernetes.container.image`. The following are supported
187
+ # prefixes and their mappings:
188
+ #
189
+ # * spark: `spark-defaults.conf`
190
+ #
191
+ # For more information, see [Cluster
192
+ # properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties).
193
+ class KubernetesSoftwareConfig
194
+ include ::Google::Protobuf::MessageExts
195
+ extend ::Google::Protobuf::MessageExts::ClassMethods
196
+
197
+ # @!attribute [rw] key
198
+ # @return [::String]
199
+ # @!attribute [rw] value
200
+ # @return [::String]
201
+ class ComponentVersionEntry
202
+ include ::Google::Protobuf::MessageExts
203
+ extend ::Google::Protobuf::MessageExts::ClassMethods
204
+ end
205
+
206
+ # @!attribute [rw] key
207
+ # @return [::String]
208
+ # @!attribute [rw] value
209
+ # @return [::String]
210
+ class PropertiesEntry
211
+ include ::Google::Protobuf::MessageExts
212
+ extend ::Google::Protobuf::MessageExts::ClassMethods
213
+ end
214
+ end
215
+
216
+ # GKE NodePools that Dataproc workloads run on.
217
+ # @!attribute [rw] node_pool
218
+ # @return [::String]
219
+ # Required. The target GKE NodePool.
220
+ # Format:
221
+ # 'projects/\\{project}/locations/\\{location}/clusters/\\{cluster}/nodePools/\\{node_pool}'
222
+ # @!attribute [rw] roles
223
+ # @return [::Array<::Google::Cloud::Dataproc::V1::GkeNodePoolTarget::Role>]
224
+ # Required. The types of role for a GKE NodePool
225
+ # @!attribute [rw] node_pool_config
226
+ # @return [::Google::Cloud::Dataproc::V1::GkeNodePoolConfig]
227
+ # Optional. The configuration for the GKE NodePool.
228
+ #
229
+ # If specified, Dataproc attempts to create a NodePool with the
230
+ # specified shape. If one with the same name already exists, it is
231
+ # verified against all specified fields. If a field differs, the
232
+ # virtual cluster creation will fail.
233
+ #
234
+ # If omitted, any NodePool with the specified name is used. If a
235
+ # NodePool with the specified name does not exist, Dataproc create a NodePool
236
+ # with default values.
237
+ class GkeNodePoolTarget
238
+ include ::Google::Protobuf::MessageExts
239
+ extend ::Google::Protobuf::MessageExts::ClassMethods
240
+
241
+ # `Role` specifies whose tasks will run on the NodePool. The roles can be
242
+ # specific to workloads. Exactly one GkeNodePoolTarget within the
243
+ # VirtualCluster must have 'default' role, which is used to run all workloads
244
+ # that are not associated with a NodePool.
245
+ module Role
246
+ # Role is unspecified.
247
+ ROLE_UNSPECIFIED = 0
248
+
249
+ # Any roles that are not directly assigned to a NodePool run on the
250
+ # `default` role's NodePool.
251
+ DEFAULT = 1
252
+
253
+ # Run controllers and webhooks.
254
+ CONTROLLER = 2
255
+
256
+ # Run spark driver.
257
+ SPARK_DRIVER = 3
258
+
259
+ # Run spark executors.
260
+ SPARK_EXECUTOR = 4
261
+ end
262
+ end
263
+
264
+ # The configuration of a GKE NodePool used by a [Dataproc-on-GKE
265
+ # cluster](https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
266
+ # @!attribute [rw] config
267
+ # @return [::Google::Cloud::Dataproc::V1::GkeNodePoolConfig::GkeNodeConfig]
268
+ # Optional. The node pool configuration.
269
+ # @!attribute [rw] locations
270
+ # @return [::Array<::String>]
271
+ # Optional. The list of Compute Engine
272
+ # [zones](https://cloud.google.com/compute/docs/zones#available) where
273
+ # NodePool's nodes will be located.
274
+ #
275
+ # **Note:** Currently, only one zone may be specified.
276
+ #
277
+ # If a location is not specified during NodePool creation, Dataproc will
278
+ # choose a location.
279
+ # @!attribute [rw] autoscaling
280
+ # @return [::Google::Cloud::Dataproc::V1::GkeNodePoolConfig::GkeNodePoolAutoscalingConfig]
281
+ # Optional. The autoscaler configuration for this NodePool. The autoscaler is enabled
282
+ # only when a valid configuration is present.
283
+ class GkeNodePoolConfig
284
+ include ::Google::Protobuf::MessageExts
285
+ extend ::Google::Protobuf::MessageExts::ClassMethods
286
+
287
+ # Parameters that describe cluster nodes.
288
+ # @!attribute [rw] machine_type
289
+ # @return [::String]
290
+ # Optional. The name of a Compute Engine [machine
291
+ # type](https://cloud.google.com/compute/docs/machine-types).
292
+ # @!attribute [rw] preemptible
293
+ # @return [::Boolean]
294
+ # Optional. Whether the nodes are created as [preemptible VM
295
+ # instances](https://cloud.google.com/compute/docs/instances/preemptible).
296
+ # @!attribute [rw] local_ssd_count
297
+ # @return [::Integer]
298
+ # Optional. The number of local SSD disks to attach to the node, which is limited by
299
+ # the maximum number of disks allowable per zone (see [Adding Local
300
+ # SSDs](https://cloud.google.com/compute/docs/disks/local-ssd)).
301
+ # @!attribute [rw] accelerators
302
+ # @return [::Array<::Google::Cloud::Dataproc::V1::GkeNodePoolConfig::GkeNodePoolAcceleratorConfig>]
303
+ # Optional. A list of [hardware
304
+ # accelerators](https://cloud.google.com/compute/docs/gpus) to attach to
305
+ # each node.
306
+ # @!attribute [rw] min_cpu_platform
307
+ # @return [::String]
308
+ # Optional. [Minimum CPU
309
+ # platform](https://cloud.google.com/compute/docs/instances/specify-min-cpu-platform)
310
+ # to be used by this instance. The instance may be scheduled on the
311
+ # specified or a newer CPU platform. Specify the friendly names of CPU
312
+ # platforms, such as "Intel Haswell"` or Intel Sandy Bridge".
313
+ class GkeNodeConfig
314
+ include ::Google::Protobuf::MessageExts
315
+ extend ::Google::Protobuf::MessageExts::ClassMethods
316
+ end
317
+
318
+ # A GkeNodeConfigAcceleratorConfig represents a Hardware Accelerator request
319
+ # for a NodePool.
320
+ # @!attribute [rw] accelerator_count
321
+ # @return [::Integer]
322
+ # The number of accelerator cards exposed to an instance.
323
+ # @!attribute [rw] accelerator_type
324
+ # @return [::String]
325
+ # The accelerator type resource namename (see GPUs on Compute Engine).
326
+ class GkeNodePoolAcceleratorConfig
327
+ include ::Google::Protobuf::MessageExts
328
+ extend ::Google::Protobuf::MessageExts::ClassMethods
329
+ end
330
+
331
+ # GkeNodePoolAutoscaling contains information the cluster autoscaler needs to
332
+ # adjust the size of the node pool to the current cluster usage.
333
+ # @!attribute [rw] min_node_count
334
+ # @return [::Integer]
335
+ # The minimum number of nodes in the NodePool. Must be >= 0 and <=
336
+ # max_node_count.
337
+ # @!attribute [rw] max_node_count
338
+ # @return [::Integer]
339
+ # The maximum number of nodes in the NodePool. Must be >= min_node_count.
340
+ # **Note:** Quota must be sufficient to scale up the cluster.
341
+ class GkeNodePoolAutoscalingConfig
342
+ include ::Google::Protobuf::MessageExts
343
+ extend ::Google::Protobuf::MessageExts::ClassMethods
344
+ end
345
+ end
346
+
138
347
  # Cluster components that can be activated.
139
348
  module Component
140
349
  # Unspecified component. Specifying this will cause Cluster creation to fail.
@@ -44,7 +44,7 @@ module Google
44
44
  # foo = any.unpack(Foo.class);
45
45
  # }
46
46
  #
47
- # Example 3: Pack and unpack a message in Python.
47
+ # Example 3: Pack and unpack a message in Python.
48
48
  #
49
49
  # foo = Foo(...)
50
50
  # any = Any()
@@ -54,7 +54,7 @@ module Google
54
54
  # any.Unpack(foo)
55
55
  # ...
56
56
  #
57
- # Example 4: Pack and unpack a message in Go
57
+ # Example 4: Pack and unpack a message in Go
58
58
  #
59
59
  # foo := &pb.Foo{...}
60
60
  # any, err := anypb.New(foo)
@@ -75,7 +75,7 @@ module Google
75
75
  #
76
76
  #
77
77
  # JSON
78
- # ====
78
+ #
79
79
  # The JSON representation of an `Any` value uses the regular
80
80
  # representation of the deserialized, embedded message, with an
81
81
  # additional field `@type` which contains the type URL. Example:
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-dataproc-v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.8.0
4
+ version: 0.11.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2022-01-11 00:00:00.000000000 Z
11
+ date: 2022-06-30 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: gapic-common
@@ -243,7 +243,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
243
243
  - !ruby/object:Gem::Version
244
244
  version: '0'
245
245
  requirements: []
246
- rubygems_version: 3.3.4
246
+ rubygems_version: 3.3.14
247
247
  signing_key:
248
248
  specification_version: 4
249
249
  summary: API Client library for the Cloud Dataproc V1 API