google-cloud-dataproc-v1 0.8.0 → 0.9.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/README.md +1 -1
- data/lib/google/cloud/dataproc/v1/cluster_controller/paths.rb +0 -19
- data/lib/google/cloud/dataproc/v1/clusters_pb.rb +13 -8
- data/lib/google/cloud/dataproc/v1/shared_pb.rb +56 -0
- data/lib/google/cloud/dataproc/v1/version.rb +1 -1
- data/lib/google/cloud/dataproc/v1/workflow_template_service/paths.rb +0 -19
- data/proto_docs/google/cloud/dataproc/v1/clusters.rb +64 -25
- data/proto_docs/google/cloud/dataproc/v1/shared.rb +209 -0
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: '096fe11db7fcb9a7ae4b9abd2deaf16fd8c1e0226c8705c40c7d70ef7bf175f5'
|
4
|
+
data.tar.gz: ff186e05094eee5034c48b4af192bea38e0e047b2d6ebb936ca52cbb1f66a9f1
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: c7c3e6d902a3c78603c7fa7b5bc92a5737dbe1e8e9b553874979914c9668a95d8366f621c4e26ea2c0d81ca1f8fa1ec32afd8b8c30244c7769299854f77cb83c
|
7
|
+
data.tar.gz: 8590a833f18fce759b1d040684ac0a43de995cce5a4b0f1174cbb209fddb6032be8b42f54eb910a2094aed8a405e0074cb7bbb1ee0d4d3189621fd2b00d097df
|
data/README.md
CHANGED
@@ -37,7 +37,7 @@ request = ::Google::Cloud::Dataproc::V1::CreateAutoscalingPolicyRequest.new # (r
|
|
37
37
|
response = client.create_autoscaling_policy request
|
38
38
|
```
|
39
39
|
|
40
|
-
View the [Client Library Documentation](https://
|
40
|
+
View the [Client Library Documentation](https://cloud.google.com/ruby/docs/reference/google-cloud-dataproc-v1/latest)
|
41
41
|
for class and method documentation.
|
42
42
|
|
43
43
|
See also the [Product Documentation](https://cloud.google.com/dataproc)
|
@@ -24,25 +24,6 @@ module Google
|
|
24
24
|
module ClusterController
|
25
25
|
# Path helper methods for the ClusterController API.
|
26
26
|
module Paths
|
27
|
-
##
|
28
|
-
# Create a fully-qualified Cluster resource string.
|
29
|
-
#
|
30
|
-
# The resource will be in the following format:
|
31
|
-
#
|
32
|
-
# `projects/{project}/locations/{location}/clusters/{cluster}`
|
33
|
-
#
|
34
|
-
# @param project [String]
|
35
|
-
# @param location [String]
|
36
|
-
# @param cluster [String]
|
37
|
-
#
|
38
|
-
# @return [::String]
|
39
|
-
def cluster_path project:, location:, cluster:
|
40
|
-
raise ::ArgumentError, "project cannot contain /" if project.to_s.include? "/"
|
41
|
-
raise ::ArgumentError, "location cannot contain /" if location.to_s.include? "/"
|
42
|
-
|
43
|
-
"projects/#{project}/locations/#{location}/clusters/#{cluster}"
|
44
|
-
end
|
45
|
-
|
46
27
|
##
|
47
28
|
# Create a fully-qualified Service resource string.
|
48
29
|
#
|
@@ -18,6 +18,7 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
18
18
|
optional :project_id, :string, 1
|
19
19
|
optional :cluster_name, :string, 2
|
20
20
|
optional :config, :message, 3, "google.cloud.dataproc.v1.ClusterConfig"
|
21
|
+
optional :virtual_cluster_config, :message, 10, "google.cloud.dataproc.v1.VirtualClusterConfig"
|
21
22
|
map :labels, :string, :string, 8
|
22
23
|
optional :status, :message, 4, "google.cloud.dataproc.v1.ClusterStatus"
|
23
24
|
repeated :status_history, :message, 7, "google.cloud.dataproc.v1.ClusterStatus"
|
@@ -39,14 +40,18 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
39
40
|
optional :lifecycle_config, :message, 17, "google.cloud.dataproc.v1.LifecycleConfig"
|
40
41
|
optional :endpoint_config, :message, 19, "google.cloud.dataproc.v1.EndpointConfig"
|
41
42
|
optional :metastore_config, :message, 20, "google.cloud.dataproc.v1.MetastoreConfig"
|
42
|
-
optional :gke_cluster_config, :message, 21, "google.cloud.dataproc.v1.GkeClusterConfig"
|
43
43
|
end
|
44
|
-
add_message "google.cloud.dataproc.v1.
|
45
|
-
optional :
|
44
|
+
add_message "google.cloud.dataproc.v1.VirtualClusterConfig" do
|
45
|
+
optional :staging_bucket, :string, 1
|
46
|
+
optional :temp_bucket, :string, 2
|
47
|
+
optional :auxiliary_services_config, :message, 7, "google.cloud.dataproc.v1.AuxiliaryServicesConfig"
|
48
|
+
oneof :infrastructure_config do
|
49
|
+
optional :kubernetes_cluster_config, :message, 6, "google.cloud.dataproc.v1.KubernetesClusterConfig"
|
50
|
+
end
|
46
51
|
end
|
47
|
-
add_message "google.cloud.dataproc.v1.
|
48
|
-
optional :
|
49
|
-
optional :
|
52
|
+
add_message "google.cloud.dataproc.v1.AuxiliaryServicesConfig" do
|
53
|
+
optional :metastore_config, :message, 1, "google.cloud.dataproc.v1.MetastoreConfig"
|
54
|
+
optional :spark_history_server_config, :message, 2, "google.cloud.dataproc.v1.SparkHistoryServerConfig"
|
50
55
|
end
|
51
56
|
add_message "google.cloud.dataproc.v1.EndpointConfig" do
|
52
57
|
map :http_ports, :string, :string, 1
|
@@ -273,8 +278,8 @@ module Google
|
|
273
278
|
module V1
|
274
279
|
Cluster = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.Cluster").msgclass
|
275
280
|
ClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.ClusterConfig").msgclass
|
276
|
-
|
277
|
-
|
281
|
+
VirtualClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.VirtualClusterConfig").msgclass
|
282
|
+
AuxiliaryServicesConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.AuxiliaryServicesConfig").msgclass
|
278
283
|
EndpointConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.EndpointConfig").msgclass
|
279
284
|
AutoscalingConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.AutoscalingConfig").msgclass
|
280
285
|
EncryptionConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.EncryptionConfig").msgclass
|
@@ -36,6 +36,53 @@ Google::Protobuf::DescriptorPool.generated_pool.build do
|
|
36
36
|
optional :output_uri, :string, 2
|
37
37
|
optional :diagnostic_output_uri, :string, 3
|
38
38
|
end
|
39
|
+
add_message "google.cloud.dataproc.v1.GkeClusterConfig" do
|
40
|
+
optional :gke_cluster_target, :string, 2
|
41
|
+
repeated :node_pool_target, :message, 3, "google.cloud.dataproc.v1.GkeNodePoolTarget"
|
42
|
+
end
|
43
|
+
add_message "google.cloud.dataproc.v1.KubernetesClusterConfig" do
|
44
|
+
optional :kubernetes_namespace, :string, 1
|
45
|
+
optional :kubernetes_software_config, :message, 3, "google.cloud.dataproc.v1.KubernetesSoftwareConfig"
|
46
|
+
oneof :config do
|
47
|
+
optional :gke_cluster_config, :message, 2, "google.cloud.dataproc.v1.GkeClusterConfig"
|
48
|
+
end
|
49
|
+
end
|
50
|
+
add_message "google.cloud.dataproc.v1.KubernetesSoftwareConfig" do
|
51
|
+
map :component_version, :string, :string, 1
|
52
|
+
map :properties, :string, :string, 2
|
53
|
+
end
|
54
|
+
add_message "google.cloud.dataproc.v1.GkeNodePoolTarget" do
|
55
|
+
optional :node_pool, :string, 1
|
56
|
+
repeated :roles, :enum, 2, "google.cloud.dataproc.v1.GkeNodePoolTarget.Role"
|
57
|
+
optional :node_pool_config, :message, 3, "google.cloud.dataproc.v1.GkeNodePoolConfig"
|
58
|
+
end
|
59
|
+
add_enum "google.cloud.dataproc.v1.GkeNodePoolTarget.Role" do
|
60
|
+
value :ROLE_UNSPECIFIED, 0
|
61
|
+
value :DEFAULT, 1
|
62
|
+
value :CONTROLLER, 2
|
63
|
+
value :SPARK_DRIVER, 3
|
64
|
+
value :SPARK_EXECUTOR, 4
|
65
|
+
end
|
66
|
+
add_message "google.cloud.dataproc.v1.GkeNodePoolConfig" do
|
67
|
+
optional :config, :message, 2, "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodeConfig"
|
68
|
+
repeated :locations, :string, 13
|
69
|
+
optional :autoscaling, :message, 4, "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAutoscalingConfig"
|
70
|
+
end
|
71
|
+
add_message "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodeConfig" do
|
72
|
+
optional :machine_type, :string, 1
|
73
|
+
optional :preemptible, :bool, 10
|
74
|
+
optional :local_ssd_count, :int32, 7
|
75
|
+
repeated :accelerators, :message, 11, "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAcceleratorConfig"
|
76
|
+
optional :min_cpu_platform, :string, 13
|
77
|
+
end
|
78
|
+
add_message "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAcceleratorConfig" do
|
79
|
+
optional :accelerator_count, :int64, 1
|
80
|
+
optional :accelerator_type, :string, 2
|
81
|
+
end
|
82
|
+
add_message "google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAutoscalingConfig" do
|
83
|
+
optional :min_node_count, :int32, 2
|
84
|
+
optional :max_node_count, :int32, 3
|
85
|
+
end
|
39
86
|
add_enum "google.cloud.dataproc.v1.Component" do
|
40
87
|
value :COMPONENT_UNSPECIFIED, 0
|
41
88
|
value :ANACONDA, 5
|
@@ -69,6 +116,15 @@ module Google
|
|
69
116
|
SparkHistoryServerConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkHistoryServerConfig").msgclass
|
70
117
|
PeripheralsConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.PeripheralsConfig").msgclass
|
71
118
|
RuntimeInfo = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.RuntimeInfo").msgclass
|
119
|
+
GkeClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeClusterConfig").msgclass
|
120
|
+
KubernetesClusterConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.KubernetesClusterConfig").msgclass
|
121
|
+
KubernetesSoftwareConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.KubernetesSoftwareConfig").msgclass
|
122
|
+
GkeNodePoolTarget = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolTarget").msgclass
|
123
|
+
GkeNodePoolTarget::Role = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolTarget.Role").enummodule
|
124
|
+
GkeNodePoolConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig").msgclass
|
125
|
+
GkeNodePoolConfig::GkeNodeConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodeConfig").msgclass
|
126
|
+
GkeNodePoolConfig::GkeNodePoolAcceleratorConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAcceleratorConfig").msgclass
|
127
|
+
GkeNodePoolConfig::GkeNodePoolAutoscalingConfig = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.GkeNodePoolConfig.GkeNodePoolAutoscalingConfig").msgclass
|
72
128
|
Component = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.Component").enummodule
|
73
129
|
FailureAction = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.FailureAction").enummodule
|
74
130
|
end
|
@@ -24,25 +24,6 @@ module Google
|
|
24
24
|
module WorkflowTemplateService
|
25
25
|
# Path helper methods for the WorkflowTemplateService API.
|
26
26
|
module Paths
|
27
|
-
##
|
28
|
-
# Create a fully-qualified Cluster resource string.
|
29
|
-
#
|
30
|
-
# The resource will be in the following format:
|
31
|
-
#
|
32
|
-
# `projects/{project}/locations/{location}/clusters/{cluster}`
|
33
|
-
#
|
34
|
-
# @param project [String]
|
35
|
-
# @param location [String]
|
36
|
-
# @param cluster [String]
|
37
|
-
#
|
38
|
-
# @return [::String]
|
39
|
-
def cluster_path project:, location:, cluster:
|
40
|
-
raise ::ArgumentError, "project cannot contain /" if project.to_s.include? "/"
|
41
|
-
raise ::ArgumentError, "location cannot contain /" if location.to_s.include? "/"
|
42
|
-
|
43
|
-
"projects/#{project}/locations/#{location}/clusters/#{cluster}"
|
44
|
-
end
|
45
|
-
|
46
27
|
##
|
47
28
|
# Create a fully-qualified Location resource string.
|
48
29
|
#
|
@@ -35,6 +35,15 @@ module Google
|
|
35
35
|
# Optional. The cluster config for a cluster of Compute Engine Instances.
|
36
36
|
# Note that Dataproc may set default values, and values may change
|
37
37
|
# when clusters are updated.
|
38
|
+
# @!attribute [rw] virtual_cluster_config
|
39
|
+
# @return [::Google::Cloud::Dataproc::V1::VirtualClusterConfig]
|
40
|
+
# Optional. The virtual cluster config, used when creating a Dataproc cluster that
|
41
|
+
# does not directly control the underlying compute resources, for example,
|
42
|
+
# when creating a [Dataproc-on-GKE
|
43
|
+
# cluster](https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
|
44
|
+
# Note that Dataproc may set default values, and values may change when
|
45
|
+
# clusters are updated. Exactly one of config or virtualClusterConfig must be
|
46
|
+
# specified.
|
38
47
|
# @!attribute [rw] labels
|
39
48
|
# @return [::Google::Protobuf::Map{::String => ::String}]
|
40
49
|
# Optional. The labels to associate with this cluster.
|
@@ -156,37 +165,63 @@ module Google
|
|
156
165
|
# @!attribute [rw] metastore_config
|
157
166
|
# @return [::Google::Cloud::Dataproc::V1::MetastoreConfig]
|
158
167
|
# Optional. Metastore configuration.
|
159
|
-
# @!attribute [rw] gke_cluster_config
|
160
|
-
# @return [::Google::Cloud::Dataproc::V1::GkeClusterConfig]
|
161
|
-
# Optional. BETA. The Kubernetes Engine config for Dataproc clusters deployed to
|
162
|
-
# Kubernetes. Setting this is considered mutually exclusive with Compute
|
163
|
-
# Engine-based options such as `gce_cluster_config`, `master_config`,
|
164
|
-
# `worker_config`, `secondary_worker_config`, and `autoscaling_config`.
|
165
168
|
class ClusterConfig
|
166
169
|
include ::Google::Protobuf::MessageExts
|
167
170
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
168
171
|
end
|
169
172
|
|
170
|
-
#
|
171
|
-
#
|
172
|
-
#
|
173
|
-
#
|
174
|
-
|
173
|
+
# Dataproc cluster config for a cluster that does not directly control the
|
174
|
+
# underlying compute resources, such as a [Dataproc-on-GKE
|
175
|
+
# cluster](https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
|
176
|
+
# @!attribute [rw] staging_bucket
|
177
|
+
# @return [::String]
|
178
|
+
# Optional. A Storage bucket used to stage job
|
179
|
+
# dependencies, config files, and job driver console output.
|
180
|
+
# If you do not specify a staging bucket, Cloud
|
181
|
+
# Dataproc will determine a Cloud Storage location (US,
|
182
|
+
# ASIA, or EU) for your cluster's staging bucket according to the
|
183
|
+
# Compute Engine zone where your cluster is deployed, and then create
|
184
|
+
# and manage this project-level, per-location bucket (see
|
185
|
+
# [Dataproc staging and temp
|
186
|
+
# buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
187
|
+
# **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
|
188
|
+
# a Cloud Storage bucket.**
|
189
|
+
# @!attribute [rw] temp_bucket
|
190
|
+
# @return [::String]
|
191
|
+
# Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data,
|
192
|
+
# such as Spark and MapReduce history files.
|
193
|
+
# If you do not specify a temp bucket,
|
194
|
+
# Dataproc will determine a Cloud Storage location (US,
|
195
|
+
# ASIA, or EU) for your cluster's temp bucket according to the
|
196
|
+
# Compute Engine zone where your cluster is deployed, and then create
|
197
|
+
# and manage this project-level, per-location bucket. The default bucket has
|
198
|
+
# a TTL of 90 days, but you can use any TTL (or none) if you specify a
|
199
|
+
# bucket (see
|
200
|
+
# [Dataproc staging and temp
|
201
|
+
# buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
202
|
+
# **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
|
203
|
+
# a Cloud Storage bucket.**
|
204
|
+
# @!attribute [rw] kubernetes_cluster_config
|
205
|
+
# @return [::Google::Cloud::Dataproc::V1::KubernetesClusterConfig]
|
206
|
+
# Required. The configuration for running the Dataproc cluster on Kubernetes.
|
207
|
+
# @!attribute [rw] auxiliary_services_config
|
208
|
+
# @return [::Google::Cloud::Dataproc::V1::AuxiliaryServicesConfig]
|
209
|
+
# Optional. Configuration of auxiliary services used by this cluster.
|
210
|
+
class VirtualClusterConfig
|
175
211
|
include ::Google::Protobuf::MessageExts
|
176
212
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
213
|
+
end
|
177
214
|
|
178
|
-
|
179
|
-
|
180
|
-
|
181
|
-
|
182
|
-
|
183
|
-
|
184
|
-
|
185
|
-
|
186
|
-
|
187
|
-
|
188
|
-
extend ::Google::Protobuf::MessageExts::ClassMethods
|
189
|
-
end
|
215
|
+
# Auxiliary services configuration for a Cluster.
|
216
|
+
# @!attribute [rw] metastore_config
|
217
|
+
# @return [::Google::Cloud::Dataproc::V1::MetastoreConfig]
|
218
|
+
# Optional. The Hive Metastore configuration for this workload.
|
219
|
+
# @!attribute [rw] spark_history_server_config
|
220
|
+
# @return [::Google::Cloud::Dataproc::V1::SparkHistoryServerConfig]
|
221
|
+
# Optional. The Spark History Server configuration for the workload.
|
222
|
+
class AuxiliaryServicesConfig
|
223
|
+
include ::Google::Protobuf::MessageExts
|
224
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
190
225
|
end
|
191
226
|
|
192
227
|
# Endpoint config for this cluster
|
@@ -594,8 +629,8 @@ module Google
|
|
594
629
|
# Optional. Interface type of local SSDs (default is "scsi").
|
595
630
|
# Valid values: "scsi" (Small Computer System Interface),
|
596
631
|
# "nvme" (Non-Volatile Memory Express).
|
597
|
-
# See [SSD
|
598
|
-
#
|
632
|
+
# See [local SSD
|
633
|
+
# performance](https://cloud.google.com/compute/docs/disks/local-ssd#performance).
|
599
634
|
class DiskConfig
|
600
635
|
include ::Google::Protobuf::MessageExts
|
601
636
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -648,6 +683,10 @@ module Google
|
|
648
683
|
CREATING = 1
|
649
684
|
|
650
685
|
# The cluster is currently running and healthy. It is ready for use.
|
686
|
+
#
|
687
|
+
# **Note:** The cluster state changes from "creating" to "running" status
|
688
|
+
# after the master node(s), first two primary worker nodes (and the last
|
689
|
+
# primary worker node if primary workers > 2) are running.
|
651
690
|
RUNNING = 2
|
652
691
|
|
653
692
|
# The cluster encountered an error. It is not ready for use.
|
@@ -135,6 +135,215 @@ module Google
|
|
135
135
|
end
|
136
136
|
end
|
137
137
|
|
138
|
+
# The cluster's GKE config.
|
139
|
+
# @!attribute [rw] gke_cluster_target
|
140
|
+
# @return [::String]
|
141
|
+
# Optional. A target GKE cluster to deploy to. It must be in the same project and
|
142
|
+
# region as the Dataproc cluster (the GKE cluster can be zonal or regional).
|
143
|
+
# Format: 'projects/\\{project}/locations/\\{location}/clusters/\\{cluster_id}'
|
144
|
+
# @!attribute [rw] node_pool_target
|
145
|
+
# @return [::Array<::Google::Cloud::Dataproc::V1::GkeNodePoolTarget>]
|
146
|
+
# Optional. GKE NodePools where workloads will be scheduled. At least one node pool
|
147
|
+
# must be assigned the 'default' role. Each role can be given to only a
|
148
|
+
# single NodePoolTarget. All NodePools must have the same location settings.
|
149
|
+
# If a nodePoolTarget is not specified, Dataproc constructs a default
|
150
|
+
# nodePoolTarget.
|
151
|
+
class GkeClusterConfig
|
152
|
+
include ::Google::Protobuf::MessageExts
|
153
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
154
|
+
end
|
155
|
+
|
156
|
+
# The configuration for running the Dataproc cluster on Kubernetes.
|
157
|
+
# @!attribute [rw] kubernetes_namespace
|
158
|
+
# @return [::String]
|
159
|
+
# Optional. A namespace within the Kubernetes cluster to deploy into. If this namespace
|
160
|
+
# does not exist, it is created. If it exists, Dataproc
|
161
|
+
# verifies that another Dataproc VirtualCluster is not installed
|
162
|
+
# into it. If not specified, the name of the Dataproc Cluster is used.
|
163
|
+
# @!attribute [rw] gke_cluster_config
|
164
|
+
# @return [::Google::Cloud::Dataproc::V1::GkeClusterConfig]
|
165
|
+
# Required. The configuration for running the Dataproc cluster on GKE.
|
166
|
+
# @!attribute [rw] kubernetes_software_config
|
167
|
+
# @return [::Google::Cloud::Dataproc::V1::KubernetesSoftwareConfig]
|
168
|
+
# Optional. The software configuration for this Dataproc cluster running on Kubernetes.
|
169
|
+
class KubernetesClusterConfig
|
170
|
+
include ::Google::Protobuf::MessageExts
|
171
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
172
|
+
end
|
173
|
+
|
174
|
+
# The software configuration for this Dataproc cluster running on Kubernetes.
|
175
|
+
# @!attribute [rw] component_version
|
176
|
+
# @return [::Google::Protobuf::Map{::String => ::String}]
|
177
|
+
# The components that should be installed in this Dataproc cluster. The key
|
178
|
+
# must be a string from the KubernetesComponent enumeration. The value is
|
179
|
+
# the version of the software to be installed.
|
180
|
+
# At least one entry must be specified.
|
181
|
+
# @!attribute [rw] properties
|
182
|
+
# @return [::Google::Protobuf::Map{::String => ::String}]
|
183
|
+
# The properties to set on daemon config files.
|
184
|
+
#
|
185
|
+
# Property keys are specified in `prefix:property` format, for example
|
186
|
+
# `spark:spark.kubernetes.container.image`. The following are supported
|
187
|
+
# prefixes and their mappings:
|
188
|
+
#
|
189
|
+
# * spark: `spark-defaults.conf`
|
190
|
+
#
|
191
|
+
# For more information, see [Cluster
|
192
|
+
# properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties).
|
193
|
+
class KubernetesSoftwareConfig
|
194
|
+
include ::Google::Protobuf::MessageExts
|
195
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
196
|
+
|
197
|
+
# @!attribute [rw] key
|
198
|
+
# @return [::String]
|
199
|
+
# @!attribute [rw] value
|
200
|
+
# @return [::String]
|
201
|
+
class ComponentVersionEntry
|
202
|
+
include ::Google::Protobuf::MessageExts
|
203
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
204
|
+
end
|
205
|
+
|
206
|
+
# @!attribute [rw] key
|
207
|
+
# @return [::String]
|
208
|
+
# @!attribute [rw] value
|
209
|
+
# @return [::String]
|
210
|
+
class PropertiesEntry
|
211
|
+
include ::Google::Protobuf::MessageExts
|
212
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
213
|
+
end
|
214
|
+
end
|
215
|
+
|
216
|
+
# GKE NodePools that Dataproc workloads run on.
|
217
|
+
# @!attribute [rw] node_pool
|
218
|
+
# @return [::String]
|
219
|
+
# Required. The target GKE NodePool.
|
220
|
+
# Format:
|
221
|
+
# 'projects/\\{project}/locations/\\{location}/clusters/\\{cluster}/nodePools/\\{node_pool}'
|
222
|
+
# @!attribute [rw] roles
|
223
|
+
# @return [::Array<::Google::Cloud::Dataproc::V1::GkeNodePoolTarget::Role>]
|
224
|
+
# Required. The types of role for a GKE NodePool
|
225
|
+
# @!attribute [rw] node_pool_config
|
226
|
+
# @return [::Google::Cloud::Dataproc::V1::GkeNodePoolConfig]
|
227
|
+
# Optional. The configuration for the GKE NodePool.
|
228
|
+
#
|
229
|
+
# If specified, Dataproc attempts to create a NodePool with the
|
230
|
+
# specified shape. If one with the same name already exists, it is
|
231
|
+
# verified against all specified fields. If a field differs, the
|
232
|
+
# virtual cluster creation will fail.
|
233
|
+
#
|
234
|
+
# If omitted, any NodePool with the specified name is used. If a
|
235
|
+
# NodePool with the specified name does not exist, Dataproc create a NodePool
|
236
|
+
# with default values.
|
237
|
+
class GkeNodePoolTarget
|
238
|
+
include ::Google::Protobuf::MessageExts
|
239
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
240
|
+
|
241
|
+
# `Role` specifies whose tasks will run on the NodePool. The roles can be
|
242
|
+
# specific to workloads. Exactly one GkeNodePoolTarget within the
|
243
|
+
# VirtualCluster must have 'default' role, which is used to run all workloads
|
244
|
+
# that are not associated with a NodePool.
|
245
|
+
module Role
|
246
|
+
# Role is unspecified.
|
247
|
+
ROLE_UNSPECIFIED = 0
|
248
|
+
|
249
|
+
# Any roles that are not directly assigned to a NodePool run on the
|
250
|
+
# `default` role's NodePool.
|
251
|
+
DEFAULT = 1
|
252
|
+
|
253
|
+
# Run controllers and webhooks.
|
254
|
+
CONTROLLER = 2
|
255
|
+
|
256
|
+
# Run spark driver.
|
257
|
+
SPARK_DRIVER = 3
|
258
|
+
|
259
|
+
# Run spark executors.
|
260
|
+
SPARK_EXECUTOR = 4
|
261
|
+
end
|
262
|
+
end
|
263
|
+
|
264
|
+
# The configuration of a GKE NodePool used by a [Dataproc-on-GKE
|
265
|
+
# cluster](https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
|
266
|
+
# @!attribute [rw] config
|
267
|
+
# @return [::Google::Cloud::Dataproc::V1::GkeNodePoolConfig::GkeNodeConfig]
|
268
|
+
# Optional. The node pool configuration.
|
269
|
+
# @!attribute [rw] locations
|
270
|
+
# @return [::Array<::String>]
|
271
|
+
# Optional. The list of Compute Engine
|
272
|
+
# [zones](https://cloud.google.com/compute/docs/zones#available) where
|
273
|
+
# NodePool's nodes will be located.
|
274
|
+
#
|
275
|
+
# **Note:** Currently, only one zone may be specified.
|
276
|
+
#
|
277
|
+
# If a location is not specified during NodePool creation, Dataproc will
|
278
|
+
# choose a location.
|
279
|
+
# @!attribute [rw] autoscaling
|
280
|
+
# @return [::Google::Cloud::Dataproc::V1::GkeNodePoolConfig::GkeNodePoolAutoscalingConfig]
|
281
|
+
# Optional. The autoscaler configuration for this NodePool. The autoscaler is enabled
|
282
|
+
# only when a valid configuration is present.
|
283
|
+
class GkeNodePoolConfig
|
284
|
+
include ::Google::Protobuf::MessageExts
|
285
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
286
|
+
|
287
|
+
# Parameters that describe cluster nodes.
|
288
|
+
# @!attribute [rw] machine_type
|
289
|
+
# @return [::String]
|
290
|
+
# Optional. The name of a Compute Engine [machine
|
291
|
+
# type](https://cloud.google.com/compute/docs/machine-types).
|
292
|
+
# @!attribute [rw] preemptible
|
293
|
+
# @return [::Boolean]
|
294
|
+
# Optional. Whether the nodes are created as [preemptible VM
|
295
|
+
# instances](https://cloud.google.com/compute/docs/instances/preemptible).
|
296
|
+
# @!attribute [rw] local_ssd_count
|
297
|
+
# @return [::Integer]
|
298
|
+
# Optional. The number of local SSD disks to attach to the node, which is limited by
|
299
|
+
# the maximum number of disks allowable per zone (see [Adding Local
|
300
|
+
# SSDs](https://cloud.google.com/compute/docs/disks/local-ssd)).
|
301
|
+
# @!attribute [rw] accelerators
|
302
|
+
# @return [::Array<::Google::Cloud::Dataproc::V1::GkeNodePoolConfig::GkeNodePoolAcceleratorConfig>]
|
303
|
+
# Optional. A list of [hardware
|
304
|
+
# accelerators](https://cloud.google.com/compute/docs/gpus) to attach to
|
305
|
+
# each node.
|
306
|
+
# @!attribute [rw] min_cpu_platform
|
307
|
+
# @return [::String]
|
308
|
+
# Optional. [Minimum CPU
|
309
|
+
# platform](https://cloud.google.com/compute/docs/instances/specify-min-cpu-platform)
|
310
|
+
# to be used by this instance. The instance may be scheduled on the
|
311
|
+
# specified or a newer CPU platform. Specify the friendly names of CPU
|
312
|
+
# platforms, such as "Intel Haswell"` or Intel Sandy Bridge".
|
313
|
+
class GkeNodeConfig
|
314
|
+
include ::Google::Protobuf::MessageExts
|
315
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
316
|
+
end
|
317
|
+
|
318
|
+
# A GkeNodeConfigAcceleratorConfig represents a Hardware Accelerator request
|
319
|
+
# for a NodePool.
|
320
|
+
# @!attribute [rw] accelerator_count
|
321
|
+
# @return [::Integer]
|
322
|
+
# The number of accelerator cards exposed to an instance.
|
323
|
+
# @!attribute [rw] accelerator_type
|
324
|
+
# @return [::String]
|
325
|
+
# The accelerator type resource namename (see GPUs on Compute Engine).
|
326
|
+
class GkeNodePoolAcceleratorConfig
|
327
|
+
include ::Google::Protobuf::MessageExts
|
328
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
329
|
+
end
|
330
|
+
|
331
|
+
# GkeNodePoolAutoscaling contains information the cluster autoscaler needs to
|
332
|
+
# adjust the size of the node pool to the current cluster usage.
|
333
|
+
# @!attribute [rw] min_node_count
|
334
|
+
# @return [::Integer]
|
335
|
+
# The minimum number of nodes in the NodePool. Must be >= 0 and <=
|
336
|
+
# max_node_count.
|
337
|
+
# @!attribute [rw] max_node_count
|
338
|
+
# @return [::Integer]
|
339
|
+
# The maximum number of nodes in the NodePool. Must be >= min_node_count.
|
340
|
+
# **Note:** Quota must be sufficient to scale up the cluster.
|
341
|
+
class GkeNodePoolAutoscalingConfig
|
342
|
+
include ::Google::Protobuf::MessageExts
|
343
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
344
|
+
end
|
345
|
+
end
|
346
|
+
|
138
347
|
# Cluster components that can be activated.
|
139
348
|
module Component
|
140
349
|
# Unspecified component. Specifying this will cause Cluster creation to fail.
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-cloud-dataproc-v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.9.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2022-
|
11
|
+
date: 2022-02-17 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: gapic-common
|
@@ -243,7 +243,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
243
243
|
- !ruby/object:Gem::Version
|
244
244
|
version: '0'
|
245
245
|
requirements: []
|
246
|
-
rubygems_version: 3.3.
|
246
|
+
rubygems_version: 3.3.5
|
247
247
|
signing_key:
|
248
248
|
specification_version: 4
|
249
249
|
summary: API Client library for the Cloud Dataproc V1 API
|