google-cloud-dataproc-v1 0.6.0 → 0.7.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/AUTHENTICATION.md +8 -8
- data/README.md +1 -1
- data/lib/google/cloud/dataproc/v1/autoscaling_policies_pb.rb +3 -2
- data/lib/google/cloud/dataproc/v1/autoscaling_policies_services_pb.rb +1 -1
- data/lib/google/cloud/dataproc/v1/autoscaling_policy_service/client.rb +48 -54
- data/lib/google/cloud/dataproc/v1/batch_controller/client.rb +637 -0
- data/lib/google/cloud/dataproc/v1/batch_controller/credentials.rb +51 -0
- data/lib/google/cloud/dataproc/v1/batch_controller/operations.rb +664 -0
- data/lib/google/cloud/dataproc/v1/batch_controller/paths.rb +69 -0
- data/lib/google/cloud/dataproc/v1/batch_controller.rb +50 -0
- data/lib/google/cloud/dataproc/v1/batches_pb.rb +123 -0
- data/lib/google/cloud/dataproc/v1/batches_services_pb.rb +52 -0
- data/lib/google/cloud/dataproc/v1/cluster_controller/client.rb +75 -80
- data/lib/google/cloud/dataproc/v1/cluster_controller/operations.rb +34 -25
- data/lib/google/cloud/dataproc/v1/clusters_pb.rb +9 -2
- data/lib/google/cloud/dataproc/v1/clusters_services_pb.rb +3 -1
- data/lib/google/cloud/dataproc/v1/job_controller/client.rb +58 -72
- data/lib/google/cloud/dataproc/v1/job_controller/operations.rb +34 -25
- data/lib/google/cloud/dataproc/v1/jobs_pb.rb +2 -2
- data/lib/google/cloud/dataproc/v1/jobs_services_pb.rb +1 -1
- data/lib/google/cloud/dataproc/v1/operations_pb.rb +18 -3
- data/lib/google/cloud/dataproc/v1/shared_pb.rb +40 -2
- data/lib/google/cloud/dataproc/v1/version.rb +1 -1
- data/lib/google/cloud/dataproc/v1/workflow_template_service/client.rb +59 -74
- data/lib/google/cloud/dataproc/v1/workflow_template_service/operations.rb +34 -25
- data/lib/google/cloud/dataproc/v1/workflow_templates_pb.rb +2 -2
- data/lib/google/cloud/dataproc/v1/workflow_templates_services_pb.rb +2 -3
- data/lib/google/cloud/dataproc/v1.rb +1 -0
- data/proto_docs/google/api/field_behavior.rb +7 -1
- data/proto_docs/google/cloud/dataproc/v1/autoscaling_policies.rb +18 -0
- data/proto_docs/google/cloud/dataproc/v1/batches.rb +339 -0
- data/proto_docs/google/cloud/dataproc/v1/clusters.rb +45 -22
- data/proto_docs/google/cloud/dataproc/v1/jobs.rb +8 -9
- data/proto_docs/google/cloud/dataproc/v1/operations.rb +48 -0
- data/proto_docs/google/cloud/dataproc/v1/shared.rb +117 -1
- data/proto_docs/google/cloud/dataproc/v1/workflow_templates.rb +11 -14
- metadata +21 -7
@@ -0,0 +1,339 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# Copyright 2021 Google LLC
|
4
|
+
#
|
5
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
6
|
+
# you may not use this file except in compliance with the License.
|
7
|
+
# You may obtain a copy of the License at
|
8
|
+
#
|
9
|
+
# https://www.apache.org/licenses/LICENSE-2.0
|
10
|
+
#
|
11
|
+
# Unless required by applicable law or agreed to in writing, software
|
12
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
13
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
14
|
+
# See the License for the specific language governing permissions and
|
15
|
+
# limitations under the License.
|
16
|
+
|
17
|
+
# Auto-generated by gapic-generator-ruby. DO NOT EDIT!
|
18
|
+
|
19
|
+
|
20
|
+
module Google
|
21
|
+
module Cloud
|
22
|
+
module Dataproc
|
23
|
+
module V1
|
24
|
+
# A request to create a batch workload.
|
25
|
+
# @!attribute [rw] parent
|
26
|
+
# @return [::String]
|
27
|
+
# Required. The parent resource where this batch will be created.
|
28
|
+
# @!attribute [rw] batch
|
29
|
+
# @return [::Google::Cloud::Dataproc::V1::Batch]
|
30
|
+
# Required. The batch to create.
|
31
|
+
# @!attribute [rw] batch_id
|
32
|
+
# @return [::String]
|
33
|
+
# Optional. The ID to use for the batch, which will become the final component of
|
34
|
+
# the batch's resource name.
|
35
|
+
#
|
36
|
+
# This value must be 4-63 characters. Valid characters are `/[a-z][0-9]-/`.
|
37
|
+
# @!attribute [rw] request_id
|
38
|
+
# @return [::String]
|
39
|
+
# Optional. A unique ID used to identify the request. If the service
|
40
|
+
# receives two
|
41
|
+
# [CreateBatchRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateBatchRequest)s
|
42
|
+
# with the same request_id, the second request is ignored and the
|
43
|
+
# Operation that corresponds to the first Batch created and stored
|
44
|
+
# in the backend is returned.
|
45
|
+
#
|
46
|
+
# Recommendation: Set this value to a
|
47
|
+
# [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
|
48
|
+
#
|
49
|
+
# The value must contain only letters (a-z, A-Z), numbers (0-9),
|
50
|
+
# underscores (_), and hyphens (-). The maximum length is 40 characters.
|
51
|
+
class CreateBatchRequest
|
52
|
+
include ::Google::Protobuf::MessageExts
|
53
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
54
|
+
end
|
55
|
+
|
56
|
+
# A request to get the resource representation for a batch workload.
|
57
|
+
# @!attribute [rw] name
|
58
|
+
# @return [::String]
|
59
|
+
# Required. The name of the batch to retrieve.
|
60
|
+
class GetBatchRequest
|
61
|
+
include ::Google::Protobuf::MessageExts
|
62
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
63
|
+
end
|
64
|
+
|
65
|
+
# A request to list batch workloads in a project.
|
66
|
+
# @!attribute [rw] parent
|
67
|
+
# @return [::String]
|
68
|
+
# Required. The parent, which owns this collection of batches.
|
69
|
+
# @!attribute [rw] page_size
|
70
|
+
# @return [::Integer]
|
71
|
+
# Optional. The maximum number of batches to return in each response.
|
72
|
+
# The service may return fewer than this value.
|
73
|
+
# The default page size is 20; the maximum page size is 1000.
|
74
|
+
# @!attribute [rw] page_token
|
75
|
+
# @return [::String]
|
76
|
+
# Optional. A page token received from a previous `ListBatches` call.
|
77
|
+
# Provide this token to retrieve the subsequent page.
|
78
|
+
class ListBatchesRequest
|
79
|
+
include ::Google::Protobuf::MessageExts
|
80
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
81
|
+
end
|
82
|
+
|
83
|
+
# A list of batch workloads.
|
84
|
+
# @!attribute [rw] batches
|
85
|
+
# @return [::Array<::Google::Cloud::Dataproc::V1::Batch>]
|
86
|
+
# The batches from the specified collection.
|
87
|
+
# @!attribute [rw] next_page_token
|
88
|
+
# @return [::String]
|
89
|
+
# A token, which can be sent as `page_token` to retrieve the next page.
|
90
|
+
# If this field is omitted, there are no subsequent pages.
|
91
|
+
class ListBatchesResponse
|
92
|
+
include ::Google::Protobuf::MessageExts
|
93
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
94
|
+
end
|
95
|
+
|
96
|
+
# A request to delete a batch workload.
|
97
|
+
# @!attribute [rw] name
|
98
|
+
# @return [::String]
|
99
|
+
# Required. The name of the batch resource to delete.
|
100
|
+
class DeleteBatchRequest
|
101
|
+
include ::Google::Protobuf::MessageExts
|
102
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
103
|
+
end
|
104
|
+
|
105
|
+
# A representation of a batch workload in the service.
|
106
|
+
# @!attribute [r] name
|
107
|
+
# @return [::String]
|
108
|
+
# Output only. The resource name of the batch.
|
109
|
+
# @!attribute [r] uuid
|
110
|
+
# @return [::String]
|
111
|
+
# Output only. A batch UUID (Unique Universal Identifier). The service
|
112
|
+
# generates this value when it creates the batch.
|
113
|
+
# @!attribute [r] create_time
|
114
|
+
# @return [::Google::Protobuf::Timestamp]
|
115
|
+
# Output only. The time when the batch was created.
|
116
|
+
# @!attribute [rw] pyspark_batch
|
117
|
+
# @return [::Google::Cloud::Dataproc::V1::PySparkBatch]
|
118
|
+
# Optional. PySpark batch config.
|
119
|
+
# @!attribute [rw] spark_batch
|
120
|
+
# @return [::Google::Cloud::Dataproc::V1::SparkBatch]
|
121
|
+
# Optional. Spark batch config.
|
122
|
+
# @!attribute [rw] spark_r_batch
|
123
|
+
# @return [::Google::Cloud::Dataproc::V1::SparkRBatch]
|
124
|
+
# Optional. SparkR batch config.
|
125
|
+
# @!attribute [rw] spark_sql_batch
|
126
|
+
# @return [::Google::Cloud::Dataproc::V1::SparkSqlBatch]
|
127
|
+
# Optional. SparkSql batch config.
|
128
|
+
# @!attribute [r] runtime_info
|
129
|
+
# @return [::Google::Cloud::Dataproc::V1::RuntimeInfo]
|
130
|
+
# Output only. Runtime information about batch execution.
|
131
|
+
# @!attribute [r] state
|
132
|
+
# @return [::Google::Cloud::Dataproc::V1::Batch::State]
|
133
|
+
# Output only. The state of the batch.
|
134
|
+
# @!attribute [r] state_message
|
135
|
+
# @return [::String]
|
136
|
+
# Output only. Batch state details, such as a failure
|
137
|
+
# description if the state is `FAILED`.
|
138
|
+
# @!attribute [r] state_time
|
139
|
+
# @return [::Google::Protobuf::Timestamp]
|
140
|
+
# Output only. The time when the batch entered a current state.
|
141
|
+
# @!attribute [r] creator
|
142
|
+
# @return [::String]
|
143
|
+
# Output only. The email address of the user who created the batch.
|
144
|
+
# @!attribute [rw] labels
|
145
|
+
# @return [::Google::Protobuf::Map{::String => ::String}]
|
146
|
+
# Optional. The labels to associate with this batch.
|
147
|
+
# Label **keys** must contain 1 to 63 characters, and must conform to
|
148
|
+
# [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
|
149
|
+
# Label **values** may be empty, but, if present, must contain 1 to 63
|
150
|
+
# characters, and must conform to [RFC
|
151
|
+
# 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
|
152
|
+
# associated with a batch.
|
153
|
+
# @!attribute [rw] runtime_config
|
154
|
+
# @return [::Google::Cloud::Dataproc::V1::RuntimeConfig]
|
155
|
+
# Optional. Runtime configuration for the batch execution.
|
156
|
+
# @!attribute [rw] environment_config
|
157
|
+
# @return [::Google::Cloud::Dataproc::V1::EnvironmentConfig]
|
158
|
+
# Optional. Environment configuration for the batch execution.
|
159
|
+
# @!attribute [r] operation
|
160
|
+
# @return [::String]
|
161
|
+
# Output only. The resource name of the operation associated with this batch.
|
162
|
+
# @!attribute [r] state_history
|
163
|
+
# @return [::Array<::Google::Cloud::Dataproc::V1::Batch::StateHistory>]
|
164
|
+
# Output only. Historical state information for the batch.
|
165
|
+
class Batch
|
166
|
+
include ::Google::Protobuf::MessageExts
|
167
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
168
|
+
|
169
|
+
# Historical state information.
|
170
|
+
# @!attribute [r] state
|
171
|
+
# @return [::Google::Cloud::Dataproc::V1::Batch::State]
|
172
|
+
# Output only. The state of the batch at this point in history.
|
173
|
+
# @!attribute [r] state_message
|
174
|
+
# @return [::String]
|
175
|
+
# Output only. Details about the state at this point in history.
|
176
|
+
# @!attribute [r] state_start_time
|
177
|
+
# @return [::Google::Protobuf::Timestamp]
|
178
|
+
# Output only. The time when the batch entered the historical state.
|
179
|
+
class StateHistory
|
180
|
+
include ::Google::Protobuf::MessageExts
|
181
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
182
|
+
end
|
183
|
+
|
184
|
+
# @!attribute [rw] key
|
185
|
+
# @return [::String]
|
186
|
+
# @!attribute [rw] value
|
187
|
+
# @return [::String]
|
188
|
+
class LabelsEntry
|
189
|
+
include ::Google::Protobuf::MessageExts
|
190
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
191
|
+
end
|
192
|
+
|
193
|
+
# The batch state.
|
194
|
+
module State
|
195
|
+
# The batch state is unknown.
|
196
|
+
STATE_UNSPECIFIED = 0
|
197
|
+
|
198
|
+
# The batch is created before running.
|
199
|
+
PENDING = 1
|
200
|
+
|
201
|
+
# The batch is running.
|
202
|
+
RUNNING = 2
|
203
|
+
|
204
|
+
# The batch is cancelling.
|
205
|
+
CANCELLING = 3
|
206
|
+
|
207
|
+
# The batch cancellation was successful.
|
208
|
+
CANCELLED = 4
|
209
|
+
|
210
|
+
# The batch completed successfully.
|
211
|
+
SUCCEEDED = 5
|
212
|
+
|
213
|
+
# The batch is no longer running due to an error.
|
214
|
+
FAILED = 6
|
215
|
+
end
|
216
|
+
end
|
217
|
+
|
218
|
+
# A configuration for running an
|
219
|
+
# [Apache
|
220
|
+
# PySpark](https://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html)
|
221
|
+
# batch workload.
|
222
|
+
# @!attribute [rw] main_python_file_uri
|
223
|
+
# @return [::String]
|
224
|
+
# Required. The HCFS URI of the main Python file to use as the Spark driver. Must
|
225
|
+
# be a .py file.
|
226
|
+
# @!attribute [rw] args
|
227
|
+
# @return [::Array<::String>]
|
228
|
+
# Optional. The arguments to pass to the driver. Do not include arguments
|
229
|
+
# that can be set as batch properties, such as `--conf`, since a collision
|
230
|
+
# can occur that causes an incorrect batch submission.
|
231
|
+
# @!attribute [rw] python_file_uris
|
232
|
+
# @return [::Array<::String>]
|
233
|
+
# Optional. HCFS file URIs of Python files to pass to the PySpark
|
234
|
+
# framework. Supported file types: `.py`, `.egg`, and `.zip`.
|
235
|
+
# @!attribute [rw] jar_file_uris
|
236
|
+
# @return [::Array<::String>]
|
237
|
+
# Optional. HCFS URIs of jar files to add to the classpath of the
|
238
|
+
# Spark driver and tasks.
|
239
|
+
# @!attribute [rw] file_uris
|
240
|
+
# @return [::Array<::String>]
|
241
|
+
# Optional. HCFS URIs of files to be placed in the working directory of
|
242
|
+
# each executor.
|
243
|
+
# @!attribute [rw] archive_uris
|
244
|
+
# @return [::Array<::String>]
|
245
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory
|
246
|
+
# of each executor. Supported file types:
|
247
|
+
# `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
|
248
|
+
class PySparkBatch
|
249
|
+
include ::Google::Protobuf::MessageExts
|
250
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
251
|
+
end
|
252
|
+
|
253
|
+
# A configuration for running an [Apache Spark](http://spark.apache.org/)
|
254
|
+
# batch workload.
|
255
|
+
# @!attribute [rw] main_jar_file_uri
|
256
|
+
# @return [::String]
|
257
|
+
# Optional. The HCFS URI of the jar file that contains the main class.
|
258
|
+
# @!attribute [rw] main_class
|
259
|
+
# @return [::String]
|
260
|
+
# Optional. The name of the driver main class. The jar file that contains the class
|
261
|
+
# must be in the classpath or specified in `jar_file_uris`.
|
262
|
+
# @!attribute [rw] args
|
263
|
+
# @return [::Array<::String>]
|
264
|
+
# Optional. The arguments to pass to the driver. Do not include arguments
|
265
|
+
# that can be set as batch properties, such as `--conf`, since a collision
|
266
|
+
# can occur that causes an incorrect batch submission.
|
267
|
+
# @!attribute [rw] jar_file_uris
|
268
|
+
# @return [::Array<::String>]
|
269
|
+
# Optional. HCFS URIs of jar files to add to the classpath of the
|
270
|
+
# Spark driver and tasks.
|
271
|
+
# @!attribute [rw] file_uris
|
272
|
+
# @return [::Array<::String>]
|
273
|
+
# Optional. HCFS URIs of files to be placed in the working directory of
|
274
|
+
# each executor.
|
275
|
+
# @!attribute [rw] archive_uris
|
276
|
+
# @return [::Array<::String>]
|
277
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory
|
278
|
+
# of each executor. Supported file types:
|
279
|
+
# `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
|
280
|
+
class SparkBatch
|
281
|
+
include ::Google::Protobuf::MessageExts
|
282
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
283
|
+
end
|
284
|
+
|
285
|
+
# A configuration for running an
|
286
|
+
# [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html)
|
287
|
+
# batch workload.
|
288
|
+
# @!attribute [rw] main_r_file_uri
|
289
|
+
# @return [::String]
|
290
|
+
# Required. The HCFS URI of the main R file to use as the driver.
|
291
|
+
# Must be a `.R` or `.r` file.
|
292
|
+
# @!attribute [rw] args
|
293
|
+
# @return [::Array<::String>]
|
294
|
+
# Optional. The arguments to pass to the Spark driver. Do not include arguments
|
295
|
+
# that can be set as batch properties, such as `--conf`, since a collision
|
296
|
+
# can occur that causes an incorrect batch submission.
|
297
|
+
# @!attribute [rw] file_uris
|
298
|
+
# @return [::Array<::String>]
|
299
|
+
# Optional. HCFS URIs of files to be placed in the working directory of
|
300
|
+
# each executor.
|
301
|
+
# @!attribute [rw] archive_uris
|
302
|
+
# @return [::Array<::String>]
|
303
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory
|
304
|
+
# of each executor. Supported file types:
|
305
|
+
# `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
|
306
|
+
class SparkRBatch
|
307
|
+
include ::Google::Protobuf::MessageExts
|
308
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
309
|
+
end
|
310
|
+
|
311
|
+
# A configuration for running
|
312
|
+
# [Apache Spark SQL](http://spark.apache.org/sql/) queries as a batch workload.
|
313
|
+
# @!attribute [rw] query_file_uri
|
314
|
+
# @return [::String]
|
315
|
+
# Required. The HCFS URI of the script that contains Spark SQL queries to execute.
|
316
|
+
# @!attribute [rw] query_variables
|
317
|
+
# @return [::Google::Protobuf::Map{::String => ::String}]
|
318
|
+
# Optional. Mapping of query variable names to values (equivalent to the
|
319
|
+
# Spark SQL command: `SET name="value";`).
|
320
|
+
# @!attribute [rw] jar_file_uris
|
321
|
+
# @return [::Array<::String>]
|
322
|
+
# Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
|
323
|
+
class SparkSqlBatch
|
324
|
+
include ::Google::Protobuf::MessageExts
|
325
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
326
|
+
|
327
|
+
# @!attribute [rw] key
|
328
|
+
# @return [::String]
|
329
|
+
# @!attribute [rw] value
|
330
|
+
# @return [::String]
|
331
|
+
class QueryVariablesEntry
|
332
|
+
include ::Google::Protobuf::MessageExts
|
333
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
334
|
+
end
|
335
|
+
end
|
336
|
+
end
|
337
|
+
end
|
338
|
+
end
|
339
|
+
end
|
@@ -22,7 +22,7 @@ module Google
|
|
22
22
|
module Dataproc
|
23
23
|
module V1
|
24
24
|
# Describes the identifying information, config, and status of
|
25
|
-
# a cluster
|
25
|
+
# a Dataproc cluster
|
26
26
|
# @!attribute [rw] project_id
|
27
27
|
# @return [::String]
|
28
28
|
# Required. The Google Cloud Platform project ID that the cluster belongs to.
|
@@ -83,10 +83,10 @@ module Google
|
|
83
83
|
# ASIA, or EU) for your cluster's staging bucket according to the
|
84
84
|
# Compute Engine zone where your cluster is deployed, and then create
|
85
85
|
# and manage this project-level, per-location bucket (see
|
86
|
-
# [Dataproc staging
|
87
|
-
#
|
88
|
-
# **This field requires a Cloud Storage bucket name, not a URI to
|
89
|
-
# Storage bucket.**
|
86
|
+
# [Dataproc staging and temp
|
87
|
+
# buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
88
|
+
# **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
|
89
|
+
# a Cloud Storage bucket.**
|
90
90
|
# @!attribute [rw] temp_bucket
|
91
91
|
# @return [::String]
|
92
92
|
# Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data,
|
@@ -97,9 +97,11 @@ module Google
|
|
97
97
|
# Compute Engine zone where your cluster is deployed, and then create
|
98
98
|
# and manage this project-level, per-location bucket. The default bucket has
|
99
99
|
# a TTL of 90 days, but you can use any TTL (or none) if you specify a
|
100
|
-
# bucket
|
101
|
-
#
|
102
|
-
#
|
100
|
+
# bucket (see
|
101
|
+
# [Dataproc staging and temp
|
102
|
+
# buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
|
103
|
+
# **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
|
104
|
+
# a Cloud Storage bucket.**
|
103
105
|
# @!attribute [rw] gce_cluster_config
|
104
106
|
# @return [::Google::Cloud::Dataproc::V1::GceClusterConfig]
|
105
107
|
# Optional. The shared Compute Engine config settings for
|
@@ -107,18 +109,18 @@ module Google
|
|
107
109
|
# @!attribute [rw] master_config
|
108
110
|
# @return [::Google::Cloud::Dataproc::V1::InstanceGroupConfig]
|
109
111
|
# Optional. The Compute Engine config settings for
|
110
|
-
# the master instance
|
112
|
+
# the cluster's master instance.
|
111
113
|
# @!attribute [rw] worker_config
|
112
114
|
# @return [::Google::Cloud::Dataproc::V1::InstanceGroupConfig]
|
113
115
|
# Optional. The Compute Engine config settings for
|
114
|
-
# worker instances
|
116
|
+
# the cluster's worker instances.
|
115
117
|
# @!attribute [rw] secondary_worker_config
|
116
118
|
# @return [::Google::Cloud::Dataproc::V1::InstanceGroupConfig]
|
117
119
|
# Optional. The Compute Engine config settings for
|
118
|
-
#
|
120
|
+
# a cluster's secondary worker instances
|
119
121
|
# @!attribute [rw] software_config
|
120
122
|
# @return [::Google::Cloud::Dataproc::V1::SoftwareConfig]
|
121
|
-
# Optional. The config settings for software
|
123
|
+
# Optional. The config settings for cluster software.
|
122
124
|
# @!attribute [rw] initialization_actions
|
123
125
|
# @return [::Array<::Google::Cloud::Dataproc::V1::NodeInitializationAction>]
|
124
126
|
# Optional. Commands to execute on each node after config is
|
@@ -334,6 +336,10 @@ module Google
|
|
334
336
|
# @return [::Google::Cloud::Dataproc::V1::ShieldedInstanceConfig]
|
335
337
|
# Optional. Shielded Instance Config for clusters using [Compute Engine Shielded
|
336
338
|
# VMs](https://cloud.google.com/security/shielded-cloud/shielded-vm).
|
339
|
+
# @!attribute [rw] confidential_instance_config
|
340
|
+
# @return [::Google::Cloud::Dataproc::V1::ConfidentialInstanceConfig]
|
341
|
+
# Optional. Confidential Instance Config for clusters using [Confidential
|
342
|
+
# VMs](https://cloud.google.com/compute/confidential-vm/docs).
|
337
343
|
class GceClusterConfig
|
338
344
|
include ::Google::Protobuf::MessageExts
|
339
345
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -406,6 +412,16 @@ module Google
|
|
406
412
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
407
413
|
end
|
408
414
|
|
415
|
+
# Confidential Instance Config for clusters using [Confidential
|
416
|
+
# VMs](https://cloud.google.com/compute/confidential-vm/docs)
|
417
|
+
# @!attribute [rw] enable_confidential_compute
|
418
|
+
# @return [::Boolean]
|
419
|
+
# Optional. Defines whether the instance should have confidential compute enabled.
|
420
|
+
class ConfidentialInstanceConfig
|
421
|
+
include ::Google::Protobuf::MessageExts
|
422
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
423
|
+
end
|
424
|
+
|
409
425
|
# The config settings for Compute Engine resources in
|
410
426
|
# an instance group, such as a master or worker group.
|
411
427
|
# @!attribute [rw] num_instances
|
@@ -629,6 +645,10 @@ module Google
|
|
629
645
|
# The cluster encountered an error. It is not ready for use.
|
630
646
|
ERROR = 3
|
631
647
|
|
648
|
+
# The cluster has encountered an error while being updated. Jobs can
|
649
|
+
# be submitted to the cluster, but the cluster cannot be updated.
|
650
|
+
ERROR_DUE_TO_UPDATE = 9
|
651
|
+
|
632
652
|
# The cluster is being deleted. It cannot be used.
|
633
653
|
DELETING = 4
|
634
654
|
|
@@ -905,7 +925,7 @@ module Google
|
|
905
925
|
# Required. The cluster to create.
|
906
926
|
# @!attribute [rw] request_id
|
907
927
|
# @return [::String]
|
908
|
-
# Optional. A unique
|
928
|
+
# Optional. A unique ID used to identify the request. If the server receives two
|
909
929
|
# [CreateClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateClusterRequest)s
|
910
930
|
# with the same id, then the second request will be ignored and the
|
911
931
|
# first {::Google::Longrunning::Operation google.longrunning.Operation} created and stored in the backend
|
@@ -914,8 +934,11 @@ module Google
|
|
914
934
|
# It is recommended to always set this value to a
|
915
935
|
# [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
|
916
936
|
#
|
917
|
-
# The
|
937
|
+
# The ID must contain only letters (a-z, A-Z), numbers (0-9),
|
918
938
|
# underscores (_), and hyphens (-). The maximum length is 40 characters.
|
939
|
+
# @!attribute [rw] action_on_failed_primary_workers
|
940
|
+
# @return [::Google::Cloud::Dataproc::V1::FailureAction]
|
941
|
+
# Optional. Failure action when primary worker creation fails.
|
919
942
|
class CreateClusterRequest
|
920
943
|
include ::Google::Protobuf::MessageExts
|
921
944
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -1001,7 +1024,7 @@ module Google
|
|
1001
1024
|
# </table>
|
1002
1025
|
# @!attribute [rw] request_id
|
1003
1026
|
# @return [::String]
|
1004
|
-
# Optional. A unique
|
1027
|
+
# Optional. A unique ID used to identify the request. If the server
|
1005
1028
|
# receives two
|
1006
1029
|
# [UpdateClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.UpdateClusterRequest)s
|
1007
1030
|
# with the same id, then the second request will be ignored and the
|
@@ -1011,7 +1034,7 @@ module Google
|
|
1011
1034
|
# It is recommended to always set this value to a
|
1012
1035
|
# [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
|
1013
1036
|
#
|
1014
|
-
# The
|
1037
|
+
# The ID must contain only letters (a-z, A-Z), numbers (0-9),
|
1015
1038
|
# underscores (_), and hyphens (-). The maximum length is 40 characters.
|
1016
1039
|
class UpdateClusterRequest
|
1017
1040
|
include ::Google::Protobuf::MessageExts
|
@@ -1035,7 +1058,7 @@ module Google
|
|
1035
1058
|
# (with error NOT_FOUND) if a cluster with the specified UUID does not exist.
|
1036
1059
|
# @!attribute [rw] request_id
|
1037
1060
|
# @return [::String]
|
1038
|
-
# Optional. A unique
|
1061
|
+
# Optional. A unique ID used to identify the request. If the server
|
1039
1062
|
# receives two
|
1040
1063
|
# [StopClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.StopClusterRequest)s
|
1041
1064
|
# with the same id, then the second request will be ignored and the
|
@@ -1045,7 +1068,7 @@ module Google
|
|
1045
1068
|
# Recommendation: Set this value to a
|
1046
1069
|
# [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
|
1047
1070
|
#
|
1048
|
-
# The
|
1071
|
+
# The ID must contain only letters (a-z, A-Z), numbers (0-9),
|
1049
1072
|
# underscores (_), and hyphens (-). The maximum length is 40 characters.
|
1050
1073
|
class StopClusterRequest
|
1051
1074
|
include ::Google::Protobuf::MessageExts
|
@@ -1069,7 +1092,7 @@ module Google
|
|
1069
1092
|
# (with error NOT_FOUND) if a cluster with the specified UUID does not exist.
|
1070
1093
|
# @!attribute [rw] request_id
|
1071
1094
|
# @return [::String]
|
1072
|
-
# Optional. A unique
|
1095
|
+
# Optional. A unique ID used to identify the request. If the server
|
1073
1096
|
# receives two
|
1074
1097
|
# [StartClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.StartClusterRequest)s
|
1075
1098
|
# with the same id, then the second request will be ignored and the
|
@@ -1079,7 +1102,7 @@ module Google
|
|
1079
1102
|
# Recommendation: Set this value to a
|
1080
1103
|
# [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
|
1081
1104
|
#
|
1082
|
-
# The
|
1105
|
+
# The ID must contain only letters (a-z, A-Z), numbers (0-9),
|
1083
1106
|
# underscores (_), and hyphens (-). The maximum length is 40 characters.
|
1084
1107
|
class StartClusterRequest
|
1085
1108
|
include ::Google::Protobuf::MessageExts
|
@@ -1103,7 +1126,7 @@ module Google
|
|
1103
1126
|
# (with error NOT_FOUND) if cluster with specified UUID does not exist.
|
1104
1127
|
# @!attribute [rw] request_id
|
1105
1128
|
# @return [::String]
|
1106
|
-
# Optional. A unique
|
1129
|
+
# Optional. A unique ID used to identify the request. If the server
|
1107
1130
|
# receives two
|
1108
1131
|
# [DeleteClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.DeleteClusterRequest)s
|
1109
1132
|
# with the same id, then the second request will be ignored and the
|
@@ -1113,7 +1136,7 @@ module Google
|
|
1113
1136
|
# It is recommended to always set this value to a
|
1114
1137
|
# [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
|
1115
1138
|
#
|
1116
|
-
# The
|
1139
|
+
# The ID must contain only letters (a-z, A-Z), numbers (0-9),
|
1117
1140
|
# underscores (_), and hyphens (-). The maximum length is 40 characters.
|
1118
1141
|
class DeleteClusterRequest
|
1119
1142
|
include ::Google::Protobuf::MessageExts
|
@@ -42,7 +42,7 @@ module Google
|
|
42
42
|
end
|
43
43
|
|
44
44
|
# The Log4j level for job execution. When running an
|
45
|
-
# [Apache Hive](
|
45
|
+
# [Apache Hive](https://hive.apache.org/) job, Cloud
|
46
46
|
# Dataproc configures the Hive client to an equivalent verbosity level.
|
47
47
|
module Level
|
48
48
|
# Level is unspecified. Use default level for log4j.
|
@@ -507,8 +507,7 @@ module Google
|
|
507
507
|
# the job is submitted.
|
508
508
|
# @!attribute [rw] cluster_labels
|
509
509
|
# @return [::Google::Protobuf::Map{::String => ::String}]
|
510
|
-
# Optional. Cluster labels to identify a cluster where the job will be
|
511
|
-
# submitted.
|
510
|
+
# Optional. Cluster labels to identify a cluster where the job will be submitted.
|
512
511
|
class JobPlacement
|
513
512
|
include ::Google::Protobuf::MessageExts
|
514
513
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
@@ -609,8 +608,8 @@ module Google
|
|
609
608
|
# Encapsulates the full scoping used to reference a job.
|
610
609
|
# @!attribute [rw] project_id
|
611
610
|
# @return [::String]
|
612
|
-
# Optional. The ID of the Google Cloud Platform project that the job belongs
|
613
|
-
#
|
611
|
+
# Optional. The ID of the Google Cloud Platform project that the job belongs to. If
|
612
|
+
# specified, must match the request project ID.
|
614
613
|
# @!attribute [rw] job_id
|
615
614
|
# @return [::String]
|
616
615
|
# Optional. The job ID, which must be unique within the project.
|
@@ -757,8 +756,8 @@ module Google
|
|
757
756
|
# may be reused over time.
|
758
757
|
# @!attribute [r] done
|
759
758
|
# @return [::Boolean]
|
760
|
-
# Output only. Indicates whether the job is completed. If the value is
|
761
|
-
#
|
759
|
+
# Output only. Indicates whether the job is completed. If the value is `false`,
|
760
|
+
# the job is still in progress. If `true`, the job is completed, and
|
762
761
|
# `status.state` field will indicate if it was successful, failed,
|
763
762
|
# or cancelled.
|
764
763
|
class Job
|
@@ -788,8 +787,8 @@ module Google
|
|
788
787
|
# Maximum value is 10.
|
789
788
|
# @!attribute [rw] max_failures_total
|
790
789
|
# @return [::Integer]
|
791
|
-
# Optional. Maximum number of times in total a driver may be restarted as a
|
792
|
-
#
|
790
|
+
# Optional. Maximum number of times in total a driver may be restarted as a result of
|
791
|
+
# driver exiting with non-zero code before job is reported failed.
|
793
792
|
# Maximum value is 240.
|
794
793
|
class JobScheduling
|
795
794
|
include ::Google::Protobuf::MessageExts
|
@@ -21,6 +21,54 @@ module Google
|
|
21
21
|
module Cloud
|
22
22
|
module Dataproc
|
23
23
|
module V1
|
24
|
+
# Metadata describing the Batch operation.
|
25
|
+
# @!attribute [rw] batch
|
26
|
+
# @return [::String]
|
27
|
+
# Name of the batch for the operation.
|
28
|
+
# @!attribute [rw] batch_uuid
|
29
|
+
# @return [::String]
|
30
|
+
# Batch UUID for the operation.
|
31
|
+
# @!attribute [rw] create_time
|
32
|
+
# @return [::Google::Protobuf::Timestamp]
|
33
|
+
# The time when the operation was created.
|
34
|
+
# @!attribute [rw] done_time
|
35
|
+
# @return [::Google::Protobuf::Timestamp]
|
36
|
+
# The time when the operation finished.
|
37
|
+
# @!attribute [rw] operation_type
|
38
|
+
# @return [::Google::Cloud::Dataproc::V1::BatchOperationMetadata::BatchOperationType]
|
39
|
+
# The operation type.
|
40
|
+
# @!attribute [rw] description
|
41
|
+
# @return [::String]
|
42
|
+
# Short description of the operation.
|
43
|
+
# @!attribute [rw] labels
|
44
|
+
# @return [::Google::Protobuf::Map{::String => ::String}]
|
45
|
+
# Labels associated with the operation.
|
46
|
+
# @!attribute [rw] warnings
|
47
|
+
# @return [::Array<::String>]
|
48
|
+
# Warnings encountered during operation execution.
|
49
|
+
class BatchOperationMetadata
|
50
|
+
include ::Google::Protobuf::MessageExts
|
51
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
52
|
+
|
53
|
+
# @!attribute [rw] key
|
54
|
+
# @return [::String]
|
55
|
+
# @!attribute [rw] value
|
56
|
+
# @return [::String]
|
57
|
+
class LabelsEntry
|
58
|
+
include ::Google::Protobuf::MessageExts
|
59
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
60
|
+
end
|
61
|
+
|
62
|
+
# Operation type for Batch resources
|
63
|
+
module BatchOperationType
|
64
|
+
# Batch operation type is unknown.
|
65
|
+
BATCH_OPERATION_TYPE_UNSPECIFIED = 0
|
66
|
+
|
67
|
+
# Batch operation type.
|
68
|
+
BATCH = 1
|
69
|
+
end
|
70
|
+
end
|
71
|
+
|
24
72
|
# The status of the operation.
|
25
73
|
# @!attribute [r] state
|
26
74
|
# @return [::Google::Cloud::Dataproc::V1::ClusterOperationStatus::State]
|