google-cloud-dataproc-v1 0.6.0 → 0.7.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. checksums.yaml +4 -4
  2. data/AUTHENTICATION.md +8 -8
  3. data/README.md +1 -1
  4. data/lib/google/cloud/dataproc/v1/autoscaling_policies_pb.rb +3 -2
  5. data/lib/google/cloud/dataproc/v1/autoscaling_policies_services_pb.rb +1 -1
  6. data/lib/google/cloud/dataproc/v1/autoscaling_policy_service/client.rb +48 -54
  7. data/lib/google/cloud/dataproc/v1/batch_controller/client.rb +637 -0
  8. data/lib/google/cloud/dataproc/v1/batch_controller/credentials.rb +51 -0
  9. data/lib/google/cloud/dataproc/v1/batch_controller/operations.rb +664 -0
  10. data/lib/google/cloud/dataproc/v1/batch_controller/paths.rb +69 -0
  11. data/lib/google/cloud/dataproc/v1/batch_controller.rb +50 -0
  12. data/lib/google/cloud/dataproc/v1/batches_pb.rb +123 -0
  13. data/lib/google/cloud/dataproc/v1/batches_services_pb.rb +52 -0
  14. data/lib/google/cloud/dataproc/v1/cluster_controller/client.rb +75 -80
  15. data/lib/google/cloud/dataproc/v1/cluster_controller/operations.rb +34 -25
  16. data/lib/google/cloud/dataproc/v1/clusters_pb.rb +9 -2
  17. data/lib/google/cloud/dataproc/v1/clusters_services_pb.rb +3 -1
  18. data/lib/google/cloud/dataproc/v1/job_controller/client.rb +58 -72
  19. data/lib/google/cloud/dataproc/v1/job_controller/operations.rb +34 -25
  20. data/lib/google/cloud/dataproc/v1/jobs_pb.rb +2 -2
  21. data/lib/google/cloud/dataproc/v1/jobs_services_pb.rb +1 -1
  22. data/lib/google/cloud/dataproc/v1/operations_pb.rb +18 -3
  23. data/lib/google/cloud/dataproc/v1/shared_pb.rb +40 -2
  24. data/lib/google/cloud/dataproc/v1/version.rb +1 -1
  25. data/lib/google/cloud/dataproc/v1/workflow_template_service/client.rb +59 -74
  26. data/lib/google/cloud/dataproc/v1/workflow_template_service/operations.rb +34 -25
  27. data/lib/google/cloud/dataproc/v1/workflow_templates_pb.rb +2 -2
  28. data/lib/google/cloud/dataproc/v1/workflow_templates_services_pb.rb +2 -3
  29. data/lib/google/cloud/dataproc/v1.rb +1 -0
  30. data/proto_docs/google/api/field_behavior.rb +7 -1
  31. data/proto_docs/google/cloud/dataproc/v1/autoscaling_policies.rb +18 -0
  32. data/proto_docs/google/cloud/dataproc/v1/batches.rb +339 -0
  33. data/proto_docs/google/cloud/dataproc/v1/clusters.rb +45 -22
  34. data/proto_docs/google/cloud/dataproc/v1/jobs.rb +8 -9
  35. data/proto_docs/google/cloud/dataproc/v1/operations.rb +48 -0
  36. data/proto_docs/google/cloud/dataproc/v1/shared.rb +117 -1
  37. data/proto_docs/google/cloud/dataproc/v1/workflow_templates.rb +11 -14
  38. metadata +21 -7
@@ -0,0 +1,339 @@
1
+ # frozen_string_literal: true
2
+
3
+ # Copyright 2021 Google LLC
4
+ #
5
+ # Licensed under the Apache License, Version 2.0 (the "License");
6
+ # you may not use this file except in compliance with the License.
7
+ # You may obtain a copy of the License at
8
+ #
9
+ # https://www.apache.org/licenses/LICENSE-2.0
10
+ #
11
+ # Unless required by applicable law or agreed to in writing, software
12
+ # distributed under the License is distributed on an "AS IS" BASIS,
13
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14
+ # See the License for the specific language governing permissions and
15
+ # limitations under the License.
16
+
17
+ # Auto-generated by gapic-generator-ruby. DO NOT EDIT!
18
+
19
+
20
+ module Google
21
+ module Cloud
22
+ module Dataproc
23
+ module V1
24
+ # A request to create a batch workload.
25
+ # @!attribute [rw] parent
26
+ # @return [::String]
27
+ # Required. The parent resource where this batch will be created.
28
+ # @!attribute [rw] batch
29
+ # @return [::Google::Cloud::Dataproc::V1::Batch]
30
+ # Required. The batch to create.
31
+ # @!attribute [rw] batch_id
32
+ # @return [::String]
33
+ # Optional. The ID to use for the batch, which will become the final component of
34
+ # the batch's resource name.
35
+ #
36
+ # This value must be 4-63 characters. Valid characters are `/[a-z][0-9]-/`.
37
+ # @!attribute [rw] request_id
38
+ # @return [::String]
39
+ # Optional. A unique ID used to identify the request. If the service
40
+ # receives two
41
+ # [CreateBatchRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateBatchRequest)s
42
+ # with the same request_id, the second request is ignored and the
43
+ # Operation that corresponds to the first Batch created and stored
44
+ # in the backend is returned.
45
+ #
46
+ # Recommendation: Set this value to a
47
+ # [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
48
+ #
49
+ # The value must contain only letters (a-z, A-Z), numbers (0-9),
50
+ # underscores (_), and hyphens (-). The maximum length is 40 characters.
51
+ class CreateBatchRequest
52
+ include ::Google::Protobuf::MessageExts
53
+ extend ::Google::Protobuf::MessageExts::ClassMethods
54
+ end
55
+
56
+ # A request to get the resource representation for a batch workload.
57
+ # @!attribute [rw] name
58
+ # @return [::String]
59
+ # Required. The name of the batch to retrieve.
60
+ class GetBatchRequest
61
+ include ::Google::Protobuf::MessageExts
62
+ extend ::Google::Protobuf::MessageExts::ClassMethods
63
+ end
64
+
65
+ # A request to list batch workloads in a project.
66
+ # @!attribute [rw] parent
67
+ # @return [::String]
68
+ # Required. The parent, which owns this collection of batches.
69
+ # @!attribute [rw] page_size
70
+ # @return [::Integer]
71
+ # Optional. The maximum number of batches to return in each response.
72
+ # The service may return fewer than this value.
73
+ # The default page size is 20; the maximum page size is 1000.
74
+ # @!attribute [rw] page_token
75
+ # @return [::String]
76
+ # Optional. A page token received from a previous `ListBatches` call.
77
+ # Provide this token to retrieve the subsequent page.
78
+ class ListBatchesRequest
79
+ include ::Google::Protobuf::MessageExts
80
+ extend ::Google::Protobuf::MessageExts::ClassMethods
81
+ end
82
+
83
+ # A list of batch workloads.
84
+ # @!attribute [rw] batches
85
+ # @return [::Array<::Google::Cloud::Dataproc::V1::Batch>]
86
+ # The batches from the specified collection.
87
+ # @!attribute [rw] next_page_token
88
+ # @return [::String]
89
+ # A token, which can be sent as `page_token` to retrieve the next page.
90
+ # If this field is omitted, there are no subsequent pages.
91
+ class ListBatchesResponse
92
+ include ::Google::Protobuf::MessageExts
93
+ extend ::Google::Protobuf::MessageExts::ClassMethods
94
+ end
95
+
96
+ # A request to delete a batch workload.
97
+ # @!attribute [rw] name
98
+ # @return [::String]
99
+ # Required. The name of the batch resource to delete.
100
+ class DeleteBatchRequest
101
+ include ::Google::Protobuf::MessageExts
102
+ extend ::Google::Protobuf::MessageExts::ClassMethods
103
+ end
104
+
105
+ # A representation of a batch workload in the service.
106
+ # @!attribute [r] name
107
+ # @return [::String]
108
+ # Output only. The resource name of the batch.
109
+ # @!attribute [r] uuid
110
+ # @return [::String]
111
+ # Output only. A batch UUID (Unique Universal Identifier). The service
112
+ # generates this value when it creates the batch.
113
+ # @!attribute [r] create_time
114
+ # @return [::Google::Protobuf::Timestamp]
115
+ # Output only. The time when the batch was created.
116
+ # @!attribute [rw] pyspark_batch
117
+ # @return [::Google::Cloud::Dataproc::V1::PySparkBatch]
118
+ # Optional. PySpark batch config.
119
+ # @!attribute [rw] spark_batch
120
+ # @return [::Google::Cloud::Dataproc::V1::SparkBatch]
121
+ # Optional. Spark batch config.
122
+ # @!attribute [rw] spark_r_batch
123
+ # @return [::Google::Cloud::Dataproc::V1::SparkRBatch]
124
+ # Optional. SparkR batch config.
125
+ # @!attribute [rw] spark_sql_batch
126
+ # @return [::Google::Cloud::Dataproc::V1::SparkSqlBatch]
127
+ # Optional. SparkSql batch config.
128
+ # @!attribute [r] runtime_info
129
+ # @return [::Google::Cloud::Dataproc::V1::RuntimeInfo]
130
+ # Output only. Runtime information about batch execution.
131
+ # @!attribute [r] state
132
+ # @return [::Google::Cloud::Dataproc::V1::Batch::State]
133
+ # Output only. The state of the batch.
134
+ # @!attribute [r] state_message
135
+ # @return [::String]
136
+ # Output only. Batch state details, such as a failure
137
+ # description if the state is `FAILED`.
138
+ # @!attribute [r] state_time
139
+ # @return [::Google::Protobuf::Timestamp]
140
+ # Output only. The time when the batch entered a current state.
141
+ # @!attribute [r] creator
142
+ # @return [::String]
143
+ # Output only. The email address of the user who created the batch.
144
+ # @!attribute [rw] labels
145
+ # @return [::Google::Protobuf::Map{::String => ::String}]
146
+ # Optional. The labels to associate with this batch.
147
+ # Label **keys** must contain 1 to 63 characters, and must conform to
148
+ # [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
149
+ # Label **values** may be empty, but, if present, must contain 1 to 63
150
+ # characters, and must conform to [RFC
151
+ # 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
152
+ # associated with a batch.
153
+ # @!attribute [rw] runtime_config
154
+ # @return [::Google::Cloud::Dataproc::V1::RuntimeConfig]
155
+ # Optional. Runtime configuration for the batch execution.
156
+ # @!attribute [rw] environment_config
157
+ # @return [::Google::Cloud::Dataproc::V1::EnvironmentConfig]
158
+ # Optional. Environment configuration for the batch execution.
159
+ # @!attribute [r] operation
160
+ # @return [::String]
161
+ # Output only. The resource name of the operation associated with this batch.
162
+ # @!attribute [r] state_history
163
+ # @return [::Array<::Google::Cloud::Dataproc::V1::Batch::StateHistory>]
164
+ # Output only. Historical state information for the batch.
165
+ class Batch
166
+ include ::Google::Protobuf::MessageExts
167
+ extend ::Google::Protobuf::MessageExts::ClassMethods
168
+
169
+ # Historical state information.
170
+ # @!attribute [r] state
171
+ # @return [::Google::Cloud::Dataproc::V1::Batch::State]
172
+ # Output only. The state of the batch at this point in history.
173
+ # @!attribute [r] state_message
174
+ # @return [::String]
175
+ # Output only. Details about the state at this point in history.
176
+ # @!attribute [r] state_start_time
177
+ # @return [::Google::Protobuf::Timestamp]
178
+ # Output only. The time when the batch entered the historical state.
179
+ class StateHistory
180
+ include ::Google::Protobuf::MessageExts
181
+ extend ::Google::Protobuf::MessageExts::ClassMethods
182
+ end
183
+
184
+ # @!attribute [rw] key
185
+ # @return [::String]
186
+ # @!attribute [rw] value
187
+ # @return [::String]
188
+ class LabelsEntry
189
+ include ::Google::Protobuf::MessageExts
190
+ extend ::Google::Protobuf::MessageExts::ClassMethods
191
+ end
192
+
193
+ # The batch state.
194
+ module State
195
+ # The batch state is unknown.
196
+ STATE_UNSPECIFIED = 0
197
+
198
+ # The batch is created before running.
199
+ PENDING = 1
200
+
201
+ # The batch is running.
202
+ RUNNING = 2
203
+
204
+ # The batch is cancelling.
205
+ CANCELLING = 3
206
+
207
+ # The batch cancellation was successful.
208
+ CANCELLED = 4
209
+
210
+ # The batch completed successfully.
211
+ SUCCEEDED = 5
212
+
213
+ # The batch is no longer running due to an error.
214
+ FAILED = 6
215
+ end
216
+ end
217
+
218
+ # A configuration for running an
219
+ # [Apache
220
+ # PySpark](https://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html)
221
+ # batch workload.
222
+ # @!attribute [rw] main_python_file_uri
223
+ # @return [::String]
224
+ # Required. The HCFS URI of the main Python file to use as the Spark driver. Must
225
+ # be a .py file.
226
+ # @!attribute [rw] args
227
+ # @return [::Array<::String>]
228
+ # Optional. The arguments to pass to the driver. Do not include arguments
229
+ # that can be set as batch properties, such as `--conf`, since a collision
230
+ # can occur that causes an incorrect batch submission.
231
+ # @!attribute [rw] python_file_uris
232
+ # @return [::Array<::String>]
233
+ # Optional. HCFS file URIs of Python files to pass to the PySpark
234
+ # framework. Supported file types: `.py`, `.egg`, and `.zip`.
235
+ # @!attribute [rw] jar_file_uris
236
+ # @return [::Array<::String>]
237
+ # Optional. HCFS URIs of jar files to add to the classpath of the
238
+ # Spark driver and tasks.
239
+ # @!attribute [rw] file_uris
240
+ # @return [::Array<::String>]
241
+ # Optional. HCFS URIs of files to be placed in the working directory of
242
+ # each executor.
243
+ # @!attribute [rw] archive_uris
244
+ # @return [::Array<::String>]
245
+ # Optional. HCFS URIs of archives to be extracted into the working directory
246
+ # of each executor. Supported file types:
247
+ # `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
248
+ class PySparkBatch
249
+ include ::Google::Protobuf::MessageExts
250
+ extend ::Google::Protobuf::MessageExts::ClassMethods
251
+ end
252
+
253
+ # A configuration for running an [Apache Spark](http://spark.apache.org/)
254
+ # batch workload.
255
+ # @!attribute [rw] main_jar_file_uri
256
+ # @return [::String]
257
+ # Optional. The HCFS URI of the jar file that contains the main class.
258
+ # @!attribute [rw] main_class
259
+ # @return [::String]
260
+ # Optional. The name of the driver main class. The jar file that contains the class
261
+ # must be in the classpath or specified in `jar_file_uris`.
262
+ # @!attribute [rw] args
263
+ # @return [::Array<::String>]
264
+ # Optional. The arguments to pass to the driver. Do not include arguments
265
+ # that can be set as batch properties, such as `--conf`, since a collision
266
+ # can occur that causes an incorrect batch submission.
267
+ # @!attribute [rw] jar_file_uris
268
+ # @return [::Array<::String>]
269
+ # Optional. HCFS URIs of jar files to add to the classpath of the
270
+ # Spark driver and tasks.
271
+ # @!attribute [rw] file_uris
272
+ # @return [::Array<::String>]
273
+ # Optional. HCFS URIs of files to be placed in the working directory of
274
+ # each executor.
275
+ # @!attribute [rw] archive_uris
276
+ # @return [::Array<::String>]
277
+ # Optional. HCFS URIs of archives to be extracted into the working directory
278
+ # of each executor. Supported file types:
279
+ # `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
280
+ class SparkBatch
281
+ include ::Google::Protobuf::MessageExts
282
+ extend ::Google::Protobuf::MessageExts::ClassMethods
283
+ end
284
+
285
+ # A configuration for running an
286
+ # [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html)
287
+ # batch workload.
288
+ # @!attribute [rw] main_r_file_uri
289
+ # @return [::String]
290
+ # Required. The HCFS URI of the main R file to use as the driver.
291
+ # Must be a `.R` or `.r` file.
292
+ # @!attribute [rw] args
293
+ # @return [::Array<::String>]
294
+ # Optional. The arguments to pass to the Spark driver. Do not include arguments
295
+ # that can be set as batch properties, such as `--conf`, since a collision
296
+ # can occur that causes an incorrect batch submission.
297
+ # @!attribute [rw] file_uris
298
+ # @return [::Array<::String>]
299
+ # Optional. HCFS URIs of files to be placed in the working directory of
300
+ # each executor.
301
+ # @!attribute [rw] archive_uris
302
+ # @return [::Array<::String>]
303
+ # Optional. HCFS URIs of archives to be extracted into the working directory
304
+ # of each executor. Supported file types:
305
+ # `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
306
+ class SparkRBatch
307
+ include ::Google::Protobuf::MessageExts
308
+ extend ::Google::Protobuf::MessageExts::ClassMethods
309
+ end
310
+
311
+ # A configuration for running
312
+ # [Apache Spark SQL](http://spark.apache.org/sql/) queries as a batch workload.
313
+ # @!attribute [rw] query_file_uri
314
+ # @return [::String]
315
+ # Required. The HCFS URI of the script that contains Spark SQL queries to execute.
316
+ # @!attribute [rw] query_variables
317
+ # @return [::Google::Protobuf::Map{::String => ::String}]
318
+ # Optional. Mapping of query variable names to values (equivalent to the
319
+ # Spark SQL command: `SET name="value";`).
320
+ # @!attribute [rw] jar_file_uris
321
+ # @return [::Array<::String>]
322
+ # Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
323
+ class SparkSqlBatch
324
+ include ::Google::Protobuf::MessageExts
325
+ extend ::Google::Protobuf::MessageExts::ClassMethods
326
+
327
+ # @!attribute [rw] key
328
+ # @return [::String]
329
+ # @!attribute [rw] value
330
+ # @return [::String]
331
+ class QueryVariablesEntry
332
+ include ::Google::Protobuf::MessageExts
333
+ extend ::Google::Protobuf::MessageExts::ClassMethods
334
+ end
335
+ end
336
+ end
337
+ end
338
+ end
339
+ end
@@ -22,7 +22,7 @@ module Google
22
22
  module Dataproc
23
23
  module V1
24
24
  # Describes the identifying information, config, and status of
25
- # a cluster of Compute Engine instances.
25
+ # a Dataproc cluster
26
26
  # @!attribute [rw] project_id
27
27
  # @return [::String]
28
28
  # Required. The Google Cloud Platform project ID that the cluster belongs to.
@@ -83,10 +83,10 @@ module Google
83
83
  # ASIA, or EU) for your cluster's staging bucket according to the
84
84
  # Compute Engine zone where your cluster is deployed, and then create
85
85
  # and manage this project-level, per-location bucket (see
86
- # [Dataproc staging
87
- # bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
88
- # **This field requires a Cloud Storage bucket name, not a URI to a Cloud
89
- # Storage bucket.**
86
+ # [Dataproc staging and temp
87
+ # buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
88
+ # **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
89
+ # a Cloud Storage bucket.**
90
90
  # @!attribute [rw] temp_bucket
91
91
  # @return [::String]
92
92
  # Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data,
@@ -97,9 +97,11 @@ module Google
97
97
  # Compute Engine zone where your cluster is deployed, and then create
98
98
  # and manage this project-level, per-location bucket. The default bucket has
99
99
  # a TTL of 90 days, but you can use any TTL (or none) if you specify a
100
- # bucket.
101
- # **This field requires a Cloud Storage bucket name, not a URI to a Cloud
102
- # Storage bucket.**
100
+ # bucket (see
101
+ # [Dataproc staging and temp
102
+ # buckets](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
103
+ # **This field requires a Cloud Storage bucket name, not a `gs://...` URI to
104
+ # a Cloud Storage bucket.**
103
105
  # @!attribute [rw] gce_cluster_config
104
106
  # @return [::Google::Cloud::Dataproc::V1::GceClusterConfig]
105
107
  # Optional. The shared Compute Engine config settings for
@@ -107,18 +109,18 @@ module Google
107
109
  # @!attribute [rw] master_config
108
110
  # @return [::Google::Cloud::Dataproc::V1::InstanceGroupConfig]
109
111
  # Optional. The Compute Engine config settings for
110
- # the master instance in a cluster.
112
+ # the cluster's master instance.
111
113
  # @!attribute [rw] worker_config
112
114
  # @return [::Google::Cloud::Dataproc::V1::InstanceGroupConfig]
113
115
  # Optional. The Compute Engine config settings for
114
- # worker instances in a cluster.
116
+ # the cluster's worker instances.
115
117
  # @!attribute [rw] secondary_worker_config
116
118
  # @return [::Google::Cloud::Dataproc::V1::InstanceGroupConfig]
117
119
  # Optional. The Compute Engine config settings for
118
- # additional worker instances in a cluster.
120
+ # a cluster's secondary worker instances
119
121
  # @!attribute [rw] software_config
120
122
  # @return [::Google::Cloud::Dataproc::V1::SoftwareConfig]
121
- # Optional. The config settings for software inside the cluster.
123
+ # Optional. The config settings for cluster software.
122
124
  # @!attribute [rw] initialization_actions
123
125
  # @return [::Array<::Google::Cloud::Dataproc::V1::NodeInitializationAction>]
124
126
  # Optional. Commands to execute on each node after config is
@@ -334,6 +336,10 @@ module Google
334
336
  # @return [::Google::Cloud::Dataproc::V1::ShieldedInstanceConfig]
335
337
  # Optional. Shielded Instance Config for clusters using [Compute Engine Shielded
336
338
  # VMs](https://cloud.google.com/security/shielded-cloud/shielded-vm).
339
+ # @!attribute [rw] confidential_instance_config
340
+ # @return [::Google::Cloud::Dataproc::V1::ConfidentialInstanceConfig]
341
+ # Optional. Confidential Instance Config for clusters using [Confidential
342
+ # VMs](https://cloud.google.com/compute/confidential-vm/docs).
337
343
  class GceClusterConfig
338
344
  include ::Google::Protobuf::MessageExts
339
345
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -406,6 +412,16 @@ module Google
406
412
  extend ::Google::Protobuf::MessageExts::ClassMethods
407
413
  end
408
414
 
415
+ # Confidential Instance Config for clusters using [Confidential
416
+ # VMs](https://cloud.google.com/compute/confidential-vm/docs)
417
+ # @!attribute [rw] enable_confidential_compute
418
+ # @return [::Boolean]
419
+ # Optional. Defines whether the instance should have confidential compute enabled.
420
+ class ConfidentialInstanceConfig
421
+ include ::Google::Protobuf::MessageExts
422
+ extend ::Google::Protobuf::MessageExts::ClassMethods
423
+ end
424
+
409
425
  # The config settings for Compute Engine resources in
410
426
  # an instance group, such as a master or worker group.
411
427
  # @!attribute [rw] num_instances
@@ -629,6 +645,10 @@ module Google
629
645
  # The cluster encountered an error. It is not ready for use.
630
646
  ERROR = 3
631
647
 
648
+ # The cluster has encountered an error while being updated. Jobs can
649
+ # be submitted to the cluster, but the cluster cannot be updated.
650
+ ERROR_DUE_TO_UPDATE = 9
651
+
632
652
  # The cluster is being deleted. It cannot be used.
633
653
  DELETING = 4
634
654
 
@@ -905,7 +925,7 @@ module Google
905
925
  # Required. The cluster to create.
906
926
  # @!attribute [rw] request_id
907
927
  # @return [::String]
908
- # Optional. A unique id used to identify the request. If the server receives two
928
+ # Optional. A unique ID used to identify the request. If the server receives two
909
929
  # [CreateClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateClusterRequest)s
910
930
  # with the same id, then the second request will be ignored and the
911
931
  # first {::Google::Longrunning::Operation google.longrunning.Operation} created and stored in the backend
@@ -914,8 +934,11 @@ module Google
914
934
  # It is recommended to always set this value to a
915
935
  # [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
916
936
  #
917
- # The id must contain only letters (a-z, A-Z), numbers (0-9),
937
+ # The ID must contain only letters (a-z, A-Z), numbers (0-9),
918
938
  # underscores (_), and hyphens (-). The maximum length is 40 characters.
939
+ # @!attribute [rw] action_on_failed_primary_workers
940
+ # @return [::Google::Cloud::Dataproc::V1::FailureAction]
941
+ # Optional. Failure action when primary worker creation fails.
919
942
  class CreateClusterRequest
920
943
  include ::Google::Protobuf::MessageExts
921
944
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -1001,7 +1024,7 @@ module Google
1001
1024
  # </table>
1002
1025
  # @!attribute [rw] request_id
1003
1026
  # @return [::String]
1004
- # Optional. A unique id used to identify the request. If the server
1027
+ # Optional. A unique ID used to identify the request. If the server
1005
1028
  # receives two
1006
1029
  # [UpdateClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.UpdateClusterRequest)s
1007
1030
  # with the same id, then the second request will be ignored and the
@@ -1011,7 +1034,7 @@ module Google
1011
1034
  # It is recommended to always set this value to a
1012
1035
  # [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
1013
1036
  #
1014
- # The id must contain only letters (a-z, A-Z), numbers (0-9),
1037
+ # The ID must contain only letters (a-z, A-Z), numbers (0-9),
1015
1038
  # underscores (_), and hyphens (-). The maximum length is 40 characters.
1016
1039
  class UpdateClusterRequest
1017
1040
  include ::Google::Protobuf::MessageExts
@@ -1035,7 +1058,7 @@ module Google
1035
1058
  # (with error NOT_FOUND) if a cluster with the specified UUID does not exist.
1036
1059
  # @!attribute [rw] request_id
1037
1060
  # @return [::String]
1038
- # Optional. A unique id used to identify the request. If the server
1061
+ # Optional. A unique ID used to identify the request. If the server
1039
1062
  # receives two
1040
1063
  # [StopClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.StopClusterRequest)s
1041
1064
  # with the same id, then the second request will be ignored and the
@@ -1045,7 +1068,7 @@ module Google
1045
1068
  # Recommendation: Set this value to a
1046
1069
  # [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
1047
1070
  #
1048
- # The id must contain only letters (a-z, A-Z), numbers (0-9),
1071
+ # The ID must contain only letters (a-z, A-Z), numbers (0-9),
1049
1072
  # underscores (_), and hyphens (-). The maximum length is 40 characters.
1050
1073
  class StopClusterRequest
1051
1074
  include ::Google::Protobuf::MessageExts
@@ -1069,7 +1092,7 @@ module Google
1069
1092
  # (with error NOT_FOUND) if a cluster with the specified UUID does not exist.
1070
1093
  # @!attribute [rw] request_id
1071
1094
  # @return [::String]
1072
- # Optional. A unique id used to identify the request. If the server
1095
+ # Optional. A unique ID used to identify the request. If the server
1073
1096
  # receives two
1074
1097
  # [StartClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.StartClusterRequest)s
1075
1098
  # with the same id, then the second request will be ignored and the
@@ -1079,7 +1102,7 @@ module Google
1079
1102
  # Recommendation: Set this value to a
1080
1103
  # [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
1081
1104
  #
1082
- # The id must contain only letters (a-z, A-Z), numbers (0-9),
1105
+ # The ID must contain only letters (a-z, A-Z), numbers (0-9),
1083
1106
  # underscores (_), and hyphens (-). The maximum length is 40 characters.
1084
1107
  class StartClusterRequest
1085
1108
  include ::Google::Protobuf::MessageExts
@@ -1103,7 +1126,7 @@ module Google
1103
1126
  # (with error NOT_FOUND) if cluster with specified UUID does not exist.
1104
1127
  # @!attribute [rw] request_id
1105
1128
  # @return [::String]
1106
- # Optional. A unique id used to identify the request. If the server
1129
+ # Optional. A unique ID used to identify the request. If the server
1107
1130
  # receives two
1108
1131
  # [DeleteClusterRequest](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#google.cloud.dataproc.v1.DeleteClusterRequest)s
1109
1132
  # with the same id, then the second request will be ignored and the
@@ -1113,7 +1136,7 @@ module Google
1113
1136
  # It is recommended to always set this value to a
1114
1137
  # [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
1115
1138
  #
1116
- # The id must contain only letters (a-z, A-Z), numbers (0-9),
1139
+ # The ID must contain only letters (a-z, A-Z), numbers (0-9),
1117
1140
  # underscores (_), and hyphens (-). The maximum length is 40 characters.
1118
1141
  class DeleteClusterRequest
1119
1142
  include ::Google::Protobuf::MessageExts
@@ -42,7 +42,7 @@ module Google
42
42
  end
43
43
 
44
44
  # The Log4j level for job execution. When running an
45
- # [Apache Hive](http://hive.apache.org/) job, Cloud
45
+ # [Apache Hive](https://hive.apache.org/) job, Cloud
46
46
  # Dataproc configures the Hive client to an equivalent verbosity level.
47
47
  module Level
48
48
  # Level is unspecified. Use default level for log4j.
@@ -507,8 +507,7 @@ module Google
507
507
  # the job is submitted.
508
508
  # @!attribute [rw] cluster_labels
509
509
  # @return [::Google::Protobuf::Map{::String => ::String}]
510
- # Optional. Cluster labels to identify a cluster where the job will be
511
- # submitted.
510
+ # Optional. Cluster labels to identify a cluster where the job will be submitted.
512
511
  class JobPlacement
513
512
  include ::Google::Protobuf::MessageExts
514
513
  extend ::Google::Protobuf::MessageExts::ClassMethods
@@ -609,8 +608,8 @@ module Google
609
608
  # Encapsulates the full scoping used to reference a job.
610
609
  # @!attribute [rw] project_id
611
610
  # @return [::String]
612
- # Optional. The ID of the Google Cloud Platform project that the job belongs
613
- # to. If specified, must match the request project ID.
611
+ # Optional. The ID of the Google Cloud Platform project that the job belongs to. If
612
+ # specified, must match the request project ID.
614
613
  # @!attribute [rw] job_id
615
614
  # @return [::String]
616
615
  # Optional. The job ID, which must be unique within the project.
@@ -757,8 +756,8 @@ module Google
757
756
  # may be reused over time.
758
757
  # @!attribute [r] done
759
758
  # @return [::Boolean]
760
- # Output only. Indicates whether the job is completed. If the value is
761
- # `false`, the job is still in progress. If `true`, the job is completed, and
759
+ # Output only. Indicates whether the job is completed. If the value is `false`,
760
+ # the job is still in progress. If `true`, the job is completed, and
762
761
  # `status.state` field will indicate if it was successful, failed,
763
762
  # or cancelled.
764
763
  class Job
@@ -788,8 +787,8 @@ module Google
788
787
  # Maximum value is 10.
789
788
  # @!attribute [rw] max_failures_total
790
789
  # @return [::Integer]
791
- # Optional. Maximum number of times in total a driver may be restarted as a
792
- # result of driver exiting with non-zero code before job is reported failed.
790
+ # Optional. Maximum number of times in total a driver may be restarted as a result of
791
+ # driver exiting with non-zero code before job is reported failed.
793
792
  # Maximum value is 240.
794
793
  class JobScheduling
795
794
  include ::Google::Protobuf::MessageExts
@@ -21,6 +21,54 @@ module Google
21
21
  module Cloud
22
22
  module Dataproc
23
23
  module V1
24
+ # Metadata describing the Batch operation.
25
+ # @!attribute [rw] batch
26
+ # @return [::String]
27
+ # Name of the batch for the operation.
28
+ # @!attribute [rw] batch_uuid
29
+ # @return [::String]
30
+ # Batch UUID for the operation.
31
+ # @!attribute [rw] create_time
32
+ # @return [::Google::Protobuf::Timestamp]
33
+ # The time when the operation was created.
34
+ # @!attribute [rw] done_time
35
+ # @return [::Google::Protobuf::Timestamp]
36
+ # The time when the operation finished.
37
+ # @!attribute [rw] operation_type
38
+ # @return [::Google::Cloud::Dataproc::V1::BatchOperationMetadata::BatchOperationType]
39
+ # The operation type.
40
+ # @!attribute [rw] description
41
+ # @return [::String]
42
+ # Short description of the operation.
43
+ # @!attribute [rw] labels
44
+ # @return [::Google::Protobuf::Map{::String => ::String}]
45
+ # Labels associated with the operation.
46
+ # @!attribute [rw] warnings
47
+ # @return [::Array<::String>]
48
+ # Warnings encountered during operation execution.
49
+ class BatchOperationMetadata
50
+ include ::Google::Protobuf::MessageExts
51
+ extend ::Google::Protobuf::MessageExts::ClassMethods
52
+
53
+ # @!attribute [rw] key
54
+ # @return [::String]
55
+ # @!attribute [rw] value
56
+ # @return [::String]
57
+ class LabelsEntry
58
+ include ::Google::Protobuf::MessageExts
59
+ extend ::Google::Protobuf::MessageExts::ClassMethods
60
+ end
61
+
62
+ # Operation type for Batch resources
63
+ module BatchOperationType
64
+ # Batch operation type is unknown.
65
+ BATCH_OPERATION_TYPE_UNSPECIFIED = 0
66
+
67
+ # Batch operation type.
68
+ BATCH = 1
69
+ end
70
+ end
71
+
24
72
  # The status of the operation.
25
73
  # @!attribute [r] state
26
74
  # @return [::Google::Cloud::Dataproc::V1::ClusterOperationStatus::State]