google-apis-dataproc_v1 0.16.0 → 0.20.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +18 -0
- data/OVERVIEW.md +1 -1
- data/lib/google/apis/dataproc_v1/classes.rb +585 -14
- data/lib/google/apis/dataproc_v1/gem_version.rb +3 -3
- data/lib/google/apis/dataproc_v1/representations.rb +226 -0
- data/lib/google/apis/dataproc_v1/service.rb +151 -2
- metadata +4 -4
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 9692cc02dabfc8b9f70502c74e471e674da59ab3b95419f12ba1b4ed3198934e
|
4
|
+
data.tar.gz: fd4d7967084c09cf50fbf53f388073366b7830218da737b561752aaf722e3e47
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: ae1b213a1699284b7b5d312cbe5b34ee36213faf34ad51eb1d80576cb21d739e3c0db6d3b7755789e5e9f5dfd2e489ed8fa5b50063a82707869daec3bef340fc
|
7
|
+
data.tar.gz: 0ab023688f4a39174b27c0b0ec4c0c1b234a9645ba3aee1097b2b1afc3167702ef87261878d1fa53b2e53c57b74e69cede862ab82b4f3c0f062add01598afc87
|
data/CHANGELOG.md
CHANGED
@@ -1,5 +1,23 @@
|
|
1
1
|
# Release history for google-apis-dataproc_v1
|
2
2
|
|
3
|
+
### v0.20.0 (2022-01-27)
|
4
|
+
|
5
|
+
* Regenerated from discovery document revision 20220122
|
6
|
+
|
7
|
+
### v0.19.0 (2022-01-14)
|
8
|
+
|
9
|
+
* Regenerated from discovery document revision 20220111
|
10
|
+
* Regenerated using generator version 0.4.1
|
11
|
+
|
12
|
+
### v0.18.0 (2022-01-07)
|
13
|
+
|
14
|
+
* Regenerated from discovery document revision 20220105
|
15
|
+
* Unspecified changes
|
16
|
+
|
17
|
+
### v0.17.0 (2021-11-10)
|
18
|
+
|
19
|
+
* Regenerated from discovery document revision 20211105
|
20
|
+
|
3
21
|
### v0.16.0 (2021-10-21)
|
4
22
|
|
5
23
|
* Regenerated from discovery document revision 20211019
|
data/OVERVIEW.md
CHANGED
@@ -51,7 +51,7 @@ require "google/apis/dataproc_v1"
|
|
51
51
|
client = Google::Apis::DataprocV1::DataprocService.new
|
52
52
|
|
53
53
|
# Authenticate calls
|
54
|
-
client.
|
54
|
+
client.authorization = # ... use the googleauth gem to create credentials
|
55
55
|
```
|
56
56
|
|
57
57
|
See the class reference docs for information on the methods you can call from a client.
|
@@ -245,6 +245,131 @@ module Google
|
|
245
245
|
end
|
246
246
|
end
|
247
247
|
|
248
|
+
# A representation of a batch workload in the service.
|
249
|
+
class Batch
|
250
|
+
include Google::Apis::Core::Hashable
|
251
|
+
|
252
|
+
# Output only. The time when the batch was created.
|
253
|
+
# Corresponds to the JSON property `createTime`
|
254
|
+
# @return [String]
|
255
|
+
attr_accessor :create_time
|
256
|
+
|
257
|
+
# Output only. The email address of the user who created the batch.
|
258
|
+
# Corresponds to the JSON property `creator`
|
259
|
+
# @return [String]
|
260
|
+
attr_accessor :creator
|
261
|
+
|
262
|
+
# Environment configuration for a workload.
|
263
|
+
# Corresponds to the JSON property `environmentConfig`
|
264
|
+
# @return [Google::Apis::DataprocV1::EnvironmentConfig]
|
265
|
+
attr_accessor :environment_config
|
266
|
+
|
267
|
+
# Optional. The labels to associate with this batch. Label keys must contain 1
|
268
|
+
# to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/
|
269
|
+
# rfc1035.txt). Label values may be empty, but, if present, must contain 1 to 63
|
270
|
+
# characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt)
|
271
|
+
# . No more than 32 labels can be associated with a batch.
|
272
|
+
# Corresponds to the JSON property `labels`
|
273
|
+
# @return [Hash<String,String>]
|
274
|
+
attr_accessor :labels
|
275
|
+
|
276
|
+
# Output only. The resource name of the batch.
|
277
|
+
# Corresponds to the JSON property `name`
|
278
|
+
# @return [String]
|
279
|
+
attr_accessor :name
|
280
|
+
|
281
|
+
# Output only. The resource name of the operation associated with this batch.
|
282
|
+
# Corresponds to the JSON property `operation`
|
283
|
+
# @return [String]
|
284
|
+
attr_accessor :operation
|
285
|
+
|
286
|
+
# A configuration for running an Apache PySpark (https://spark.apache.org/docs/
|
287
|
+
# latest/api/python/getting_started/quickstart.html) batch workload.
|
288
|
+
# Corresponds to the JSON property `pysparkBatch`
|
289
|
+
# @return [Google::Apis::DataprocV1::PySparkBatch]
|
290
|
+
attr_accessor :pyspark_batch
|
291
|
+
|
292
|
+
# Runtime configuration for a workload.
|
293
|
+
# Corresponds to the JSON property `runtimeConfig`
|
294
|
+
# @return [Google::Apis::DataprocV1::RuntimeConfig]
|
295
|
+
attr_accessor :runtime_config
|
296
|
+
|
297
|
+
# Runtime information about workload execution.
|
298
|
+
# Corresponds to the JSON property `runtimeInfo`
|
299
|
+
# @return [Google::Apis::DataprocV1::RuntimeInfo]
|
300
|
+
attr_accessor :runtime_info
|
301
|
+
|
302
|
+
# A configuration for running an Apache Spark (https://spark.apache.org/) batch
|
303
|
+
# workload.
|
304
|
+
# Corresponds to the JSON property `sparkBatch`
|
305
|
+
# @return [Google::Apis::DataprocV1::SparkBatch]
|
306
|
+
attr_accessor :spark_batch
|
307
|
+
|
308
|
+
# A configuration for running an Apache SparkR (https://spark.apache.org/docs/
|
309
|
+
# latest/sparkr.html) batch workload.
|
310
|
+
# Corresponds to the JSON property `sparkRBatch`
|
311
|
+
# @return [Google::Apis::DataprocV1::SparkRBatch]
|
312
|
+
attr_accessor :spark_r_batch
|
313
|
+
|
314
|
+
# A configuration for running Apache Spark SQL (https://spark.apache.org/sql/)
|
315
|
+
# queries as a batch workload.
|
316
|
+
# Corresponds to the JSON property `sparkSqlBatch`
|
317
|
+
# @return [Google::Apis::DataprocV1::SparkSqlBatch]
|
318
|
+
attr_accessor :spark_sql_batch
|
319
|
+
|
320
|
+
# Output only. The state of the batch.
|
321
|
+
# Corresponds to the JSON property `state`
|
322
|
+
# @return [String]
|
323
|
+
attr_accessor :state
|
324
|
+
|
325
|
+
# Output only. Historical state information for the batch.
|
326
|
+
# Corresponds to the JSON property `stateHistory`
|
327
|
+
# @return [Array<Google::Apis::DataprocV1::StateHistory>]
|
328
|
+
attr_accessor :state_history
|
329
|
+
|
330
|
+
# Output only. Batch state details, such as a failure description if the state
|
331
|
+
# is FAILED.
|
332
|
+
# Corresponds to the JSON property `stateMessage`
|
333
|
+
# @return [String]
|
334
|
+
attr_accessor :state_message
|
335
|
+
|
336
|
+
# Output only. The time when the batch entered a current state.
|
337
|
+
# Corresponds to the JSON property `stateTime`
|
338
|
+
# @return [String]
|
339
|
+
attr_accessor :state_time
|
340
|
+
|
341
|
+
# Output only. A batch UUID (Unique Universal Identifier). The service generates
|
342
|
+
# this value when it creates the batch.
|
343
|
+
# Corresponds to the JSON property `uuid`
|
344
|
+
# @return [String]
|
345
|
+
attr_accessor :uuid
|
346
|
+
|
347
|
+
def initialize(**args)
|
348
|
+
update!(**args)
|
349
|
+
end
|
350
|
+
|
351
|
+
# Update properties of this object
|
352
|
+
def update!(**args)
|
353
|
+
@create_time = args[:create_time] if args.key?(:create_time)
|
354
|
+
@creator = args[:creator] if args.key?(:creator)
|
355
|
+
@environment_config = args[:environment_config] if args.key?(:environment_config)
|
356
|
+
@labels = args[:labels] if args.key?(:labels)
|
357
|
+
@name = args[:name] if args.key?(:name)
|
358
|
+
@operation = args[:operation] if args.key?(:operation)
|
359
|
+
@pyspark_batch = args[:pyspark_batch] if args.key?(:pyspark_batch)
|
360
|
+
@runtime_config = args[:runtime_config] if args.key?(:runtime_config)
|
361
|
+
@runtime_info = args[:runtime_info] if args.key?(:runtime_info)
|
362
|
+
@spark_batch = args[:spark_batch] if args.key?(:spark_batch)
|
363
|
+
@spark_r_batch = args[:spark_r_batch] if args.key?(:spark_r_batch)
|
364
|
+
@spark_sql_batch = args[:spark_sql_batch] if args.key?(:spark_sql_batch)
|
365
|
+
@state = args[:state] if args.key?(:state)
|
366
|
+
@state_history = args[:state_history] if args.key?(:state_history)
|
367
|
+
@state_message = args[:state_message] if args.key?(:state_message)
|
368
|
+
@state_time = args[:state_time] if args.key?(:state_time)
|
369
|
+
@uuid = args[:uuid] if args.key?(:uuid)
|
370
|
+
end
|
371
|
+
end
|
372
|
+
|
248
373
|
# Metadata describing the Batch operation.
|
249
374
|
class BatchOperationMetadata
|
250
375
|
include Google::Apis::Core::Hashable
|
@@ -884,6 +1009,14 @@ module Google
|
|
884
1009
|
# @return [String]
|
885
1010
|
attr_accessor :boot_disk_type
|
886
1011
|
|
1012
|
+
# Optional. Interface type of local SSDs (default is "scsi"). Valid values: "
|
1013
|
+
# scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express).
|
1014
|
+
# See local SSD performance (https://cloud.google.com/compute/docs/disks/local-
|
1015
|
+
# ssd#performance).
|
1016
|
+
# Corresponds to the JSON property `localSsdInterface`
|
1017
|
+
# @return [String]
|
1018
|
+
attr_accessor :local_ssd_interface
|
1019
|
+
|
887
1020
|
# Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not
|
888
1021
|
# attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.
|
889
1022
|
# apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are
|
@@ -901,6 +1034,7 @@ module Google
|
|
901
1034
|
def update!(**args)
|
902
1035
|
@boot_disk_size_gb = args[:boot_disk_size_gb] if args.key?(:boot_disk_size_gb)
|
903
1036
|
@boot_disk_type = args[:boot_disk_type] if args.key?(:boot_disk_type)
|
1037
|
+
@local_ssd_interface = args[:local_ssd_interface] if args.key?(:local_ssd_interface)
|
904
1038
|
@num_local_ssds = args[:num_local_ssds] if args.key?(:num_local_ssds)
|
905
1039
|
end
|
906
1040
|
end
|
@@ -970,6 +1104,74 @@ module Google
|
|
970
1104
|
end
|
971
1105
|
end
|
972
1106
|
|
1107
|
+
# Environment configuration for a workload.
|
1108
|
+
class EnvironmentConfig
|
1109
|
+
include Google::Apis::Core::Hashable
|
1110
|
+
|
1111
|
+
# Execution configuration for a workload.
|
1112
|
+
# Corresponds to the JSON property `executionConfig`
|
1113
|
+
# @return [Google::Apis::DataprocV1::ExecutionConfig]
|
1114
|
+
attr_accessor :execution_config
|
1115
|
+
|
1116
|
+
# Auxiliary services configuration for a workload.
|
1117
|
+
# Corresponds to the JSON property `peripheralsConfig`
|
1118
|
+
# @return [Google::Apis::DataprocV1::PeripheralsConfig]
|
1119
|
+
attr_accessor :peripherals_config
|
1120
|
+
|
1121
|
+
def initialize(**args)
|
1122
|
+
update!(**args)
|
1123
|
+
end
|
1124
|
+
|
1125
|
+
# Update properties of this object
|
1126
|
+
def update!(**args)
|
1127
|
+
@execution_config = args[:execution_config] if args.key?(:execution_config)
|
1128
|
+
@peripherals_config = args[:peripherals_config] if args.key?(:peripherals_config)
|
1129
|
+
end
|
1130
|
+
end
|
1131
|
+
|
1132
|
+
# Execution configuration for a workload.
|
1133
|
+
class ExecutionConfig
|
1134
|
+
include Google::Apis::Core::Hashable
|
1135
|
+
|
1136
|
+
# Optional. The Cloud KMS key to use for encryption.
|
1137
|
+
# Corresponds to the JSON property `kmsKey`
|
1138
|
+
# @return [String]
|
1139
|
+
attr_accessor :kms_key
|
1140
|
+
|
1141
|
+
# Optional. Tags used for network traffic control.
|
1142
|
+
# Corresponds to the JSON property `networkTags`
|
1143
|
+
# @return [Array<String>]
|
1144
|
+
attr_accessor :network_tags
|
1145
|
+
|
1146
|
+
# Optional. Network URI to connect workload to.
|
1147
|
+
# Corresponds to the JSON property `networkUri`
|
1148
|
+
# @return [String]
|
1149
|
+
attr_accessor :network_uri
|
1150
|
+
|
1151
|
+
# Optional. Service account that used to execute workload.
|
1152
|
+
# Corresponds to the JSON property `serviceAccount`
|
1153
|
+
# @return [String]
|
1154
|
+
attr_accessor :service_account
|
1155
|
+
|
1156
|
+
# Optional. Subnetwork URI to connect workload to.
|
1157
|
+
# Corresponds to the JSON property `subnetworkUri`
|
1158
|
+
# @return [String]
|
1159
|
+
attr_accessor :subnetwork_uri
|
1160
|
+
|
1161
|
+
def initialize(**args)
|
1162
|
+
update!(**args)
|
1163
|
+
end
|
1164
|
+
|
1165
|
+
# Update properties of this object
|
1166
|
+
def update!(**args)
|
1167
|
+
@kms_key = args[:kms_key] if args.key?(:kms_key)
|
1168
|
+
@network_tags = args[:network_tags] if args.key?(:network_tags)
|
1169
|
+
@network_uri = args[:network_uri] if args.key?(:network_uri)
|
1170
|
+
@service_account = args[:service_account] if args.key?(:service_account)
|
1171
|
+
@subnetwork_uri = args[:subnetwork_uri] if args.key?(:subnetwork_uri)
|
1172
|
+
end
|
1173
|
+
end
|
1174
|
+
|
973
1175
|
# Represents a textual expression in the Common Expression Language (CEL) syntax.
|
974
1176
|
# CEL is a C-like expression language. The syntax and semantics of CEL are
|
975
1177
|
# documented at https://github.com/google/cel-spec.Example (Comparison): title: "
|
@@ -1181,12 +1383,16 @@ module Google
|
|
1181
1383
|
class GetPolicyOptions
|
1182
1384
|
include Google::Apis::Core::Hashable
|
1183
1385
|
|
1184
|
-
# Optional. The policy
|
1185
|
-
# 3. Requests specifying an invalid value will be
|
1186
|
-
# with any conditional bindings must specify
|
1187
|
-
# conditional bindings may specify any valid
|
1188
|
-
#
|
1189
|
-
#
|
1386
|
+
# Optional. The maximum policy version that will be used to format the policy.
|
1387
|
+
# Valid values are 0, 1, and 3. Requests specifying an invalid value will be
|
1388
|
+
# rejected.Requests for policies with any conditional role bindings must specify
|
1389
|
+
# version 3. Policies with no conditional role bindings may specify any valid
|
1390
|
+
# value or leave the field unset.The policy in the response might use the policy
|
1391
|
+
# version that you specified, or it might use a lower policy version. For
|
1392
|
+
# example, if you specify version 3, but the policy has no conditional role
|
1393
|
+
# bindings, the response uses version 1.To learn which resources support
|
1394
|
+
# conditions in their IAM policies, see the IAM documentation (https://cloud.
|
1395
|
+
# google.com/iam/help/conditions/resource-policies).
|
1190
1396
|
# Corresponds to the JSON property `requestedPolicyVersion`
|
1191
1397
|
# @return [Fixnum]
|
1192
1398
|
attr_accessor :requested_policy_version
|
@@ -1725,7 +1931,7 @@ module Google
|
|
1725
1931
|
# @return [Google::Apis::DataprocV1::JobScheduling]
|
1726
1932
|
attr_accessor :scheduling
|
1727
1933
|
|
1728
|
-
# A Dataproc job for running Apache Spark (
|
1934
|
+
# A Dataproc job for running Apache Spark (https://spark.apache.org/)
|
1729
1935
|
# applications on YARN.
|
1730
1936
|
# Corresponds to the JSON property `sparkJob`
|
1731
1937
|
# @return [Google::Apis::DataprocV1::SparkJob]
|
@@ -1737,7 +1943,7 @@ module Google
|
|
1737
1943
|
# @return [Google::Apis::DataprocV1::SparkRJob]
|
1738
1944
|
attr_accessor :spark_r_job
|
1739
1945
|
|
1740
|
-
# A Dataproc job for running Apache Spark SQL (
|
1946
|
+
# A Dataproc job for running Apache Spark SQL (https://spark.apache.org/sql/)
|
1741
1947
|
# queries.
|
1742
1948
|
# Corresponds to the JSON property `sparkSqlJob`
|
1743
1949
|
# @return [Google::Apis::DataprocV1::SparkSqlJob]
|
@@ -1893,14 +2099,19 @@ module Google
|
|
1893
2099
|
# Optional. Maximum number of times per hour a driver may be restarted as a
|
1894
2100
|
# result of driver exiting with non-zero code before job is reported failed.A
|
1895
2101
|
# job may be reported as thrashing if driver exits with non-zero code 4 times
|
1896
|
-
# within 10 minute window.Maximum value is 10.
|
2102
|
+
# within 10 minute window.Maximum value is 10.Note: Currently, this restartable
|
2103
|
+
# job option is not supported in Dataproc workflow template (https://cloud.
|
2104
|
+
# google.com/dataproc/docs/concepts/workflows/using-workflows#
|
2105
|
+
# adding_jobs_to_a_template) jobs.
|
1897
2106
|
# Corresponds to the JSON property `maxFailuresPerHour`
|
1898
2107
|
# @return [Fixnum]
|
1899
2108
|
attr_accessor :max_failures_per_hour
|
1900
2109
|
|
1901
2110
|
# Optional. Maximum number of times in total a driver may be restarted as a
|
1902
2111
|
# result of driver exiting with non-zero code before job is reported failed.
|
1903
|
-
# Maximum value is 240.
|
2112
|
+
# Maximum value is 240.Note: Currently, this restartable job option is not
|
2113
|
+
# supported in Dataproc workflow template (https://cloud.google.com/dataproc/
|
2114
|
+
# docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
|
1904
2115
|
# Corresponds to the JSON property `maxFailuresTotal`
|
1905
2116
|
# @return [Fixnum]
|
1906
2117
|
attr_accessor :max_failures_total
|
@@ -2150,6 +2361,32 @@ module Google
|
|
2150
2361
|
end
|
2151
2362
|
end
|
2152
2363
|
|
2364
|
+
# A list of batch workloads.
|
2365
|
+
class ListBatchesResponse
|
2366
|
+
include Google::Apis::Core::Hashable
|
2367
|
+
|
2368
|
+
# The batches from the specified collection.
|
2369
|
+
# Corresponds to the JSON property `batches`
|
2370
|
+
# @return [Array<Google::Apis::DataprocV1::Batch>]
|
2371
|
+
attr_accessor :batches
|
2372
|
+
|
2373
|
+
# A token, which can be sent as page_token to retrieve the next page. If this
|
2374
|
+
# field is omitted, there are no subsequent pages.
|
2375
|
+
# Corresponds to the JSON property `nextPageToken`
|
2376
|
+
# @return [String]
|
2377
|
+
attr_accessor :next_page_token
|
2378
|
+
|
2379
|
+
def initialize(**args)
|
2380
|
+
update!(**args)
|
2381
|
+
end
|
2382
|
+
|
2383
|
+
# Update properties of this object
|
2384
|
+
def update!(**args)
|
2385
|
+
@batches = args[:batches] if args.key?(:batches)
|
2386
|
+
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
|
2387
|
+
end
|
2388
|
+
end
|
2389
|
+
|
2153
2390
|
# The list of all clusters in a project.
|
2154
2391
|
class ListClustersResponse
|
2155
2392
|
include Google::Apis::Core::Hashable
|
@@ -2562,7 +2799,7 @@ module Google
|
|
2562
2799
|
# @return [Google::Apis::DataprocV1::JobScheduling]
|
2563
2800
|
attr_accessor :scheduling
|
2564
2801
|
|
2565
|
-
# A Dataproc job for running Apache Spark (
|
2802
|
+
# A Dataproc job for running Apache Spark (https://spark.apache.org/)
|
2566
2803
|
# applications on YARN.
|
2567
2804
|
# Corresponds to the JSON property `sparkJob`
|
2568
2805
|
# @return [Google::Apis::DataprocV1::SparkJob]
|
@@ -2574,7 +2811,7 @@ module Google
|
|
2574
2811
|
# @return [Google::Apis::DataprocV1::SparkRJob]
|
2575
2812
|
attr_accessor :spark_r_job
|
2576
2813
|
|
2577
|
-
# A Dataproc job for running Apache Spark SQL (
|
2814
|
+
# A Dataproc job for running Apache Spark SQL (https://spark.apache.org/sql/)
|
2578
2815
|
# queries.
|
2579
2816
|
# Corresponds to the JSON property `sparkSqlJob`
|
2580
2817
|
# @return [Google::Apis::DataprocV1::SparkSqlJob]
|
@@ -2636,6 +2873,32 @@ module Google
|
|
2636
2873
|
end
|
2637
2874
|
end
|
2638
2875
|
|
2876
|
+
# Auxiliary services configuration for a workload.
|
2877
|
+
class PeripheralsConfig
|
2878
|
+
include Google::Apis::Core::Hashable
|
2879
|
+
|
2880
|
+
# Optional. Resource name of an existing Dataproc Metastore service.Example:
|
2881
|
+
# projects/[project_id]/locations/[region]/services/[service_id]
|
2882
|
+
# Corresponds to the JSON property `metastoreService`
|
2883
|
+
# @return [String]
|
2884
|
+
attr_accessor :metastore_service
|
2885
|
+
|
2886
|
+
# Spark History Server configuration for the workload.
|
2887
|
+
# Corresponds to the JSON property `sparkHistoryServerConfig`
|
2888
|
+
# @return [Google::Apis::DataprocV1::SparkHistoryServerConfig]
|
2889
|
+
attr_accessor :spark_history_server_config
|
2890
|
+
|
2891
|
+
def initialize(**args)
|
2892
|
+
update!(**args)
|
2893
|
+
end
|
2894
|
+
|
2895
|
+
# Update properties of this object
|
2896
|
+
def update!(**args)
|
2897
|
+
@metastore_service = args[:metastore_service] if args.key?(:metastore_service)
|
2898
|
+
@spark_history_server_config = args[:spark_history_server_config] if args.key?(:spark_history_server_config)
|
2899
|
+
end
|
2900
|
+
end
|
2901
|
+
|
2639
2902
|
# A Dataproc job for running Apache Pig (https://pig.apache.org/) queries on
|
2640
2903
|
# YARN.
|
2641
2904
|
class PigJob
|
@@ -2851,6 +3114,63 @@ module Google
|
|
2851
3114
|
end
|
2852
3115
|
end
|
2853
3116
|
|
3117
|
+
# A configuration for running an Apache PySpark (https://spark.apache.org/docs/
|
3118
|
+
# latest/api/python/getting_started/quickstart.html) batch workload.
|
3119
|
+
class PySparkBatch
|
3120
|
+
include Google::Apis::Core::Hashable
|
3121
|
+
|
3122
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory of
|
3123
|
+
# each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
|
3124
|
+
# Corresponds to the JSON property `archiveUris`
|
3125
|
+
# @return [Array<String>]
|
3126
|
+
attr_accessor :archive_uris
|
3127
|
+
|
3128
|
+
# Optional. The arguments to pass to the driver. Do not include arguments that
|
3129
|
+
# can be set as batch properties, such as --conf, since a collision can occur
|
3130
|
+
# that causes an incorrect batch submission.
|
3131
|
+
# Corresponds to the JSON property `args`
|
3132
|
+
# @return [Array<String>]
|
3133
|
+
attr_accessor :args
|
3134
|
+
|
3135
|
+
# Optional. HCFS URIs of files to be placed in the working directory of each
|
3136
|
+
# executor.
|
3137
|
+
# Corresponds to the JSON property `fileUris`
|
3138
|
+
# @return [Array<String>]
|
3139
|
+
attr_accessor :file_uris
|
3140
|
+
|
3141
|
+
# Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
|
3142
|
+
# and tasks.
|
3143
|
+
# Corresponds to the JSON property `jarFileUris`
|
3144
|
+
# @return [Array<String>]
|
3145
|
+
attr_accessor :jar_file_uris
|
3146
|
+
|
3147
|
+
# Required. The HCFS URI of the main Python file to use as the Spark driver.
|
3148
|
+
# Must be a .py file.
|
3149
|
+
# Corresponds to the JSON property `mainPythonFileUri`
|
3150
|
+
# @return [String]
|
3151
|
+
attr_accessor :main_python_file_uri
|
3152
|
+
|
3153
|
+
# Optional. HCFS file URIs of Python files to pass to the PySpark framework.
|
3154
|
+
# Supported file types: .py, .egg, and .zip.
|
3155
|
+
# Corresponds to the JSON property `pythonFileUris`
|
3156
|
+
# @return [Array<String>]
|
3157
|
+
attr_accessor :python_file_uris
|
3158
|
+
|
3159
|
+
def initialize(**args)
|
3160
|
+
update!(**args)
|
3161
|
+
end
|
3162
|
+
|
3163
|
+
# Update properties of this object
|
3164
|
+
def update!(**args)
|
3165
|
+
@archive_uris = args[:archive_uris] if args.key?(:archive_uris)
|
3166
|
+
@args = args[:args] if args.key?(:args)
|
3167
|
+
@file_uris = args[:file_uris] if args.key?(:file_uris)
|
3168
|
+
@jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
|
3169
|
+
@main_python_file_uri = args[:main_python_file_uri] if args.key?(:main_python_file_uri)
|
3170
|
+
@python_file_uris = args[:python_file_uris] if args.key?(:python_file_uris)
|
3171
|
+
end
|
3172
|
+
end
|
3173
|
+
|
2854
3174
|
# A Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/
|
2855
3175
|
# python-programming-guide.html) applications on YARN.
|
2856
3176
|
class PySparkJob
|
@@ -3030,6 +3350,72 @@ module Google
|
|
3030
3350
|
end
|
3031
3351
|
end
|
3032
3352
|
|
3353
|
+
# Runtime configuration for a workload.
|
3354
|
+
class RuntimeConfig
|
3355
|
+
include Google::Apis::Core::Hashable
|
3356
|
+
|
3357
|
+
# Optional. Optional custom container image for the job runtime environment. If
|
3358
|
+
# not specified, a default container image will be used.
|
3359
|
+
# Corresponds to the JSON property `containerImage`
|
3360
|
+
# @return [String]
|
3361
|
+
attr_accessor :container_image
|
3362
|
+
|
3363
|
+
# Optional. A mapping of property names to values, which are used to configure
|
3364
|
+
# workload execution.
|
3365
|
+
# Corresponds to the JSON property `properties`
|
3366
|
+
# @return [Hash<String,String>]
|
3367
|
+
attr_accessor :properties
|
3368
|
+
|
3369
|
+
# Optional. Version of the batch runtime.
|
3370
|
+
# Corresponds to the JSON property `version`
|
3371
|
+
# @return [String]
|
3372
|
+
attr_accessor :version
|
3373
|
+
|
3374
|
+
def initialize(**args)
|
3375
|
+
update!(**args)
|
3376
|
+
end
|
3377
|
+
|
3378
|
+
# Update properties of this object
|
3379
|
+
def update!(**args)
|
3380
|
+
@container_image = args[:container_image] if args.key?(:container_image)
|
3381
|
+
@properties = args[:properties] if args.key?(:properties)
|
3382
|
+
@version = args[:version] if args.key?(:version)
|
3383
|
+
end
|
3384
|
+
end
|
3385
|
+
|
3386
|
+
# Runtime information about workload execution.
|
3387
|
+
class RuntimeInfo
|
3388
|
+
include Google::Apis::Core::Hashable
|
3389
|
+
|
3390
|
+
# Output only. A URI pointing to the location of the diagnostics tarball.
|
3391
|
+
# Corresponds to the JSON property `diagnosticOutputUri`
|
3392
|
+
# @return [String]
|
3393
|
+
attr_accessor :diagnostic_output_uri
|
3394
|
+
|
3395
|
+
# Output only. Map of remote access endpoints (such as web interfaces and APIs)
|
3396
|
+
# to their URIs.
|
3397
|
+
# Corresponds to the JSON property `endpoints`
|
3398
|
+
# @return [Hash<String,String>]
|
3399
|
+
attr_accessor :endpoints
|
3400
|
+
|
3401
|
+
# Output only. A URI pointing to the location of the stdout and stderr of the
|
3402
|
+
# workload.
|
3403
|
+
# Corresponds to the JSON property `outputUri`
|
3404
|
+
# @return [String]
|
3405
|
+
attr_accessor :output_uri
|
3406
|
+
|
3407
|
+
def initialize(**args)
|
3408
|
+
update!(**args)
|
3409
|
+
end
|
3410
|
+
|
3411
|
+
# Update properties of this object
|
3412
|
+
def update!(**args)
|
3413
|
+
@diagnostic_output_uri = args[:diagnostic_output_uri] if args.key?(:diagnostic_output_uri)
|
3414
|
+
@endpoints = args[:endpoints] if args.key?(:endpoints)
|
3415
|
+
@output_uri = args[:output_uri] if args.key?(:output_uri)
|
3416
|
+
end
|
3417
|
+
end
|
3418
|
+
|
3033
3419
|
# Security related configuration, including encryption, Kerberos, etc.
|
3034
3420
|
class SecurityConfig
|
3035
3421
|
include Google::Apis::Core::Hashable
|
@@ -3240,7 +3626,84 @@ module Google
|
|
3240
3626
|
end
|
3241
3627
|
end
|
3242
3628
|
|
3243
|
-
# A
|
3629
|
+
# A configuration for running an Apache Spark (https://spark.apache.org/) batch
|
3630
|
+
# workload.
|
3631
|
+
class SparkBatch
|
3632
|
+
include Google::Apis::Core::Hashable
|
3633
|
+
|
3634
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory of
|
3635
|
+
# each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
|
3636
|
+
# Corresponds to the JSON property `archiveUris`
|
3637
|
+
# @return [Array<String>]
|
3638
|
+
attr_accessor :archive_uris
|
3639
|
+
|
3640
|
+
# Optional. The arguments to pass to the driver. Do not include arguments that
|
3641
|
+
# can be set as batch properties, such as --conf, since a collision can occur
|
3642
|
+
# that causes an incorrect batch submission.
|
3643
|
+
# Corresponds to the JSON property `args`
|
3644
|
+
# @return [Array<String>]
|
3645
|
+
attr_accessor :args
|
3646
|
+
|
3647
|
+
# Optional. HCFS URIs of files to be placed in the working directory of each
|
3648
|
+
# executor.
|
3649
|
+
# Corresponds to the JSON property `fileUris`
|
3650
|
+
# @return [Array<String>]
|
3651
|
+
attr_accessor :file_uris
|
3652
|
+
|
3653
|
+
# Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
|
3654
|
+
# and tasks.
|
3655
|
+
# Corresponds to the JSON property `jarFileUris`
|
3656
|
+
# @return [Array<String>]
|
3657
|
+
attr_accessor :jar_file_uris
|
3658
|
+
|
3659
|
+
# Optional. The name of the driver main class. The jar file that contains the
|
3660
|
+
# class must be in the classpath or specified in jar_file_uris.
|
3661
|
+
# Corresponds to the JSON property `mainClass`
|
3662
|
+
# @return [String]
|
3663
|
+
attr_accessor :main_class
|
3664
|
+
|
3665
|
+
# Optional. The HCFS URI of the jar file that contains the main class.
|
3666
|
+
# Corresponds to the JSON property `mainJarFileUri`
|
3667
|
+
# @return [String]
|
3668
|
+
attr_accessor :main_jar_file_uri
|
3669
|
+
|
3670
|
+
def initialize(**args)
|
3671
|
+
update!(**args)
|
3672
|
+
end
|
3673
|
+
|
3674
|
+
# Update properties of this object
|
3675
|
+
def update!(**args)
|
3676
|
+
@archive_uris = args[:archive_uris] if args.key?(:archive_uris)
|
3677
|
+
@args = args[:args] if args.key?(:args)
|
3678
|
+
@file_uris = args[:file_uris] if args.key?(:file_uris)
|
3679
|
+
@jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
|
3680
|
+
@main_class = args[:main_class] if args.key?(:main_class)
|
3681
|
+
@main_jar_file_uri = args[:main_jar_file_uri] if args.key?(:main_jar_file_uri)
|
3682
|
+
end
|
3683
|
+
end
|
3684
|
+
|
3685
|
+
# Spark History Server configuration for the workload.
|
3686
|
+
class SparkHistoryServerConfig
|
3687
|
+
include Google::Apis::Core::Hashable
|
3688
|
+
|
3689
|
+
# Optional. Resource name of an existing Dataproc Cluster to act as a Spark
|
3690
|
+
# History Server for the workload.Example: projects/[project_id]/regions/[region]
|
3691
|
+
# /clusters/[cluster_name]
|
3692
|
+
# Corresponds to the JSON property `dataprocCluster`
|
3693
|
+
# @return [String]
|
3694
|
+
attr_accessor :dataproc_cluster
|
3695
|
+
|
3696
|
+
def initialize(**args)
|
3697
|
+
update!(**args)
|
3698
|
+
end
|
3699
|
+
|
3700
|
+
# Update properties of this object
|
3701
|
+
def update!(**args)
|
3702
|
+
@dataproc_cluster = args[:dataproc_cluster] if args.key?(:dataproc_cluster)
|
3703
|
+
end
|
3704
|
+
end
|
3705
|
+
|
3706
|
+
# A Dataproc job for running Apache Spark (https://spark.apache.org/)
|
3244
3707
|
# applications on YARN.
|
3245
3708
|
class SparkJob
|
3246
3709
|
include Google::Apis::Core::Hashable
|
@@ -3311,6 +3774,49 @@ module Google
|
|
3311
3774
|
end
|
3312
3775
|
end
|
3313
3776
|
|
3777
|
+
# A configuration for running an Apache SparkR (https://spark.apache.org/docs/
|
3778
|
+
# latest/sparkr.html) batch workload.
|
3779
|
+
class SparkRBatch
|
3780
|
+
include Google::Apis::Core::Hashable
|
3781
|
+
|
3782
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory of
|
3783
|
+
# each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
|
3784
|
+
# Corresponds to the JSON property `archiveUris`
|
3785
|
+
# @return [Array<String>]
|
3786
|
+
attr_accessor :archive_uris
|
3787
|
+
|
3788
|
+
# Optional. The arguments to pass to the Spark driver. Do not include arguments
|
3789
|
+
# that can be set as batch properties, such as --conf, since a collision can
|
3790
|
+
# occur that causes an incorrect batch submission.
|
3791
|
+
# Corresponds to the JSON property `args`
|
3792
|
+
# @return [Array<String>]
|
3793
|
+
attr_accessor :args
|
3794
|
+
|
3795
|
+
# Optional. HCFS URIs of files to be placed in the working directory of each
|
3796
|
+
# executor.
|
3797
|
+
# Corresponds to the JSON property `fileUris`
|
3798
|
+
# @return [Array<String>]
|
3799
|
+
attr_accessor :file_uris
|
3800
|
+
|
3801
|
+
# Required. The HCFS URI of the main R file to use as the driver. Must be a .R
|
3802
|
+
# or .r file.
|
3803
|
+
# Corresponds to the JSON property `mainRFileUri`
|
3804
|
+
# @return [String]
|
3805
|
+
attr_accessor :main_r_file_uri
|
3806
|
+
|
3807
|
+
def initialize(**args)
|
3808
|
+
update!(**args)
|
3809
|
+
end
|
3810
|
+
|
3811
|
+
# Update properties of this object
|
3812
|
+
def update!(**args)
|
3813
|
+
@archive_uris = args[:archive_uris] if args.key?(:archive_uris)
|
3814
|
+
@args = args[:args] if args.key?(:args)
|
3815
|
+
@file_uris = args[:file_uris] if args.key?(:file_uris)
|
3816
|
+
@main_r_file_uri = args[:main_r_file_uri] if args.key?(:main_r_file_uri)
|
3817
|
+
end
|
3818
|
+
end
|
3819
|
+
|
3314
3820
|
# A Dataproc job for running Apache SparkR (https://spark.apache.org/docs/latest/
|
3315
3821
|
# sparkr.html) applications on YARN.
|
3316
3822
|
class SparkRJob
|
@@ -3369,7 +3875,41 @@ module Google
|
|
3369
3875
|
end
|
3370
3876
|
end
|
3371
3877
|
|
3372
|
-
# A
|
3878
|
+
# A configuration for running Apache Spark SQL (https://spark.apache.org/sql/)
|
3879
|
+
# queries as a batch workload.
|
3880
|
+
class SparkSqlBatch
|
3881
|
+
include Google::Apis::Core::Hashable
|
3882
|
+
|
3883
|
+
# Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
|
3884
|
+
# Corresponds to the JSON property `jarFileUris`
|
3885
|
+
# @return [Array<String>]
|
3886
|
+
attr_accessor :jar_file_uris
|
3887
|
+
|
3888
|
+
# Required. The HCFS URI of the script that contains Spark SQL queries to
|
3889
|
+
# execute.
|
3890
|
+
# Corresponds to the JSON property `queryFileUri`
|
3891
|
+
# @return [String]
|
3892
|
+
attr_accessor :query_file_uri
|
3893
|
+
|
3894
|
+
# Optional. Mapping of query variable names to values (equivalent to the Spark
|
3895
|
+
# SQL command: SET name="value";).
|
3896
|
+
# Corresponds to the JSON property `queryVariables`
|
3897
|
+
# @return [Hash<String,String>]
|
3898
|
+
attr_accessor :query_variables
|
3899
|
+
|
3900
|
+
def initialize(**args)
|
3901
|
+
update!(**args)
|
3902
|
+
end
|
3903
|
+
|
3904
|
+
# Update properties of this object
|
3905
|
+
def update!(**args)
|
3906
|
+
@jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
|
3907
|
+
@query_file_uri = args[:query_file_uri] if args.key?(:query_file_uri)
|
3908
|
+
@query_variables = args[:query_variables] if args.key?(:query_variables)
|
3909
|
+
end
|
3910
|
+
end
|
3911
|
+
|
3912
|
+
# A Dataproc job for running Apache Spark SQL (https://spark.apache.org/sql/)
|
3373
3913
|
# queries.
|
3374
3914
|
class SparkSqlJob
|
3375
3915
|
include Google::Apis::Core::Hashable
|
@@ -3518,6 +4058,37 @@ module Google
|
|
3518
4058
|
end
|
3519
4059
|
end
|
3520
4060
|
|
4061
|
+
# Historical state information.
|
4062
|
+
class StateHistory
|
4063
|
+
include Google::Apis::Core::Hashable
|
4064
|
+
|
4065
|
+
# Output only. The state of the batch at this point in history.
|
4066
|
+
# Corresponds to the JSON property `state`
|
4067
|
+
# @return [String]
|
4068
|
+
attr_accessor :state
|
4069
|
+
|
4070
|
+
# Output only. Details about the state at this point in history.
|
4071
|
+
# Corresponds to the JSON property `stateMessage`
|
4072
|
+
# @return [String]
|
4073
|
+
attr_accessor :state_message
|
4074
|
+
|
4075
|
+
# Output only. The time when the batch entered the historical state.
|
4076
|
+
# Corresponds to the JSON property `stateStartTime`
|
4077
|
+
# @return [String]
|
4078
|
+
attr_accessor :state_start_time
|
4079
|
+
|
4080
|
+
def initialize(**args)
|
4081
|
+
update!(**args)
|
4082
|
+
end
|
4083
|
+
|
4084
|
+
# Update properties of this object
|
4085
|
+
def update!(**args)
|
4086
|
+
@state = args[:state] if args.key?(:state)
|
4087
|
+
@state_message = args[:state_message] if args.key?(:state_message)
|
4088
|
+
@state_start_time = args[:state_start_time] if args.key?(:state_start_time)
|
4089
|
+
end
|
4090
|
+
end
|
4091
|
+
|
3521
4092
|
# The Status type defines a logical error model that is suitable for different
|
3522
4093
|
# programming environments, including REST APIs and RPC APIs. It is used by gRPC
|
3523
4094
|
# (https://github.com/grpc). Each Status message contains three pieces of data:
|
@@ -16,13 +16,13 @@ module Google
|
|
16
16
|
module Apis
|
17
17
|
module DataprocV1
|
18
18
|
# Version of the google-apis-dataproc_v1 gem
|
19
|
-
GEM_VERSION = "0.
|
19
|
+
GEM_VERSION = "0.20.0"
|
20
20
|
|
21
21
|
# Version of the code generator used to generate this client
|
22
|
-
GENERATOR_VERSION = "0.4.
|
22
|
+
GENERATOR_VERSION = "0.4.1"
|
23
23
|
|
24
24
|
# Revision of the discovery document this client was generated from
|
25
|
-
REVISION = "
|
25
|
+
REVISION = "20220122"
|
26
26
|
end
|
27
27
|
end
|
28
28
|
end
|
@@ -52,6 +52,12 @@ module Google
|
|
52
52
|
include Google::Apis::Core::JsonObjectSupport
|
53
53
|
end
|
54
54
|
|
55
|
+
class Batch
|
56
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
57
|
+
|
58
|
+
include Google::Apis::Core::JsonObjectSupport
|
59
|
+
end
|
60
|
+
|
55
61
|
class BatchOperationMetadata
|
56
62
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
57
63
|
|
@@ -160,6 +166,18 @@ module Google
|
|
160
166
|
include Google::Apis::Core::JsonObjectSupport
|
161
167
|
end
|
162
168
|
|
169
|
+
class EnvironmentConfig
|
170
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
171
|
+
|
172
|
+
include Google::Apis::Core::JsonObjectSupport
|
173
|
+
end
|
174
|
+
|
175
|
+
class ExecutionConfig
|
176
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
177
|
+
|
178
|
+
include Google::Apis::Core::JsonObjectSupport
|
179
|
+
end
|
180
|
+
|
163
181
|
class Expr
|
164
182
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
165
183
|
|
@@ -292,6 +310,12 @@ module Google
|
|
292
310
|
include Google::Apis::Core::JsonObjectSupport
|
293
311
|
end
|
294
312
|
|
313
|
+
class ListBatchesResponse
|
314
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
315
|
+
|
316
|
+
include Google::Apis::Core::JsonObjectSupport
|
317
|
+
end
|
318
|
+
|
295
319
|
class ListClustersResponse
|
296
320
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
297
321
|
|
@@ -376,6 +400,12 @@ module Google
|
|
376
400
|
include Google::Apis::Core::JsonObjectSupport
|
377
401
|
end
|
378
402
|
|
403
|
+
class PeripheralsConfig
|
404
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
405
|
+
|
406
|
+
include Google::Apis::Core::JsonObjectSupport
|
407
|
+
end
|
408
|
+
|
379
409
|
class PigJob
|
380
410
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
381
411
|
|
@@ -394,6 +424,12 @@ module Google
|
|
394
424
|
include Google::Apis::Core::JsonObjectSupport
|
395
425
|
end
|
396
426
|
|
427
|
+
class PySparkBatch
|
428
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
429
|
+
|
430
|
+
include Google::Apis::Core::JsonObjectSupport
|
431
|
+
end
|
432
|
+
|
397
433
|
class PySparkJob
|
398
434
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
399
435
|
|
@@ -424,6 +460,18 @@ module Google
|
|
424
460
|
include Google::Apis::Core::JsonObjectSupport
|
425
461
|
end
|
426
462
|
|
463
|
+
class RuntimeConfig
|
464
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
465
|
+
|
466
|
+
include Google::Apis::Core::JsonObjectSupport
|
467
|
+
end
|
468
|
+
|
469
|
+
class RuntimeInfo
|
470
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
471
|
+
|
472
|
+
include Google::Apis::Core::JsonObjectSupport
|
473
|
+
end
|
474
|
+
|
427
475
|
class SecurityConfig
|
428
476
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
429
477
|
|
@@ -454,18 +502,42 @@ module Google
|
|
454
502
|
include Google::Apis::Core::JsonObjectSupport
|
455
503
|
end
|
456
504
|
|
505
|
+
class SparkBatch
|
506
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
507
|
+
|
508
|
+
include Google::Apis::Core::JsonObjectSupport
|
509
|
+
end
|
510
|
+
|
511
|
+
class SparkHistoryServerConfig
|
512
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
513
|
+
|
514
|
+
include Google::Apis::Core::JsonObjectSupport
|
515
|
+
end
|
516
|
+
|
457
517
|
class SparkJob
|
458
518
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
459
519
|
|
460
520
|
include Google::Apis::Core::JsonObjectSupport
|
461
521
|
end
|
462
522
|
|
523
|
+
class SparkRBatch
|
524
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
525
|
+
|
526
|
+
include Google::Apis::Core::JsonObjectSupport
|
527
|
+
end
|
528
|
+
|
463
529
|
class SparkRJob
|
464
530
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
465
531
|
|
466
532
|
include Google::Apis::Core::JsonObjectSupport
|
467
533
|
end
|
468
534
|
|
535
|
+
class SparkSqlBatch
|
536
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
537
|
+
|
538
|
+
include Google::Apis::Core::JsonObjectSupport
|
539
|
+
end
|
540
|
+
|
469
541
|
class SparkSqlJob
|
470
542
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
471
543
|
|
@@ -484,6 +556,12 @@ module Google
|
|
484
556
|
include Google::Apis::Core::JsonObjectSupport
|
485
557
|
end
|
486
558
|
|
559
|
+
class StateHistory
|
560
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
561
|
+
|
562
|
+
include Google::Apis::Core::JsonObjectSupport
|
563
|
+
end
|
564
|
+
|
487
565
|
class Status
|
488
566
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
489
567
|
|
@@ -614,6 +692,37 @@ module Google
|
|
614
692
|
end
|
615
693
|
end
|
616
694
|
|
695
|
+
class Batch
|
696
|
+
# @private
|
697
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
698
|
+
property :create_time, as: 'createTime'
|
699
|
+
property :creator, as: 'creator'
|
700
|
+
property :environment_config, as: 'environmentConfig', class: Google::Apis::DataprocV1::EnvironmentConfig, decorator: Google::Apis::DataprocV1::EnvironmentConfig::Representation
|
701
|
+
|
702
|
+
hash :labels, as: 'labels'
|
703
|
+
property :name, as: 'name'
|
704
|
+
property :operation, as: 'operation'
|
705
|
+
property :pyspark_batch, as: 'pysparkBatch', class: Google::Apis::DataprocV1::PySparkBatch, decorator: Google::Apis::DataprocV1::PySparkBatch::Representation
|
706
|
+
|
707
|
+
property :runtime_config, as: 'runtimeConfig', class: Google::Apis::DataprocV1::RuntimeConfig, decorator: Google::Apis::DataprocV1::RuntimeConfig::Representation
|
708
|
+
|
709
|
+
property :runtime_info, as: 'runtimeInfo', class: Google::Apis::DataprocV1::RuntimeInfo, decorator: Google::Apis::DataprocV1::RuntimeInfo::Representation
|
710
|
+
|
711
|
+
property :spark_batch, as: 'sparkBatch', class: Google::Apis::DataprocV1::SparkBatch, decorator: Google::Apis::DataprocV1::SparkBatch::Representation
|
712
|
+
|
713
|
+
property :spark_r_batch, as: 'sparkRBatch', class: Google::Apis::DataprocV1::SparkRBatch, decorator: Google::Apis::DataprocV1::SparkRBatch::Representation
|
714
|
+
|
715
|
+
property :spark_sql_batch, as: 'sparkSqlBatch', class: Google::Apis::DataprocV1::SparkSqlBatch, decorator: Google::Apis::DataprocV1::SparkSqlBatch::Representation
|
716
|
+
|
717
|
+
property :state, as: 'state'
|
718
|
+
collection :state_history, as: 'stateHistory', class: Google::Apis::DataprocV1::StateHistory, decorator: Google::Apis::DataprocV1::StateHistory::Representation
|
719
|
+
|
720
|
+
property :state_message, as: 'stateMessage'
|
721
|
+
property :state_time, as: 'stateTime'
|
722
|
+
property :uuid, as: 'uuid'
|
723
|
+
end
|
724
|
+
end
|
725
|
+
|
617
726
|
class BatchOperationMetadata
|
618
727
|
# @private
|
619
728
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -782,6 +891,7 @@ module Google
|
|
782
891
|
class Representation < Google::Apis::Core::JsonRepresentation
|
783
892
|
property :boot_disk_size_gb, as: 'bootDiskSizeGb'
|
784
893
|
property :boot_disk_type, as: 'bootDiskType'
|
894
|
+
property :local_ssd_interface, as: 'localSsdInterface'
|
785
895
|
property :num_local_ssds, as: 'numLocalSsds'
|
786
896
|
end
|
787
897
|
end
|
@@ -807,6 +917,27 @@ module Google
|
|
807
917
|
end
|
808
918
|
end
|
809
919
|
|
920
|
+
class EnvironmentConfig
|
921
|
+
# @private
|
922
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
923
|
+
property :execution_config, as: 'executionConfig', class: Google::Apis::DataprocV1::ExecutionConfig, decorator: Google::Apis::DataprocV1::ExecutionConfig::Representation
|
924
|
+
|
925
|
+
property :peripherals_config, as: 'peripheralsConfig', class: Google::Apis::DataprocV1::PeripheralsConfig, decorator: Google::Apis::DataprocV1::PeripheralsConfig::Representation
|
926
|
+
|
927
|
+
end
|
928
|
+
end
|
929
|
+
|
930
|
+
class ExecutionConfig
|
931
|
+
# @private
|
932
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
933
|
+
property :kms_key, as: 'kmsKey'
|
934
|
+
collection :network_tags, as: 'networkTags'
|
935
|
+
property :network_uri, as: 'networkUri'
|
936
|
+
property :service_account, as: 'serviceAccount'
|
937
|
+
property :subnetwork_uri, as: 'subnetworkUri'
|
938
|
+
end
|
939
|
+
end
|
940
|
+
|
810
941
|
class Expr
|
811
942
|
# @private
|
812
943
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1080,6 +1211,15 @@ module Google
|
|
1080
1211
|
end
|
1081
1212
|
end
|
1082
1213
|
|
1214
|
+
class ListBatchesResponse
|
1215
|
+
# @private
|
1216
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1217
|
+
collection :batches, as: 'batches', class: Google::Apis::DataprocV1::Batch, decorator: Google::Apis::DataprocV1::Batch::Representation
|
1218
|
+
|
1219
|
+
property :next_page_token, as: 'nextPageToken'
|
1220
|
+
end
|
1221
|
+
end
|
1222
|
+
|
1083
1223
|
class ListClustersResponse
|
1084
1224
|
# @private
|
1085
1225
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1220,6 +1360,15 @@ module Google
|
|
1220
1360
|
end
|
1221
1361
|
end
|
1222
1362
|
|
1363
|
+
class PeripheralsConfig
|
1364
|
+
# @private
|
1365
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1366
|
+
property :metastore_service, as: 'metastoreService'
|
1367
|
+
property :spark_history_server_config, as: 'sparkHistoryServerConfig', class: Google::Apis::DataprocV1::SparkHistoryServerConfig, decorator: Google::Apis::DataprocV1::SparkHistoryServerConfig::Representation
|
1368
|
+
|
1369
|
+
end
|
1370
|
+
end
|
1371
|
+
|
1223
1372
|
class PigJob
|
1224
1373
|
# @private
|
1225
1374
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1260,6 +1409,18 @@ module Google
|
|
1260
1409
|
end
|
1261
1410
|
end
|
1262
1411
|
|
1412
|
+
class PySparkBatch
|
1413
|
+
# @private
|
1414
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1415
|
+
collection :archive_uris, as: 'archiveUris'
|
1416
|
+
collection :args, as: 'args'
|
1417
|
+
collection :file_uris, as: 'fileUris'
|
1418
|
+
collection :jar_file_uris, as: 'jarFileUris'
|
1419
|
+
property :main_python_file_uri, as: 'mainPythonFileUri'
|
1420
|
+
collection :python_file_uris, as: 'pythonFileUris'
|
1421
|
+
end
|
1422
|
+
end
|
1423
|
+
|
1263
1424
|
class PySparkJob
|
1264
1425
|
# @private
|
1265
1426
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1306,6 +1467,24 @@ module Google
|
|
1306
1467
|
end
|
1307
1468
|
end
|
1308
1469
|
|
1470
|
+
class RuntimeConfig
|
1471
|
+
# @private
|
1472
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1473
|
+
property :container_image, as: 'containerImage'
|
1474
|
+
hash :properties, as: 'properties'
|
1475
|
+
property :version, as: 'version'
|
1476
|
+
end
|
1477
|
+
end
|
1478
|
+
|
1479
|
+
class RuntimeInfo
|
1480
|
+
# @private
|
1481
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1482
|
+
property :diagnostic_output_uri, as: 'diagnosticOutputUri'
|
1483
|
+
hash :endpoints, as: 'endpoints'
|
1484
|
+
property :output_uri, as: 'outputUri'
|
1485
|
+
end
|
1486
|
+
end
|
1487
|
+
|
1309
1488
|
class SecurityConfig
|
1310
1489
|
# @private
|
1311
1490
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1356,6 +1535,25 @@ module Google
|
|
1356
1535
|
end
|
1357
1536
|
end
|
1358
1537
|
|
1538
|
+
class SparkBatch
|
1539
|
+
# @private
|
1540
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1541
|
+
collection :archive_uris, as: 'archiveUris'
|
1542
|
+
collection :args, as: 'args'
|
1543
|
+
collection :file_uris, as: 'fileUris'
|
1544
|
+
collection :jar_file_uris, as: 'jarFileUris'
|
1545
|
+
property :main_class, as: 'mainClass'
|
1546
|
+
property :main_jar_file_uri, as: 'mainJarFileUri'
|
1547
|
+
end
|
1548
|
+
end
|
1549
|
+
|
1550
|
+
class SparkHistoryServerConfig
|
1551
|
+
# @private
|
1552
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1553
|
+
property :dataproc_cluster, as: 'dataprocCluster'
|
1554
|
+
end
|
1555
|
+
end
|
1556
|
+
|
1359
1557
|
class SparkJob
|
1360
1558
|
# @private
|
1361
1559
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1371,6 +1569,16 @@ module Google
|
|
1371
1569
|
end
|
1372
1570
|
end
|
1373
1571
|
|
1572
|
+
class SparkRBatch
|
1573
|
+
# @private
|
1574
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1575
|
+
collection :archive_uris, as: 'archiveUris'
|
1576
|
+
collection :args, as: 'args'
|
1577
|
+
collection :file_uris, as: 'fileUris'
|
1578
|
+
property :main_r_file_uri, as: 'mainRFileUri'
|
1579
|
+
end
|
1580
|
+
end
|
1581
|
+
|
1374
1582
|
class SparkRJob
|
1375
1583
|
# @private
|
1376
1584
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1384,6 +1592,15 @@ module Google
|
|
1384
1592
|
end
|
1385
1593
|
end
|
1386
1594
|
|
1595
|
+
class SparkSqlBatch
|
1596
|
+
# @private
|
1597
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1598
|
+
collection :jar_file_uris, as: 'jarFileUris'
|
1599
|
+
property :query_file_uri, as: 'queryFileUri'
|
1600
|
+
hash :query_variables, as: 'queryVariables'
|
1601
|
+
end
|
1602
|
+
end
|
1603
|
+
|
1387
1604
|
class SparkSqlJob
|
1388
1605
|
# @private
|
1389
1606
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1417,6 +1634,15 @@ module Google
|
|
1417
1634
|
end
|
1418
1635
|
end
|
1419
1636
|
|
1637
|
+
class StateHistory
|
1638
|
+
# @private
|
1639
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1640
|
+
property :state, as: 'state'
|
1641
|
+
property :state_message, as: 'stateMessage'
|
1642
|
+
property :state_start_time, as: 'stateStartTime'
|
1643
|
+
end
|
1644
|
+
end
|
1645
|
+
|
1420
1646
|
class Status
|
1421
1647
|
# @private
|
1422
1648
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -352,11 +352,160 @@ module Google
|
|
352
352
|
execute_or_queue_command(command, &block)
|
353
353
|
end
|
354
354
|
|
355
|
+
# Creates a batch workload that executes asynchronously.
|
356
|
+
# @param [String] parent
|
357
|
+
# Required. The parent resource where this batch will be created.
|
358
|
+
# @param [Google::Apis::DataprocV1::Batch] batch_object
|
359
|
+
# @param [String] batch_id
|
360
|
+
# Optional. The ID to use for the batch, which will become the final component
|
361
|
+
# of the batch's resource name.This value must be 4-63 characters. Valid
|
362
|
+
# characters are /[a-z][0-9]-/.
|
363
|
+
# @param [String] request_id
|
364
|
+
# Optional. A unique ID used to identify the request. If the service receives
|
365
|
+
# two CreateBatchRequest (https://cloud.google.com/dataproc/docs/reference/rpc/
|
366
|
+
# google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateBatchRequest)s with
|
367
|
+
# the same request_id, the second request is ignored and the Operation that
|
368
|
+
# corresponds to the first Batch created and stored in the backend is returned.
|
369
|
+
# Recommendation: Set this value to a UUID (https://en.wikipedia.org/wiki/
|
370
|
+
# Universally_unique_identifier).The value must contain only letters (a-z, A-Z),
|
371
|
+
# numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40
|
372
|
+
# characters.
|
373
|
+
# @param [String] fields
|
374
|
+
# Selector specifying which fields to include in a partial response.
|
375
|
+
# @param [String] quota_user
|
376
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
377
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
378
|
+
# @param [Google::Apis::RequestOptions] options
|
379
|
+
# Request-specific options
|
380
|
+
#
|
381
|
+
# @yield [result, err] Result & error if block supplied
|
382
|
+
# @yieldparam result [Google::Apis::DataprocV1::Operation] parsed result object
|
383
|
+
# @yieldparam err [StandardError] error object if request failed
|
384
|
+
#
|
385
|
+
# @return [Google::Apis::DataprocV1::Operation]
|
386
|
+
#
|
387
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
388
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
389
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
390
|
+
def create_project_location_batch(parent, batch_object = nil, batch_id: nil, request_id: nil, fields: nil, quota_user: nil, options: nil, &block)
|
391
|
+
command = make_simple_command(:post, 'v1/{+parent}/batches', options)
|
392
|
+
command.request_representation = Google::Apis::DataprocV1::Batch::Representation
|
393
|
+
command.request_object = batch_object
|
394
|
+
command.response_representation = Google::Apis::DataprocV1::Operation::Representation
|
395
|
+
command.response_class = Google::Apis::DataprocV1::Operation
|
396
|
+
command.params['parent'] = parent unless parent.nil?
|
397
|
+
command.query['batchId'] = batch_id unless batch_id.nil?
|
398
|
+
command.query['requestId'] = request_id unless request_id.nil?
|
399
|
+
command.query['fields'] = fields unless fields.nil?
|
400
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
401
|
+
execute_or_queue_command(command, &block)
|
402
|
+
end
|
403
|
+
|
404
|
+
# Deletes the batch workload resource. If the batch is not in terminal state,
|
405
|
+
# the delete fails and the response returns FAILED_PRECONDITION.
|
406
|
+
# @param [String] name
|
407
|
+
# Required. The name of the batch resource to delete.
|
408
|
+
# @param [String] fields
|
409
|
+
# Selector specifying which fields to include in a partial response.
|
410
|
+
# @param [String] quota_user
|
411
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
412
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
413
|
+
# @param [Google::Apis::RequestOptions] options
|
414
|
+
# Request-specific options
|
415
|
+
#
|
416
|
+
# @yield [result, err] Result & error if block supplied
|
417
|
+
# @yieldparam result [Google::Apis::DataprocV1::Empty] parsed result object
|
418
|
+
# @yieldparam err [StandardError] error object if request failed
|
419
|
+
#
|
420
|
+
# @return [Google::Apis::DataprocV1::Empty]
|
421
|
+
#
|
422
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
423
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
424
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
425
|
+
def delete_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
|
426
|
+
command = make_simple_command(:delete, 'v1/{+name}', options)
|
427
|
+
command.response_representation = Google::Apis::DataprocV1::Empty::Representation
|
428
|
+
command.response_class = Google::Apis::DataprocV1::Empty
|
429
|
+
command.params['name'] = name unless name.nil?
|
430
|
+
command.query['fields'] = fields unless fields.nil?
|
431
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
432
|
+
execute_or_queue_command(command, &block)
|
433
|
+
end
|
434
|
+
|
435
|
+
# Gets the batch workload resource representation.
|
436
|
+
# @param [String] name
|
437
|
+
# Required. The name of the batch to retrieve.
|
438
|
+
# @param [String] fields
|
439
|
+
# Selector specifying which fields to include in a partial response.
|
440
|
+
# @param [String] quota_user
|
441
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
442
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
443
|
+
# @param [Google::Apis::RequestOptions] options
|
444
|
+
# Request-specific options
|
445
|
+
#
|
446
|
+
# @yield [result, err] Result & error if block supplied
|
447
|
+
# @yieldparam result [Google::Apis::DataprocV1::Batch] parsed result object
|
448
|
+
# @yieldparam err [StandardError] error object if request failed
|
449
|
+
#
|
450
|
+
# @return [Google::Apis::DataprocV1::Batch]
|
451
|
+
#
|
452
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
453
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
454
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
455
|
+
def get_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
|
456
|
+
command = make_simple_command(:get, 'v1/{+name}', options)
|
457
|
+
command.response_representation = Google::Apis::DataprocV1::Batch::Representation
|
458
|
+
command.response_class = Google::Apis::DataprocV1::Batch
|
459
|
+
command.params['name'] = name unless name.nil?
|
460
|
+
command.query['fields'] = fields unless fields.nil?
|
461
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
462
|
+
execute_or_queue_command(command, &block)
|
463
|
+
end
|
464
|
+
|
465
|
+
# Lists batch workloads.
|
466
|
+
# @param [String] parent
|
467
|
+
# Required. The parent, which owns this collection of batches.
|
468
|
+
# @param [Fixnum] page_size
|
469
|
+
# Optional. The maximum number of batches to return in each response. The
|
470
|
+
# service may return fewer than this value. The default page size is 20; the
|
471
|
+
# maximum page size is 1000.
|
472
|
+
# @param [String] page_token
|
473
|
+
# Optional. A page token received from a previous ListBatches call. Provide this
|
474
|
+
# token to retrieve the subsequent page.
|
475
|
+
# @param [String] fields
|
476
|
+
# Selector specifying which fields to include in a partial response.
|
477
|
+
# @param [String] quota_user
|
478
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
479
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
480
|
+
# @param [Google::Apis::RequestOptions] options
|
481
|
+
# Request-specific options
|
482
|
+
#
|
483
|
+
# @yield [result, err] Result & error if block supplied
|
484
|
+
# @yieldparam result [Google::Apis::DataprocV1::ListBatchesResponse] parsed result object
|
485
|
+
# @yieldparam err [StandardError] error object if request failed
|
486
|
+
#
|
487
|
+
# @return [Google::Apis::DataprocV1::ListBatchesResponse]
|
488
|
+
#
|
489
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
490
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
491
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
492
|
+
def list_project_location_batches(parent, page_size: nil, page_token: nil, fields: nil, quota_user: nil, options: nil, &block)
|
493
|
+
command = make_simple_command(:get, 'v1/{+parent}/batches', options)
|
494
|
+
command.response_representation = Google::Apis::DataprocV1::ListBatchesResponse::Representation
|
495
|
+
command.response_class = Google::Apis::DataprocV1::ListBatchesResponse
|
496
|
+
command.params['parent'] = parent unless parent.nil?
|
497
|
+
command.query['pageSize'] = page_size unless page_size.nil?
|
498
|
+
command.query['pageToken'] = page_token unless page_token.nil?
|
499
|
+
command.query['fields'] = fields unless fields.nil?
|
500
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
501
|
+
execute_or_queue_command(command, &block)
|
502
|
+
end
|
503
|
+
|
355
504
|
# Creates new workflow template.
|
356
505
|
# @param [String] parent
|
357
506
|
# Required. The resource name of the region or location, as described in https://
|
358
507
|
# cloud.google.com/apis/design/resource_names. For projects.regions.
|
359
|
-
# workflowTemplates
|
508
|
+
# workflowTemplates.create, the resource name of the region has the following
|
360
509
|
# format: projects/`project_id`/regions/`region` For projects.locations.
|
361
510
|
# workflowTemplates.create, the resource name of the location has the following
|
362
511
|
# format: projects/`project_id`/locations/`location`
|
@@ -2317,7 +2466,7 @@ module Google
|
|
2317
2466
|
# @param [String] parent
|
2318
2467
|
# Required. The resource name of the region or location, as described in https://
|
2319
2468
|
# cloud.google.com/apis/design/resource_names. For projects.regions.
|
2320
|
-
# workflowTemplates
|
2469
|
+
# workflowTemplates.create, the resource name of the region has the following
|
2321
2470
|
# format: projects/`project_id`/regions/`region` For projects.locations.
|
2322
2471
|
# workflowTemplates.create, the resource name of the location has the following
|
2323
2472
|
# format: projects/`project_id`/locations/`location`
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-apis-dataproc_v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.20.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2022-01-31 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: google-apis-core
|
@@ -58,7 +58,7 @@ licenses:
|
|
58
58
|
metadata:
|
59
59
|
bug_tracker_uri: https://github.com/googleapis/google-api-ruby-client/issues
|
60
60
|
changelog_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1/CHANGELOG.md
|
61
|
-
documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.
|
61
|
+
documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.20.0
|
62
62
|
source_code_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1
|
63
63
|
post_install_message:
|
64
64
|
rdoc_options: []
|
@@ -75,7 +75,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
75
75
|
- !ruby/object:Gem::Version
|
76
76
|
version: '0'
|
77
77
|
requirements: []
|
78
|
-
rubygems_version: 3.
|
78
|
+
rubygems_version: 3.3.5
|
79
79
|
signing_key:
|
80
80
|
specification_version: 4
|
81
81
|
summary: Simple REST client for Cloud Dataproc API V1
|