google-apis-dataproc_v1 0.18.0 → 0.19.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: b49e139fab6c34e0bd6707069e53c4105077e7abe1dc22ba91fe76124fc9410e
|
4
|
+
data.tar.gz: 4f11bed732911ec54b23d5d6eb1189e789e35b858c9abb20fe1b9f7bc4b8ac8b
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 74cd3145e6b509ee3ce020ad5d016f2ffc30e53505ac4c70b70b910ae336f694861c29331e4e681bb5f7efdb3126aa4dd027631ab2b29f417dfcdbeedc49f362
|
7
|
+
data.tar.gz: 0dc5eccb08bb472d8adcabbb54510b287c02eb923f359c365526c92326ccfc17331b19582180bde9b1f45767868af314c1e1b6350ff953040bedd9f351a42451
|
data/CHANGELOG.md
CHANGED
@@ -1,5 +1,10 @@
|
|
1
1
|
# Release history for google-apis-dataproc_v1
|
2
2
|
|
3
|
+
### v0.19.0 (2022-01-14)
|
4
|
+
|
5
|
+
* Regenerated from discovery document revision 20220111
|
6
|
+
* Regenerated using generator version 0.4.1
|
7
|
+
|
3
8
|
### v0.18.0 (2022-01-07)
|
4
9
|
|
5
10
|
* Regenerated from discovery document revision 20220105
|
@@ -245,6 +245,131 @@ module Google
|
|
245
245
|
end
|
246
246
|
end
|
247
247
|
|
248
|
+
# A representation of a batch workload in the service.
|
249
|
+
class Batch
|
250
|
+
include Google::Apis::Core::Hashable
|
251
|
+
|
252
|
+
# Output only. The time when the batch was created.
|
253
|
+
# Corresponds to the JSON property `createTime`
|
254
|
+
# @return [String]
|
255
|
+
attr_accessor :create_time
|
256
|
+
|
257
|
+
# Output only. The email address of the user who created the batch.
|
258
|
+
# Corresponds to the JSON property `creator`
|
259
|
+
# @return [String]
|
260
|
+
attr_accessor :creator
|
261
|
+
|
262
|
+
# Environment configuration for a workload.
|
263
|
+
# Corresponds to the JSON property `environmentConfig`
|
264
|
+
# @return [Google::Apis::DataprocV1::EnvironmentConfig]
|
265
|
+
attr_accessor :environment_config
|
266
|
+
|
267
|
+
# Optional. The labels to associate with this batch. Label keys must contain 1
|
268
|
+
# to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/
|
269
|
+
# rfc1035.txt). Label values may be empty, but, if present, must contain 1 to 63
|
270
|
+
# characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt)
|
271
|
+
# . No more than 32 labels can be associated with a batch.
|
272
|
+
# Corresponds to the JSON property `labels`
|
273
|
+
# @return [Hash<String,String>]
|
274
|
+
attr_accessor :labels
|
275
|
+
|
276
|
+
# Output only. The resource name of the batch.
|
277
|
+
# Corresponds to the JSON property `name`
|
278
|
+
# @return [String]
|
279
|
+
attr_accessor :name
|
280
|
+
|
281
|
+
# Output only. The resource name of the operation associated with this batch.
|
282
|
+
# Corresponds to the JSON property `operation`
|
283
|
+
# @return [String]
|
284
|
+
attr_accessor :operation
|
285
|
+
|
286
|
+
# A configuration for running an Apache PySpark (https://spark.apache.org/docs/
|
287
|
+
# latest/api/python/getting_started/quickstart.html) batch workload.
|
288
|
+
# Corresponds to the JSON property `pysparkBatch`
|
289
|
+
# @return [Google::Apis::DataprocV1::PySparkBatch]
|
290
|
+
attr_accessor :pyspark_batch
|
291
|
+
|
292
|
+
# Runtime configuration for a workload.
|
293
|
+
# Corresponds to the JSON property `runtimeConfig`
|
294
|
+
# @return [Google::Apis::DataprocV1::RuntimeConfig]
|
295
|
+
attr_accessor :runtime_config
|
296
|
+
|
297
|
+
# Runtime information about workload execution.
|
298
|
+
# Corresponds to the JSON property `runtimeInfo`
|
299
|
+
# @return [Google::Apis::DataprocV1::RuntimeInfo]
|
300
|
+
attr_accessor :runtime_info
|
301
|
+
|
302
|
+
# A configuration for running an Apache Spark (http://spark.apache.org/) batch
|
303
|
+
# workload.
|
304
|
+
# Corresponds to the JSON property `sparkBatch`
|
305
|
+
# @return [Google::Apis::DataprocV1::SparkBatch]
|
306
|
+
attr_accessor :spark_batch
|
307
|
+
|
308
|
+
# A configuration for running an Apache SparkR (https://spark.apache.org/docs/
|
309
|
+
# latest/sparkr.html) batch workload.
|
310
|
+
# Corresponds to the JSON property `sparkRBatch`
|
311
|
+
# @return [Google::Apis::DataprocV1::SparkRBatch]
|
312
|
+
attr_accessor :spark_r_batch
|
313
|
+
|
314
|
+
# A configuration for running Apache Spark SQL (http://spark.apache.org/sql/)
|
315
|
+
# queries as a batch workload.
|
316
|
+
# Corresponds to the JSON property `sparkSqlBatch`
|
317
|
+
# @return [Google::Apis::DataprocV1::SparkSqlBatch]
|
318
|
+
attr_accessor :spark_sql_batch
|
319
|
+
|
320
|
+
# Output only. The state of the batch.
|
321
|
+
# Corresponds to the JSON property `state`
|
322
|
+
# @return [String]
|
323
|
+
attr_accessor :state
|
324
|
+
|
325
|
+
# Output only. Historical state information for the batch.
|
326
|
+
# Corresponds to the JSON property `stateHistory`
|
327
|
+
# @return [Array<Google::Apis::DataprocV1::StateHistory>]
|
328
|
+
attr_accessor :state_history
|
329
|
+
|
330
|
+
# Output only. Batch state details, such as a failure description if the state
|
331
|
+
# is FAILED.
|
332
|
+
# Corresponds to the JSON property `stateMessage`
|
333
|
+
# @return [String]
|
334
|
+
attr_accessor :state_message
|
335
|
+
|
336
|
+
# Output only. The time when the batch entered a current state.
|
337
|
+
# Corresponds to the JSON property `stateTime`
|
338
|
+
# @return [String]
|
339
|
+
attr_accessor :state_time
|
340
|
+
|
341
|
+
# Output only. A batch UUID (Unique Universal Identifier). The service generates
|
342
|
+
# this value when it creates the batch.
|
343
|
+
# Corresponds to the JSON property `uuid`
|
344
|
+
# @return [String]
|
345
|
+
attr_accessor :uuid
|
346
|
+
|
347
|
+
def initialize(**args)
|
348
|
+
update!(**args)
|
349
|
+
end
|
350
|
+
|
351
|
+
# Update properties of this object
|
352
|
+
def update!(**args)
|
353
|
+
@create_time = args[:create_time] if args.key?(:create_time)
|
354
|
+
@creator = args[:creator] if args.key?(:creator)
|
355
|
+
@environment_config = args[:environment_config] if args.key?(:environment_config)
|
356
|
+
@labels = args[:labels] if args.key?(:labels)
|
357
|
+
@name = args[:name] if args.key?(:name)
|
358
|
+
@operation = args[:operation] if args.key?(:operation)
|
359
|
+
@pyspark_batch = args[:pyspark_batch] if args.key?(:pyspark_batch)
|
360
|
+
@runtime_config = args[:runtime_config] if args.key?(:runtime_config)
|
361
|
+
@runtime_info = args[:runtime_info] if args.key?(:runtime_info)
|
362
|
+
@spark_batch = args[:spark_batch] if args.key?(:spark_batch)
|
363
|
+
@spark_r_batch = args[:spark_r_batch] if args.key?(:spark_r_batch)
|
364
|
+
@spark_sql_batch = args[:spark_sql_batch] if args.key?(:spark_sql_batch)
|
365
|
+
@state = args[:state] if args.key?(:state)
|
366
|
+
@state_history = args[:state_history] if args.key?(:state_history)
|
367
|
+
@state_message = args[:state_message] if args.key?(:state_message)
|
368
|
+
@state_time = args[:state_time] if args.key?(:state_time)
|
369
|
+
@uuid = args[:uuid] if args.key?(:uuid)
|
370
|
+
end
|
371
|
+
end
|
372
|
+
|
248
373
|
# Metadata describing the Batch operation.
|
249
374
|
class BatchOperationMetadata
|
250
375
|
include Google::Apis::Core::Hashable
|
@@ -979,6 +1104,74 @@ module Google
|
|
979
1104
|
end
|
980
1105
|
end
|
981
1106
|
|
1107
|
+
# Environment configuration for a workload.
|
1108
|
+
class EnvironmentConfig
|
1109
|
+
include Google::Apis::Core::Hashable
|
1110
|
+
|
1111
|
+
# Execution configuration for a workload.
|
1112
|
+
# Corresponds to the JSON property `executionConfig`
|
1113
|
+
# @return [Google::Apis::DataprocV1::ExecutionConfig]
|
1114
|
+
attr_accessor :execution_config
|
1115
|
+
|
1116
|
+
# Auxiliary services configuration for a workload.
|
1117
|
+
# Corresponds to the JSON property `peripheralsConfig`
|
1118
|
+
# @return [Google::Apis::DataprocV1::PeripheralsConfig]
|
1119
|
+
attr_accessor :peripherals_config
|
1120
|
+
|
1121
|
+
def initialize(**args)
|
1122
|
+
update!(**args)
|
1123
|
+
end
|
1124
|
+
|
1125
|
+
# Update properties of this object
|
1126
|
+
def update!(**args)
|
1127
|
+
@execution_config = args[:execution_config] if args.key?(:execution_config)
|
1128
|
+
@peripherals_config = args[:peripherals_config] if args.key?(:peripherals_config)
|
1129
|
+
end
|
1130
|
+
end
|
1131
|
+
|
1132
|
+
# Execution configuration for a workload.
|
1133
|
+
class ExecutionConfig
|
1134
|
+
include Google::Apis::Core::Hashable
|
1135
|
+
|
1136
|
+
# Optional. The Cloud KMS key to use for encryption.
|
1137
|
+
# Corresponds to the JSON property `kmsKey`
|
1138
|
+
# @return [String]
|
1139
|
+
attr_accessor :kms_key
|
1140
|
+
|
1141
|
+
# Optional. Tags used for network traffic control.
|
1142
|
+
# Corresponds to the JSON property `networkTags`
|
1143
|
+
# @return [Array<String>]
|
1144
|
+
attr_accessor :network_tags
|
1145
|
+
|
1146
|
+
# Optional. Network URI to connect workload to.
|
1147
|
+
# Corresponds to the JSON property `networkUri`
|
1148
|
+
# @return [String]
|
1149
|
+
attr_accessor :network_uri
|
1150
|
+
|
1151
|
+
# Optional. Service account that used to execute workload.
|
1152
|
+
# Corresponds to the JSON property `serviceAccount`
|
1153
|
+
# @return [String]
|
1154
|
+
attr_accessor :service_account
|
1155
|
+
|
1156
|
+
# Optional. Subnetwork URI to connect workload to.
|
1157
|
+
# Corresponds to the JSON property `subnetworkUri`
|
1158
|
+
# @return [String]
|
1159
|
+
attr_accessor :subnetwork_uri
|
1160
|
+
|
1161
|
+
def initialize(**args)
|
1162
|
+
update!(**args)
|
1163
|
+
end
|
1164
|
+
|
1165
|
+
# Update properties of this object
|
1166
|
+
def update!(**args)
|
1167
|
+
@kms_key = args[:kms_key] if args.key?(:kms_key)
|
1168
|
+
@network_tags = args[:network_tags] if args.key?(:network_tags)
|
1169
|
+
@network_uri = args[:network_uri] if args.key?(:network_uri)
|
1170
|
+
@service_account = args[:service_account] if args.key?(:service_account)
|
1171
|
+
@subnetwork_uri = args[:subnetwork_uri] if args.key?(:subnetwork_uri)
|
1172
|
+
end
|
1173
|
+
end
|
1174
|
+
|
982
1175
|
# Represents a textual expression in the Common Expression Language (CEL) syntax.
|
983
1176
|
# CEL is a C-like expression language. The syntax and semantics of CEL are
|
984
1177
|
# documented at https://github.com/google/cel-spec.Example (Comparison): title: "
|
@@ -2168,6 +2361,32 @@ module Google
|
|
2168
2361
|
end
|
2169
2362
|
end
|
2170
2363
|
|
2364
|
+
# A list of batch workloads.
|
2365
|
+
class ListBatchesResponse
|
2366
|
+
include Google::Apis::Core::Hashable
|
2367
|
+
|
2368
|
+
# The batches from the specified collection.
|
2369
|
+
# Corresponds to the JSON property `batches`
|
2370
|
+
# @return [Array<Google::Apis::DataprocV1::Batch>]
|
2371
|
+
attr_accessor :batches
|
2372
|
+
|
2373
|
+
# A token, which can be sent as page_token to retrieve the next page. If this
|
2374
|
+
# field is omitted, there are no subsequent pages.
|
2375
|
+
# Corresponds to the JSON property `nextPageToken`
|
2376
|
+
# @return [String]
|
2377
|
+
attr_accessor :next_page_token
|
2378
|
+
|
2379
|
+
def initialize(**args)
|
2380
|
+
update!(**args)
|
2381
|
+
end
|
2382
|
+
|
2383
|
+
# Update properties of this object
|
2384
|
+
def update!(**args)
|
2385
|
+
@batches = args[:batches] if args.key?(:batches)
|
2386
|
+
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
|
2387
|
+
end
|
2388
|
+
end
|
2389
|
+
|
2171
2390
|
# The list of all clusters in a project.
|
2172
2391
|
class ListClustersResponse
|
2173
2392
|
include Google::Apis::Core::Hashable
|
@@ -2654,6 +2873,32 @@ module Google
|
|
2654
2873
|
end
|
2655
2874
|
end
|
2656
2875
|
|
2876
|
+
# Auxiliary services configuration for a workload.
|
2877
|
+
class PeripheralsConfig
|
2878
|
+
include Google::Apis::Core::Hashable
|
2879
|
+
|
2880
|
+
# Optional. Resource name of an existing Dataproc Metastore service.Example:
|
2881
|
+
# projects/[project_id]/locations/[region]/services/[service_id]
|
2882
|
+
# Corresponds to the JSON property `metastoreService`
|
2883
|
+
# @return [String]
|
2884
|
+
attr_accessor :metastore_service
|
2885
|
+
|
2886
|
+
# Spark History Server configuration for the workload.
|
2887
|
+
# Corresponds to the JSON property `sparkHistoryServerConfig`
|
2888
|
+
# @return [Google::Apis::DataprocV1::SparkHistoryServerConfig]
|
2889
|
+
attr_accessor :spark_history_server_config
|
2890
|
+
|
2891
|
+
def initialize(**args)
|
2892
|
+
update!(**args)
|
2893
|
+
end
|
2894
|
+
|
2895
|
+
# Update properties of this object
|
2896
|
+
def update!(**args)
|
2897
|
+
@metastore_service = args[:metastore_service] if args.key?(:metastore_service)
|
2898
|
+
@spark_history_server_config = args[:spark_history_server_config] if args.key?(:spark_history_server_config)
|
2899
|
+
end
|
2900
|
+
end
|
2901
|
+
|
2657
2902
|
# A Dataproc job for running Apache Pig (https://pig.apache.org/) queries on
|
2658
2903
|
# YARN.
|
2659
2904
|
class PigJob
|
@@ -2869,6 +3114,63 @@ module Google
|
|
2869
3114
|
end
|
2870
3115
|
end
|
2871
3116
|
|
3117
|
+
# A configuration for running an Apache PySpark (https://spark.apache.org/docs/
|
3118
|
+
# latest/api/python/getting_started/quickstart.html) batch workload.
|
3119
|
+
class PySparkBatch
|
3120
|
+
include Google::Apis::Core::Hashable
|
3121
|
+
|
3122
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory of
|
3123
|
+
# each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
|
3124
|
+
# Corresponds to the JSON property `archiveUris`
|
3125
|
+
# @return [Array<String>]
|
3126
|
+
attr_accessor :archive_uris
|
3127
|
+
|
3128
|
+
# Optional. The arguments to pass to the driver. Do not include arguments that
|
3129
|
+
# can be set as batch properties, such as --conf, since a collision can occur
|
3130
|
+
# that causes an incorrect batch submission.
|
3131
|
+
# Corresponds to the JSON property `args`
|
3132
|
+
# @return [Array<String>]
|
3133
|
+
attr_accessor :args
|
3134
|
+
|
3135
|
+
# Optional. HCFS URIs of files to be placed in the working directory of each
|
3136
|
+
# executor.
|
3137
|
+
# Corresponds to the JSON property `fileUris`
|
3138
|
+
# @return [Array<String>]
|
3139
|
+
attr_accessor :file_uris
|
3140
|
+
|
3141
|
+
# Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
|
3142
|
+
# and tasks.
|
3143
|
+
# Corresponds to the JSON property `jarFileUris`
|
3144
|
+
# @return [Array<String>]
|
3145
|
+
attr_accessor :jar_file_uris
|
3146
|
+
|
3147
|
+
# Required. The HCFS URI of the main Python file to use as the Spark driver.
|
3148
|
+
# Must be a .py file.
|
3149
|
+
# Corresponds to the JSON property `mainPythonFileUri`
|
3150
|
+
# @return [String]
|
3151
|
+
attr_accessor :main_python_file_uri
|
3152
|
+
|
3153
|
+
# Optional. HCFS file URIs of Python files to pass to the PySpark framework.
|
3154
|
+
# Supported file types: .py, .egg, and .zip.
|
3155
|
+
# Corresponds to the JSON property `pythonFileUris`
|
3156
|
+
# @return [Array<String>]
|
3157
|
+
attr_accessor :python_file_uris
|
3158
|
+
|
3159
|
+
def initialize(**args)
|
3160
|
+
update!(**args)
|
3161
|
+
end
|
3162
|
+
|
3163
|
+
# Update properties of this object
|
3164
|
+
def update!(**args)
|
3165
|
+
@archive_uris = args[:archive_uris] if args.key?(:archive_uris)
|
3166
|
+
@args = args[:args] if args.key?(:args)
|
3167
|
+
@file_uris = args[:file_uris] if args.key?(:file_uris)
|
3168
|
+
@jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
|
3169
|
+
@main_python_file_uri = args[:main_python_file_uri] if args.key?(:main_python_file_uri)
|
3170
|
+
@python_file_uris = args[:python_file_uris] if args.key?(:python_file_uris)
|
3171
|
+
end
|
3172
|
+
end
|
3173
|
+
|
2872
3174
|
# A Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/
|
2873
3175
|
# python-programming-guide.html) applications on YARN.
|
2874
3176
|
class PySparkJob
|
@@ -3048,6 +3350,72 @@ module Google
|
|
3048
3350
|
end
|
3049
3351
|
end
|
3050
3352
|
|
3353
|
+
# Runtime configuration for a workload.
|
3354
|
+
class RuntimeConfig
|
3355
|
+
include Google::Apis::Core::Hashable
|
3356
|
+
|
3357
|
+
# Optional. Optional custom container image for the job runtime environment. If
|
3358
|
+
# not specified, a default container image will be used.
|
3359
|
+
# Corresponds to the JSON property `containerImage`
|
3360
|
+
# @return [String]
|
3361
|
+
attr_accessor :container_image
|
3362
|
+
|
3363
|
+
# Optional. A mapping of property names to values, which are used to configure
|
3364
|
+
# workload execution.
|
3365
|
+
# Corresponds to the JSON property `properties`
|
3366
|
+
# @return [Hash<String,String>]
|
3367
|
+
attr_accessor :properties
|
3368
|
+
|
3369
|
+
# Optional. Version of the batch runtime.
|
3370
|
+
# Corresponds to the JSON property `version`
|
3371
|
+
# @return [String]
|
3372
|
+
attr_accessor :version
|
3373
|
+
|
3374
|
+
def initialize(**args)
|
3375
|
+
update!(**args)
|
3376
|
+
end
|
3377
|
+
|
3378
|
+
# Update properties of this object
|
3379
|
+
def update!(**args)
|
3380
|
+
@container_image = args[:container_image] if args.key?(:container_image)
|
3381
|
+
@properties = args[:properties] if args.key?(:properties)
|
3382
|
+
@version = args[:version] if args.key?(:version)
|
3383
|
+
end
|
3384
|
+
end
|
3385
|
+
|
3386
|
+
# Runtime information about workload execution.
|
3387
|
+
class RuntimeInfo
|
3388
|
+
include Google::Apis::Core::Hashable
|
3389
|
+
|
3390
|
+
# Output only. A URI pointing to the location of the diagnostics tarball.
|
3391
|
+
# Corresponds to the JSON property `diagnosticOutputUri`
|
3392
|
+
# @return [String]
|
3393
|
+
attr_accessor :diagnostic_output_uri
|
3394
|
+
|
3395
|
+
# Output only. Map of remote access endpoints (such as web interfaces and APIs)
|
3396
|
+
# to their URIs.
|
3397
|
+
# Corresponds to the JSON property `endpoints`
|
3398
|
+
# @return [Hash<String,String>]
|
3399
|
+
attr_accessor :endpoints
|
3400
|
+
|
3401
|
+
# Output only. A URI pointing to the location of the stdout and stderr of the
|
3402
|
+
# workload.
|
3403
|
+
# Corresponds to the JSON property `outputUri`
|
3404
|
+
# @return [String]
|
3405
|
+
attr_accessor :output_uri
|
3406
|
+
|
3407
|
+
def initialize(**args)
|
3408
|
+
update!(**args)
|
3409
|
+
end
|
3410
|
+
|
3411
|
+
# Update properties of this object
|
3412
|
+
def update!(**args)
|
3413
|
+
@diagnostic_output_uri = args[:diagnostic_output_uri] if args.key?(:diagnostic_output_uri)
|
3414
|
+
@endpoints = args[:endpoints] if args.key?(:endpoints)
|
3415
|
+
@output_uri = args[:output_uri] if args.key?(:output_uri)
|
3416
|
+
end
|
3417
|
+
end
|
3418
|
+
|
3051
3419
|
# Security related configuration, including encryption, Kerberos, etc.
|
3052
3420
|
class SecurityConfig
|
3053
3421
|
include Google::Apis::Core::Hashable
|
@@ -3258,6 +3626,83 @@ module Google
|
|
3258
3626
|
end
|
3259
3627
|
end
|
3260
3628
|
|
3629
|
+
# A configuration for running an Apache Spark (http://spark.apache.org/) batch
|
3630
|
+
# workload.
|
3631
|
+
class SparkBatch
|
3632
|
+
include Google::Apis::Core::Hashable
|
3633
|
+
|
3634
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory of
|
3635
|
+
# each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
|
3636
|
+
# Corresponds to the JSON property `archiveUris`
|
3637
|
+
# @return [Array<String>]
|
3638
|
+
attr_accessor :archive_uris
|
3639
|
+
|
3640
|
+
# Optional. The arguments to pass to the driver. Do not include arguments that
|
3641
|
+
# can be set as batch properties, such as --conf, since a collision can occur
|
3642
|
+
# that causes an incorrect batch submission.
|
3643
|
+
# Corresponds to the JSON property `args`
|
3644
|
+
# @return [Array<String>]
|
3645
|
+
attr_accessor :args
|
3646
|
+
|
3647
|
+
# Optional. HCFS URIs of files to be placed in the working directory of each
|
3648
|
+
# executor.
|
3649
|
+
# Corresponds to the JSON property `fileUris`
|
3650
|
+
# @return [Array<String>]
|
3651
|
+
attr_accessor :file_uris
|
3652
|
+
|
3653
|
+
# Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
|
3654
|
+
# and tasks.
|
3655
|
+
# Corresponds to the JSON property `jarFileUris`
|
3656
|
+
# @return [Array<String>]
|
3657
|
+
attr_accessor :jar_file_uris
|
3658
|
+
|
3659
|
+
# Optional. The name of the driver main class. The jar file that contains the
|
3660
|
+
# class must be in the classpath or specified in jar_file_uris.
|
3661
|
+
# Corresponds to the JSON property `mainClass`
|
3662
|
+
# @return [String]
|
3663
|
+
attr_accessor :main_class
|
3664
|
+
|
3665
|
+
# Optional. The HCFS URI of the jar file that contains the main class.
|
3666
|
+
# Corresponds to the JSON property `mainJarFileUri`
|
3667
|
+
# @return [String]
|
3668
|
+
attr_accessor :main_jar_file_uri
|
3669
|
+
|
3670
|
+
def initialize(**args)
|
3671
|
+
update!(**args)
|
3672
|
+
end
|
3673
|
+
|
3674
|
+
# Update properties of this object
|
3675
|
+
def update!(**args)
|
3676
|
+
@archive_uris = args[:archive_uris] if args.key?(:archive_uris)
|
3677
|
+
@args = args[:args] if args.key?(:args)
|
3678
|
+
@file_uris = args[:file_uris] if args.key?(:file_uris)
|
3679
|
+
@jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
|
3680
|
+
@main_class = args[:main_class] if args.key?(:main_class)
|
3681
|
+
@main_jar_file_uri = args[:main_jar_file_uri] if args.key?(:main_jar_file_uri)
|
3682
|
+
end
|
3683
|
+
end
|
3684
|
+
|
3685
|
+
# Spark History Server configuration for the workload.
|
3686
|
+
class SparkHistoryServerConfig
|
3687
|
+
include Google::Apis::Core::Hashable
|
3688
|
+
|
3689
|
+
# Optional. Resource name of an existing Dataproc Cluster to act as a Spark
|
3690
|
+
# History Server for the workload.Example: projects/[project_id]/regions/[region]
|
3691
|
+
# /clusters/[cluster_name]
|
3692
|
+
# Corresponds to the JSON property `dataprocCluster`
|
3693
|
+
# @return [String]
|
3694
|
+
attr_accessor :dataproc_cluster
|
3695
|
+
|
3696
|
+
def initialize(**args)
|
3697
|
+
update!(**args)
|
3698
|
+
end
|
3699
|
+
|
3700
|
+
# Update properties of this object
|
3701
|
+
def update!(**args)
|
3702
|
+
@dataproc_cluster = args[:dataproc_cluster] if args.key?(:dataproc_cluster)
|
3703
|
+
end
|
3704
|
+
end
|
3705
|
+
|
3261
3706
|
# A Dataproc job for running Apache Spark (http://spark.apache.org/)
|
3262
3707
|
# applications on YARN.
|
3263
3708
|
class SparkJob
|
@@ -3329,6 +3774,49 @@ module Google
|
|
3329
3774
|
end
|
3330
3775
|
end
|
3331
3776
|
|
3777
|
+
# A configuration for running an Apache SparkR (https://spark.apache.org/docs/
|
3778
|
+
# latest/sparkr.html) batch workload.
|
3779
|
+
class SparkRBatch
|
3780
|
+
include Google::Apis::Core::Hashable
|
3781
|
+
|
3782
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory of
|
3783
|
+
# each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
|
3784
|
+
# Corresponds to the JSON property `archiveUris`
|
3785
|
+
# @return [Array<String>]
|
3786
|
+
attr_accessor :archive_uris
|
3787
|
+
|
3788
|
+
# Optional. The arguments to pass to the Spark driver. Do not include arguments
|
3789
|
+
# that can be set as batch properties, such as --conf, since a collision can
|
3790
|
+
# occur that causes an incorrect batch submission.
|
3791
|
+
# Corresponds to the JSON property `args`
|
3792
|
+
# @return [Array<String>]
|
3793
|
+
attr_accessor :args
|
3794
|
+
|
3795
|
+
# Optional. HCFS URIs of files to be placed in the working directory of each
|
3796
|
+
# executor.
|
3797
|
+
# Corresponds to the JSON property `fileUris`
|
3798
|
+
# @return [Array<String>]
|
3799
|
+
attr_accessor :file_uris
|
3800
|
+
|
3801
|
+
# Required. The HCFS URI of the main R file to use as the driver. Must be a .R
|
3802
|
+
# or .r file.
|
3803
|
+
# Corresponds to the JSON property `mainRFileUri`
|
3804
|
+
# @return [String]
|
3805
|
+
attr_accessor :main_r_file_uri
|
3806
|
+
|
3807
|
+
def initialize(**args)
|
3808
|
+
update!(**args)
|
3809
|
+
end
|
3810
|
+
|
3811
|
+
# Update properties of this object
|
3812
|
+
def update!(**args)
|
3813
|
+
@archive_uris = args[:archive_uris] if args.key?(:archive_uris)
|
3814
|
+
@args = args[:args] if args.key?(:args)
|
3815
|
+
@file_uris = args[:file_uris] if args.key?(:file_uris)
|
3816
|
+
@main_r_file_uri = args[:main_r_file_uri] if args.key?(:main_r_file_uri)
|
3817
|
+
end
|
3818
|
+
end
|
3819
|
+
|
3332
3820
|
# A Dataproc job for running Apache SparkR (https://spark.apache.org/docs/latest/
|
3333
3821
|
# sparkr.html) applications on YARN.
|
3334
3822
|
class SparkRJob
|
@@ -3387,6 +3875,40 @@ module Google
|
|
3387
3875
|
end
|
3388
3876
|
end
|
3389
3877
|
|
3878
|
+
# A configuration for running Apache Spark SQL (http://spark.apache.org/sql/)
|
3879
|
+
# queries as a batch workload.
|
3880
|
+
class SparkSqlBatch
|
3881
|
+
include Google::Apis::Core::Hashable
|
3882
|
+
|
3883
|
+
# Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
|
3884
|
+
# Corresponds to the JSON property `jarFileUris`
|
3885
|
+
# @return [Array<String>]
|
3886
|
+
attr_accessor :jar_file_uris
|
3887
|
+
|
3888
|
+
# Required. The HCFS URI of the script that contains Spark SQL queries to
|
3889
|
+
# execute.
|
3890
|
+
# Corresponds to the JSON property `queryFileUri`
|
3891
|
+
# @return [String]
|
3892
|
+
attr_accessor :query_file_uri
|
3893
|
+
|
3894
|
+
# Optional. Mapping of query variable names to values (equivalent to the Spark
|
3895
|
+
# SQL command: SET name="value";).
|
3896
|
+
# Corresponds to the JSON property `queryVariables`
|
3897
|
+
# @return [Hash<String,String>]
|
3898
|
+
attr_accessor :query_variables
|
3899
|
+
|
3900
|
+
def initialize(**args)
|
3901
|
+
update!(**args)
|
3902
|
+
end
|
3903
|
+
|
3904
|
+
# Update properties of this object
|
3905
|
+
def update!(**args)
|
3906
|
+
@jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
|
3907
|
+
@query_file_uri = args[:query_file_uri] if args.key?(:query_file_uri)
|
3908
|
+
@query_variables = args[:query_variables] if args.key?(:query_variables)
|
3909
|
+
end
|
3910
|
+
end
|
3911
|
+
|
3390
3912
|
# A Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/)
|
3391
3913
|
# queries.
|
3392
3914
|
class SparkSqlJob
|
@@ -3536,6 +4058,37 @@ module Google
|
|
3536
4058
|
end
|
3537
4059
|
end
|
3538
4060
|
|
4061
|
+
# Historical state information.
|
4062
|
+
class StateHistory
|
4063
|
+
include Google::Apis::Core::Hashable
|
4064
|
+
|
4065
|
+
# Output only. The state of the batch at this point in history.
|
4066
|
+
# Corresponds to the JSON property `state`
|
4067
|
+
# @return [String]
|
4068
|
+
attr_accessor :state
|
4069
|
+
|
4070
|
+
# Output only. Details about the state at this point in history.
|
4071
|
+
# Corresponds to the JSON property `stateMessage`
|
4072
|
+
# @return [String]
|
4073
|
+
attr_accessor :state_message
|
4074
|
+
|
4075
|
+
# Output only. The time when the batch entered the historical state.
|
4076
|
+
# Corresponds to the JSON property `stateStartTime`
|
4077
|
+
# @return [String]
|
4078
|
+
attr_accessor :state_start_time
|
4079
|
+
|
4080
|
+
def initialize(**args)
|
4081
|
+
update!(**args)
|
4082
|
+
end
|
4083
|
+
|
4084
|
+
# Update properties of this object
|
4085
|
+
def update!(**args)
|
4086
|
+
@state = args[:state] if args.key?(:state)
|
4087
|
+
@state_message = args[:state_message] if args.key?(:state_message)
|
4088
|
+
@state_start_time = args[:state_start_time] if args.key?(:state_start_time)
|
4089
|
+
end
|
4090
|
+
end
|
4091
|
+
|
3539
4092
|
# The Status type defines a logical error model that is suitable for different
|
3540
4093
|
# programming environments, including REST APIs and RPC APIs. It is used by gRPC
|
3541
4094
|
# (https://github.com/grpc). Each Status message contains three pieces of data:
|
@@ -16,13 +16,13 @@ module Google
|
|
16
16
|
module Apis
|
17
17
|
module DataprocV1
|
18
18
|
# Version of the google-apis-dataproc_v1 gem
|
19
|
-
GEM_VERSION = "0.
|
19
|
+
GEM_VERSION = "0.19.0"
|
20
20
|
|
21
21
|
# Version of the code generator used to generate this client
|
22
|
-
GENERATOR_VERSION = "0.4.
|
22
|
+
GENERATOR_VERSION = "0.4.1"
|
23
23
|
|
24
24
|
# Revision of the discovery document this client was generated from
|
25
|
-
REVISION = "
|
25
|
+
REVISION = "20220111"
|
26
26
|
end
|
27
27
|
end
|
28
28
|
end
|
@@ -52,6 +52,12 @@ module Google
|
|
52
52
|
include Google::Apis::Core::JsonObjectSupport
|
53
53
|
end
|
54
54
|
|
55
|
+
class Batch
|
56
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
57
|
+
|
58
|
+
include Google::Apis::Core::JsonObjectSupport
|
59
|
+
end
|
60
|
+
|
55
61
|
class BatchOperationMetadata
|
56
62
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
57
63
|
|
@@ -160,6 +166,18 @@ module Google
|
|
160
166
|
include Google::Apis::Core::JsonObjectSupport
|
161
167
|
end
|
162
168
|
|
169
|
+
class EnvironmentConfig
|
170
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
171
|
+
|
172
|
+
include Google::Apis::Core::JsonObjectSupport
|
173
|
+
end
|
174
|
+
|
175
|
+
class ExecutionConfig
|
176
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
177
|
+
|
178
|
+
include Google::Apis::Core::JsonObjectSupport
|
179
|
+
end
|
180
|
+
|
163
181
|
class Expr
|
164
182
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
165
183
|
|
@@ -292,6 +310,12 @@ module Google
|
|
292
310
|
include Google::Apis::Core::JsonObjectSupport
|
293
311
|
end
|
294
312
|
|
313
|
+
class ListBatchesResponse
|
314
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
315
|
+
|
316
|
+
include Google::Apis::Core::JsonObjectSupport
|
317
|
+
end
|
318
|
+
|
295
319
|
class ListClustersResponse
|
296
320
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
297
321
|
|
@@ -376,6 +400,12 @@ module Google
|
|
376
400
|
include Google::Apis::Core::JsonObjectSupport
|
377
401
|
end
|
378
402
|
|
403
|
+
class PeripheralsConfig
|
404
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
405
|
+
|
406
|
+
include Google::Apis::Core::JsonObjectSupport
|
407
|
+
end
|
408
|
+
|
379
409
|
class PigJob
|
380
410
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
381
411
|
|
@@ -394,6 +424,12 @@ module Google
|
|
394
424
|
include Google::Apis::Core::JsonObjectSupport
|
395
425
|
end
|
396
426
|
|
427
|
+
class PySparkBatch
|
428
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
429
|
+
|
430
|
+
include Google::Apis::Core::JsonObjectSupport
|
431
|
+
end
|
432
|
+
|
397
433
|
class PySparkJob
|
398
434
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
399
435
|
|
@@ -424,6 +460,18 @@ module Google
|
|
424
460
|
include Google::Apis::Core::JsonObjectSupport
|
425
461
|
end
|
426
462
|
|
463
|
+
class RuntimeConfig
|
464
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
465
|
+
|
466
|
+
include Google::Apis::Core::JsonObjectSupport
|
467
|
+
end
|
468
|
+
|
469
|
+
class RuntimeInfo
|
470
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
471
|
+
|
472
|
+
include Google::Apis::Core::JsonObjectSupport
|
473
|
+
end
|
474
|
+
|
427
475
|
class SecurityConfig
|
428
476
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
429
477
|
|
@@ -454,18 +502,42 @@ module Google
|
|
454
502
|
include Google::Apis::Core::JsonObjectSupport
|
455
503
|
end
|
456
504
|
|
505
|
+
class SparkBatch
|
506
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
507
|
+
|
508
|
+
include Google::Apis::Core::JsonObjectSupport
|
509
|
+
end
|
510
|
+
|
511
|
+
class SparkHistoryServerConfig
|
512
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
513
|
+
|
514
|
+
include Google::Apis::Core::JsonObjectSupport
|
515
|
+
end
|
516
|
+
|
457
517
|
class SparkJob
|
458
518
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
459
519
|
|
460
520
|
include Google::Apis::Core::JsonObjectSupport
|
461
521
|
end
|
462
522
|
|
523
|
+
class SparkRBatch
|
524
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
525
|
+
|
526
|
+
include Google::Apis::Core::JsonObjectSupport
|
527
|
+
end
|
528
|
+
|
463
529
|
class SparkRJob
|
464
530
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
465
531
|
|
466
532
|
include Google::Apis::Core::JsonObjectSupport
|
467
533
|
end
|
468
534
|
|
535
|
+
class SparkSqlBatch
|
536
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
537
|
+
|
538
|
+
include Google::Apis::Core::JsonObjectSupport
|
539
|
+
end
|
540
|
+
|
469
541
|
class SparkSqlJob
|
470
542
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
471
543
|
|
@@ -484,6 +556,12 @@ module Google
|
|
484
556
|
include Google::Apis::Core::JsonObjectSupport
|
485
557
|
end
|
486
558
|
|
559
|
+
class StateHistory
|
560
|
+
class Representation < Google::Apis::Core::JsonRepresentation; end
|
561
|
+
|
562
|
+
include Google::Apis::Core::JsonObjectSupport
|
563
|
+
end
|
564
|
+
|
487
565
|
class Status
|
488
566
|
class Representation < Google::Apis::Core::JsonRepresentation; end
|
489
567
|
|
@@ -614,6 +692,37 @@ module Google
|
|
614
692
|
end
|
615
693
|
end
|
616
694
|
|
695
|
+
class Batch
|
696
|
+
# @private
|
697
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
698
|
+
property :create_time, as: 'createTime'
|
699
|
+
property :creator, as: 'creator'
|
700
|
+
property :environment_config, as: 'environmentConfig', class: Google::Apis::DataprocV1::EnvironmentConfig, decorator: Google::Apis::DataprocV1::EnvironmentConfig::Representation
|
701
|
+
|
702
|
+
hash :labels, as: 'labels'
|
703
|
+
property :name, as: 'name'
|
704
|
+
property :operation, as: 'operation'
|
705
|
+
property :pyspark_batch, as: 'pysparkBatch', class: Google::Apis::DataprocV1::PySparkBatch, decorator: Google::Apis::DataprocV1::PySparkBatch::Representation
|
706
|
+
|
707
|
+
property :runtime_config, as: 'runtimeConfig', class: Google::Apis::DataprocV1::RuntimeConfig, decorator: Google::Apis::DataprocV1::RuntimeConfig::Representation
|
708
|
+
|
709
|
+
property :runtime_info, as: 'runtimeInfo', class: Google::Apis::DataprocV1::RuntimeInfo, decorator: Google::Apis::DataprocV1::RuntimeInfo::Representation
|
710
|
+
|
711
|
+
property :spark_batch, as: 'sparkBatch', class: Google::Apis::DataprocV1::SparkBatch, decorator: Google::Apis::DataprocV1::SparkBatch::Representation
|
712
|
+
|
713
|
+
property :spark_r_batch, as: 'sparkRBatch', class: Google::Apis::DataprocV1::SparkRBatch, decorator: Google::Apis::DataprocV1::SparkRBatch::Representation
|
714
|
+
|
715
|
+
property :spark_sql_batch, as: 'sparkSqlBatch', class: Google::Apis::DataprocV1::SparkSqlBatch, decorator: Google::Apis::DataprocV1::SparkSqlBatch::Representation
|
716
|
+
|
717
|
+
property :state, as: 'state'
|
718
|
+
collection :state_history, as: 'stateHistory', class: Google::Apis::DataprocV1::StateHistory, decorator: Google::Apis::DataprocV1::StateHistory::Representation
|
719
|
+
|
720
|
+
property :state_message, as: 'stateMessage'
|
721
|
+
property :state_time, as: 'stateTime'
|
722
|
+
property :uuid, as: 'uuid'
|
723
|
+
end
|
724
|
+
end
|
725
|
+
|
617
726
|
class BatchOperationMetadata
|
618
727
|
# @private
|
619
728
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -808,6 +917,27 @@ module Google
|
|
808
917
|
end
|
809
918
|
end
|
810
919
|
|
920
|
+
class EnvironmentConfig
|
921
|
+
# @private
|
922
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
923
|
+
property :execution_config, as: 'executionConfig', class: Google::Apis::DataprocV1::ExecutionConfig, decorator: Google::Apis::DataprocV1::ExecutionConfig::Representation
|
924
|
+
|
925
|
+
property :peripherals_config, as: 'peripheralsConfig', class: Google::Apis::DataprocV1::PeripheralsConfig, decorator: Google::Apis::DataprocV1::PeripheralsConfig::Representation
|
926
|
+
|
927
|
+
end
|
928
|
+
end
|
929
|
+
|
930
|
+
class ExecutionConfig
|
931
|
+
# @private
|
932
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
933
|
+
property :kms_key, as: 'kmsKey'
|
934
|
+
collection :network_tags, as: 'networkTags'
|
935
|
+
property :network_uri, as: 'networkUri'
|
936
|
+
property :service_account, as: 'serviceAccount'
|
937
|
+
property :subnetwork_uri, as: 'subnetworkUri'
|
938
|
+
end
|
939
|
+
end
|
940
|
+
|
811
941
|
class Expr
|
812
942
|
# @private
|
813
943
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1081,6 +1211,15 @@ module Google
|
|
1081
1211
|
end
|
1082
1212
|
end
|
1083
1213
|
|
1214
|
+
class ListBatchesResponse
|
1215
|
+
# @private
|
1216
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1217
|
+
collection :batches, as: 'batches', class: Google::Apis::DataprocV1::Batch, decorator: Google::Apis::DataprocV1::Batch::Representation
|
1218
|
+
|
1219
|
+
property :next_page_token, as: 'nextPageToken'
|
1220
|
+
end
|
1221
|
+
end
|
1222
|
+
|
1084
1223
|
class ListClustersResponse
|
1085
1224
|
# @private
|
1086
1225
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1221,6 +1360,15 @@ module Google
|
|
1221
1360
|
end
|
1222
1361
|
end
|
1223
1362
|
|
1363
|
+
class PeripheralsConfig
|
1364
|
+
# @private
|
1365
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1366
|
+
property :metastore_service, as: 'metastoreService'
|
1367
|
+
property :spark_history_server_config, as: 'sparkHistoryServerConfig', class: Google::Apis::DataprocV1::SparkHistoryServerConfig, decorator: Google::Apis::DataprocV1::SparkHistoryServerConfig::Representation
|
1368
|
+
|
1369
|
+
end
|
1370
|
+
end
|
1371
|
+
|
1224
1372
|
class PigJob
|
1225
1373
|
# @private
|
1226
1374
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1261,6 +1409,18 @@ module Google
|
|
1261
1409
|
end
|
1262
1410
|
end
|
1263
1411
|
|
1412
|
+
class PySparkBatch
|
1413
|
+
# @private
|
1414
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1415
|
+
collection :archive_uris, as: 'archiveUris'
|
1416
|
+
collection :args, as: 'args'
|
1417
|
+
collection :file_uris, as: 'fileUris'
|
1418
|
+
collection :jar_file_uris, as: 'jarFileUris'
|
1419
|
+
property :main_python_file_uri, as: 'mainPythonFileUri'
|
1420
|
+
collection :python_file_uris, as: 'pythonFileUris'
|
1421
|
+
end
|
1422
|
+
end
|
1423
|
+
|
1264
1424
|
class PySparkJob
|
1265
1425
|
# @private
|
1266
1426
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1307,6 +1467,24 @@ module Google
|
|
1307
1467
|
end
|
1308
1468
|
end
|
1309
1469
|
|
1470
|
+
class RuntimeConfig
|
1471
|
+
# @private
|
1472
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1473
|
+
property :container_image, as: 'containerImage'
|
1474
|
+
hash :properties, as: 'properties'
|
1475
|
+
property :version, as: 'version'
|
1476
|
+
end
|
1477
|
+
end
|
1478
|
+
|
1479
|
+
class RuntimeInfo
|
1480
|
+
# @private
|
1481
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1482
|
+
property :diagnostic_output_uri, as: 'diagnosticOutputUri'
|
1483
|
+
hash :endpoints, as: 'endpoints'
|
1484
|
+
property :output_uri, as: 'outputUri'
|
1485
|
+
end
|
1486
|
+
end
|
1487
|
+
|
1310
1488
|
class SecurityConfig
|
1311
1489
|
# @private
|
1312
1490
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1357,6 +1535,25 @@ module Google
|
|
1357
1535
|
end
|
1358
1536
|
end
|
1359
1537
|
|
1538
|
+
class SparkBatch
|
1539
|
+
# @private
|
1540
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1541
|
+
collection :archive_uris, as: 'archiveUris'
|
1542
|
+
collection :args, as: 'args'
|
1543
|
+
collection :file_uris, as: 'fileUris'
|
1544
|
+
collection :jar_file_uris, as: 'jarFileUris'
|
1545
|
+
property :main_class, as: 'mainClass'
|
1546
|
+
property :main_jar_file_uri, as: 'mainJarFileUri'
|
1547
|
+
end
|
1548
|
+
end
|
1549
|
+
|
1550
|
+
class SparkHistoryServerConfig
|
1551
|
+
# @private
|
1552
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1553
|
+
property :dataproc_cluster, as: 'dataprocCluster'
|
1554
|
+
end
|
1555
|
+
end
|
1556
|
+
|
1360
1557
|
class SparkJob
|
1361
1558
|
# @private
|
1362
1559
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1372,6 +1569,16 @@ module Google
|
|
1372
1569
|
end
|
1373
1570
|
end
|
1374
1571
|
|
1572
|
+
class SparkRBatch
|
1573
|
+
# @private
|
1574
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1575
|
+
collection :archive_uris, as: 'archiveUris'
|
1576
|
+
collection :args, as: 'args'
|
1577
|
+
collection :file_uris, as: 'fileUris'
|
1578
|
+
property :main_r_file_uri, as: 'mainRFileUri'
|
1579
|
+
end
|
1580
|
+
end
|
1581
|
+
|
1375
1582
|
class SparkRJob
|
1376
1583
|
# @private
|
1377
1584
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1385,6 +1592,15 @@ module Google
|
|
1385
1592
|
end
|
1386
1593
|
end
|
1387
1594
|
|
1595
|
+
class SparkSqlBatch
|
1596
|
+
# @private
|
1597
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1598
|
+
collection :jar_file_uris, as: 'jarFileUris'
|
1599
|
+
property :query_file_uri, as: 'queryFileUri'
|
1600
|
+
hash :query_variables, as: 'queryVariables'
|
1601
|
+
end
|
1602
|
+
end
|
1603
|
+
|
1388
1604
|
class SparkSqlJob
|
1389
1605
|
# @private
|
1390
1606
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -1418,6 +1634,15 @@ module Google
|
|
1418
1634
|
end
|
1419
1635
|
end
|
1420
1636
|
|
1637
|
+
class StateHistory
|
1638
|
+
# @private
|
1639
|
+
class Representation < Google::Apis::Core::JsonRepresentation
|
1640
|
+
property :state, as: 'state'
|
1641
|
+
property :state_message, as: 'stateMessage'
|
1642
|
+
property :state_start_time, as: 'stateStartTime'
|
1643
|
+
end
|
1644
|
+
end
|
1645
|
+
|
1421
1646
|
class Status
|
1422
1647
|
# @private
|
1423
1648
|
class Representation < Google::Apis::Core::JsonRepresentation
|
@@ -352,6 +352,155 @@ module Google
|
|
352
352
|
execute_or_queue_command(command, &block)
|
353
353
|
end
|
354
354
|
|
355
|
+
# Creates a batch workload that executes asynchronously.
|
356
|
+
# @param [String] parent
|
357
|
+
# Required. The parent resource where this batch will be created.
|
358
|
+
# @param [Google::Apis::DataprocV1::Batch] batch_object
|
359
|
+
# @param [String] batch_id
|
360
|
+
# Optional. The ID to use for the batch, which will become the final component
|
361
|
+
# of the batch's resource name.This value must be 4-63 characters. Valid
|
362
|
+
# characters are /[a-z][0-9]-/.
|
363
|
+
# @param [String] request_id
|
364
|
+
# Optional. A unique ID used to identify the request. If the service receives
|
365
|
+
# two CreateBatchRequest (https://cloud.google.com/dataproc/docs/reference/rpc/
|
366
|
+
# google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateBatchRequest)s with
|
367
|
+
# the same request_id, the second request is ignored and the Operation that
|
368
|
+
# corresponds to the first Batch created and stored in the backend is returned.
|
369
|
+
# Recommendation: Set this value to a UUID (https://en.wikipedia.org/wiki/
|
370
|
+
# Universally_unique_identifier).The value must contain only letters (a-z, A-Z),
|
371
|
+
# numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40
|
372
|
+
# characters.
|
373
|
+
# @param [String] fields
|
374
|
+
# Selector specifying which fields to include in a partial response.
|
375
|
+
# @param [String] quota_user
|
376
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
377
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
378
|
+
# @param [Google::Apis::RequestOptions] options
|
379
|
+
# Request-specific options
|
380
|
+
#
|
381
|
+
# @yield [result, err] Result & error if block supplied
|
382
|
+
# @yieldparam result [Google::Apis::DataprocV1::Operation] parsed result object
|
383
|
+
# @yieldparam err [StandardError] error object if request failed
|
384
|
+
#
|
385
|
+
# @return [Google::Apis::DataprocV1::Operation]
|
386
|
+
#
|
387
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
388
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
389
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
390
|
+
def create_project_location_batch(parent, batch_object = nil, batch_id: nil, request_id: nil, fields: nil, quota_user: nil, options: nil, &block)
|
391
|
+
command = make_simple_command(:post, 'v1/{+parent}/batches', options)
|
392
|
+
command.request_representation = Google::Apis::DataprocV1::Batch::Representation
|
393
|
+
command.request_object = batch_object
|
394
|
+
command.response_representation = Google::Apis::DataprocV1::Operation::Representation
|
395
|
+
command.response_class = Google::Apis::DataprocV1::Operation
|
396
|
+
command.params['parent'] = parent unless parent.nil?
|
397
|
+
command.query['batchId'] = batch_id unless batch_id.nil?
|
398
|
+
command.query['requestId'] = request_id unless request_id.nil?
|
399
|
+
command.query['fields'] = fields unless fields.nil?
|
400
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
401
|
+
execute_or_queue_command(command, &block)
|
402
|
+
end
|
403
|
+
|
404
|
+
# Deletes the batch workload resource. If the batch is not in terminal state,
|
405
|
+
# the delete fails and the response returns FAILED_PRECONDITION.
|
406
|
+
# @param [String] name
|
407
|
+
# Required. The name of the batch resource to delete.
|
408
|
+
# @param [String] fields
|
409
|
+
# Selector specifying which fields to include in a partial response.
|
410
|
+
# @param [String] quota_user
|
411
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
412
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
413
|
+
# @param [Google::Apis::RequestOptions] options
|
414
|
+
# Request-specific options
|
415
|
+
#
|
416
|
+
# @yield [result, err] Result & error if block supplied
|
417
|
+
# @yieldparam result [Google::Apis::DataprocV1::Empty] parsed result object
|
418
|
+
# @yieldparam err [StandardError] error object if request failed
|
419
|
+
#
|
420
|
+
# @return [Google::Apis::DataprocV1::Empty]
|
421
|
+
#
|
422
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
423
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
424
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
425
|
+
def delete_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
|
426
|
+
command = make_simple_command(:delete, 'v1/{+name}', options)
|
427
|
+
command.response_representation = Google::Apis::DataprocV1::Empty::Representation
|
428
|
+
command.response_class = Google::Apis::DataprocV1::Empty
|
429
|
+
command.params['name'] = name unless name.nil?
|
430
|
+
command.query['fields'] = fields unless fields.nil?
|
431
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
432
|
+
execute_or_queue_command(command, &block)
|
433
|
+
end
|
434
|
+
|
435
|
+
# Gets the batch workload resource representation.
|
436
|
+
# @param [String] name
|
437
|
+
# Required. The name of the batch to retrieve.
|
438
|
+
# @param [String] fields
|
439
|
+
# Selector specifying which fields to include in a partial response.
|
440
|
+
# @param [String] quota_user
|
441
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
442
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
443
|
+
# @param [Google::Apis::RequestOptions] options
|
444
|
+
# Request-specific options
|
445
|
+
#
|
446
|
+
# @yield [result, err] Result & error if block supplied
|
447
|
+
# @yieldparam result [Google::Apis::DataprocV1::Batch] parsed result object
|
448
|
+
# @yieldparam err [StandardError] error object if request failed
|
449
|
+
#
|
450
|
+
# @return [Google::Apis::DataprocV1::Batch]
|
451
|
+
#
|
452
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
453
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
454
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
455
|
+
def get_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
|
456
|
+
command = make_simple_command(:get, 'v1/{+name}', options)
|
457
|
+
command.response_representation = Google::Apis::DataprocV1::Batch::Representation
|
458
|
+
command.response_class = Google::Apis::DataprocV1::Batch
|
459
|
+
command.params['name'] = name unless name.nil?
|
460
|
+
command.query['fields'] = fields unless fields.nil?
|
461
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
462
|
+
execute_or_queue_command(command, &block)
|
463
|
+
end
|
464
|
+
|
465
|
+
# Lists batch workloads.
|
466
|
+
# @param [String] parent
|
467
|
+
# Required. The parent, which owns this collection of batches.
|
468
|
+
# @param [Fixnum] page_size
|
469
|
+
# Optional. The maximum number of batches to return in each response. The
|
470
|
+
# service may return fewer than this value. The default page size is 20; the
|
471
|
+
# maximum page size is 1000.
|
472
|
+
# @param [String] page_token
|
473
|
+
# Optional. A page token received from a previous ListBatches call. Provide this
|
474
|
+
# token to retrieve the subsequent page.
|
475
|
+
# @param [String] fields
|
476
|
+
# Selector specifying which fields to include in a partial response.
|
477
|
+
# @param [String] quota_user
|
478
|
+
# Available to use for quota purposes for server-side applications. Can be any
|
479
|
+
# arbitrary string assigned to a user, but should not exceed 40 characters.
|
480
|
+
# @param [Google::Apis::RequestOptions] options
|
481
|
+
# Request-specific options
|
482
|
+
#
|
483
|
+
# @yield [result, err] Result & error if block supplied
|
484
|
+
# @yieldparam result [Google::Apis::DataprocV1::ListBatchesResponse] parsed result object
|
485
|
+
# @yieldparam err [StandardError] error object if request failed
|
486
|
+
#
|
487
|
+
# @return [Google::Apis::DataprocV1::ListBatchesResponse]
|
488
|
+
#
|
489
|
+
# @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
|
490
|
+
# @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
|
491
|
+
# @raise [Google::Apis::AuthorizationError] Authorization is required
|
492
|
+
def list_project_location_batches(parent, page_size: nil, page_token: nil, fields: nil, quota_user: nil, options: nil, &block)
|
493
|
+
command = make_simple_command(:get, 'v1/{+parent}/batches', options)
|
494
|
+
command.response_representation = Google::Apis::DataprocV1::ListBatchesResponse::Representation
|
495
|
+
command.response_class = Google::Apis::DataprocV1::ListBatchesResponse
|
496
|
+
command.params['parent'] = parent unless parent.nil?
|
497
|
+
command.query['pageSize'] = page_size unless page_size.nil?
|
498
|
+
command.query['pageToken'] = page_token unless page_token.nil?
|
499
|
+
command.query['fields'] = fields unless fields.nil?
|
500
|
+
command.query['quotaUser'] = quota_user unless quota_user.nil?
|
501
|
+
execute_or_queue_command(command, &block)
|
502
|
+
end
|
503
|
+
|
355
504
|
# Creates new workflow template.
|
356
505
|
# @param [String] parent
|
357
506
|
# Required. The resource name of the region or location, as described in https://
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: google-apis-dataproc_v1
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.19.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Google LLC
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2022-01-
|
11
|
+
date: 2022-01-17 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: google-apis-core
|
@@ -58,7 +58,7 @@ licenses:
|
|
58
58
|
metadata:
|
59
59
|
bug_tracker_uri: https://github.com/googleapis/google-api-ruby-client/issues
|
60
60
|
changelog_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1/CHANGELOG.md
|
61
|
-
documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.
|
61
|
+
documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.19.0
|
62
62
|
source_code_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1
|
63
63
|
post_install_message:
|
64
64
|
rdoc_options: []
|
@@ -75,7 +75,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
75
75
|
- !ruby/object:Gem::Version
|
76
76
|
version: '0'
|
77
77
|
requirements: []
|
78
|
-
rubygems_version: 3.3.
|
78
|
+
rubygems_version: 3.3.5
|
79
79
|
signing_key:
|
80
80
|
specification_version: 4
|
81
81
|
summary: Simple REST client for Cloud Dataproc API V1
|