google-apis-dataproc_v1 0.18.0 → 0.19.0

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: bf1feca2f1c11ebedec266a2b5e0b508e97ccd77ca7f5f4c1720803516c37a1f
4
- data.tar.gz: b10349942e5e629ec7af13d9493adf5f206085a526fd04a1ea74ee8adfbced92
3
+ metadata.gz: b49e139fab6c34e0bd6707069e53c4105077e7abe1dc22ba91fe76124fc9410e
4
+ data.tar.gz: 4f11bed732911ec54b23d5d6eb1189e789e35b858c9abb20fe1b9f7bc4b8ac8b
5
5
  SHA512:
6
- metadata.gz: 2b3125d24f0eaac4cbeb6e082f85ad7f8a907754038cd6ed3e841c3cf8df6a91ef96e4af5c6b61afd4aff1e6a09fc062ee0aa7801a6a2a64b66f7dd63eaa1647
7
- data.tar.gz: aee8dbedaef3365d6f9c651a07aea954dfc5ca330daaaa78b4c29838643d454fc4a23bf99d32d6f63852695176dbcde36f53c6b91dc8b6a135ef592d4bc34e96
6
+ metadata.gz: 74cd3145e6b509ee3ce020ad5d016f2ffc30e53505ac4c70b70b910ae336f694861c29331e4e681bb5f7efdb3126aa4dd027631ab2b29f417dfcdbeedc49f362
7
+ data.tar.gz: 0dc5eccb08bb472d8adcabbb54510b287c02eb923f359c365526c92326ccfc17331b19582180bde9b1f45767868af314c1e1b6350ff953040bedd9f351a42451
data/CHANGELOG.md CHANGED
@@ -1,5 +1,10 @@
1
1
  # Release history for google-apis-dataproc_v1
2
2
 
3
+ ### v0.19.0 (2022-01-14)
4
+
5
+ * Regenerated from discovery document revision 20220111
6
+ * Regenerated using generator version 0.4.1
7
+
3
8
  ### v0.18.0 (2022-01-07)
4
9
 
5
10
  * Regenerated from discovery document revision 20220105
@@ -245,6 +245,131 @@ module Google
245
245
  end
246
246
  end
247
247
 
248
+ # A representation of a batch workload in the service.
249
+ class Batch
250
+ include Google::Apis::Core::Hashable
251
+
252
+ # Output only. The time when the batch was created.
253
+ # Corresponds to the JSON property `createTime`
254
+ # @return [String]
255
+ attr_accessor :create_time
256
+
257
+ # Output only. The email address of the user who created the batch.
258
+ # Corresponds to the JSON property `creator`
259
+ # @return [String]
260
+ attr_accessor :creator
261
+
262
+ # Environment configuration for a workload.
263
+ # Corresponds to the JSON property `environmentConfig`
264
+ # @return [Google::Apis::DataprocV1::EnvironmentConfig]
265
+ attr_accessor :environment_config
266
+
267
+ # Optional. The labels to associate with this batch. Label keys must contain 1
268
+ # to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/
269
+ # rfc1035.txt). Label values may be empty, but, if present, must contain 1 to 63
270
+ # characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt)
271
+ # . No more than 32 labels can be associated with a batch.
272
+ # Corresponds to the JSON property `labels`
273
+ # @return [Hash<String,String>]
274
+ attr_accessor :labels
275
+
276
+ # Output only. The resource name of the batch.
277
+ # Corresponds to the JSON property `name`
278
+ # @return [String]
279
+ attr_accessor :name
280
+
281
+ # Output only. The resource name of the operation associated with this batch.
282
+ # Corresponds to the JSON property `operation`
283
+ # @return [String]
284
+ attr_accessor :operation
285
+
286
+ # A configuration for running an Apache PySpark (https://spark.apache.org/docs/
287
+ # latest/api/python/getting_started/quickstart.html) batch workload.
288
+ # Corresponds to the JSON property `pysparkBatch`
289
+ # @return [Google::Apis::DataprocV1::PySparkBatch]
290
+ attr_accessor :pyspark_batch
291
+
292
+ # Runtime configuration for a workload.
293
+ # Corresponds to the JSON property `runtimeConfig`
294
+ # @return [Google::Apis::DataprocV1::RuntimeConfig]
295
+ attr_accessor :runtime_config
296
+
297
+ # Runtime information about workload execution.
298
+ # Corresponds to the JSON property `runtimeInfo`
299
+ # @return [Google::Apis::DataprocV1::RuntimeInfo]
300
+ attr_accessor :runtime_info
301
+
302
+ # A configuration for running an Apache Spark (http://spark.apache.org/) batch
303
+ # workload.
304
+ # Corresponds to the JSON property `sparkBatch`
305
+ # @return [Google::Apis::DataprocV1::SparkBatch]
306
+ attr_accessor :spark_batch
307
+
308
+ # A configuration for running an Apache SparkR (https://spark.apache.org/docs/
309
+ # latest/sparkr.html) batch workload.
310
+ # Corresponds to the JSON property `sparkRBatch`
311
+ # @return [Google::Apis::DataprocV1::SparkRBatch]
312
+ attr_accessor :spark_r_batch
313
+
314
+ # A configuration for running Apache Spark SQL (http://spark.apache.org/sql/)
315
+ # queries as a batch workload.
316
+ # Corresponds to the JSON property `sparkSqlBatch`
317
+ # @return [Google::Apis::DataprocV1::SparkSqlBatch]
318
+ attr_accessor :spark_sql_batch
319
+
320
+ # Output only. The state of the batch.
321
+ # Corresponds to the JSON property `state`
322
+ # @return [String]
323
+ attr_accessor :state
324
+
325
+ # Output only. Historical state information for the batch.
326
+ # Corresponds to the JSON property `stateHistory`
327
+ # @return [Array<Google::Apis::DataprocV1::StateHistory>]
328
+ attr_accessor :state_history
329
+
330
+ # Output only. Batch state details, such as a failure description if the state
331
+ # is FAILED.
332
+ # Corresponds to the JSON property `stateMessage`
333
+ # @return [String]
334
+ attr_accessor :state_message
335
+
336
+ # Output only. The time when the batch entered a current state.
337
+ # Corresponds to the JSON property `stateTime`
338
+ # @return [String]
339
+ attr_accessor :state_time
340
+
341
+ # Output only. A batch UUID (Unique Universal Identifier). The service generates
342
+ # this value when it creates the batch.
343
+ # Corresponds to the JSON property `uuid`
344
+ # @return [String]
345
+ attr_accessor :uuid
346
+
347
+ def initialize(**args)
348
+ update!(**args)
349
+ end
350
+
351
+ # Update properties of this object
352
+ def update!(**args)
353
+ @create_time = args[:create_time] if args.key?(:create_time)
354
+ @creator = args[:creator] if args.key?(:creator)
355
+ @environment_config = args[:environment_config] if args.key?(:environment_config)
356
+ @labels = args[:labels] if args.key?(:labels)
357
+ @name = args[:name] if args.key?(:name)
358
+ @operation = args[:operation] if args.key?(:operation)
359
+ @pyspark_batch = args[:pyspark_batch] if args.key?(:pyspark_batch)
360
+ @runtime_config = args[:runtime_config] if args.key?(:runtime_config)
361
+ @runtime_info = args[:runtime_info] if args.key?(:runtime_info)
362
+ @spark_batch = args[:spark_batch] if args.key?(:spark_batch)
363
+ @spark_r_batch = args[:spark_r_batch] if args.key?(:spark_r_batch)
364
+ @spark_sql_batch = args[:spark_sql_batch] if args.key?(:spark_sql_batch)
365
+ @state = args[:state] if args.key?(:state)
366
+ @state_history = args[:state_history] if args.key?(:state_history)
367
+ @state_message = args[:state_message] if args.key?(:state_message)
368
+ @state_time = args[:state_time] if args.key?(:state_time)
369
+ @uuid = args[:uuid] if args.key?(:uuid)
370
+ end
371
+ end
372
+
248
373
  # Metadata describing the Batch operation.
249
374
  class BatchOperationMetadata
250
375
  include Google::Apis::Core::Hashable
@@ -979,6 +1104,74 @@ module Google
979
1104
  end
980
1105
  end
981
1106
 
1107
+ # Environment configuration for a workload.
1108
+ class EnvironmentConfig
1109
+ include Google::Apis::Core::Hashable
1110
+
1111
+ # Execution configuration for a workload.
1112
+ # Corresponds to the JSON property `executionConfig`
1113
+ # @return [Google::Apis::DataprocV1::ExecutionConfig]
1114
+ attr_accessor :execution_config
1115
+
1116
+ # Auxiliary services configuration for a workload.
1117
+ # Corresponds to the JSON property `peripheralsConfig`
1118
+ # @return [Google::Apis::DataprocV1::PeripheralsConfig]
1119
+ attr_accessor :peripherals_config
1120
+
1121
+ def initialize(**args)
1122
+ update!(**args)
1123
+ end
1124
+
1125
+ # Update properties of this object
1126
+ def update!(**args)
1127
+ @execution_config = args[:execution_config] if args.key?(:execution_config)
1128
+ @peripherals_config = args[:peripherals_config] if args.key?(:peripherals_config)
1129
+ end
1130
+ end
1131
+
1132
+ # Execution configuration for a workload.
1133
+ class ExecutionConfig
1134
+ include Google::Apis::Core::Hashable
1135
+
1136
+ # Optional. The Cloud KMS key to use for encryption.
1137
+ # Corresponds to the JSON property `kmsKey`
1138
+ # @return [String]
1139
+ attr_accessor :kms_key
1140
+
1141
+ # Optional. Tags used for network traffic control.
1142
+ # Corresponds to the JSON property `networkTags`
1143
+ # @return [Array<String>]
1144
+ attr_accessor :network_tags
1145
+
1146
+ # Optional. Network URI to connect workload to.
1147
+ # Corresponds to the JSON property `networkUri`
1148
+ # @return [String]
1149
+ attr_accessor :network_uri
1150
+
1151
+ # Optional. Service account that used to execute workload.
1152
+ # Corresponds to the JSON property `serviceAccount`
1153
+ # @return [String]
1154
+ attr_accessor :service_account
1155
+
1156
+ # Optional. Subnetwork URI to connect workload to.
1157
+ # Corresponds to the JSON property `subnetworkUri`
1158
+ # @return [String]
1159
+ attr_accessor :subnetwork_uri
1160
+
1161
+ def initialize(**args)
1162
+ update!(**args)
1163
+ end
1164
+
1165
+ # Update properties of this object
1166
+ def update!(**args)
1167
+ @kms_key = args[:kms_key] if args.key?(:kms_key)
1168
+ @network_tags = args[:network_tags] if args.key?(:network_tags)
1169
+ @network_uri = args[:network_uri] if args.key?(:network_uri)
1170
+ @service_account = args[:service_account] if args.key?(:service_account)
1171
+ @subnetwork_uri = args[:subnetwork_uri] if args.key?(:subnetwork_uri)
1172
+ end
1173
+ end
1174
+
982
1175
  # Represents a textual expression in the Common Expression Language (CEL) syntax.
983
1176
  # CEL is a C-like expression language. The syntax and semantics of CEL are
984
1177
  # documented at https://github.com/google/cel-spec.Example (Comparison): title: "
@@ -2168,6 +2361,32 @@ module Google
2168
2361
  end
2169
2362
  end
2170
2363
 
2364
+ # A list of batch workloads.
2365
+ class ListBatchesResponse
2366
+ include Google::Apis::Core::Hashable
2367
+
2368
+ # The batches from the specified collection.
2369
+ # Corresponds to the JSON property `batches`
2370
+ # @return [Array<Google::Apis::DataprocV1::Batch>]
2371
+ attr_accessor :batches
2372
+
2373
+ # A token, which can be sent as page_token to retrieve the next page. If this
2374
+ # field is omitted, there are no subsequent pages.
2375
+ # Corresponds to the JSON property `nextPageToken`
2376
+ # @return [String]
2377
+ attr_accessor :next_page_token
2378
+
2379
+ def initialize(**args)
2380
+ update!(**args)
2381
+ end
2382
+
2383
+ # Update properties of this object
2384
+ def update!(**args)
2385
+ @batches = args[:batches] if args.key?(:batches)
2386
+ @next_page_token = args[:next_page_token] if args.key?(:next_page_token)
2387
+ end
2388
+ end
2389
+
2171
2390
  # The list of all clusters in a project.
2172
2391
  class ListClustersResponse
2173
2392
  include Google::Apis::Core::Hashable
@@ -2654,6 +2873,32 @@ module Google
2654
2873
  end
2655
2874
  end
2656
2875
 
2876
+ # Auxiliary services configuration for a workload.
2877
+ class PeripheralsConfig
2878
+ include Google::Apis::Core::Hashable
2879
+
2880
+ # Optional. Resource name of an existing Dataproc Metastore service.Example:
2881
+ # projects/[project_id]/locations/[region]/services/[service_id]
2882
+ # Corresponds to the JSON property `metastoreService`
2883
+ # @return [String]
2884
+ attr_accessor :metastore_service
2885
+
2886
+ # Spark History Server configuration for the workload.
2887
+ # Corresponds to the JSON property `sparkHistoryServerConfig`
2888
+ # @return [Google::Apis::DataprocV1::SparkHistoryServerConfig]
2889
+ attr_accessor :spark_history_server_config
2890
+
2891
+ def initialize(**args)
2892
+ update!(**args)
2893
+ end
2894
+
2895
+ # Update properties of this object
2896
+ def update!(**args)
2897
+ @metastore_service = args[:metastore_service] if args.key?(:metastore_service)
2898
+ @spark_history_server_config = args[:spark_history_server_config] if args.key?(:spark_history_server_config)
2899
+ end
2900
+ end
2901
+
2657
2902
  # A Dataproc job for running Apache Pig (https://pig.apache.org/) queries on
2658
2903
  # YARN.
2659
2904
  class PigJob
@@ -2869,6 +3114,63 @@ module Google
2869
3114
  end
2870
3115
  end
2871
3116
 
3117
+ # A configuration for running an Apache PySpark (https://spark.apache.org/docs/
3118
+ # latest/api/python/getting_started/quickstart.html) batch workload.
3119
+ class PySparkBatch
3120
+ include Google::Apis::Core::Hashable
3121
+
3122
+ # Optional. HCFS URIs of archives to be extracted into the working directory of
3123
+ # each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
3124
+ # Corresponds to the JSON property `archiveUris`
3125
+ # @return [Array<String>]
3126
+ attr_accessor :archive_uris
3127
+
3128
+ # Optional. The arguments to pass to the driver. Do not include arguments that
3129
+ # can be set as batch properties, such as --conf, since a collision can occur
3130
+ # that causes an incorrect batch submission.
3131
+ # Corresponds to the JSON property `args`
3132
+ # @return [Array<String>]
3133
+ attr_accessor :args
3134
+
3135
+ # Optional. HCFS URIs of files to be placed in the working directory of each
3136
+ # executor.
3137
+ # Corresponds to the JSON property `fileUris`
3138
+ # @return [Array<String>]
3139
+ attr_accessor :file_uris
3140
+
3141
+ # Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
3142
+ # and tasks.
3143
+ # Corresponds to the JSON property `jarFileUris`
3144
+ # @return [Array<String>]
3145
+ attr_accessor :jar_file_uris
3146
+
3147
+ # Required. The HCFS URI of the main Python file to use as the Spark driver.
3148
+ # Must be a .py file.
3149
+ # Corresponds to the JSON property `mainPythonFileUri`
3150
+ # @return [String]
3151
+ attr_accessor :main_python_file_uri
3152
+
3153
+ # Optional. HCFS file URIs of Python files to pass to the PySpark framework.
3154
+ # Supported file types: .py, .egg, and .zip.
3155
+ # Corresponds to the JSON property `pythonFileUris`
3156
+ # @return [Array<String>]
3157
+ attr_accessor :python_file_uris
3158
+
3159
+ def initialize(**args)
3160
+ update!(**args)
3161
+ end
3162
+
3163
+ # Update properties of this object
3164
+ def update!(**args)
3165
+ @archive_uris = args[:archive_uris] if args.key?(:archive_uris)
3166
+ @args = args[:args] if args.key?(:args)
3167
+ @file_uris = args[:file_uris] if args.key?(:file_uris)
3168
+ @jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
3169
+ @main_python_file_uri = args[:main_python_file_uri] if args.key?(:main_python_file_uri)
3170
+ @python_file_uris = args[:python_file_uris] if args.key?(:python_file_uris)
3171
+ end
3172
+ end
3173
+
2872
3174
  # A Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/
2873
3175
  # python-programming-guide.html) applications on YARN.
2874
3176
  class PySparkJob
@@ -3048,6 +3350,72 @@ module Google
3048
3350
  end
3049
3351
  end
3050
3352
 
3353
+ # Runtime configuration for a workload.
3354
+ class RuntimeConfig
3355
+ include Google::Apis::Core::Hashable
3356
+
3357
+ # Optional. Optional custom container image for the job runtime environment. If
3358
+ # not specified, a default container image will be used.
3359
+ # Corresponds to the JSON property `containerImage`
3360
+ # @return [String]
3361
+ attr_accessor :container_image
3362
+
3363
+ # Optional. A mapping of property names to values, which are used to configure
3364
+ # workload execution.
3365
+ # Corresponds to the JSON property `properties`
3366
+ # @return [Hash<String,String>]
3367
+ attr_accessor :properties
3368
+
3369
+ # Optional. Version of the batch runtime.
3370
+ # Corresponds to the JSON property `version`
3371
+ # @return [String]
3372
+ attr_accessor :version
3373
+
3374
+ def initialize(**args)
3375
+ update!(**args)
3376
+ end
3377
+
3378
+ # Update properties of this object
3379
+ def update!(**args)
3380
+ @container_image = args[:container_image] if args.key?(:container_image)
3381
+ @properties = args[:properties] if args.key?(:properties)
3382
+ @version = args[:version] if args.key?(:version)
3383
+ end
3384
+ end
3385
+
3386
+ # Runtime information about workload execution.
3387
+ class RuntimeInfo
3388
+ include Google::Apis::Core::Hashable
3389
+
3390
+ # Output only. A URI pointing to the location of the diagnostics tarball.
3391
+ # Corresponds to the JSON property `diagnosticOutputUri`
3392
+ # @return [String]
3393
+ attr_accessor :diagnostic_output_uri
3394
+
3395
+ # Output only. Map of remote access endpoints (such as web interfaces and APIs)
3396
+ # to their URIs.
3397
+ # Corresponds to the JSON property `endpoints`
3398
+ # @return [Hash<String,String>]
3399
+ attr_accessor :endpoints
3400
+
3401
+ # Output only. A URI pointing to the location of the stdout and stderr of the
3402
+ # workload.
3403
+ # Corresponds to the JSON property `outputUri`
3404
+ # @return [String]
3405
+ attr_accessor :output_uri
3406
+
3407
+ def initialize(**args)
3408
+ update!(**args)
3409
+ end
3410
+
3411
+ # Update properties of this object
3412
+ def update!(**args)
3413
+ @diagnostic_output_uri = args[:diagnostic_output_uri] if args.key?(:diagnostic_output_uri)
3414
+ @endpoints = args[:endpoints] if args.key?(:endpoints)
3415
+ @output_uri = args[:output_uri] if args.key?(:output_uri)
3416
+ end
3417
+ end
3418
+
3051
3419
  # Security related configuration, including encryption, Kerberos, etc.
3052
3420
  class SecurityConfig
3053
3421
  include Google::Apis::Core::Hashable
@@ -3258,6 +3626,83 @@ module Google
3258
3626
  end
3259
3627
  end
3260
3628
 
3629
+ # A configuration for running an Apache Spark (http://spark.apache.org/) batch
3630
+ # workload.
3631
+ class SparkBatch
3632
+ include Google::Apis::Core::Hashable
3633
+
3634
+ # Optional. HCFS URIs of archives to be extracted into the working directory of
3635
+ # each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
3636
+ # Corresponds to the JSON property `archiveUris`
3637
+ # @return [Array<String>]
3638
+ attr_accessor :archive_uris
3639
+
3640
+ # Optional. The arguments to pass to the driver. Do not include arguments that
3641
+ # can be set as batch properties, such as --conf, since a collision can occur
3642
+ # that causes an incorrect batch submission.
3643
+ # Corresponds to the JSON property `args`
3644
+ # @return [Array<String>]
3645
+ attr_accessor :args
3646
+
3647
+ # Optional. HCFS URIs of files to be placed in the working directory of each
3648
+ # executor.
3649
+ # Corresponds to the JSON property `fileUris`
3650
+ # @return [Array<String>]
3651
+ attr_accessor :file_uris
3652
+
3653
+ # Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
3654
+ # and tasks.
3655
+ # Corresponds to the JSON property `jarFileUris`
3656
+ # @return [Array<String>]
3657
+ attr_accessor :jar_file_uris
3658
+
3659
+ # Optional. The name of the driver main class. The jar file that contains the
3660
+ # class must be in the classpath or specified in jar_file_uris.
3661
+ # Corresponds to the JSON property `mainClass`
3662
+ # @return [String]
3663
+ attr_accessor :main_class
3664
+
3665
+ # Optional. The HCFS URI of the jar file that contains the main class.
3666
+ # Corresponds to the JSON property `mainJarFileUri`
3667
+ # @return [String]
3668
+ attr_accessor :main_jar_file_uri
3669
+
3670
+ def initialize(**args)
3671
+ update!(**args)
3672
+ end
3673
+
3674
+ # Update properties of this object
3675
+ def update!(**args)
3676
+ @archive_uris = args[:archive_uris] if args.key?(:archive_uris)
3677
+ @args = args[:args] if args.key?(:args)
3678
+ @file_uris = args[:file_uris] if args.key?(:file_uris)
3679
+ @jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
3680
+ @main_class = args[:main_class] if args.key?(:main_class)
3681
+ @main_jar_file_uri = args[:main_jar_file_uri] if args.key?(:main_jar_file_uri)
3682
+ end
3683
+ end
3684
+
3685
+ # Spark History Server configuration for the workload.
3686
+ class SparkHistoryServerConfig
3687
+ include Google::Apis::Core::Hashable
3688
+
3689
+ # Optional. Resource name of an existing Dataproc Cluster to act as a Spark
3690
+ # History Server for the workload.Example: projects/[project_id]/regions/[region]
3691
+ # /clusters/[cluster_name]
3692
+ # Corresponds to the JSON property `dataprocCluster`
3693
+ # @return [String]
3694
+ attr_accessor :dataproc_cluster
3695
+
3696
+ def initialize(**args)
3697
+ update!(**args)
3698
+ end
3699
+
3700
+ # Update properties of this object
3701
+ def update!(**args)
3702
+ @dataproc_cluster = args[:dataproc_cluster] if args.key?(:dataproc_cluster)
3703
+ end
3704
+ end
3705
+
3261
3706
  # A Dataproc job for running Apache Spark (http://spark.apache.org/)
3262
3707
  # applications on YARN.
3263
3708
  class SparkJob
@@ -3329,6 +3774,49 @@ module Google
3329
3774
  end
3330
3775
  end
3331
3776
 
3777
+ # A configuration for running an Apache SparkR (https://spark.apache.org/docs/
3778
+ # latest/sparkr.html) batch workload.
3779
+ class SparkRBatch
3780
+ include Google::Apis::Core::Hashable
3781
+
3782
+ # Optional. HCFS URIs of archives to be extracted into the working directory of
3783
+ # each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
3784
+ # Corresponds to the JSON property `archiveUris`
3785
+ # @return [Array<String>]
3786
+ attr_accessor :archive_uris
3787
+
3788
+ # Optional. The arguments to pass to the Spark driver. Do not include arguments
3789
+ # that can be set as batch properties, such as --conf, since a collision can
3790
+ # occur that causes an incorrect batch submission.
3791
+ # Corresponds to the JSON property `args`
3792
+ # @return [Array<String>]
3793
+ attr_accessor :args
3794
+
3795
+ # Optional. HCFS URIs of files to be placed in the working directory of each
3796
+ # executor.
3797
+ # Corresponds to the JSON property `fileUris`
3798
+ # @return [Array<String>]
3799
+ attr_accessor :file_uris
3800
+
3801
+ # Required. The HCFS URI of the main R file to use as the driver. Must be a .R
3802
+ # or .r file.
3803
+ # Corresponds to the JSON property `mainRFileUri`
3804
+ # @return [String]
3805
+ attr_accessor :main_r_file_uri
3806
+
3807
+ def initialize(**args)
3808
+ update!(**args)
3809
+ end
3810
+
3811
+ # Update properties of this object
3812
+ def update!(**args)
3813
+ @archive_uris = args[:archive_uris] if args.key?(:archive_uris)
3814
+ @args = args[:args] if args.key?(:args)
3815
+ @file_uris = args[:file_uris] if args.key?(:file_uris)
3816
+ @main_r_file_uri = args[:main_r_file_uri] if args.key?(:main_r_file_uri)
3817
+ end
3818
+ end
3819
+
3332
3820
  # A Dataproc job for running Apache SparkR (https://spark.apache.org/docs/latest/
3333
3821
  # sparkr.html) applications on YARN.
3334
3822
  class SparkRJob
@@ -3387,6 +3875,40 @@ module Google
3387
3875
  end
3388
3876
  end
3389
3877
 
3878
+ # A configuration for running Apache Spark SQL (http://spark.apache.org/sql/)
3879
+ # queries as a batch workload.
3880
+ class SparkSqlBatch
3881
+ include Google::Apis::Core::Hashable
3882
+
3883
+ # Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
3884
+ # Corresponds to the JSON property `jarFileUris`
3885
+ # @return [Array<String>]
3886
+ attr_accessor :jar_file_uris
3887
+
3888
+ # Required. The HCFS URI of the script that contains Spark SQL queries to
3889
+ # execute.
3890
+ # Corresponds to the JSON property `queryFileUri`
3891
+ # @return [String]
3892
+ attr_accessor :query_file_uri
3893
+
3894
+ # Optional. Mapping of query variable names to values (equivalent to the Spark
3895
+ # SQL command: SET name="value";).
3896
+ # Corresponds to the JSON property `queryVariables`
3897
+ # @return [Hash<String,String>]
3898
+ attr_accessor :query_variables
3899
+
3900
+ def initialize(**args)
3901
+ update!(**args)
3902
+ end
3903
+
3904
+ # Update properties of this object
3905
+ def update!(**args)
3906
+ @jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
3907
+ @query_file_uri = args[:query_file_uri] if args.key?(:query_file_uri)
3908
+ @query_variables = args[:query_variables] if args.key?(:query_variables)
3909
+ end
3910
+ end
3911
+
3390
3912
  # A Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/)
3391
3913
  # queries.
3392
3914
  class SparkSqlJob
@@ -3536,6 +4058,37 @@ module Google
3536
4058
  end
3537
4059
  end
3538
4060
 
4061
+ # Historical state information.
4062
+ class StateHistory
4063
+ include Google::Apis::Core::Hashable
4064
+
4065
+ # Output only. The state of the batch at this point in history.
4066
+ # Corresponds to the JSON property `state`
4067
+ # @return [String]
4068
+ attr_accessor :state
4069
+
4070
+ # Output only. Details about the state at this point in history.
4071
+ # Corresponds to the JSON property `stateMessage`
4072
+ # @return [String]
4073
+ attr_accessor :state_message
4074
+
4075
+ # Output only. The time when the batch entered the historical state.
4076
+ # Corresponds to the JSON property `stateStartTime`
4077
+ # @return [String]
4078
+ attr_accessor :state_start_time
4079
+
4080
+ def initialize(**args)
4081
+ update!(**args)
4082
+ end
4083
+
4084
+ # Update properties of this object
4085
+ def update!(**args)
4086
+ @state = args[:state] if args.key?(:state)
4087
+ @state_message = args[:state_message] if args.key?(:state_message)
4088
+ @state_start_time = args[:state_start_time] if args.key?(:state_start_time)
4089
+ end
4090
+ end
4091
+
3539
4092
  # The Status type defines a logical error model that is suitable for different
3540
4093
  # programming environments, including REST APIs and RPC APIs. It is used by gRPC
3541
4094
  # (https://github.com/grpc). Each Status message contains three pieces of data:
@@ -16,13 +16,13 @@ module Google
16
16
  module Apis
17
17
  module DataprocV1
18
18
  # Version of the google-apis-dataproc_v1 gem
19
- GEM_VERSION = "0.18.0"
19
+ GEM_VERSION = "0.19.0"
20
20
 
21
21
  # Version of the code generator used to generate this client
22
- GENERATOR_VERSION = "0.4.0"
22
+ GENERATOR_VERSION = "0.4.1"
23
23
 
24
24
  # Revision of the discovery document this client was generated from
25
- REVISION = "20220105"
25
+ REVISION = "20220111"
26
26
  end
27
27
  end
28
28
  end
@@ -52,6 +52,12 @@ module Google
52
52
  include Google::Apis::Core::JsonObjectSupport
53
53
  end
54
54
 
55
+ class Batch
56
+ class Representation < Google::Apis::Core::JsonRepresentation; end
57
+
58
+ include Google::Apis::Core::JsonObjectSupport
59
+ end
60
+
55
61
  class BatchOperationMetadata
56
62
  class Representation < Google::Apis::Core::JsonRepresentation; end
57
63
 
@@ -160,6 +166,18 @@ module Google
160
166
  include Google::Apis::Core::JsonObjectSupport
161
167
  end
162
168
 
169
+ class EnvironmentConfig
170
+ class Representation < Google::Apis::Core::JsonRepresentation; end
171
+
172
+ include Google::Apis::Core::JsonObjectSupport
173
+ end
174
+
175
+ class ExecutionConfig
176
+ class Representation < Google::Apis::Core::JsonRepresentation; end
177
+
178
+ include Google::Apis::Core::JsonObjectSupport
179
+ end
180
+
163
181
  class Expr
164
182
  class Representation < Google::Apis::Core::JsonRepresentation; end
165
183
 
@@ -292,6 +310,12 @@ module Google
292
310
  include Google::Apis::Core::JsonObjectSupport
293
311
  end
294
312
 
313
+ class ListBatchesResponse
314
+ class Representation < Google::Apis::Core::JsonRepresentation; end
315
+
316
+ include Google::Apis::Core::JsonObjectSupport
317
+ end
318
+
295
319
  class ListClustersResponse
296
320
  class Representation < Google::Apis::Core::JsonRepresentation; end
297
321
 
@@ -376,6 +400,12 @@ module Google
376
400
  include Google::Apis::Core::JsonObjectSupport
377
401
  end
378
402
 
403
+ class PeripheralsConfig
404
+ class Representation < Google::Apis::Core::JsonRepresentation; end
405
+
406
+ include Google::Apis::Core::JsonObjectSupport
407
+ end
408
+
379
409
  class PigJob
380
410
  class Representation < Google::Apis::Core::JsonRepresentation; end
381
411
 
@@ -394,6 +424,12 @@ module Google
394
424
  include Google::Apis::Core::JsonObjectSupport
395
425
  end
396
426
 
427
+ class PySparkBatch
428
+ class Representation < Google::Apis::Core::JsonRepresentation; end
429
+
430
+ include Google::Apis::Core::JsonObjectSupport
431
+ end
432
+
397
433
  class PySparkJob
398
434
  class Representation < Google::Apis::Core::JsonRepresentation; end
399
435
 
@@ -424,6 +460,18 @@ module Google
424
460
  include Google::Apis::Core::JsonObjectSupport
425
461
  end
426
462
 
463
+ class RuntimeConfig
464
+ class Representation < Google::Apis::Core::JsonRepresentation; end
465
+
466
+ include Google::Apis::Core::JsonObjectSupport
467
+ end
468
+
469
+ class RuntimeInfo
470
+ class Representation < Google::Apis::Core::JsonRepresentation; end
471
+
472
+ include Google::Apis::Core::JsonObjectSupport
473
+ end
474
+
427
475
  class SecurityConfig
428
476
  class Representation < Google::Apis::Core::JsonRepresentation; end
429
477
 
@@ -454,18 +502,42 @@ module Google
454
502
  include Google::Apis::Core::JsonObjectSupport
455
503
  end
456
504
 
505
+ class SparkBatch
506
+ class Representation < Google::Apis::Core::JsonRepresentation; end
507
+
508
+ include Google::Apis::Core::JsonObjectSupport
509
+ end
510
+
511
+ class SparkHistoryServerConfig
512
+ class Representation < Google::Apis::Core::JsonRepresentation; end
513
+
514
+ include Google::Apis::Core::JsonObjectSupport
515
+ end
516
+
457
517
  class SparkJob
458
518
  class Representation < Google::Apis::Core::JsonRepresentation; end
459
519
 
460
520
  include Google::Apis::Core::JsonObjectSupport
461
521
  end
462
522
 
523
+ class SparkRBatch
524
+ class Representation < Google::Apis::Core::JsonRepresentation; end
525
+
526
+ include Google::Apis::Core::JsonObjectSupport
527
+ end
528
+
463
529
  class SparkRJob
464
530
  class Representation < Google::Apis::Core::JsonRepresentation; end
465
531
 
466
532
  include Google::Apis::Core::JsonObjectSupport
467
533
  end
468
534
 
535
+ class SparkSqlBatch
536
+ class Representation < Google::Apis::Core::JsonRepresentation; end
537
+
538
+ include Google::Apis::Core::JsonObjectSupport
539
+ end
540
+
469
541
  class SparkSqlJob
470
542
  class Representation < Google::Apis::Core::JsonRepresentation; end
471
543
 
@@ -484,6 +556,12 @@ module Google
484
556
  include Google::Apis::Core::JsonObjectSupport
485
557
  end
486
558
 
559
+ class StateHistory
560
+ class Representation < Google::Apis::Core::JsonRepresentation; end
561
+
562
+ include Google::Apis::Core::JsonObjectSupport
563
+ end
564
+
487
565
  class Status
488
566
  class Representation < Google::Apis::Core::JsonRepresentation; end
489
567
 
@@ -614,6 +692,37 @@ module Google
614
692
  end
615
693
  end
616
694
 
695
+ class Batch
696
+ # @private
697
+ class Representation < Google::Apis::Core::JsonRepresentation
698
+ property :create_time, as: 'createTime'
699
+ property :creator, as: 'creator'
700
+ property :environment_config, as: 'environmentConfig', class: Google::Apis::DataprocV1::EnvironmentConfig, decorator: Google::Apis::DataprocV1::EnvironmentConfig::Representation
701
+
702
+ hash :labels, as: 'labels'
703
+ property :name, as: 'name'
704
+ property :operation, as: 'operation'
705
+ property :pyspark_batch, as: 'pysparkBatch', class: Google::Apis::DataprocV1::PySparkBatch, decorator: Google::Apis::DataprocV1::PySparkBatch::Representation
706
+
707
+ property :runtime_config, as: 'runtimeConfig', class: Google::Apis::DataprocV1::RuntimeConfig, decorator: Google::Apis::DataprocV1::RuntimeConfig::Representation
708
+
709
+ property :runtime_info, as: 'runtimeInfo', class: Google::Apis::DataprocV1::RuntimeInfo, decorator: Google::Apis::DataprocV1::RuntimeInfo::Representation
710
+
711
+ property :spark_batch, as: 'sparkBatch', class: Google::Apis::DataprocV1::SparkBatch, decorator: Google::Apis::DataprocV1::SparkBatch::Representation
712
+
713
+ property :spark_r_batch, as: 'sparkRBatch', class: Google::Apis::DataprocV1::SparkRBatch, decorator: Google::Apis::DataprocV1::SparkRBatch::Representation
714
+
715
+ property :spark_sql_batch, as: 'sparkSqlBatch', class: Google::Apis::DataprocV1::SparkSqlBatch, decorator: Google::Apis::DataprocV1::SparkSqlBatch::Representation
716
+
717
+ property :state, as: 'state'
718
+ collection :state_history, as: 'stateHistory', class: Google::Apis::DataprocV1::StateHistory, decorator: Google::Apis::DataprocV1::StateHistory::Representation
719
+
720
+ property :state_message, as: 'stateMessage'
721
+ property :state_time, as: 'stateTime'
722
+ property :uuid, as: 'uuid'
723
+ end
724
+ end
725
+
617
726
  class BatchOperationMetadata
618
727
  # @private
619
728
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -808,6 +917,27 @@ module Google
808
917
  end
809
918
  end
810
919
 
920
+ class EnvironmentConfig
921
+ # @private
922
+ class Representation < Google::Apis::Core::JsonRepresentation
923
+ property :execution_config, as: 'executionConfig', class: Google::Apis::DataprocV1::ExecutionConfig, decorator: Google::Apis::DataprocV1::ExecutionConfig::Representation
924
+
925
+ property :peripherals_config, as: 'peripheralsConfig', class: Google::Apis::DataprocV1::PeripheralsConfig, decorator: Google::Apis::DataprocV1::PeripheralsConfig::Representation
926
+
927
+ end
928
+ end
929
+
930
+ class ExecutionConfig
931
+ # @private
932
+ class Representation < Google::Apis::Core::JsonRepresentation
933
+ property :kms_key, as: 'kmsKey'
934
+ collection :network_tags, as: 'networkTags'
935
+ property :network_uri, as: 'networkUri'
936
+ property :service_account, as: 'serviceAccount'
937
+ property :subnetwork_uri, as: 'subnetworkUri'
938
+ end
939
+ end
940
+
811
941
  class Expr
812
942
  # @private
813
943
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1081,6 +1211,15 @@ module Google
1081
1211
  end
1082
1212
  end
1083
1213
 
1214
+ class ListBatchesResponse
1215
+ # @private
1216
+ class Representation < Google::Apis::Core::JsonRepresentation
1217
+ collection :batches, as: 'batches', class: Google::Apis::DataprocV1::Batch, decorator: Google::Apis::DataprocV1::Batch::Representation
1218
+
1219
+ property :next_page_token, as: 'nextPageToken'
1220
+ end
1221
+ end
1222
+
1084
1223
  class ListClustersResponse
1085
1224
  # @private
1086
1225
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1221,6 +1360,15 @@ module Google
1221
1360
  end
1222
1361
  end
1223
1362
 
1363
+ class PeripheralsConfig
1364
+ # @private
1365
+ class Representation < Google::Apis::Core::JsonRepresentation
1366
+ property :metastore_service, as: 'metastoreService'
1367
+ property :spark_history_server_config, as: 'sparkHistoryServerConfig', class: Google::Apis::DataprocV1::SparkHistoryServerConfig, decorator: Google::Apis::DataprocV1::SparkHistoryServerConfig::Representation
1368
+
1369
+ end
1370
+ end
1371
+
1224
1372
  class PigJob
1225
1373
  # @private
1226
1374
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1261,6 +1409,18 @@ module Google
1261
1409
  end
1262
1410
  end
1263
1411
 
1412
+ class PySparkBatch
1413
+ # @private
1414
+ class Representation < Google::Apis::Core::JsonRepresentation
1415
+ collection :archive_uris, as: 'archiveUris'
1416
+ collection :args, as: 'args'
1417
+ collection :file_uris, as: 'fileUris'
1418
+ collection :jar_file_uris, as: 'jarFileUris'
1419
+ property :main_python_file_uri, as: 'mainPythonFileUri'
1420
+ collection :python_file_uris, as: 'pythonFileUris'
1421
+ end
1422
+ end
1423
+
1264
1424
  class PySparkJob
1265
1425
  # @private
1266
1426
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1307,6 +1467,24 @@ module Google
1307
1467
  end
1308
1468
  end
1309
1469
 
1470
+ class RuntimeConfig
1471
+ # @private
1472
+ class Representation < Google::Apis::Core::JsonRepresentation
1473
+ property :container_image, as: 'containerImage'
1474
+ hash :properties, as: 'properties'
1475
+ property :version, as: 'version'
1476
+ end
1477
+ end
1478
+
1479
+ class RuntimeInfo
1480
+ # @private
1481
+ class Representation < Google::Apis::Core::JsonRepresentation
1482
+ property :diagnostic_output_uri, as: 'diagnosticOutputUri'
1483
+ hash :endpoints, as: 'endpoints'
1484
+ property :output_uri, as: 'outputUri'
1485
+ end
1486
+ end
1487
+
1310
1488
  class SecurityConfig
1311
1489
  # @private
1312
1490
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1357,6 +1535,25 @@ module Google
1357
1535
  end
1358
1536
  end
1359
1537
 
1538
+ class SparkBatch
1539
+ # @private
1540
+ class Representation < Google::Apis::Core::JsonRepresentation
1541
+ collection :archive_uris, as: 'archiveUris'
1542
+ collection :args, as: 'args'
1543
+ collection :file_uris, as: 'fileUris'
1544
+ collection :jar_file_uris, as: 'jarFileUris'
1545
+ property :main_class, as: 'mainClass'
1546
+ property :main_jar_file_uri, as: 'mainJarFileUri'
1547
+ end
1548
+ end
1549
+
1550
+ class SparkHistoryServerConfig
1551
+ # @private
1552
+ class Representation < Google::Apis::Core::JsonRepresentation
1553
+ property :dataproc_cluster, as: 'dataprocCluster'
1554
+ end
1555
+ end
1556
+
1360
1557
  class SparkJob
1361
1558
  # @private
1362
1559
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1372,6 +1569,16 @@ module Google
1372
1569
  end
1373
1570
  end
1374
1571
 
1572
+ class SparkRBatch
1573
+ # @private
1574
+ class Representation < Google::Apis::Core::JsonRepresentation
1575
+ collection :archive_uris, as: 'archiveUris'
1576
+ collection :args, as: 'args'
1577
+ collection :file_uris, as: 'fileUris'
1578
+ property :main_r_file_uri, as: 'mainRFileUri'
1579
+ end
1580
+ end
1581
+
1375
1582
  class SparkRJob
1376
1583
  # @private
1377
1584
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1385,6 +1592,15 @@ module Google
1385
1592
  end
1386
1593
  end
1387
1594
 
1595
+ class SparkSqlBatch
1596
+ # @private
1597
+ class Representation < Google::Apis::Core::JsonRepresentation
1598
+ collection :jar_file_uris, as: 'jarFileUris'
1599
+ property :query_file_uri, as: 'queryFileUri'
1600
+ hash :query_variables, as: 'queryVariables'
1601
+ end
1602
+ end
1603
+
1388
1604
  class SparkSqlJob
1389
1605
  # @private
1390
1606
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1418,6 +1634,15 @@ module Google
1418
1634
  end
1419
1635
  end
1420
1636
 
1637
+ class StateHistory
1638
+ # @private
1639
+ class Representation < Google::Apis::Core::JsonRepresentation
1640
+ property :state, as: 'state'
1641
+ property :state_message, as: 'stateMessage'
1642
+ property :state_start_time, as: 'stateStartTime'
1643
+ end
1644
+ end
1645
+
1421
1646
  class Status
1422
1647
  # @private
1423
1648
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -352,6 +352,155 @@ module Google
352
352
  execute_or_queue_command(command, &block)
353
353
  end
354
354
 
355
+ # Creates a batch workload that executes asynchronously.
356
+ # @param [String] parent
357
+ # Required. The parent resource where this batch will be created.
358
+ # @param [Google::Apis::DataprocV1::Batch] batch_object
359
+ # @param [String] batch_id
360
+ # Optional. The ID to use for the batch, which will become the final component
361
+ # of the batch's resource name.This value must be 4-63 characters. Valid
362
+ # characters are /[a-z][0-9]-/.
363
+ # @param [String] request_id
364
+ # Optional. A unique ID used to identify the request. If the service receives
365
+ # two CreateBatchRequest (https://cloud.google.com/dataproc/docs/reference/rpc/
366
+ # google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateBatchRequest)s with
367
+ # the same request_id, the second request is ignored and the Operation that
368
+ # corresponds to the first Batch created and stored in the backend is returned.
369
+ # Recommendation: Set this value to a UUID (https://en.wikipedia.org/wiki/
370
+ # Universally_unique_identifier).The value must contain only letters (a-z, A-Z),
371
+ # numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40
372
+ # characters.
373
+ # @param [String] fields
374
+ # Selector specifying which fields to include in a partial response.
375
+ # @param [String] quota_user
376
+ # Available to use for quota purposes for server-side applications. Can be any
377
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
378
+ # @param [Google::Apis::RequestOptions] options
379
+ # Request-specific options
380
+ #
381
+ # @yield [result, err] Result & error if block supplied
382
+ # @yieldparam result [Google::Apis::DataprocV1::Operation] parsed result object
383
+ # @yieldparam err [StandardError] error object if request failed
384
+ #
385
+ # @return [Google::Apis::DataprocV1::Operation]
386
+ #
387
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
388
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
389
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
390
+ def create_project_location_batch(parent, batch_object = nil, batch_id: nil, request_id: nil, fields: nil, quota_user: nil, options: nil, &block)
391
+ command = make_simple_command(:post, 'v1/{+parent}/batches', options)
392
+ command.request_representation = Google::Apis::DataprocV1::Batch::Representation
393
+ command.request_object = batch_object
394
+ command.response_representation = Google::Apis::DataprocV1::Operation::Representation
395
+ command.response_class = Google::Apis::DataprocV1::Operation
396
+ command.params['parent'] = parent unless parent.nil?
397
+ command.query['batchId'] = batch_id unless batch_id.nil?
398
+ command.query['requestId'] = request_id unless request_id.nil?
399
+ command.query['fields'] = fields unless fields.nil?
400
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
401
+ execute_or_queue_command(command, &block)
402
+ end
403
+
404
+ # Deletes the batch workload resource. If the batch is not in terminal state,
405
+ # the delete fails and the response returns FAILED_PRECONDITION.
406
+ # @param [String] name
407
+ # Required. The name of the batch resource to delete.
408
+ # @param [String] fields
409
+ # Selector specifying which fields to include in a partial response.
410
+ # @param [String] quota_user
411
+ # Available to use for quota purposes for server-side applications. Can be any
412
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
413
+ # @param [Google::Apis::RequestOptions] options
414
+ # Request-specific options
415
+ #
416
+ # @yield [result, err] Result & error if block supplied
417
+ # @yieldparam result [Google::Apis::DataprocV1::Empty] parsed result object
418
+ # @yieldparam err [StandardError] error object if request failed
419
+ #
420
+ # @return [Google::Apis::DataprocV1::Empty]
421
+ #
422
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
423
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
424
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
425
+ def delete_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
426
+ command = make_simple_command(:delete, 'v1/{+name}', options)
427
+ command.response_representation = Google::Apis::DataprocV1::Empty::Representation
428
+ command.response_class = Google::Apis::DataprocV1::Empty
429
+ command.params['name'] = name unless name.nil?
430
+ command.query['fields'] = fields unless fields.nil?
431
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
432
+ execute_or_queue_command(command, &block)
433
+ end
434
+
435
+ # Gets the batch workload resource representation.
436
+ # @param [String] name
437
+ # Required. The name of the batch to retrieve.
438
+ # @param [String] fields
439
+ # Selector specifying which fields to include in a partial response.
440
+ # @param [String] quota_user
441
+ # Available to use for quota purposes for server-side applications. Can be any
442
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
443
+ # @param [Google::Apis::RequestOptions] options
444
+ # Request-specific options
445
+ #
446
+ # @yield [result, err] Result & error if block supplied
447
+ # @yieldparam result [Google::Apis::DataprocV1::Batch] parsed result object
448
+ # @yieldparam err [StandardError] error object if request failed
449
+ #
450
+ # @return [Google::Apis::DataprocV1::Batch]
451
+ #
452
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
453
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
454
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
455
+ def get_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
456
+ command = make_simple_command(:get, 'v1/{+name}', options)
457
+ command.response_representation = Google::Apis::DataprocV1::Batch::Representation
458
+ command.response_class = Google::Apis::DataprocV1::Batch
459
+ command.params['name'] = name unless name.nil?
460
+ command.query['fields'] = fields unless fields.nil?
461
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
462
+ execute_or_queue_command(command, &block)
463
+ end
464
+
465
+ # Lists batch workloads.
466
+ # @param [String] parent
467
+ # Required. The parent, which owns this collection of batches.
468
+ # @param [Fixnum] page_size
469
+ # Optional. The maximum number of batches to return in each response. The
470
+ # service may return fewer than this value. The default page size is 20; the
471
+ # maximum page size is 1000.
472
+ # @param [String] page_token
473
+ # Optional. A page token received from a previous ListBatches call. Provide this
474
+ # token to retrieve the subsequent page.
475
+ # @param [String] fields
476
+ # Selector specifying which fields to include in a partial response.
477
+ # @param [String] quota_user
478
+ # Available to use for quota purposes for server-side applications. Can be any
479
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
480
+ # @param [Google::Apis::RequestOptions] options
481
+ # Request-specific options
482
+ #
483
+ # @yield [result, err] Result & error if block supplied
484
+ # @yieldparam result [Google::Apis::DataprocV1::ListBatchesResponse] parsed result object
485
+ # @yieldparam err [StandardError] error object if request failed
486
+ #
487
+ # @return [Google::Apis::DataprocV1::ListBatchesResponse]
488
+ #
489
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
490
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
491
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
492
+ def list_project_location_batches(parent, page_size: nil, page_token: nil, fields: nil, quota_user: nil, options: nil, &block)
493
+ command = make_simple_command(:get, 'v1/{+parent}/batches', options)
494
+ command.response_representation = Google::Apis::DataprocV1::ListBatchesResponse::Representation
495
+ command.response_class = Google::Apis::DataprocV1::ListBatchesResponse
496
+ command.params['parent'] = parent unless parent.nil?
497
+ command.query['pageSize'] = page_size unless page_size.nil?
498
+ command.query['pageToken'] = page_token unless page_token.nil?
499
+ command.query['fields'] = fields unless fields.nil?
500
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
501
+ execute_or_queue_command(command, &block)
502
+ end
503
+
355
504
  # Creates new workflow template.
356
505
  # @param [String] parent
357
506
  # Required. The resource name of the region or location, as described in https://
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-apis-dataproc_v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.18.0
4
+ version: 0.19.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2022-01-10 00:00:00.000000000 Z
11
+ date: 2022-01-17 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: google-apis-core
@@ -58,7 +58,7 @@ licenses:
58
58
  metadata:
59
59
  bug_tracker_uri: https://github.com/googleapis/google-api-ruby-client/issues
60
60
  changelog_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1/CHANGELOG.md
61
- documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.18.0
61
+ documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.19.0
62
62
  source_code_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1
63
63
  post_install_message:
64
64
  rdoc_options: []
@@ -75,7 +75,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
75
75
  - !ruby/object:Gem::Version
76
76
  version: '0'
77
77
  requirements: []
78
- rubygems_version: 3.3.4
78
+ rubygems_version: 3.3.5
79
79
  signing_key:
80
80
  specification_version: 4
81
81
  summary: Simple REST client for Cloud Dataproc API V1