google-apis-dataproc_v1 0.18.0 → 0.19.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: bf1feca2f1c11ebedec266a2b5e0b508e97ccd77ca7f5f4c1720803516c37a1f
4
- data.tar.gz: b10349942e5e629ec7af13d9493adf5f206085a526fd04a1ea74ee8adfbced92
3
+ metadata.gz: b49e139fab6c34e0bd6707069e53c4105077e7abe1dc22ba91fe76124fc9410e
4
+ data.tar.gz: 4f11bed732911ec54b23d5d6eb1189e789e35b858c9abb20fe1b9f7bc4b8ac8b
5
5
  SHA512:
6
- metadata.gz: 2b3125d24f0eaac4cbeb6e082f85ad7f8a907754038cd6ed3e841c3cf8df6a91ef96e4af5c6b61afd4aff1e6a09fc062ee0aa7801a6a2a64b66f7dd63eaa1647
7
- data.tar.gz: aee8dbedaef3365d6f9c651a07aea954dfc5ca330daaaa78b4c29838643d454fc4a23bf99d32d6f63852695176dbcde36f53c6b91dc8b6a135ef592d4bc34e96
6
+ metadata.gz: 74cd3145e6b509ee3ce020ad5d016f2ffc30e53505ac4c70b70b910ae336f694861c29331e4e681bb5f7efdb3126aa4dd027631ab2b29f417dfcdbeedc49f362
7
+ data.tar.gz: 0dc5eccb08bb472d8adcabbb54510b287c02eb923f359c365526c92326ccfc17331b19582180bde9b1f45767868af314c1e1b6350ff953040bedd9f351a42451
data/CHANGELOG.md CHANGED
@@ -1,5 +1,10 @@
1
1
  # Release history for google-apis-dataproc_v1
2
2
 
3
+ ### v0.19.0 (2022-01-14)
4
+
5
+ * Regenerated from discovery document revision 20220111
6
+ * Regenerated using generator version 0.4.1
7
+
3
8
  ### v0.18.0 (2022-01-07)
4
9
 
5
10
  * Regenerated from discovery document revision 20220105
@@ -245,6 +245,131 @@ module Google
245
245
  end
246
246
  end
247
247
 
248
+ # A representation of a batch workload in the service.
249
+ class Batch
250
+ include Google::Apis::Core::Hashable
251
+
252
+ # Output only. The time when the batch was created.
253
+ # Corresponds to the JSON property `createTime`
254
+ # @return [String]
255
+ attr_accessor :create_time
256
+
257
+ # Output only. The email address of the user who created the batch.
258
+ # Corresponds to the JSON property `creator`
259
+ # @return [String]
260
+ attr_accessor :creator
261
+
262
+ # Environment configuration for a workload.
263
+ # Corresponds to the JSON property `environmentConfig`
264
+ # @return [Google::Apis::DataprocV1::EnvironmentConfig]
265
+ attr_accessor :environment_config
266
+
267
+ # Optional. The labels to associate with this batch. Label keys must contain 1
268
+ # to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/
269
+ # rfc1035.txt). Label values may be empty, but, if present, must contain 1 to 63
270
+ # characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt)
271
+ # . No more than 32 labels can be associated with a batch.
272
+ # Corresponds to the JSON property `labels`
273
+ # @return [Hash<String,String>]
274
+ attr_accessor :labels
275
+
276
+ # Output only. The resource name of the batch.
277
+ # Corresponds to the JSON property `name`
278
+ # @return [String]
279
+ attr_accessor :name
280
+
281
+ # Output only. The resource name of the operation associated with this batch.
282
+ # Corresponds to the JSON property `operation`
283
+ # @return [String]
284
+ attr_accessor :operation
285
+
286
+ # A configuration for running an Apache PySpark (https://spark.apache.org/docs/
287
+ # latest/api/python/getting_started/quickstart.html) batch workload.
288
+ # Corresponds to the JSON property `pysparkBatch`
289
+ # @return [Google::Apis::DataprocV1::PySparkBatch]
290
+ attr_accessor :pyspark_batch
291
+
292
+ # Runtime configuration for a workload.
293
+ # Corresponds to the JSON property `runtimeConfig`
294
+ # @return [Google::Apis::DataprocV1::RuntimeConfig]
295
+ attr_accessor :runtime_config
296
+
297
+ # Runtime information about workload execution.
298
+ # Corresponds to the JSON property `runtimeInfo`
299
+ # @return [Google::Apis::DataprocV1::RuntimeInfo]
300
+ attr_accessor :runtime_info
301
+
302
+ # A configuration for running an Apache Spark (http://spark.apache.org/) batch
303
+ # workload.
304
+ # Corresponds to the JSON property `sparkBatch`
305
+ # @return [Google::Apis::DataprocV1::SparkBatch]
306
+ attr_accessor :spark_batch
307
+
308
+ # A configuration for running an Apache SparkR (https://spark.apache.org/docs/
309
+ # latest/sparkr.html) batch workload.
310
+ # Corresponds to the JSON property `sparkRBatch`
311
+ # @return [Google::Apis::DataprocV1::SparkRBatch]
312
+ attr_accessor :spark_r_batch
313
+
314
+ # A configuration for running Apache Spark SQL (http://spark.apache.org/sql/)
315
+ # queries as a batch workload.
316
+ # Corresponds to the JSON property `sparkSqlBatch`
317
+ # @return [Google::Apis::DataprocV1::SparkSqlBatch]
318
+ attr_accessor :spark_sql_batch
319
+
320
+ # Output only. The state of the batch.
321
+ # Corresponds to the JSON property `state`
322
+ # @return [String]
323
+ attr_accessor :state
324
+
325
+ # Output only. Historical state information for the batch.
326
+ # Corresponds to the JSON property `stateHistory`
327
+ # @return [Array<Google::Apis::DataprocV1::StateHistory>]
328
+ attr_accessor :state_history
329
+
330
+ # Output only. Batch state details, such as a failure description if the state
331
+ # is FAILED.
332
+ # Corresponds to the JSON property `stateMessage`
333
+ # @return [String]
334
+ attr_accessor :state_message
335
+
336
+ # Output only. The time when the batch entered a current state.
337
+ # Corresponds to the JSON property `stateTime`
338
+ # @return [String]
339
+ attr_accessor :state_time
340
+
341
+ # Output only. A batch UUID (Unique Universal Identifier). The service generates
342
+ # this value when it creates the batch.
343
+ # Corresponds to the JSON property `uuid`
344
+ # @return [String]
345
+ attr_accessor :uuid
346
+
347
+ def initialize(**args)
348
+ update!(**args)
349
+ end
350
+
351
+ # Update properties of this object
352
+ def update!(**args)
353
+ @create_time = args[:create_time] if args.key?(:create_time)
354
+ @creator = args[:creator] if args.key?(:creator)
355
+ @environment_config = args[:environment_config] if args.key?(:environment_config)
356
+ @labels = args[:labels] if args.key?(:labels)
357
+ @name = args[:name] if args.key?(:name)
358
+ @operation = args[:operation] if args.key?(:operation)
359
+ @pyspark_batch = args[:pyspark_batch] if args.key?(:pyspark_batch)
360
+ @runtime_config = args[:runtime_config] if args.key?(:runtime_config)
361
+ @runtime_info = args[:runtime_info] if args.key?(:runtime_info)
362
+ @spark_batch = args[:spark_batch] if args.key?(:spark_batch)
363
+ @spark_r_batch = args[:spark_r_batch] if args.key?(:spark_r_batch)
364
+ @spark_sql_batch = args[:spark_sql_batch] if args.key?(:spark_sql_batch)
365
+ @state = args[:state] if args.key?(:state)
366
+ @state_history = args[:state_history] if args.key?(:state_history)
367
+ @state_message = args[:state_message] if args.key?(:state_message)
368
+ @state_time = args[:state_time] if args.key?(:state_time)
369
+ @uuid = args[:uuid] if args.key?(:uuid)
370
+ end
371
+ end
372
+
248
373
  # Metadata describing the Batch operation.
249
374
  class BatchOperationMetadata
250
375
  include Google::Apis::Core::Hashable
@@ -979,6 +1104,74 @@ module Google
979
1104
  end
980
1105
  end
981
1106
 
1107
+ # Environment configuration for a workload.
1108
+ class EnvironmentConfig
1109
+ include Google::Apis::Core::Hashable
1110
+
1111
+ # Execution configuration for a workload.
1112
+ # Corresponds to the JSON property `executionConfig`
1113
+ # @return [Google::Apis::DataprocV1::ExecutionConfig]
1114
+ attr_accessor :execution_config
1115
+
1116
+ # Auxiliary services configuration for a workload.
1117
+ # Corresponds to the JSON property `peripheralsConfig`
1118
+ # @return [Google::Apis::DataprocV1::PeripheralsConfig]
1119
+ attr_accessor :peripherals_config
1120
+
1121
+ def initialize(**args)
1122
+ update!(**args)
1123
+ end
1124
+
1125
+ # Update properties of this object
1126
+ def update!(**args)
1127
+ @execution_config = args[:execution_config] if args.key?(:execution_config)
1128
+ @peripherals_config = args[:peripherals_config] if args.key?(:peripherals_config)
1129
+ end
1130
+ end
1131
+
1132
+ # Execution configuration for a workload.
1133
+ class ExecutionConfig
1134
+ include Google::Apis::Core::Hashable
1135
+
1136
+ # Optional. The Cloud KMS key to use for encryption.
1137
+ # Corresponds to the JSON property `kmsKey`
1138
+ # @return [String]
1139
+ attr_accessor :kms_key
1140
+
1141
+ # Optional. Tags used for network traffic control.
1142
+ # Corresponds to the JSON property `networkTags`
1143
+ # @return [Array<String>]
1144
+ attr_accessor :network_tags
1145
+
1146
+ # Optional. Network URI to connect workload to.
1147
+ # Corresponds to the JSON property `networkUri`
1148
+ # @return [String]
1149
+ attr_accessor :network_uri
1150
+
1151
+ # Optional. Service account that used to execute workload.
1152
+ # Corresponds to the JSON property `serviceAccount`
1153
+ # @return [String]
1154
+ attr_accessor :service_account
1155
+
1156
+ # Optional. Subnetwork URI to connect workload to.
1157
+ # Corresponds to the JSON property `subnetworkUri`
1158
+ # @return [String]
1159
+ attr_accessor :subnetwork_uri
1160
+
1161
+ def initialize(**args)
1162
+ update!(**args)
1163
+ end
1164
+
1165
+ # Update properties of this object
1166
+ def update!(**args)
1167
+ @kms_key = args[:kms_key] if args.key?(:kms_key)
1168
+ @network_tags = args[:network_tags] if args.key?(:network_tags)
1169
+ @network_uri = args[:network_uri] if args.key?(:network_uri)
1170
+ @service_account = args[:service_account] if args.key?(:service_account)
1171
+ @subnetwork_uri = args[:subnetwork_uri] if args.key?(:subnetwork_uri)
1172
+ end
1173
+ end
1174
+
982
1175
  # Represents a textual expression in the Common Expression Language (CEL) syntax.
983
1176
  # CEL is a C-like expression language. The syntax and semantics of CEL are
984
1177
  # documented at https://github.com/google/cel-spec.Example (Comparison): title: "
@@ -2168,6 +2361,32 @@ module Google
2168
2361
  end
2169
2362
  end
2170
2363
 
2364
+ # A list of batch workloads.
2365
+ class ListBatchesResponse
2366
+ include Google::Apis::Core::Hashable
2367
+
2368
+ # The batches from the specified collection.
2369
+ # Corresponds to the JSON property `batches`
2370
+ # @return [Array<Google::Apis::DataprocV1::Batch>]
2371
+ attr_accessor :batches
2372
+
2373
+ # A token, which can be sent as page_token to retrieve the next page. If this
2374
+ # field is omitted, there are no subsequent pages.
2375
+ # Corresponds to the JSON property `nextPageToken`
2376
+ # @return [String]
2377
+ attr_accessor :next_page_token
2378
+
2379
+ def initialize(**args)
2380
+ update!(**args)
2381
+ end
2382
+
2383
+ # Update properties of this object
2384
+ def update!(**args)
2385
+ @batches = args[:batches] if args.key?(:batches)
2386
+ @next_page_token = args[:next_page_token] if args.key?(:next_page_token)
2387
+ end
2388
+ end
2389
+
2171
2390
  # The list of all clusters in a project.
2172
2391
  class ListClustersResponse
2173
2392
  include Google::Apis::Core::Hashable
@@ -2654,6 +2873,32 @@ module Google
2654
2873
  end
2655
2874
  end
2656
2875
 
2876
+ # Auxiliary services configuration for a workload.
2877
+ class PeripheralsConfig
2878
+ include Google::Apis::Core::Hashable
2879
+
2880
+ # Optional. Resource name of an existing Dataproc Metastore service.Example:
2881
+ # projects/[project_id]/locations/[region]/services/[service_id]
2882
+ # Corresponds to the JSON property `metastoreService`
2883
+ # @return [String]
2884
+ attr_accessor :metastore_service
2885
+
2886
+ # Spark History Server configuration for the workload.
2887
+ # Corresponds to the JSON property `sparkHistoryServerConfig`
2888
+ # @return [Google::Apis::DataprocV1::SparkHistoryServerConfig]
2889
+ attr_accessor :spark_history_server_config
2890
+
2891
+ def initialize(**args)
2892
+ update!(**args)
2893
+ end
2894
+
2895
+ # Update properties of this object
2896
+ def update!(**args)
2897
+ @metastore_service = args[:metastore_service] if args.key?(:metastore_service)
2898
+ @spark_history_server_config = args[:spark_history_server_config] if args.key?(:spark_history_server_config)
2899
+ end
2900
+ end
2901
+
2657
2902
  # A Dataproc job for running Apache Pig (https://pig.apache.org/) queries on
2658
2903
  # YARN.
2659
2904
  class PigJob
@@ -2869,6 +3114,63 @@ module Google
2869
3114
  end
2870
3115
  end
2871
3116
 
3117
+ # A configuration for running an Apache PySpark (https://spark.apache.org/docs/
3118
+ # latest/api/python/getting_started/quickstart.html) batch workload.
3119
+ class PySparkBatch
3120
+ include Google::Apis::Core::Hashable
3121
+
3122
+ # Optional. HCFS URIs of archives to be extracted into the working directory of
3123
+ # each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
3124
+ # Corresponds to the JSON property `archiveUris`
3125
+ # @return [Array<String>]
3126
+ attr_accessor :archive_uris
3127
+
3128
+ # Optional. The arguments to pass to the driver. Do not include arguments that
3129
+ # can be set as batch properties, such as --conf, since a collision can occur
3130
+ # that causes an incorrect batch submission.
3131
+ # Corresponds to the JSON property `args`
3132
+ # @return [Array<String>]
3133
+ attr_accessor :args
3134
+
3135
+ # Optional. HCFS URIs of files to be placed in the working directory of each
3136
+ # executor.
3137
+ # Corresponds to the JSON property `fileUris`
3138
+ # @return [Array<String>]
3139
+ attr_accessor :file_uris
3140
+
3141
+ # Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
3142
+ # and tasks.
3143
+ # Corresponds to the JSON property `jarFileUris`
3144
+ # @return [Array<String>]
3145
+ attr_accessor :jar_file_uris
3146
+
3147
+ # Required. The HCFS URI of the main Python file to use as the Spark driver.
3148
+ # Must be a .py file.
3149
+ # Corresponds to the JSON property `mainPythonFileUri`
3150
+ # @return [String]
3151
+ attr_accessor :main_python_file_uri
3152
+
3153
+ # Optional. HCFS file URIs of Python files to pass to the PySpark framework.
3154
+ # Supported file types: .py, .egg, and .zip.
3155
+ # Corresponds to the JSON property `pythonFileUris`
3156
+ # @return [Array<String>]
3157
+ attr_accessor :python_file_uris
3158
+
3159
+ def initialize(**args)
3160
+ update!(**args)
3161
+ end
3162
+
3163
+ # Update properties of this object
3164
+ def update!(**args)
3165
+ @archive_uris = args[:archive_uris] if args.key?(:archive_uris)
3166
+ @args = args[:args] if args.key?(:args)
3167
+ @file_uris = args[:file_uris] if args.key?(:file_uris)
3168
+ @jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
3169
+ @main_python_file_uri = args[:main_python_file_uri] if args.key?(:main_python_file_uri)
3170
+ @python_file_uris = args[:python_file_uris] if args.key?(:python_file_uris)
3171
+ end
3172
+ end
3173
+
2872
3174
  # A Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/
2873
3175
  # python-programming-guide.html) applications on YARN.
2874
3176
  class PySparkJob
@@ -3048,6 +3350,72 @@ module Google
3048
3350
  end
3049
3351
  end
3050
3352
 
3353
+ # Runtime configuration for a workload.
3354
+ class RuntimeConfig
3355
+ include Google::Apis::Core::Hashable
3356
+
3357
+ # Optional. Optional custom container image for the job runtime environment. If
3358
+ # not specified, a default container image will be used.
3359
+ # Corresponds to the JSON property `containerImage`
3360
+ # @return [String]
3361
+ attr_accessor :container_image
3362
+
3363
+ # Optional. A mapping of property names to values, which are used to configure
3364
+ # workload execution.
3365
+ # Corresponds to the JSON property `properties`
3366
+ # @return [Hash<String,String>]
3367
+ attr_accessor :properties
3368
+
3369
+ # Optional. Version of the batch runtime.
3370
+ # Corresponds to the JSON property `version`
3371
+ # @return [String]
3372
+ attr_accessor :version
3373
+
3374
+ def initialize(**args)
3375
+ update!(**args)
3376
+ end
3377
+
3378
+ # Update properties of this object
3379
+ def update!(**args)
3380
+ @container_image = args[:container_image] if args.key?(:container_image)
3381
+ @properties = args[:properties] if args.key?(:properties)
3382
+ @version = args[:version] if args.key?(:version)
3383
+ end
3384
+ end
3385
+
3386
+ # Runtime information about workload execution.
3387
+ class RuntimeInfo
3388
+ include Google::Apis::Core::Hashable
3389
+
3390
+ # Output only. A URI pointing to the location of the diagnostics tarball.
3391
+ # Corresponds to the JSON property `diagnosticOutputUri`
3392
+ # @return [String]
3393
+ attr_accessor :diagnostic_output_uri
3394
+
3395
+ # Output only. Map of remote access endpoints (such as web interfaces and APIs)
3396
+ # to their URIs.
3397
+ # Corresponds to the JSON property `endpoints`
3398
+ # @return [Hash<String,String>]
3399
+ attr_accessor :endpoints
3400
+
3401
+ # Output only. A URI pointing to the location of the stdout and stderr of the
3402
+ # workload.
3403
+ # Corresponds to the JSON property `outputUri`
3404
+ # @return [String]
3405
+ attr_accessor :output_uri
3406
+
3407
+ def initialize(**args)
3408
+ update!(**args)
3409
+ end
3410
+
3411
+ # Update properties of this object
3412
+ def update!(**args)
3413
+ @diagnostic_output_uri = args[:diagnostic_output_uri] if args.key?(:diagnostic_output_uri)
3414
+ @endpoints = args[:endpoints] if args.key?(:endpoints)
3415
+ @output_uri = args[:output_uri] if args.key?(:output_uri)
3416
+ end
3417
+ end
3418
+
3051
3419
  # Security related configuration, including encryption, Kerberos, etc.
3052
3420
  class SecurityConfig
3053
3421
  include Google::Apis::Core::Hashable
@@ -3258,6 +3626,83 @@ module Google
3258
3626
  end
3259
3627
  end
3260
3628
 
3629
+ # A configuration for running an Apache Spark (http://spark.apache.org/) batch
3630
+ # workload.
3631
+ class SparkBatch
3632
+ include Google::Apis::Core::Hashable
3633
+
3634
+ # Optional. HCFS URIs of archives to be extracted into the working directory of
3635
+ # each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
3636
+ # Corresponds to the JSON property `archiveUris`
3637
+ # @return [Array<String>]
3638
+ attr_accessor :archive_uris
3639
+
3640
+ # Optional. The arguments to pass to the driver. Do not include arguments that
3641
+ # can be set as batch properties, such as --conf, since a collision can occur
3642
+ # that causes an incorrect batch submission.
3643
+ # Corresponds to the JSON property `args`
3644
+ # @return [Array<String>]
3645
+ attr_accessor :args
3646
+
3647
+ # Optional. HCFS URIs of files to be placed in the working directory of each
3648
+ # executor.
3649
+ # Corresponds to the JSON property `fileUris`
3650
+ # @return [Array<String>]
3651
+ attr_accessor :file_uris
3652
+
3653
+ # Optional. HCFS URIs of jar files to add to the classpath of the Spark driver
3654
+ # and tasks.
3655
+ # Corresponds to the JSON property `jarFileUris`
3656
+ # @return [Array<String>]
3657
+ attr_accessor :jar_file_uris
3658
+
3659
+ # Optional. The name of the driver main class. The jar file that contains the
3660
+ # class must be in the classpath or specified in jar_file_uris.
3661
+ # Corresponds to the JSON property `mainClass`
3662
+ # @return [String]
3663
+ attr_accessor :main_class
3664
+
3665
+ # Optional. The HCFS URI of the jar file that contains the main class.
3666
+ # Corresponds to the JSON property `mainJarFileUri`
3667
+ # @return [String]
3668
+ attr_accessor :main_jar_file_uri
3669
+
3670
+ def initialize(**args)
3671
+ update!(**args)
3672
+ end
3673
+
3674
+ # Update properties of this object
3675
+ def update!(**args)
3676
+ @archive_uris = args[:archive_uris] if args.key?(:archive_uris)
3677
+ @args = args[:args] if args.key?(:args)
3678
+ @file_uris = args[:file_uris] if args.key?(:file_uris)
3679
+ @jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
3680
+ @main_class = args[:main_class] if args.key?(:main_class)
3681
+ @main_jar_file_uri = args[:main_jar_file_uri] if args.key?(:main_jar_file_uri)
3682
+ end
3683
+ end
3684
+
3685
+ # Spark History Server configuration for the workload.
3686
+ class SparkHistoryServerConfig
3687
+ include Google::Apis::Core::Hashable
3688
+
3689
+ # Optional. Resource name of an existing Dataproc Cluster to act as a Spark
3690
+ # History Server for the workload.Example: projects/[project_id]/regions/[region]
3691
+ # /clusters/[cluster_name]
3692
+ # Corresponds to the JSON property `dataprocCluster`
3693
+ # @return [String]
3694
+ attr_accessor :dataproc_cluster
3695
+
3696
+ def initialize(**args)
3697
+ update!(**args)
3698
+ end
3699
+
3700
+ # Update properties of this object
3701
+ def update!(**args)
3702
+ @dataproc_cluster = args[:dataproc_cluster] if args.key?(:dataproc_cluster)
3703
+ end
3704
+ end
3705
+
3261
3706
  # A Dataproc job for running Apache Spark (http://spark.apache.org/)
3262
3707
  # applications on YARN.
3263
3708
  class SparkJob
@@ -3329,6 +3774,49 @@ module Google
3329
3774
  end
3330
3775
  end
3331
3776
 
3777
+ # A configuration for running an Apache SparkR (https://spark.apache.org/docs/
3778
+ # latest/sparkr.html) batch workload.
3779
+ class SparkRBatch
3780
+ include Google::Apis::Core::Hashable
3781
+
3782
+ # Optional. HCFS URIs of archives to be extracted into the working directory of
3783
+ # each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
3784
+ # Corresponds to the JSON property `archiveUris`
3785
+ # @return [Array<String>]
3786
+ attr_accessor :archive_uris
3787
+
3788
+ # Optional. The arguments to pass to the Spark driver. Do not include arguments
3789
+ # that can be set as batch properties, such as --conf, since a collision can
3790
+ # occur that causes an incorrect batch submission.
3791
+ # Corresponds to the JSON property `args`
3792
+ # @return [Array<String>]
3793
+ attr_accessor :args
3794
+
3795
+ # Optional. HCFS URIs of files to be placed in the working directory of each
3796
+ # executor.
3797
+ # Corresponds to the JSON property `fileUris`
3798
+ # @return [Array<String>]
3799
+ attr_accessor :file_uris
3800
+
3801
+ # Required. The HCFS URI of the main R file to use as the driver. Must be a .R
3802
+ # or .r file.
3803
+ # Corresponds to the JSON property `mainRFileUri`
3804
+ # @return [String]
3805
+ attr_accessor :main_r_file_uri
3806
+
3807
+ def initialize(**args)
3808
+ update!(**args)
3809
+ end
3810
+
3811
+ # Update properties of this object
3812
+ def update!(**args)
3813
+ @archive_uris = args[:archive_uris] if args.key?(:archive_uris)
3814
+ @args = args[:args] if args.key?(:args)
3815
+ @file_uris = args[:file_uris] if args.key?(:file_uris)
3816
+ @main_r_file_uri = args[:main_r_file_uri] if args.key?(:main_r_file_uri)
3817
+ end
3818
+ end
3819
+
3332
3820
  # A Dataproc job for running Apache SparkR (https://spark.apache.org/docs/latest/
3333
3821
  # sparkr.html) applications on YARN.
3334
3822
  class SparkRJob
@@ -3387,6 +3875,40 @@ module Google
3387
3875
  end
3388
3876
  end
3389
3877
 
3878
+ # A configuration for running Apache Spark SQL (http://spark.apache.org/sql/)
3879
+ # queries as a batch workload.
3880
+ class SparkSqlBatch
3881
+ include Google::Apis::Core::Hashable
3882
+
3883
+ # Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
3884
+ # Corresponds to the JSON property `jarFileUris`
3885
+ # @return [Array<String>]
3886
+ attr_accessor :jar_file_uris
3887
+
3888
+ # Required. The HCFS URI of the script that contains Spark SQL queries to
3889
+ # execute.
3890
+ # Corresponds to the JSON property `queryFileUri`
3891
+ # @return [String]
3892
+ attr_accessor :query_file_uri
3893
+
3894
+ # Optional. Mapping of query variable names to values (equivalent to the Spark
3895
+ # SQL command: SET name="value";).
3896
+ # Corresponds to the JSON property `queryVariables`
3897
+ # @return [Hash<String,String>]
3898
+ attr_accessor :query_variables
3899
+
3900
+ def initialize(**args)
3901
+ update!(**args)
3902
+ end
3903
+
3904
+ # Update properties of this object
3905
+ def update!(**args)
3906
+ @jar_file_uris = args[:jar_file_uris] if args.key?(:jar_file_uris)
3907
+ @query_file_uri = args[:query_file_uri] if args.key?(:query_file_uri)
3908
+ @query_variables = args[:query_variables] if args.key?(:query_variables)
3909
+ end
3910
+ end
3911
+
3390
3912
  # A Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/)
3391
3913
  # queries.
3392
3914
  class SparkSqlJob
@@ -3536,6 +4058,37 @@ module Google
3536
4058
  end
3537
4059
  end
3538
4060
 
4061
+ # Historical state information.
4062
+ class StateHistory
4063
+ include Google::Apis::Core::Hashable
4064
+
4065
+ # Output only. The state of the batch at this point in history.
4066
+ # Corresponds to the JSON property `state`
4067
+ # @return [String]
4068
+ attr_accessor :state
4069
+
4070
+ # Output only. Details about the state at this point in history.
4071
+ # Corresponds to the JSON property `stateMessage`
4072
+ # @return [String]
4073
+ attr_accessor :state_message
4074
+
4075
+ # Output only. The time when the batch entered the historical state.
4076
+ # Corresponds to the JSON property `stateStartTime`
4077
+ # @return [String]
4078
+ attr_accessor :state_start_time
4079
+
4080
+ def initialize(**args)
4081
+ update!(**args)
4082
+ end
4083
+
4084
+ # Update properties of this object
4085
+ def update!(**args)
4086
+ @state = args[:state] if args.key?(:state)
4087
+ @state_message = args[:state_message] if args.key?(:state_message)
4088
+ @state_start_time = args[:state_start_time] if args.key?(:state_start_time)
4089
+ end
4090
+ end
4091
+
3539
4092
  # The Status type defines a logical error model that is suitable for different
3540
4093
  # programming environments, including REST APIs and RPC APIs. It is used by gRPC
3541
4094
  # (https://github.com/grpc). Each Status message contains three pieces of data:
@@ -16,13 +16,13 @@ module Google
16
16
  module Apis
17
17
  module DataprocV1
18
18
  # Version of the google-apis-dataproc_v1 gem
19
- GEM_VERSION = "0.18.0"
19
+ GEM_VERSION = "0.19.0"
20
20
 
21
21
  # Version of the code generator used to generate this client
22
- GENERATOR_VERSION = "0.4.0"
22
+ GENERATOR_VERSION = "0.4.1"
23
23
 
24
24
  # Revision of the discovery document this client was generated from
25
- REVISION = "20220105"
25
+ REVISION = "20220111"
26
26
  end
27
27
  end
28
28
  end
@@ -52,6 +52,12 @@ module Google
52
52
  include Google::Apis::Core::JsonObjectSupport
53
53
  end
54
54
 
55
+ class Batch
56
+ class Representation < Google::Apis::Core::JsonRepresentation; end
57
+
58
+ include Google::Apis::Core::JsonObjectSupport
59
+ end
60
+
55
61
  class BatchOperationMetadata
56
62
  class Representation < Google::Apis::Core::JsonRepresentation; end
57
63
 
@@ -160,6 +166,18 @@ module Google
160
166
  include Google::Apis::Core::JsonObjectSupport
161
167
  end
162
168
 
169
+ class EnvironmentConfig
170
+ class Representation < Google::Apis::Core::JsonRepresentation; end
171
+
172
+ include Google::Apis::Core::JsonObjectSupport
173
+ end
174
+
175
+ class ExecutionConfig
176
+ class Representation < Google::Apis::Core::JsonRepresentation; end
177
+
178
+ include Google::Apis::Core::JsonObjectSupport
179
+ end
180
+
163
181
  class Expr
164
182
  class Representation < Google::Apis::Core::JsonRepresentation; end
165
183
 
@@ -292,6 +310,12 @@ module Google
292
310
  include Google::Apis::Core::JsonObjectSupport
293
311
  end
294
312
 
313
+ class ListBatchesResponse
314
+ class Representation < Google::Apis::Core::JsonRepresentation; end
315
+
316
+ include Google::Apis::Core::JsonObjectSupport
317
+ end
318
+
295
319
  class ListClustersResponse
296
320
  class Representation < Google::Apis::Core::JsonRepresentation; end
297
321
 
@@ -376,6 +400,12 @@ module Google
376
400
  include Google::Apis::Core::JsonObjectSupport
377
401
  end
378
402
 
403
+ class PeripheralsConfig
404
+ class Representation < Google::Apis::Core::JsonRepresentation; end
405
+
406
+ include Google::Apis::Core::JsonObjectSupport
407
+ end
408
+
379
409
  class PigJob
380
410
  class Representation < Google::Apis::Core::JsonRepresentation; end
381
411
 
@@ -394,6 +424,12 @@ module Google
394
424
  include Google::Apis::Core::JsonObjectSupport
395
425
  end
396
426
 
427
+ class PySparkBatch
428
+ class Representation < Google::Apis::Core::JsonRepresentation; end
429
+
430
+ include Google::Apis::Core::JsonObjectSupport
431
+ end
432
+
397
433
  class PySparkJob
398
434
  class Representation < Google::Apis::Core::JsonRepresentation; end
399
435
 
@@ -424,6 +460,18 @@ module Google
424
460
  include Google::Apis::Core::JsonObjectSupport
425
461
  end
426
462
 
463
+ class RuntimeConfig
464
+ class Representation < Google::Apis::Core::JsonRepresentation; end
465
+
466
+ include Google::Apis::Core::JsonObjectSupport
467
+ end
468
+
469
+ class RuntimeInfo
470
+ class Representation < Google::Apis::Core::JsonRepresentation; end
471
+
472
+ include Google::Apis::Core::JsonObjectSupport
473
+ end
474
+
427
475
  class SecurityConfig
428
476
  class Representation < Google::Apis::Core::JsonRepresentation; end
429
477
 
@@ -454,18 +502,42 @@ module Google
454
502
  include Google::Apis::Core::JsonObjectSupport
455
503
  end
456
504
 
505
+ class SparkBatch
506
+ class Representation < Google::Apis::Core::JsonRepresentation; end
507
+
508
+ include Google::Apis::Core::JsonObjectSupport
509
+ end
510
+
511
+ class SparkHistoryServerConfig
512
+ class Representation < Google::Apis::Core::JsonRepresentation; end
513
+
514
+ include Google::Apis::Core::JsonObjectSupport
515
+ end
516
+
457
517
  class SparkJob
458
518
  class Representation < Google::Apis::Core::JsonRepresentation; end
459
519
 
460
520
  include Google::Apis::Core::JsonObjectSupport
461
521
  end
462
522
 
523
+ class SparkRBatch
524
+ class Representation < Google::Apis::Core::JsonRepresentation; end
525
+
526
+ include Google::Apis::Core::JsonObjectSupport
527
+ end
528
+
463
529
  class SparkRJob
464
530
  class Representation < Google::Apis::Core::JsonRepresentation; end
465
531
 
466
532
  include Google::Apis::Core::JsonObjectSupport
467
533
  end
468
534
 
535
+ class SparkSqlBatch
536
+ class Representation < Google::Apis::Core::JsonRepresentation; end
537
+
538
+ include Google::Apis::Core::JsonObjectSupport
539
+ end
540
+
469
541
  class SparkSqlJob
470
542
  class Representation < Google::Apis::Core::JsonRepresentation; end
471
543
 
@@ -484,6 +556,12 @@ module Google
484
556
  include Google::Apis::Core::JsonObjectSupport
485
557
  end
486
558
 
559
+ class StateHistory
560
+ class Representation < Google::Apis::Core::JsonRepresentation; end
561
+
562
+ include Google::Apis::Core::JsonObjectSupport
563
+ end
564
+
487
565
  class Status
488
566
  class Representation < Google::Apis::Core::JsonRepresentation; end
489
567
 
@@ -614,6 +692,37 @@ module Google
614
692
  end
615
693
  end
616
694
 
695
+ class Batch
696
+ # @private
697
+ class Representation < Google::Apis::Core::JsonRepresentation
698
+ property :create_time, as: 'createTime'
699
+ property :creator, as: 'creator'
700
+ property :environment_config, as: 'environmentConfig', class: Google::Apis::DataprocV1::EnvironmentConfig, decorator: Google::Apis::DataprocV1::EnvironmentConfig::Representation
701
+
702
+ hash :labels, as: 'labels'
703
+ property :name, as: 'name'
704
+ property :operation, as: 'operation'
705
+ property :pyspark_batch, as: 'pysparkBatch', class: Google::Apis::DataprocV1::PySparkBatch, decorator: Google::Apis::DataprocV1::PySparkBatch::Representation
706
+
707
+ property :runtime_config, as: 'runtimeConfig', class: Google::Apis::DataprocV1::RuntimeConfig, decorator: Google::Apis::DataprocV1::RuntimeConfig::Representation
708
+
709
+ property :runtime_info, as: 'runtimeInfo', class: Google::Apis::DataprocV1::RuntimeInfo, decorator: Google::Apis::DataprocV1::RuntimeInfo::Representation
710
+
711
+ property :spark_batch, as: 'sparkBatch', class: Google::Apis::DataprocV1::SparkBatch, decorator: Google::Apis::DataprocV1::SparkBatch::Representation
712
+
713
+ property :spark_r_batch, as: 'sparkRBatch', class: Google::Apis::DataprocV1::SparkRBatch, decorator: Google::Apis::DataprocV1::SparkRBatch::Representation
714
+
715
+ property :spark_sql_batch, as: 'sparkSqlBatch', class: Google::Apis::DataprocV1::SparkSqlBatch, decorator: Google::Apis::DataprocV1::SparkSqlBatch::Representation
716
+
717
+ property :state, as: 'state'
718
+ collection :state_history, as: 'stateHistory', class: Google::Apis::DataprocV1::StateHistory, decorator: Google::Apis::DataprocV1::StateHistory::Representation
719
+
720
+ property :state_message, as: 'stateMessage'
721
+ property :state_time, as: 'stateTime'
722
+ property :uuid, as: 'uuid'
723
+ end
724
+ end
725
+
617
726
  class BatchOperationMetadata
618
727
  # @private
619
728
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -808,6 +917,27 @@ module Google
808
917
  end
809
918
  end
810
919
 
920
+ class EnvironmentConfig
921
+ # @private
922
+ class Representation < Google::Apis::Core::JsonRepresentation
923
+ property :execution_config, as: 'executionConfig', class: Google::Apis::DataprocV1::ExecutionConfig, decorator: Google::Apis::DataprocV1::ExecutionConfig::Representation
924
+
925
+ property :peripherals_config, as: 'peripheralsConfig', class: Google::Apis::DataprocV1::PeripheralsConfig, decorator: Google::Apis::DataprocV1::PeripheralsConfig::Representation
926
+
927
+ end
928
+ end
929
+
930
+ class ExecutionConfig
931
+ # @private
932
+ class Representation < Google::Apis::Core::JsonRepresentation
933
+ property :kms_key, as: 'kmsKey'
934
+ collection :network_tags, as: 'networkTags'
935
+ property :network_uri, as: 'networkUri'
936
+ property :service_account, as: 'serviceAccount'
937
+ property :subnetwork_uri, as: 'subnetworkUri'
938
+ end
939
+ end
940
+
811
941
  class Expr
812
942
  # @private
813
943
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1081,6 +1211,15 @@ module Google
1081
1211
  end
1082
1212
  end
1083
1213
 
1214
+ class ListBatchesResponse
1215
+ # @private
1216
+ class Representation < Google::Apis::Core::JsonRepresentation
1217
+ collection :batches, as: 'batches', class: Google::Apis::DataprocV1::Batch, decorator: Google::Apis::DataprocV1::Batch::Representation
1218
+
1219
+ property :next_page_token, as: 'nextPageToken'
1220
+ end
1221
+ end
1222
+
1084
1223
  class ListClustersResponse
1085
1224
  # @private
1086
1225
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1221,6 +1360,15 @@ module Google
1221
1360
  end
1222
1361
  end
1223
1362
 
1363
+ class PeripheralsConfig
1364
+ # @private
1365
+ class Representation < Google::Apis::Core::JsonRepresentation
1366
+ property :metastore_service, as: 'metastoreService'
1367
+ property :spark_history_server_config, as: 'sparkHistoryServerConfig', class: Google::Apis::DataprocV1::SparkHistoryServerConfig, decorator: Google::Apis::DataprocV1::SparkHistoryServerConfig::Representation
1368
+
1369
+ end
1370
+ end
1371
+
1224
1372
  class PigJob
1225
1373
  # @private
1226
1374
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1261,6 +1409,18 @@ module Google
1261
1409
  end
1262
1410
  end
1263
1411
 
1412
+ class PySparkBatch
1413
+ # @private
1414
+ class Representation < Google::Apis::Core::JsonRepresentation
1415
+ collection :archive_uris, as: 'archiveUris'
1416
+ collection :args, as: 'args'
1417
+ collection :file_uris, as: 'fileUris'
1418
+ collection :jar_file_uris, as: 'jarFileUris'
1419
+ property :main_python_file_uri, as: 'mainPythonFileUri'
1420
+ collection :python_file_uris, as: 'pythonFileUris'
1421
+ end
1422
+ end
1423
+
1264
1424
  class PySparkJob
1265
1425
  # @private
1266
1426
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1307,6 +1467,24 @@ module Google
1307
1467
  end
1308
1468
  end
1309
1469
 
1470
+ class RuntimeConfig
1471
+ # @private
1472
+ class Representation < Google::Apis::Core::JsonRepresentation
1473
+ property :container_image, as: 'containerImage'
1474
+ hash :properties, as: 'properties'
1475
+ property :version, as: 'version'
1476
+ end
1477
+ end
1478
+
1479
+ class RuntimeInfo
1480
+ # @private
1481
+ class Representation < Google::Apis::Core::JsonRepresentation
1482
+ property :diagnostic_output_uri, as: 'diagnosticOutputUri'
1483
+ hash :endpoints, as: 'endpoints'
1484
+ property :output_uri, as: 'outputUri'
1485
+ end
1486
+ end
1487
+
1310
1488
  class SecurityConfig
1311
1489
  # @private
1312
1490
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1357,6 +1535,25 @@ module Google
1357
1535
  end
1358
1536
  end
1359
1537
 
1538
+ class SparkBatch
1539
+ # @private
1540
+ class Representation < Google::Apis::Core::JsonRepresentation
1541
+ collection :archive_uris, as: 'archiveUris'
1542
+ collection :args, as: 'args'
1543
+ collection :file_uris, as: 'fileUris'
1544
+ collection :jar_file_uris, as: 'jarFileUris'
1545
+ property :main_class, as: 'mainClass'
1546
+ property :main_jar_file_uri, as: 'mainJarFileUri'
1547
+ end
1548
+ end
1549
+
1550
+ class SparkHistoryServerConfig
1551
+ # @private
1552
+ class Representation < Google::Apis::Core::JsonRepresentation
1553
+ property :dataproc_cluster, as: 'dataprocCluster'
1554
+ end
1555
+ end
1556
+
1360
1557
  class SparkJob
1361
1558
  # @private
1362
1559
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1372,6 +1569,16 @@ module Google
1372
1569
  end
1373
1570
  end
1374
1571
 
1572
+ class SparkRBatch
1573
+ # @private
1574
+ class Representation < Google::Apis::Core::JsonRepresentation
1575
+ collection :archive_uris, as: 'archiveUris'
1576
+ collection :args, as: 'args'
1577
+ collection :file_uris, as: 'fileUris'
1578
+ property :main_r_file_uri, as: 'mainRFileUri'
1579
+ end
1580
+ end
1581
+
1375
1582
  class SparkRJob
1376
1583
  # @private
1377
1584
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1385,6 +1592,15 @@ module Google
1385
1592
  end
1386
1593
  end
1387
1594
 
1595
+ class SparkSqlBatch
1596
+ # @private
1597
+ class Representation < Google::Apis::Core::JsonRepresentation
1598
+ collection :jar_file_uris, as: 'jarFileUris'
1599
+ property :query_file_uri, as: 'queryFileUri'
1600
+ hash :query_variables, as: 'queryVariables'
1601
+ end
1602
+ end
1603
+
1388
1604
  class SparkSqlJob
1389
1605
  # @private
1390
1606
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -1418,6 +1634,15 @@ module Google
1418
1634
  end
1419
1635
  end
1420
1636
 
1637
+ class StateHistory
1638
+ # @private
1639
+ class Representation < Google::Apis::Core::JsonRepresentation
1640
+ property :state, as: 'state'
1641
+ property :state_message, as: 'stateMessage'
1642
+ property :state_start_time, as: 'stateStartTime'
1643
+ end
1644
+ end
1645
+
1421
1646
  class Status
1422
1647
  # @private
1423
1648
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -352,6 +352,155 @@ module Google
352
352
  execute_or_queue_command(command, &block)
353
353
  end
354
354
 
355
+ # Creates a batch workload that executes asynchronously.
356
+ # @param [String] parent
357
+ # Required. The parent resource where this batch will be created.
358
+ # @param [Google::Apis::DataprocV1::Batch] batch_object
359
+ # @param [String] batch_id
360
+ # Optional. The ID to use for the batch, which will become the final component
361
+ # of the batch's resource name.This value must be 4-63 characters. Valid
362
+ # characters are /[a-z][0-9]-/.
363
+ # @param [String] request_id
364
+ # Optional. A unique ID used to identify the request. If the service receives
365
+ # two CreateBatchRequest (https://cloud.google.com/dataproc/docs/reference/rpc/
366
+ # google.cloud.dataproc.v1#google.cloud.dataproc.v1.CreateBatchRequest)s with
367
+ # the same request_id, the second request is ignored and the Operation that
368
+ # corresponds to the first Batch created and stored in the backend is returned.
369
+ # Recommendation: Set this value to a UUID (https://en.wikipedia.org/wiki/
370
+ # Universally_unique_identifier).The value must contain only letters (a-z, A-Z),
371
+ # numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40
372
+ # characters.
373
+ # @param [String] fields
374
+ # Selector specifying which fields to include in a partial response.
375
+ # @param [String] quota_user
376
+ # Available to use for quota purposes for server-side applications. Can be any
377
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
378
+ # @param [Google::Apis::RequestOptions] options
379
+ # Request-specific options
380
+ #
381
+ # @yield [result, err] Result & error if block supplied
382
+ # @yieldparam result [Google::Apis::DataprocV1::Operation] parsed result object
383
+ # @yieldparam err [StandardError] error object if request failed
384
+ #
385
+ # @return [Google::Apis::DataprocV1::Operation]
386
+ #
387
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
388
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
389
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
390
+ def create_project_location_batch(parent, batch_object = nil, batch_id: nil, request_id: nil, fields: nil, quota_user: nil, options: nil, &block)
391
+ command = make_simple_command(:post, 'v1/{+parent}/batches', options)
392
+ command.request_representation = Google::Apis::DataprocV1::Batch::Representation
393
+ command.request_object = batch_object
394
+ command.response_representation = Google::Apis::DataprocV1::Operation::Representation
395
+ command.response_class = Google::Apis::DataprocV1::Operation
396
+ command.params['parent'] = parent unless parent.nil?
397
+ command.query['batchId'] = batch_id unless batch_id.nil?
398
+ command.query['requestId'] = request_id unless request_id.nil?
399
+ command.query['fields'] = fields unless fields.nil?
400
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
401
+ execute_or_queue_command(command, &block)
402
+ end
403
+
404
+ # Deletes the batch workload resource. If the batch is not in terminal state,
405
+ # the delete fails and the response returns FAILED_PRECONDITION.
406
+ # @param [String] name
407
+ # Required. The name of the batch resource to delete.
408
+ # @param [String] fields
409
+ # Selector specifying which fields to include in a partial response.
410
+ # @param [String] quota_user
411
+ # Available to use for quota purposes for server-side applications. Can be any
412
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
413
+ # @param [Google::Apis::RequestOptions] options
414
+ # Request-specific options
415
+ #
416
+ # @yield [result, err] Result & error if block supplied
417
+ # @yieldparam result [Google::Apis::DataprocV1::Empty] parsed result object
418
+ # @yieldparam err [StandardError] error object if request failed
419
+ #
420
+ # @return [Google::Apis::DataprocV1::Empty]
421
+ #
422
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
423
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
424
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
425
+ def delete_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
426
+ command = make_simple_command(:delete, 'v1/{+name}', options)
427
+ command.response_representation = Google::Apis::DataprocV1::Empty::Representation
428
+ command.response_class = Google::Apis::DataprocV1::Empty
429
+ command.params['name'] = name unless name.nil?
430
+ command.query['fields'] = fields unless fields.nil?
431
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
432
+ execute_or_queue_command(command, &block)
433
+ end
434
+
435
+ # Gets the batch workload resource representation.
436
+ # @param [String] name
437
+ # Required. The name of the batch to retrieve.
438
+ # @param [String] fields
439
+ # Selector specifying which fields to include in a partial response.
440
+ # @param [String] quota_user
441
+ # Available to use for quota purposes for server-side applications. Can be any
442
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
443
+ # @param [Google::Apis::RequestOptions] options
444
+ # Request-specific options
445
+ #
446
+ # @yield [result, err] Result & error if block supplied
447
+ # @yieldparam result [Google::Apis::DataprocV1::Batch] parsed result object
448
+ # @yieldparam err [StandardError] error object if request failed
449
+ #
450
+ # @return [Google::Apis::DataprocV1::Batch]
451
+ #
452
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
453
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
454
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
455
+ def get_project_location_batch(name, fields: nil, quota_user: nil, options: nil, &block)
456
+ command = make_simple_command(:get, 'v1/{+name}', options)
457
+ command.response_representation = Google::Apis::DataprocV1::Batch::Representation
458
+ command.response_class = Google::Apis::DataprocV1::Batch
459
+ command.params['name'] = name unless name.nil?
460
+ command.query['fields'] = fields unless fields.nil?
461
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
462
+ execute_or_queue_command(command, &block)
463
+ end
464
+
465
+ # Lists batch workloads.
466
+ # @param [String] parent
467
+ # Required. The parent, which owns this collection of batches.
468
+ # @param [Fixnum] page_size
469
+ # Optional. The maximum number of batches to return in each response. The
470
+ # service may return fewer than this value. The default page size is 20; the
471
+ # maximum page size is 1000.
472
+ # @param [String] page_token
473
+ # Optional. A page token received from a previous ListBatches call. Provide this
474
+ # token to retrieve the subsequent page.
475
+ # @param [String] fields
476
+ # Selector specifying which fields to include in a partial response.
477
+ # @param [String] quota_user
478
+ # Available to use for quota purposes for server-side applications. Can be any
479
+ # arbitrary string assigned to a user, but should not exceed 40 characters.
480
+ # @param [Google::Apis::RequestOptions] options
481
+ # Request-specific options
482
+ #
483
+ # @yield [result, err] Result & error if block supplied
484
+ # @yieldparam result [Google::Apis::DataprocV1::ListBatchesResponse] parsed result object
485
+ # @yieldparam err [StandardError] error object if request failed
486
+ #
487
+ # @return [Google::Apis::DataprocV1::ListBatchesResponse]
488
+ #
489
+ # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried
490
+ # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification
491
+ # @raise [Google::Apis::AuthorizationError] Authorization is required
492
+ def list_project_location_batches(parent, page_size: nil, page_token: nil, fields: nil, quota_user: nil, options: nil, &block)
493
+ command = make_simple_command(:get, 'v1/{+parent}/batches', options)
494
+ command.response_representation = Google::Apis::DataprocV1::ListBatchesResponse::Representation
495
+ command.response_class = Google::Apis::DataprocV1::ListBatchesResponse
496
+ command.params['parent'] = parent unless parent.nil?
497
+ command.query['pageSize'] = page_size unless page_size.nil?
498
+ command.query['pageToken'] = page_token unless page_token.nil?
499
+ command.query['fields'] = fields unless fields.nil?
500
+ command.query['quotaUser'] = quota_user unless quota_user.nil?
501
+ execute_or_queue_command(command, &block)
502
+ end
503
+
355
504
  # Creates new workflow template.
356
505
  # @param [String] parent
357
506
  # Required. The resource name of the region or location, as described in https://
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-apis-dataproc_v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.18.0
4
+ version: 0.19.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2022-01-10 00:00:00.000000000 Z
11
+ date: 2022-01-17 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: google-apis-core
@@ -58,7 +58,7 @@ licenses:
58
58
  metadata:
59
59
  bug_tracker_uri: https://github.com/googleapis/google-api-ruby-client/issues
60
60
  changelog_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1/CHANGELOG.md
61
- documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.18.0
61
+ documentation_uri: https://googleapis.dev/ruby/google-apis-dataproc_v1/v0.19.0
62
62
  source_code_uri: https://github.com/googleapis/google-api-ruby-client/tree/main/generated/google-apis-dataproc_v1
63
63
  post_install_message:
64
64
  rdoc_options: []
@@ -75,7 +75,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
75
75
  - !ruby/object:Gem::Version
76
76
  version: '0'
77
77
  requirements: []
78
- rubygems_version: 3.3.4
78
+ rubygems_version: 3.3.5
79
79
  signing_key:
80
80
  specification_version: 4
81
81
  summary: Simple REST client for Cloud Dataproc API V1