google-cloud-dataproc-v1 1.11.0 → 1.12.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: 27303e4bc8144f2dc1c53c914c4578644a112d5f3a4258dad0c01f34a674c6ef
|
|
4
|
+
data.tar.gz: 274803c54dfcbe3aaaf26d2d4f427f10f55670c58eddb596070cd53ab46f6c0a
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 5dc0fec27d452861498e66ed4352ce296a8f20f1f16b41ccb75f86c0c8a4d1b72428eb7fcdaf5a1b26a8fcdc449c780a7e8cec27d3b34bc40e4d1343388f3542
|
|
7
|
+
data.tar.gz: 38021feb7934623efef2981d218854b92f129694fe840cab86493833c362b2e5944193735d68bcea0015556580cac96403147d270d76e60d7c959e3559671263
|
|
@@ -14,7 +14,7 @@ require 'google/protobuf/empty_pb'
|
|
|
14
14
|
require 'google/protobuf/timestamp_pb'
|
|
15
15
|
|
|
16
16
|
|
|
17
|
-
descriptor_data = "\n&google/cloud/dataproc/v1/batches.proto\x12\x18google.cloud.dataproc.v1\x1a\x1cgoogle/api/annotations.proto\x1a\x17google/api/client.proto\x1a\x1fgoogle/api/field_behavior.proto\x1a\x19google/api/resource.proto\x1a%google/cloud/dataproc/v1/shared.proto\x1a#google/longrunning/operations.proto\x1a\x1bgoogle/protobuf/empty.proto\x1a\x1fgoogle/protobuf/timestamp.proto\"\xb0\x01\n\x12\x43reateBatchRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x33\n\x05\x62\x61tch\x18\x02 \x01(\x0b\x32\x1f.google.cloud.dataproc.v1.BatchB\x03\xe0\x41\x02\x12\x15\n\x08\x62\x61tch_id\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x17\n\nrequest_id\x18\x04 \x01(\tB\x03\xe0\x41\x01\"F\n\x0fGetBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\xa8\x01\n\x12ListBatchesRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x16\n\tpage_size\x18\x02 \x01(\x05\x42\x03\xe0\x41\x01\x12\x17\n\npage_token\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x13\n\x06\x66ilter\x18\x04 \x01(\tB\x03\xe0\x41\x01\x12\x15\n\x08order_by\x18\x05 \x01(\tB\x03\xe0\x41\x01\"z\n\x13ListBatchesResponse\x12\x30\n\x07\x62\x61tches\x18\x01 \x03(\x0b\x32\x1f.google.cloud.dataproc.v1.Batch\x12\x17\n\x0fnext_page_token\x18\x02 \x01(\t\x12\x18\n\x0bunreachable\x18\x03 \x03(\tB\x03\xe0\x41\x03\"I\n\x12\x44\x65leteBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\
|
|
17
|
+
descriptor_data = "\n&google/cloud/dataproc/v1/batches.proto\x12\x18google.cloud.dataproc.v1\x1a\x1cgoogle/api/annotations.proto\x1a\x17google/api/client.proto\x1a\x1fgoogle/api/field_behavior.proto\x1a\x19google/api/resource.proto\x1a%google/cloud/dataproc/v1/shared.proto\x1a#google/longrunning/operations.proto\x1a\x1bgoogle/protobuf/empty.proto\x1a\x1fgoogle/protobuf/timestamp.proto\"\xb0\x01\n\x12\x43reateBatchRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x33\n\x05\x62\x61tch\x18\x02 \x01(\x0b\x32\x1f.google.cloud.dataproc.v1.BatchB\x03\xe0\x41\x02\x12\x15\n\x08\x62\x61tch_id\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x17\n\nrequest_id\x18\x04 \x01(\tB\x03\xe0\x41\x01\"F\n\x0fGetBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\xa8\x01\n\x12ListBatchesRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x16\n\tpage_size\x18\x02 \x01(\x05\x42\x03\xe0\x41\x01\x12\x17\n\npage_token\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x13\n\x06\x66ilter\x18\x04 \x01(\tB\x03\xe0\x41\x01\x12\x15\n\x08order_by\x18\x05 \x01(\tB\x03\xe0\x41\x01\"z\n\x13ListBatchesResponse\x12\x30\n\x07\x62\x61tches\x18\x01 \x03(\x0b\x32\x1f.google.cloud.dataproc.v1.Batch\x12\x17\n\x0fnext_page_token\x18\x02 \x01(\t\x12\x18\n\x0bunreachable\x18\x03 \x03(\tB\x03\xe0\x41\x03\"I\n\x12\x44\x65leteBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\x9f\x0b\n\x05\x42\x61tch\x12\x11\n\x04name\x18\x01 \x01(\tB\x03\xe0\x41\x03\x12\x11\n\x04uuid\x18\x02 \x01(\tB\x03\xe0\x41\x03\x12\x34\n\x0b\x63reate_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x12\x44\n\rpyspark_batch\x18\x04 \x01(\x0b\x32&.google.cloud.dataproc.v1.PySparkBatchB\x03\xe0\x41\x01H\x00\x12@\n\x0bspark_batch\x18\x05 \x01(\x0b\x32$.google.cloud.dataproc.v1.SparkBatchB\x03\xe0\x41\x01H\x00\x12\x43\n\rspark_r_batch\x18\x06 \x01(\x0b\x32%.google.cloud.dataproc.v1.SparkRBatchB\x03\xe0\x41\x01H\x00\x12G\n\x0fspark_sql_batch\x18\x07 \x01(\x0b\x32\'.google.cloud.dataproc.v1.SparkSqlBatchB\x03\xe0\x41\x01H\x00\x12U\n\x16pyspark_notebook_batch\x18\x13 \x01(\x0b\x32..google.cloud.dataproc.v1.PySparkNotebookBatchB\x03\xe0\x41\x01H\x00\x12@\n\x0cruntime_info\x18\x08 \x01(\x0b\x32%.google.cloud.dataproc.v1.RuntimeInfoB\x03\xe0\x41\x03\x12\x39\n\x05state\x18\t \x01(\x0e\x32%.google.cloud.dataproc.v1.Batch.StateB\x03\xe0\x41\x03\x12\x1a\n\rstate_message\x18\n \x01(\tB\x03\xe0\x41\x03\x12\x33\n\nstate_time\x18\x0b \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x12\x14\n\x07\x63reator\x18\x0c \x01(\tB\x03\xe0\x41\x03\x12@\n\x06labels\x18\r \x03(\x0b\x32+.google.cloud.dataproc.v1.Batch.LabelsEntryB\x03\xe0\x41\x01\x12\x44\n\x0eruntime_config\x18\x0e \x01(\x0b\x32\'.google.cloud.dataproc.v1.RuntimeConfigB\x03\xe0\x41\x01\x12L\n\x12\x65nvironment_config\x18\x0f \x01(\x0b\x32+.google.cloud.dataproc.v1.EnvironmentConfigB\x03\xe0\x41\x01\x12\x16\n\toperation\x18\x10 \x01(\tB\x03\xe0\x41\x03\x12H\n\rstate_history\x18\x11 \x03(\x0b\x32,.google.cloud.dataproc.v1.Batch.StateHistoryB\x03\xe0\x41\x03\x1a\xa0\x01\n\x0cStateHistory\x12\x39\n\x05state\x18\x01 \x01(\x0e\x32%.google.cloud.dataproc.v1.Batch.StateB\x03\xe0\x41\x03\x12\x1a\n\rstate_message\x18\x02 \x01(\tB\x03\xe0\x41\x03\x12\x39\n\x10state_start_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x1a-\n\x0bLabelsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"r\n\x05State\x12\x15\n\x11STATE_UNSPECIFIED\x10\x00\x12\x0b\n\x07PENDING\x10\x01\x12\x0b\n\x07RUNNING\x10\x02\x12\x0e\n\nCANCELLING\x10\x03\x12\r\n\tCANCELLED\x10\x04\x12\r\n\tSUCCEEDED\x10\x05\x12\n\n\x06\x46\x41ILED\x10\x06:[\xea\x41X\n\x1d\x64\x61taproc.googleapis.com/Batch\x12\x37projects/{project}/locations/{location}/batches/{batch}B\x0e\n\x0c\x62\x61tch_config\"\xb2\x01\n\x0cPySparkBatch\x12!\n\x14main_python_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x11\n\x04\x61rgs\x18\x02 \x03(\tB\x03\xe0\x41\x01\x12\x1d\n\x10python_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\"\xb5\x01\n\nSparkBatch\x12 \n\x11main_jar_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x01H\x00\x12\x19\n\nmain_class\x18\x02 \x01(\tB\x03\xe0\x41\x01H\x00\x12\x11\n\x04\x61rgs\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\x42\x08\n\x06\x64river\"q\n\x0bSparkRBatch\x12\x1c\n\x0fmain_r_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x11\n\x04\x61rgs\x18\x02 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\"\xda\x01\n\rSparkSqlBatch\x12\x1b\n\x0equery_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12Y\n\x0fquery_variables\x18\x02 \x03(\x0b\x32;.google.cloud.dataproc.v1.SparkSqlBatch.QueryVariablesEntryB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x1a\x35\n\x13QueryVariablesEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"\xa4\x02\n\x14PySparkNotebookBatch\x12\x1e\n\x11notebook_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12O\n\x06params\x18\x02 \x03(\x0b\x32:.google.cloud.dataproc.v1.PySparkNotebookBatch.ParamsEntryB\x03\xe0\x41\x01\x12\x1d\n\x10python_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\x1a-\n\x0bParamsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\x32\x9d\x06\n\x0f\x42\x61tchController\x12\xea\x01\n\x0b\x43reateBatch\x12,.google.cloud.dataproc.v1.CreateBatchRequest\x1a\x1d.google.longrunning.Operation\"\x8d\x01\xca\x41\x38\n\x05\x42\x61tch\x12/google.cloud.dataproc.v1.BatchOperationMetadata\xda\x41\x15parent,batch,batch_id\x82\xd3\xe4\x93\x02\x34\"+/v1/{parent=projects/*/locations/*}/batches:\x05\x62\x61tch\x12\x92\x01\n\x08GetBatch\x12).google.cloud.dataproc.v1.GetBatchRequest\x1a\x1f.google.cloud.dataproc.v1.Batch\":\xda\x41\x04name\x82\xd3\xe4\x93\x02-\x12+/v1/{name=projects/*/locations/*/batches/*}\x12\xa8\x01\n\x0bListBatches\x12,.google.cloud.dataproc.v1.ListBatchesRequest\x1a-.google.cloud.dataproc.v1.ListBatchesResponse\"<\xda\x41\x06parent\x82\xd3\xe4\x93\x02-\x12+/v1/{parent=projects/*/locations/*}/batches\x12\x8f\x01\n\x0b\x44\x65leteBatch\x12,.google.cloud.dataproc.v1.DeleteBatchRequest\x1a\x16.google.protobuf.Empty\":\xda\x41\x04name\x82\xd3\xe4\x93\x02-*+/v1/{name=projects/*/locations/*/batches/*}\x1aK\xca\x41\x17\x64\x61taproc.googleapis.com\xd2\x41.https://www.googleapis.com/auth/cloud-platformBk\n\x1c\x63om.google.cloud.dataproc.v1B\x0c\x42\x61tchesProtoP\x01Z;cloud.google.com/go/dataproc/v2/apiv1/dataprocpb;dataprocpbb\x06proto3"
|
|
18
18
|
|
|
19
19
|
pool = ::Google::Protobuf::DescriptorPool.generated_pool
|
|
20
20
|
pool.add_serialized_file(descriptor_data)
|
|
@@ -35,6 +35,7 @@ module Google
|
|
|
35
35
|
SparkBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkBatch").msgclass
|
|
36
36
|
SparkRBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkRBatch").msgclass
|
|
37
37
|
SparkSqlBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkSqlBatch").msgclass
|
|
38
|
+
PySparkNotebookBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.PySparkNotebookBatch").msgclass
|
|
38
39
|
end
|
|
39
40
|
end
|
|
40
41
|
end
|
|
@@ -148,22 +148,27 @@ module Google
|
|
|
148
148
|
# @return [::Google::Cloud::Dataproc::V1::PySparkBatch]
|
|
149
149
|
# Optional. PySpark batch config.
|
|
150
150
|
#
|
|
151
|
-
# Note: The following fields are mutually exclusive: `pyspark_batch`, `spark_batch`, `spark_r_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
151
|
+
# Note: The following fields are mutually exclusive: `pyspark_batch`, `spark_batch`, `spark_r_batch`, `spark_sql_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
152
152
|
# @!attribute [rw] spark_batch
|
|
153
153
|
# @return [::Google::Cloud::Dataproc::V1::SparkBatch]
|
|
154
154
|
# Optional. Spark batch config.
|
|
155
155
|
#
|
|
156
|
-
# Note: The following fields are mutually exclusive: `spark_batch`, `pyspark_batch`, `spark_r_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
156
|
+
# Note: The following fields are mutually exclusive: `spark_batch`, `pyspark_batch`, `spark_r_batch`, `spark_sql_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
157
157
|
# @!attribute [rw] spark_r_batch
|
|
158
158
|
# @return [::Google::Cloud::Dataproc::V1::SparkRBatch]
|
|
159
159
|
# Optional. SparkR batch config.
|
|
160
160
|
#
|
|
161
|
-
# Note: The following fields are mutually exclusive: `spark_r_batch`, `pyspark_batch`, `spark_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
161
|
+
# Note: The following fields are mutually exclusive: `spark_r_batch`, `pyspark_batch`, `spark_batch`, `spark_sql_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
162
162
|
# @!attribute [rw] spark_sql_batch
|
|
163
163
|
# @return [::Google::Cloud::Dataproc::V1::SparkSqlBatch]
|
|
164
164
|
# Optional. SparkSql batch config.
|
|
165
165
|
#
|
|
166
|
-
# Note: The following fields are mutually exclusive: `spark_sql_batch`, `pyspark_batch`, `spark_batch`, `spark_r_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
166
|
+
# Note: The following fields are mutually exclusive: `spark_sql_batch`, `pyspark_batch`, `spark_batch`, `spark_r_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
167
|
+
# @!attribute [rw] pyspark_notebook_batch
|
|
168
|
+
# @return [::Google::Cloud::Dataproc::V1::PySparkNotebookBatch]
|
|
169
|
+
# Optional. PySpark notebook batch config.
|
|
170
|
+
#
|
|
171
|
+
# Note: The following fields are mutually exclusive: `pyspark_notebook_batch`, `pyspark_batch`, `spark_batch`, `spark_r_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
|
|
167
172
|
# @!attribute [r] runtime_info
|
|
168
173
|
# @return [::Google::Cloud::Dataproc::V1::RuntimeInfo]
|
|
169
174
|
# Output only. Runtime information about batch execution.
|
|
@@ -378,6 +383,42 @@ module Google
|
|
|
378
383
|
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
379
384
|
end
|
|
380
385
|
end
|
|
386
|
+
|
|
387
|
+
# A configuration for running a PySpark Notebook batch workload.
|
|
388
|
+
# @!attribute [rw] notebook_file_uri
|
|
389
|
+
# @return [::String]
|
|
390
|
+
# Required. The HCFS URI of the notebook file to execute.
|
|
391
|
+
# @!attribute [rw] params
|
|
392
|
+
# @return [::Google::Protobuf::Map{::String => ::String}]
|
|
393
|
+
# Optional. The parameters to pass to the notebook.
|
|
394
|
+
# @!attribute [rw] python_file_uris
|
|
395
|
+
# @return [::Array<::String>]
|
|
396
|
+
# Optional. HCFS URIs of Python files to pass to the PySpark framework.
|
|
397
|
+
# @!attribute [rw] jar_file_uris
|
|
398
|
+
# @return [::Array<::String>]
|
|
399
|
+
# Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
|
|
400
|
+
# @!attribute [rw] file_uris
|
|
401
|
+
# @return [::Array<::String>]
|
|
402
|
+
# Optional. HCFS URIs of files to be placed in the working directory of
|
|
403
|
+
# each executor
|
|
404
|
+
# @!attribute [rw] archive_uris
|
|
405
|
+
# @return [::Array<::String>]
|
|
406
|
+
# Optional. HCFS URIs of archives to be extracted into the working directory
|
|
407
|
+
# of each executor. Supported file types:
|
|
408
|
+
# `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
|
|
409
|
+
class PySparkNotebookBatch
|
|
410
|
+
include ::Google::Protobuf::MessageExts
|
|
411
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
412
|
+
|
|
413
|
+
# @!attribute [rw] key
|
|
414
|
+
# @return [::String]
|
|
415
|
+
# @!attribute [rw] value
|
|
416
|
+
# @return [::String]
|
|
417
|
+
class ParamsEntry
|
|
418
|
+
include ::Google::Protobuf::MessageExts
|
|
419
|
+
extend ::Google::Protobuf::MessageExts::ClassMethods
|
|
420
|
+
end
|
|
421
|
+
end
|
|
381
422
|
end
|
|
382
423
|
end
|
|
383
424
|
end
|