google-cloud-dataproc-v1 1.11.0 → 1.12.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 37932bd2bb6207e3c495595122bfdb61141cd3f83f7f2b7a0aa4ddbf25bf88db
4
- data.tar.gz: 9013eb67fb73fffd203b5a8fe952632ca8fa550c4df9726e5ac47e56cb1f64f6
3
+ metadata.gz: 27303e4bc8144f2dc1c53c914c4578644a112d5f3a4258dad0c01f34a674c6ef
4
+ data.tar.gz: 274803c54dfcbe3aaaf26d2d4f427f10f55670c58eddb596070cd53ab46f6c0a
5
5
  SHA512:
6
- metadata.gz: 16eae9d7d2a2ad94b0e2fe03edd9385f57a5b74cd07094a589f63d5d1c54f5527bec5876478ebc26a21d22dc2df7d182d5ea33b6660bcaa3c87c9bb30c882005
7
- data.tar.gz: 5787840e898c51092d9690ecb33bddfb73f225b7456db3c5254e6e36027e3cd5ed6d8d107acb8c4b73b5e8ca0124a93562ecf3ff069f076f9a089e8f88c3f7ef
6
+ metadata.gz: 5dc0fec27d452861498e66ed4352ce296a8f20f1f16b41ccb75f86c0c8a4d1b72428eb7fcdaf5a1b26a8fcdc449c780a7e8cec27d3b34bc40e4d1343388f3542
7
+ data.tar.gz: 38021feb7934623efef2981d218854b92f129694fe840cab86493833c362b2e5944193735d68bcea0015556580cac96403147d270d76e60d7c959e3559671263
@@ -14,7 +14,7 @@ require 'google/protobuf/empty_pb'
14
14
  require 'google/protobuf/timestamp_pb'
15
15
 
16
16
 
17
- descriptor_data = "\n&google/cloud/dataproc/v1/batches.proto\x12\x18google.cloud.dataproc.v1\x1a\x1cgoogle/api/annotations.proto\x1a\x17google/api/client.proto\x1a\x1fgoogle/api/field_behavior.proto\x1a\x19google/api/resource.proto\x1a%google/cloud/dataproc/v1/shared.proto\x1a#google/longrunning/operations.proto\x1a\x1bgoogle/protobuf/empty.proto\x1a\x1fgoogle/protobuf/timestamp.proto\"\xb0\x01\n\x12\x43reateBatchRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x33\n\x05\x62\x61tch\x18\x02 \x01(\x0b\x32\x1f.google.cloud.dataproc.v1.BatchB\x03\xe0\x41\x02\x12\x15\n\x08\x62\x61tch_id\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x17\n\nrequest_id\x18\x04 \x01(\tB\x03\xe0\x41\x01\"F\n\x0fGetBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\xa8\x01\n\x12ListBatchesRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x16\n\tpage_size\x18\x02 \x01(\x05\x42\x03\xe0\x41\x01\x12\x17\n\npage_token\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x13\n\x06\x66ilter\x18\x04 \x01(\tB\x03\xe0\x41\x01\x12\x15\n\x08order_by\x18\x05 \x01(\tB\x03\xe0\x41\x01\"z\n\x13ListBatchesResponse\x12\x30\n\x07\x62\x61tches\x18\x01 \x03(\x0b\x32\x1f.google.cloud.dataproc.v1.Batch\x12\x17\n\x0fnext_page_token\x18\x02 \x01(\t\x12\x18\n\x0bunreachable\x18\x03 \x03(\tB\x03\xe0\x41\x03\"I\n\x12\x44\x65leteBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\xc8\n\n\x05\x42\x61tch\x12\x11\n\x04name\x18\x01 \x01(\tB\x03\xe0\x41\x03\x12\x11\n\x04uuid\x18\x02 \x01(\tB\x03\xe0\x41\x03\x12\x34\n\x0b\x63reate_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x12\x44\n\rpyspark_batch\x18\x04 \x01(\x0b\x32&.google.cloud.dataproc.v1.PySparkBatchB\x03\xe0\x41\x01H\x00\x12@\n\x0bspark_batch\x18\x05 \x01(\x0b\x32$.google.cloud.dataproc.v1.SparkBatchB\x03\xe0\x41\x01H\x00\x12\x43\n\rspark_r_batch\x18\x06 \x01(\x0b\x32%.google.cloud.dataproc.v1.SparkRBatchB\x03\xe0\x41\x01H\x00\x12G\n\x0fspark_sql_batch\x18\x07 \x01(\x0b\x32\'.google.cloud.dataproc.v1.SparkSqlBatchB\x03\xe0\x41\x01H\x00\x12@\n\x0cruntime_info\x18\x08 \x01(\x0b\x32%.google.cloud.dataproc.v1.RuntimeInfoB\x03\xe0\x41\x03\x12\x39\n\x05state\x18\t \x01(\x0e\x32%.google.cloud.dataproc.v1.Batch.StateB\x03\xe0\x41\x03\x12\x1a\n\rstate_message\x18\n \x01(\tB\x03\xe0\x41\x03\x12\x33\n\nstate_time\x18\x0b \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x12\x14\n\x07\x63reator\x18\x0c \x01(\tB\x03\xe0\x41\x03\x12@\n\x06labels\x18\r \x03(\x0b\x32+.google.cloud.dataproc.v1.Batch.LabelsEntryB\x03\xe0\x41\x01\x12\x44\n\x0eruntime_config\x18\x0e \x01(\x0b\x32\'.google.cloud.dataproc.v1.RuntimeConfigB\x03\xe0\x41\x01\x12L\n\x12\x65nvironment_config\x18\x0f \x01(\x0b\x32+.google.cloud.dataproc.v1.EnvironmentConfigB\x03\xe0\x41\x01\x12\x16\n\toperation\x18\x10 \x01(\tB\x03\xe0\x41\x03\x12H\n\rstate_history\x18\x11 \x03(\x0b\x32,.google.cloud.dataproc.v1.Batch.StateHistoryB\x03\xe0\x41\x03\x1a\xa0\x01\n\x0cStateHistory\x12\x39\n\x05state\x18\x01 \x01(\x0e\x32%.google.cloud.dataproc.v1.Batch.StateB\x03\xe0\x41\x03\x12\x1a\n\rstate_message\x18\x02 \x01(\tB\x03\xe0\x41\x03\x12\x39\n\x10state_start_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x1a-\n\x0bLabelsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"r\n\x05State\x12\x15\n\x11STATE_UNSPECIFIED\x10\x00\x12\x0b\n\x07PENDING\x10\x01\x12\x0b\n\x07RUNNING\x10\x02\x12\x0e\n\nCANCELLING\x10\x03\x12\r\n\tCANCELLED\x10\x04\x12\r\n\tSUCCEEDED\x10\x05\x12\n\n\x06\x46\x41ILED\x10\x06:[\xea\x41X\n\x1d\x64\x61taproc.googleapis.com/Batch\x12\x37projects/{project}/locations/{location}/batches/{batch}B\x0e\n\x0c\x62\x61tch_config\"\xb2\x01\n\x0cPySparkBatch\x12!\n\x14main_python_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x11\n\x04\x61rgs\x18\x02 \x03(\tB\x03\xe0\x41\x01\x12\x1d\n\x10python_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\"\xb5\x01\n\nSparkBatch\x12 \n\x11main_jar_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x01H\x00\x12\x19\n\nmain_class\x18\x02 \x01(\tB\x03\xe0\x41\x01H\x00\x12\x11\n\x04\x61rgs\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\x42\x08\n\x06\x64river\"q\n\x0bSparkRBatch\x12\x1c\n\x0fmain_r_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x11\n\x04\x61rgs\x18\x02 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\"\xda\x01\n\rSparkSqlBatch\x12\x1b\n\x0equery_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12Y\n\x0fquery_variables\x18\x02 \x03(\x0b\x32;.google.cloud.dataproc.v1.SparkSqlBatch.QueryVariablesEntryB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x1a\x35\n\x13QueryVariablesEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\x32\x9d\x06\n\x0f\x42\x61tchController\x12\xea\x01\n\x0b\x43reateBatch\x12,.google.cloud.dataproc.v1.CreateBatchRequest\x1a\x1d.google.longrunning.Operation\"\x8d\x01\xca\x41\x38\n\x05\x42\x61tch\x12/google.cloud.dataproc.v1.BatchOperationMetadata\xda\x41\x15parent,batch,batch_id\x82\xd3\xe4\x93\x02\x34\"+/v1/{parent=projects/*/locations/*}/batches:\x05\x62\x61tch\x12\x92\x01\n\x08GetBatch\x12).google.cloud.dataproc.v1.GetBatchRequest\x1a\x1f.google.cloud.dataproc.v1.Batch\":\xda\x41\x04name\x82\xd3\xe4\x93\x02-\x12+/v1/{name=projects/*/locations/*/batches/*}\x12\xa8\x01\n\x0bListBatches\x12,.google.cloud.dataproc.v1.ListBatchesRequest\x1a-.google.cloud.dataproc.v1.ListBatchesResponse\"<\xda\x41\x06parent\x82\xd3\xe4\x93\x02-\x12+/v1/{parent=projects/*/locations/*}/batches\x12\x8f\x01\n\x0b\x44\x65leteBatch\x12,.google.cloud.dataproc.v1.DeleteBatchRequest\x1a\x16.google.protobuf.Empty\":\xda\x41\x04name\x82\xd3\xe4\x93\x02-*+/v1/{name=projects/*/locations/*/batches/*}\x1aK\xca\x41\x17\x64\x61taproc.googleapis.com\xd2\x41.https://www.googleapis.com/auth/cloud-platformBk\n\x1c\x63om.google.cloud.dataproc.v1B\x0c\x42\x61tchesProtoP\x01Z;cloud.google.com/go/dataproc/v2/apiv1/dataprocpb;dataprocpbb\x06proto3"
17
+ descriptor_data = "\n&google/cloud/dataproc/v1/batches.proto\x12\x18google.cloud.dataproc.v1\x1a\x1cgoogle/api/annotations.proto\x1a\x17google/api/client.proto\x1a\x1fgoogle/api/field_behavior.proto\x1a\x19google/api/resource.proto\x1a%google/cloud/dataproc/v1/shared.proto\x1a#google/longrunning/operations.proto\x1a\x1bgoogle/protobuf/empty.proto\x1a\x1fgoogle/protobuf/timestamp.proto\"\xb0\x01\n\x12\x43reateBatchRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x33\n\x05\x62\x61tch\x18\x02 \x01(\x0b\x32\x1f.google.cloud.dataproc.v1.BatchB\x03\xe0\x41\x02\x12\x15\n\x08\x62\x61tch_id\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x17\n\nrequest_id\x18\x04 \x01(\tB\x03\xe0\x41\x01\"F\n\x0fGetBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\xa8\x01\n\x12ListBatchesRequest\x12\x35\n\x06parent\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\x12\x1d\x64\x61taproc.googleapis.com/Batch\x12\x16\n\tpage_size\x18\x02 \x01(\x05\x42\x03\xe0\x41\x01\x12\x17\n\npage_token\x18\x03 \x01(\tB\x03\xe0\x41\x01\x12\x13\n\x06\x66ilter\x18\x04 \x01(\tB\x03\xe0\x41\x01\x12\x15\n\x08order_by\x18\x05 \x01(\tB\x03\xe0\x41\x01\"z\n\x13ListBatchesResponse\x12\x30\n\x07\x62\x61tches\x18\x01 \x03(\x0b\x32\x1f.google.cloud.dataproc.v1.Batch\x12\x17\n\x0fnext_page_token\x18\x02 \x01(\t\x12\x18\n\x0bunreachable\x18\x03 \x03(\tB\x03\xe0\x41\x03\"I\n\x12\x44\x65leteBatchRequest\x12\x33\n\x04name\x18\x01 \x01(\tB%\xe0\x41\x02\xfa\x41\x1f\n\x1d\x64\x61taproc.googleapis.com/Batch\"\x9f\x0b\n\x05\x42\x61tch\x12\x11\n\x04name\x18\x01 \x01(\tB\x03\xe0\x41\x03\x12\x11\n\x04uuid\x18\x02 \x01(\tB\x03\xe0\x41\x03\x12\x34\n\x0b\x63reate_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x12\x44\n\rpyspark_batch\x18\x04 \x01(\x0b\x32&.google.cloud.dataproc.v1.PySparkBatchB\x03\xe0\x41\x01H\x00\x12@\n\x0bspark_batch\x18\x05 \x01(\x0b\x32$.google.cloud.dataproc.v1.SparkBatchB\x03\xe0\x41\x01H\x00\x12\x43\n\rspark_r_batch\x18\x06 \x01(\x0b\x32%.google.cloud.dataproc.v1.SparkRBatchB\x03\xe0\x41\x01H\x00\x12G\n\x0fspark_sql_batch\x18\x07 \x01(\x0b\x32\'.google.cloud.dataproc.v1.SparkSqlBatchB\x03\xe0\x41\x01H\x00\x12U\n\x16pyspark_notebook_batch\x18\x13 \x01(\x0b\x32..google.cloud.dataproc.v1.PySparkNotebookBatchB\x03\xe0\x41\x01H\x00\x12@\n\x0cruntime_info\x18\x08 \x01(\x0b\x32%.google.cloud.dataproc.v1.RuntimeInfoB\x03\xe0\x41\x03\x12\x39\n\x05state\x18\t \x01(\x0e\x32%.google.cloud.dataproc.v1.Batch.StateB\x03\xe0\x41\x03\x12\x1a\n\rstate_message\x18\n \x01(\tB\x03\xe0\x41\x03\x12\x33\n\nstate_time\x18\x0b \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x12\x14\n\x07\x63reator\x18\x0c \x01(\tB\x03\xe0\x41\x03\x12@\n\x06labels\x18\r \x03(\x0b\x32+.google.cloud.dataproc.v1.Batch.LabelsEntryB\x03\xe0\x41\x01\x12\x44\n\x0eruntime_config\x18\x0e \x01(\x0b\x32\'.google.cloud.dataproc.v1.RuntimeConfigB\x03\xe0\x41\x01\x12L\n\x12\x65nvironment_config\x18\x0f \x01(\x0b\x32+.google.cloud.dataproc.v1.EnvironmentConfigB\x03\xe0\x41\x01\x12\x16\n\toperation\x18\x10 \x01(\tB\x03\xe0\x41\x03\x12H\n\rstate_history\x18\x11 \x03(\x0b\x32,.google.cloud.dataproc.v1.Batch.StateHistoryB\x03\xe0\x41\x03\x1a\xa0\x01\n\x0cStateHistory\x12\x39\n\x05state\x18\x01 \x01(\x0e\x32%.google.cloud.dataproc.v1.Batch.StateB\x03\xe0\x41\x03\x12\x1a\n\rstate_message\x18\x02 \x01(\tB\x03\xe0\x41\x03\x12\x39\n\x10state_start_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\x03\xe0\x41\x03\x1a-\n\x0bLabelsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"r\n\x05State\x12\x15\n\x11STATE_UNSPECIFIED\x10\x00\x12\x0b\n\x07PENDING\x10\x01\x12\x0b\n\x07RUNNING\x10\x02\x12\x0e\n\nCANCELLING\x10\x03\x12\r\n\tCANCELLED\x10\x04\x12\r\n\tSUCCEEDED\x10\x05\x12\n\n\x06\x46\x41ILED\x10\x06:[\xea\x41X\n\x1d\x64\x61taproc.googleapis.com/Batch\x12\x37projects/{project}/locations/{location}/batches/{batch}B\x0e\n\x0c\x62\x61tch_config\"\xb2\x01\n\x0cPySparkBatch\x12!\n\x14main_python_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x11\n\x04\x61rgs\x18\x02 \x03(\tB\x03\xe0\x41\x01\x12\x1d\n\x10python_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\"\xb5\x01\n\nSparkBatch\x12 \n\x11main_jar_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x01H\x00\x12\x19\n\nmain_class\x18\x02 \x01(\tB\x03\xe0\x41\x01H\x00\x12\x11\n\x04\x61rgs\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\x42\x08\n\x06\x64river\"q\n\x0bSparkRBatch\x12\x1c\n\x0fmain_r_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x11\n\x04\x61rgs\x18\x02 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\"\xda\x01\n\rSparkSqlBatch\x12\x1b\n\x0equery_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12Y\n\x0fquery_variables\x18\x02 \x03(\x0b\x32;.google.cloud.dataproc.v1.SparkSqlBatch.QueryVariablesEntryB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x1a\x35\n\x13QueryVariablesEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\"\xa4\x02\n\x14PySparkNotebookBatch\x12\x1e\n\x11notebook_file_uri\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12O\n\x06params\x18\x02 \x03(\x0b\x32:.google.cloud.dataproc.v1.PySparkNotebookBatch.ParamsEntryB\x03\xe0\x41\x01\x12\x1d\n\x10python_file_uris\x18\x03 \x03(\tB\x03\xe0\x41\x01\x12\x1a\n\rjar_file_uris\x18\x04 \x03(\tB\x03\xe0\x41\x01\x12\x16\n\tfile_uris\x18\x05 \x03(\tB\x03\xe0\x41\x01\x12\x19\n\x0c\x61rchive_uris\x18\x06 \x03(\tB\x03\xe0\x41\x01\x1a-\n\x0bParamsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t:\x02\x38\x01\x32\x9d\x06\n\x0f\x42\x61tchController\x12\xea\x01\n\x0b\x43reateBatch\x12,.google.cloud.dataproc.v1.CreateBatchRequest\x1a\x1d.google.longrunning.Operation\"\x8d\x01\xca\x41\x38\n\x05\x42\x61tch\x12/google.cloud.dataproc.v1.BatchOperationMetadata\xda\x41\x15parent,batch,batch_id\x82\xd3\xe4\x93\x02\x34\"+/v1/{parent=projects/*/locations/*}/batches:\x05\x62\x61tch\x12\x92\x01\n\x08GetBatch\x12).google.cloud.dataproc.v1.GetBatchRequest\x1a\x1f.google.cloud.dataproc.v1.Batch\":\xda\x41\x04name\x82\xd3\xe4\x93\x02-\x12+/v1/{name=projects/*/locations/*/batches/*}\x12\xa8\x01\n\x0bListBatches\x12,.google.cloud.dataproc.v1.ListBatchesRequest\x1a-.google.cloud.dataproc.v1.ListBatchesResponse\"<\xda\x41\x06parent\x82\xd3\xe4\x93\x02-\x12+/v1/{parent=projects/*/locations/*}/batches\x12\x8f\x01\n\x0b\x44\x65leteBatch\x12,.google.cloud.dataproc.v1.DeleteBatchRequest\x1a\x16.google.protobuf.Empty\":\xda\x41\x04name\x82\xd3\xe4\x93\x02-*+/v1/{name=projects/*/locations/*/batches/*}\x1aK\xca\x41\x17\x64\x61taproc.googleapis.com\xd2\x41.https://www.googleapis.com/auth/cloud-platformBk\n\x1c\x63om.google.cloud.dataproc.v1B\x0c\x42\x61tchesProtoP\x01Z;cloud.google.com/go/dataproc/v2/apiv1/dataprocpb;dataprocpbb\x06proto3"
18
18
 
19
19
  pool = ::Google::Protobuf::DescriptorPool.generated_pool
20
20
  pool.add_serialized_file(descriptor_data)
@@ -35,6 +35,7 @@ module Google
35
35
  SparkBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkBatch").msgclass
36
36
  SparkRBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkRBatch").msgclass
37
37
  SparkSqlBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkSqlBatch").msgclass
38
+ PySparkNotebookBatch = ::Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.PySparkNotebookBatch").msgclass
38
39
  end
39
40
  end
40
41
  end
@@ -21,7 +21,7 @@ module Google
21
21
  module Cloud
22
22
  module Dataproc
23
23
  module V1
24
- VERSION = "1.11.0"
24
+ VERSION = "1.12.0"
25
25
  end
26
26
  end
27
27
  end
@@ -148,22 +148,27 @@ module Google
148
148
  # @return [::Google::Cloud::Dataproc::V1::PySparkBatch]
149
149
  # Optional. PySpark batch config.
150
150
  #
151
- # Note: The following fields are mutually exclusive: `pyspark_batch`, `spark_batch`, `spark_r_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
151
+ # Note: The following fields are mutually exclusive: `pyspark_batch`, `spark_batch`, `spark_r_batch`, `spark_sql_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
152
152
  # @!attribute [rw] spark_batch
153
153
  # @return [::Google::Cloud::Dataproc::V1::SparkBatch]
154
154
  # Optional. Spark batch config.
155
155
  #
156
- # Note: The following fields are mutually exclusive: `spark_batch`, `pyspark_batch`, `spark_r_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
156
+ # Note: The following fields are mutually exclusive: `spark_batch`, `pyspark_batch`, `spark_r_batch`, `spark_sql_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
157
157
  # @!attribute [rw] spark_r_batch
158
158
  # @return [::Google::Cloud::Dataproc::V1::SparkRBatch]
159
159
  # Optional. SparkR batch config.
160
160
  #
161
- # Note: The following fields are mutually exclusive: `spark_r_batch`, `pyspark_batch`, `spark_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
161
+ # Note: The following fields are mutually exclusive: `spark_r_batch`, `pyspark_batch`, `spark_batch`, `spark_sql_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
162
162
  # @!attribute [rw] spark_sql_batch
163
163
  # @return [::Google::Cloud::Dataproc::V1::SparkSqlBatch]
164
164
  # Optional. SparkSql batch config.
165
165
  #
166
- # Note: The following fields are mutually exclusive: `spark_sql_batch`, `pyspark_batch`, `spark_batch`, `spark_r_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
166
+ # Note: The following fields are mutually exclusive: `spark_sql_batch`, `pyspark_batch`, `spark_batch`, `spark_r_batch`, `pyspark_notebook_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
167
+ # @!attribute [rw] pyspark_notebook_batch
168
+ # @return [::Google::Cloud::Dataproc::V1::PySparkNotebookBatch]
169
+ # Optional. PySpark notebook batch config.
170
+ #
171
+ # Note: The following fields are mutually exclusive: `pyspark_notebook_batch`, `pyspark_batch`, `spark_batch`, `spark_r_batch`, `spark_sql_batch`. If a field in that set is populated, all other fields in the set will automatically be cleared.
167
172
  # @!attribute [r] runtime_info
168
173
  # @return [::Google::Cloud::Dataproc::V1::RuntimeInfo]
169
174
  # Output only. Runtime information about batch execution.
@@ -378,6 +383,42 @@ module Google
378
383
  extend ::Google::Protobuf::MessageExts::ClassMethods
379
384
  end
380
385
  end
386
+
387
+ # A configuration for running a PySpark Notebook batch workload.
388
+ # @!attribute [rw] notebook_file_uri
389
+ # @return [::String]
390
+ # Required. The HCFS URI of the notebook file to execute.
391
+ # @!attribute [rw] params
392
+ # @return [::Google::Protobuf::Map{::String => ::String}]
393
+ # Optional. The parameters to pass to the notebook.
394
+ # @!attribute [rw] python_file_uris
395
+ # @return [::Array<::String>]
396
+ # Optional. HCFS URIs of Python files to pass to the PySpark framework.
397
+ # @!attribute [rw] jar_file_uris
398
+ # @return [::Array<::String>]
399
+ # Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
400
+ # @!attribute [rw] file_uris
401
+ # @return [::Array<::String>]
402
+ # Optional. HCFS URIs of files to be placed in the working directory of
403
+ # each executor
404
+ # @!attribute [rw] archive_uris
405
+ # @return [::Array<::String>]
406
+ # Optional. HCFS URIs of archives to be extracted into the working directory
407
+ # of each executor. Supported file types:
408
+ # `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.
409
+ class PySparkNotebookBatch
410
+ include ::Google::Protobuf::MessageExts
411
+ extend ::Google::Protobuf::MessageExts::ClassMethods
412
+
413
+ # @!attribute [rw] key
414
+ # @return [::String]
415
+ # @!attribute [rw] value
416
+ # @return [::String]
417
+ class ParamsEntry
418
+ include ::Google::Protobuf::MessageExts
419
+ extend ::Google::Protobuf::MessageExts::ClassMethods
420
+ end
421
+ end
381
422
  end
382
423
  end
383
424
  end
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: google-cloud-dataproc-v1
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.11.0
4
+ version: 1.12.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Google LLC