yandexcloud 0.338.0__py3-none-any.whl → 0.340.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- yandex/cloud/ai/assistants/v1/common_pb2.py +32 -27
- yandex/cloud/ai/assistants/v1/common_pb2.pyi +29 -2
- yandex/cloud/ai/dataset/v1/dataset_service_pb2.py +32 -32
- yandex/cloud/ai/dataset/v1/dataset_service_pb2.pyi +7 -1
- yandex/cloud/ai/tuning/v1/tuning_service_pb2.py +39 -3
- yandex/cloud/ai/tuning/v1/tuning_service_pb2.pyi +220 -0
- yandex/cloud/ai/tuning/v1/tuning_service_pb2_grpc.py +172 -0
- yandex/cloud/ai/tuning/v1/tuning_service_pb2_grpc.pyi +68 -0
- yandex/cloud/ai/tuning/v1/tuning_task_pb2.py +3 -3
- yandex/cloud/ai/tuning/v1/tuning_task_pb2.pyi +2 -0
- yandex/cloud/apploadbalancer/v1/load_balancer_pb2.py +50 -50
- yandex/cloud/apploadbalancer/v1/load_balancer_pb2.pyi +10 -2
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2.py +86 -66
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2.pyi +98 -2
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2_grpc.py +88 -0
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2_grpc.pyi +40 -0
- yandex/cloud/backup/v1/policy_pb2.py +56 -52
- yandex/cloud/backup/v1/policy_pb2.pyi +36 -5
- yandex/cloud/backup/v1/resource_pb2.py +21 -19
- yandex/cloud/backup/v1/resource_pb2.pyi +32 -2
- yandex/cloud/backup/v1/resource_service_pb2.py +30 -30
- yandex/cloud/backup/v1/resource_service_pb2.pyi +5 -1
- yandex/cloud/baremetal/__init__.py +0 -0
- yandex/cloud/baremetal/v1alpha/__init__.py +0 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2.py +42 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2.pyi +109 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2.py +49 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2.pyi +111 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2_grpc.py +148 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2_grpc.pyi +80 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2.py +37 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2.pyi +33 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2.py +39 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2.pyi +68 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2.py +37 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2.pyi +36 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2.py +49 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2.pyi +87 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2_grpc.py +148 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2_grpc.pyi +80 -0
- yandex/cloud/baremetal/v1alpha/image_pb2.py +44 -0
- yandex/cloud/baremetal/v1alpha/image_pb2.pyi +125 -0
- yandex/cloud/baremetal/v1alpha/image_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/image_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2.py +108 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2.pyi +364 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2.py +47 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2.pyi +174 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2.py +112 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2.pyi +443 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2.py +45 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2.pyi +143 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2.py +106 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2.pyi +382 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/server_pb2.py +54 -0
- yandex/cloud/baremetal/v1alpha/server_pb2.pyi +282 -0
- yandex/cloud/baremetal/v1alpha/server_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/server_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2.py +217 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2.pyi +850 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2_grpc.py +589 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2_grpc.pyi +281 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2.py +46 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2.pyi +165 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2.py +51 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2.pyi +91 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2_grpc.py +145 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2_grpc.pyi +70 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2.py +42 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2.pyi +77 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2.py +106 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2.pyi +357 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2.py +37 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2.pyi +29 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2.py +49 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2.pyi +87 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2_grpc.py +148 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2_grpc.pyi +80 -0
- yandex/cloud/cloudregistry/v1/registry_service_pb2.py +40 -40
- yandex/cloud/dataproc/manager/v1/job_pb2.pyi +26 -26
- yandex/cloud/dataproc/manager/v1/job_service_pb2.pyi +7 -7
- yandex/cloud/dataproc/manager/v1/job_service_pb2_grpc.py +2 -2
- yandex/cloud/dataproc/manager/v1/job_service_pb2_grpc.pyi +6 -6
- yandex/cloud/dataproc/manager/v1/manager_service_pb2.pyi +2 -2
- yandex/cloud/dataproc/manager/v1/manager_service_pb2_grpc.py +3 -3
- yandex/cloud/dataproc/manager/v1/manager_service_pb2_grpc.pyi +3 -3
- yandex/cloud/dataproc/v1/cluster_pb2.pyi +5 -5
- yandex/cloud/dataproc/v1/cluster_service_pb2.pyi +13 -13
- yandex/cloud/dataproc/v1/cluster_service_pb2_grpc.py +4 -4
- yandex/cloud/dataproc/v1/cluster_service_pb2_grpc.pyi +6 -6
- yandex/cloud/dataproc/v1/job_pb2.pyi +15 -15
- yandex/cloud/dataproc/v1/job_service_pb2.pyi +3 -3
- yandex/cloud/dataproc/v1/job_service_pb2_grpc.py +4 -4
- yandex/cloud/dataproc/v1/job_service_pb2_grpc.pyi +6 -6
- yandex/cloud/dataproc/v1/resource_preset_pb2.pyi +2 -2
- yandex/cloud/dataproc/v1/subcluster_pb2.pyi +11 -11
- yandex/cloud/dataproc/v1/subcluster_service_pb2.pyi +3 -3
- yandex/cloud/dataproc/v1/subcluster_service_pb2_grpc.py +3 -3
- yandex/cloud/dataproc/v1/subcluster_service_pb2_grpc.pyi +3 -3
- yandex/cloud/iam/v1/api_key_pb2.pyi +3 -1
- yandex/cloud/iam/v1/api_key_service_pb2.pyi +3 -1
- yandex/cloud/iam/v1/workload/oidc/federation_service_pb2.py +32 -25
- yandex/cloud/iam/v1/workload/oidc/federation_service_pb2_grpc.py +133 -0
- yandex/cloud/iam/v1/workload/oidc/federation_service_pb2_grpc.pyi +61 -0
- yandex/cloud/k8s/v1/cluster_pb2.py +40 -30
- yandex/cloud/k8s/v1/cluster_pb2.pyi +94 -2
- yandex/cloud/k8s/v1/cluster_service_pb2.py +54 -42
- yandex/cloud/k8s/v1/cluster_service_pb2.pyi +70 -4
- yandex/cloud/k8s/v1/resource_preset_pb2.py +37 -0
- yandex/cloud/k8s/v1/resource_preset_pb2.pyi +43 -0
- yandex/cloud/k8s/v1/resource_preset_pb2_grpc.py +24 -0
- yandex/cloud/k8s/v1/resource_preset_pb2_grpc.pyi +17 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2.py +53 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2.pyi +81 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2_grpc.py +146 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2_grpc.pyi +71 -0
- yandex/cloud/mdb/mysql/v1/backup_pb2.py +6 -6
- yandex/cloud/mdb/mysql/v1/backup_pb2.pyi +5 -1
- yandex/cloud/mdb/mysql/v1/config/mysql5_7_pb2.py +26 -20
- yandex/cloud/mdb/mysql/v1/config/mysql5_7_pb2.pyi +73 -2
- yandex/cloud/mdb/mysql/v1/config/mysql8_0_pb2.py +26 -20
- yandex/cloud/mdb/mysql/v1/config/mysql8_0_pb2.pyi +100 -2
- yandex/cloud/mdb/opensearch/v1/backup_pb2.py +30 -3
- yandex/cloud/mdb/opensearch/v1/backup_pb2.pyi +162 -0
- yandex/cloud/mdb/opensearch/v1/cluster_pb2.py +46 -45
- yandex/cloud/mdb/opensearch/v1/cluster_pb2.pyi +9 -2
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2.py +78 -78
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2.pyi +16 -4
- yandex/cloud/mdb/postgresql/v1/config/host10_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host10_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host11_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host11_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host12_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host12_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host13_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host13_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host14_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host14_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host15_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host15_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host16_1c_pb2.py +25 -25
- yandex/cloud/mdb/postgresql/v1/config/host16_pb2.py +25 -25
- yandex/cloud/mdb/postgresql/v1/config/host9_6_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/postgresql10_1c_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql10_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql11_1c_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql11_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql12_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql12_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql13_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql13_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql14_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql14_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql15_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql15_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql16_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql16_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql9_6_pb2.py +29 -29
- yandex/cloud/monitoring/v3/unit_format_pb2.py +2 -2
- yandex/cloud/monitoring/v3/unit_format_pb2.pyi +16 -0
- yandex/cloud/searchapi/v2/gen_search_service_pb2.py +28 -22
- yandex/cloud/spark/__init__.py +0 -0
- yandex/cloud/spark/v1/__init__.py +0 -0
- yandex/cloud/spark/v1/cluster_pb2.py +100 -0
- yandex/cloud/spark/v1/cluster_pb2.pyi +444 -0
- yandex/cloud/spark/v1/cluster_pb2_grpc.py +24 -0
- yandex/cloud/spark/v1/cluster_pb2_grpc.pyi +17 -0
- yandex/cloud/spark/v1/cluster_service_pb2.py +124 -0
- yandex/cloud/spark/v1/cluster_service_pb2.pyi +333 -0
- yandex/cloud/spark/v1/cluster_service_pb2_grpc.py +366 -0
- yandex/cloud/spark/v1/cluster_service_pb2_grpc.pyi +169 -0
- yandex/cloud/spark/v1/job_pb2.py +56 -0
- yandex/cloud/spark/v1/job_pb2.pyi +346 -0
- yandex/cloud/spark/v1/job_pb2_grpc.py +24 -0
- yandex/cloud/spark/v1/job_pb2_grpc.pyi +17 -0
- yandex/cloud/spark/v1/job_service_pb2.py +97 -0
- yandex/cloud/spark/v1/job_service_pb2.pyi +224 -0
- yandex/cloud/spark/v1/job_service_pb2_grpc.py +279 -0
- yandex/cloud/spark/v1/job_service_pb2_grpc.pyi +132 -0
- yandex/cloud/spark/v1/maintenance_pb2.py +53 -0
- yandex/cloud/spark/v1/maintenance_pb2.pyi +123 -0
- yandex/cloud/spark/v1/maintenance_pb2_grpc.py +24 -0
- yandex/cloud/spark/v1/maintenance_pb2_grpc.pyi +17 -0
- yandexcloud/__init__.py +1 -1
- yandexcloud/_sdk.py +1 -0
- yandexcloud/_wrappers/__init__.py +15 -0
- yandexcloud/_wrappers/spark/__init__.py +603 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/METADATA +2 -1
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/RECORD +219 -106
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/AUTHORS +0 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/LICENSE +0 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/WHEEL +0 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/top_level.txt +0 -0
|
@@ -0,0 +1,346 @@
|
|
|
1
|
+
"""
|
|
2
|
+
@generated by mypy-protobuf. Do not edit manually!
|
|
3
|
+
isort:skip_file
|
|
4
|
+
"""
|
|
5
|
+
|
|
6
|
+
import builtins
|
|
7
|
+
import collections.abc
|
|
8
|
+
import google.protobuf.descriptor
|
|
9
|
+
import google.protobuf.internal.containers
|
|
10
|
+
import google.protobuf.internal.enum_type_wrapper
|
|
11
|
+
import google.protobuf.message
|
|
12
|
+
import google.protobuf.timestamp_pb2
|
|
13
|
+
import sys
|
|
14
|
+
import typing
|
|
15
|
+
|
|
16
|
+
if sys.version_info >= (3, 10):
|
|
17
|
+
import typing as typing_extensions
|
|
18
|
+
else:
|
|
19
|
+
import typing_extensions
|
|
20
|
+
|
|
21
|
+
DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
22
|
+
|
|
23
|
+
@typing.final
|
|
24
|
+
class Job(google.protobuf.message.Message):
|
|
25
|
+
"""Spark job."""
|
|
26
|
+
|
|
27
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
28
|
+
|
|
29
|
+
class _Status:
|
|
30
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
31
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
32
|
+
|
|
33
|
+
class _StatusEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[Job._Status.ValueType], builtins.type):
|
|
34
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
35
|
+
STATUS_UNSPECIFIED: Job._Status.ValueType # 0
|
|
36
|
+
PROVISIONING: Job._Status.ValueType # 1
|
|
37
|
+
"""Job created and is waiting to acquire."""
|
|
38
|
+
PENDING: Job._Status.ValueType # 2
|
|
39
|
+
"""Job acquired and is waiting for execution."""
|
|
40
|
+
RUNNING: Job._Status.ValueType # 3
|
|
41
|
+
"""Job is running."""
|
|
42
|
+
ERROR: Job._Status.ValueType # 4
|
|
43
|
+
"""Job failed."""
|
|
44
|
+
DONE: Job._Status.ValueType # 5
|
|
45
|
+
"""Job finished."""
|
|
46
|
+
CANCELLED: Job._Status.ValueType # 6
|
|
47
|
+
"""Job cancelled."""
|
|
48
|
+
CANCELLING: Job._Status.ValueType # 7
|
|
49
|
+
"""Job is waiting for cancellation."""
|
|
50
|
+
|
|
51
|
+
class Status(_Status, metaclass=_StatusEnumTypeWrapper): ...
|
|
52
|
+
STATUS_UNSPECIFIED: Job.Status.ValueType # 0
|
|
53
|
+
PROVISIONING: Job.Status.ValueType # 1
|
|
54
|
+
"""Job created and is waiting to acquire."""
|
|
55
|
+
PENDING: Job.Status.ValueType # 2
|
|
56
|
+
"""Job acquired and is waiting for execution."""
|
|
57
|
+
RUNNING: Job.Status.ValueType # 3
|
|
58
|
+
"""Job is running."""
|
|
59
|
+
ERROR: Job.Status.ValueType # 4
|
|
60
|
+
"""Job failed."""
|
|
61
|
+
DONE: Job.Status.ValueType # 5
|
|
62
|
+
"""Job finished."""
|
|
63
|
+
CANCELLED: Job.Status.ValueType # 6
|
|
64
|
+
"""Job cancelled."""
|
|
65
|
+
CANCELLING: Job.Status.ValueType # 7
|
|
66
|
+
"""Job is waiting for cancellation."""
|
|
67
|
+
|
|
68
|
+
ID_FIELD_NUMBER: builtins.int
|
|
69
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
70
|
+
CREATED_AT_FIELD_NUMBER: builtins.int
|
|
71
|
+
STARTED_AT_FIELD_NUMBER: builtins.int
|
|
72
|
+
FINISHED_AT_FIELD_NUMBER: builtins.int
|
|
73
|
+
NAME_FIELD_NUMBER: builtins.int
|
|
74
|
+
CREATED_BY_FIELD_NUMBER: builtins.int
|
|
75
|
+
STATUS_FIELD_NUMBER: builtins.int
|
|
76
|
+
SPARK_JOB_FIELD_NUMBER: builtins.int
|
|
77
|
+
PYSPARK_JOB_FIELD_NUMBER: builtins.int
|
|
78
|
+
APPLICATION_INFO_FIELD_NUMBER: builtins.int
|
|
79
|
+
UI_URL_FIELD_NUMBER: builtins.int
|
|
80
|
+
id: builtins.str
|
|
81
|
+
"""Required. Unique ID of the Spark job.
|
|
82
|
+
This ID is assigned by MDB in the process of creating Spark job.
|
|
83
|
+
"""
|
|
84
|
+
cluster_id: builtins.str
|
|
85
|
+
"""Required. Unique ID of the Spark cluster."""
|
|
86
|
+
name: builtins.str
|
|
87
|
+
"""Name of the Spark job."""
|
|
88
|
+
created_by: builtins.str
|
|
89
|
+
"""The id of the user who created the job"""
|
|
90
|
+
status: global___Job.Status.ValueType
|
|
91
|
+
"""Status."""
|
|
92
|
+
ui_url: builtins.str
|
|
93
|
+
"""Spark UI Url."""
|
|
94
|
+
@property
|
|
95
|
+
def created_at(self) -> google.protobuf.timestamp_pb2.Timestamp:
|
|
96
|
+
"""The time when the Spark job was created."""
|
|
97
|
+
|
|
98
|
+
@property
|
|
99
|
+
def started_at(self) -> google.protobuf.timestamp_pb2.Timestamp:
|
|
100
|
+
"""The time when the Spark job was started."""
|
|
101
|
+
|
|
102
|
+
@property
|
|
103
|
+
def finished_at(self) -> google.protobuf.timestamp_pb2.Timestamp:
|
|
104
|
+
"""The time when the Spark job was finished."""
|
|
105
|
+
|
|
106
|
+
@property
|
|
107
|
+
def spark_job(self) -> global___SparkJob: ...
|
|
108
|
+
@property
|
|
109
|
+
def pyspark_job(self) -> global___PysparkJob: ...
|
|
110
|
+
@property
|
|
111
|
+
def application_info(self) -> global___ApplicationInfo:
|
|
112
|
+
"""Attributes of application."""
|
|
113
|
+
|
|
114
|
+
def __init__(
|
|
115
|
+
self,
|
|
116
|
+
*,
|
|
117
|
+
id: builtins.str = ...,
|
|
118
|
+
cluster_id: builtins.str = ...,
|
|
119
|
+
created_at: google.protobuf.timestamp_pb2.Timestamp | None = ...,
|
|
120
|
+
started_at: google.protobuf.timestamp_pb2.Timestamp | None = ...,
|
|
121
|
+
finished_at: google.protobuf.timestamp_pb2.Timestamp | None = ...,
|
|
122
|
+
name: builtins.str = ...,
|
|
123
|
+
created_by: builtins.str = ...,
|
|
124
|
+
status: global___Job.Status.ValueType = ...,
|
|
125
|
+
spark_job: global___SparkJob | None = ...,
|
|
126
|
+
pyspark_job: global___PysparkJob | None = ...,
|
|
127
|
+
application_info: global___ApplicationInfo | None = ...,
|
|
128
|
+
ui_url: builtins.str = ...,
|
|
129
|
+
) -> None: ...
|
|
130
|
+
def HasField(self, field_name: typing.Literal["application_info", b"application_info", "created_at", b"created_at", "finished_at", b"finished_at", "job_spec", b"job_spec", "pyspark_job", b"pyspark_job", "spark_job", b"spark_job", "started_at", b"started_at"]) -> builtins.bool: ...
|
|
131
|
+
def ClearField(self, field_name: typing.Literal["application_info", b"application_info", "cluster_id", b"cluster_id", "created_at", b"created_at", "created_by", b"created_by", "finished_at", b"finished_at", "id", b"id", "job_spec", b"job_spec", "name", b"name", "pyspark_job", b"pyspark_job", "spark_job", b"spark_job", "started_at", b"started_at", "status", b"status", "ui_url", b"ui_url"]) -> None: ...
|
|
132
|
+
def WhichOneof(self, oneof_group: typing.Literal["job_spec", b"job_spec"]) -> typing.Literal["spark_job", "pyspark_job"] | None: ...
|
|
133
|
+
|
|
134
|
+
global___Job = Job
|
|
135
|
+
|
|
136
|
+
@typing.final
|
|
137
|
+
class ApplicationAttempt(google.protobuf.message.Message):
|
|
138
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
139
|
+
|
|
140
|
+
ID_FIELD_NUMBER: builtins.int
|
|
141
|
+
AM_CONTAINER_ID_FIELD_NUMBER: builtins.int
|
|
142
|
+
id: builtins.str
|
|
143
|
+
"""ID of application attempt"""
|
|
144
|
+
am_container_id: builtins.str
|
|
145
|
+
"""ID of Application Master container"""
|
|
146
|
+
def __init__(
|
|
147
|
+
self,
|
|
148
|
+
*,
|
|
149
|
+
id: builtins.str = ...,
|
|
150
|
+
am_container_id: builtins.str = ...,
|
|
151
|
+
) -> None: ...
|
|
152
|
+
def ClearField(self, field_name: typing.Literal["am_container_id", b"am_container_id", "id", b"id"]) -> None: ...
|
|
153
|
+
|
|
154
|
+
global___ApplicationAttempt = ApplicationAttempt
|
|
155
|
+
|
|
156
|
+
@typing.final
|
|
157
|
+
class ApplicationInfo(google.protobuf.message.Message):
|
|
158
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
159
|
+
|
|
160
|
+
ID_FIELD_NUMBER: builtins.int
|
|
161
|
+
APPLICATION_ATTEMPTS_FIELD_NUMBER: builtins.int
|
|
162
|
+
id: builtins.str
|
|
163
|
+
"""ID of application"""
|
|
164
|
+
@property
|
|
165
|
+
def application_attempts(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ApplicationAttempt]:
|
|
166
|
+
"""Application attempts"""
|
|
167
|
+
|
|
168
|
+
def __init__(
|
|
169
|
+
self,
|
|
170
|
+
*,
|
|
171
|
+
id: builtins.str = ...,
|
|
172
|
+
application_attempts: collections.abc.Iterable[global___ApplicationAttempt] | None = ...,
|
|
173
|
+
) -> None: ...
|
|
174
|
+
def ClearField(self, field_name: typing.Literal["application_attempts", b"application_attempts", "id", b"id"]) -> None: ...
|
|
175
|
+
|
|
176
|
+
global___ApplicationInfo = ApplicationInfo
|
|
177
|
+
|
|
178
|
+
@typing.final
|
|
179
|
+
class SparkJob(google.protobuf.message.Message):
|
|
180
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
181
|
+
|
|
182
|
+
@typing.final
|
|
183
|
+
class PropertiesEntry(google.protobuf.message.Message):
|
|
184
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
185
|
+
|
|
186
|
+
KEY_FIELD_NUMBER: builtins.int
|
|
187
|
+
VALUE_FIELD_NUMBER: builtins.int
|
|
188
|
+
key: builtins.str
|
|
189
|
+
value: builtins.str
|
|
190
|
+
def __init__(
|
|
191
|
+
self,
|
|
192
|
+
*,
|
|
193
|
+
key: builtins.str = ...,
|
|
194
|
+
value: builtins.str = ...,
|
|
195
|
+
) -> None: ...
|
|
196
|
+
def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
|
|
197
|
+
|
|
198
|
+
ARGS_FIELD_NUMBER: builtins.int
|
|
199
|
+
JAR_FILE_URIS_FIELD_NUMBER: builtins.int
|
|
200
|
+
FILE_URIS_FIELD_NUMBER: builtins.int
|
|
201
|
+
ARCHIVE_URIS_FIELD_NUMBER: builtins.int
|
|
202
|
+
PROPERTIES_FIELD_NUMBER: builtins.int
|
|
203
|
+
MAIN_JAR_FILE_URI_FIELD_NUMBER: builtins.int
|
|
204
|
+
MAIN_CLASS_FIELD_NUMBER: builtins.int
|
|
205
|
+
PACKAGES_FIELD_NUMBER: builtins.int
|
|
206
|
+
REPOSITORIES_FIELD_NUMBER: builtins.int
|
|
207
|
+
EXCLUDE_PACKAGES_FIELD_NUMBER: builtins.int
|
|
208
|
+
main_jar_file_uri: builtins.str
|
|
209
|
+
"""The HCFS URI of the jar file containing the main class."""
|
|
210
|
+
main_class: builtins.str
|
|
211
|
+
"""The name of the driver's main class."""
|
|
212
|
+
@property
|
|
213
|
+
def args(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
214
|
+
"""Optional arguments to pass to the driver."""
|
|
215
|
+
|
|
216
|
+
@property
|
|
217
|
+
def jar_file_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
218
|
+
"""Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks."""
|
|
219
|
+
|
|
220
|
+
@property
|
|
221
|
+
def file_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
222
|
+
"""URIs of files to be copied to the working directory of Spark drivers and distributed tasks."""
|
|
223
|
+
|
|
224
|
+
@property
|
|
225
|
+
def archive_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
226
|
+
"""URIs of archives to be extracted in the working directory of Spark drivers and tasks."""
|
|
227
|
+
|
|
228
|
+
@property
|
|
229
|
+
def properties(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.str]:
|
|
230
|
+
"""A mapping of property names to values, used to configure Spark."""
|
|
231
|
+
|
|
232
|
+
@property
|
|
233
|
+
def packages(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
234
|
+
"""List of maven coordinates of jars to include on the driver and executor classpaths."""
|
|
235
|
+
|
|
236
|
+
@property
|
|
237
|
+
def repositories(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
238
|
+
"""List of additional remote repositories to search for the maven coordinates given with --packages."""
|
|
239
|
+
|
|
240
|
+
@property
|
|
241
|
+
def exclude_packages(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
242
|
+
"""List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts."""
|
|
243
|
+
|
|
244
|
+
def __init__(
|
|
245
|
+
self,
|
|
246
|
+
*,
|
|
247
|
+
args: collections.abc.Iterable[builtins.str] | None = ...,
|
|
248
|
+
jar_file_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
249
|
+
file_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
250
|
+
archive_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
251
|
+
properties: collections.abc.Mapping[builtins.str, builtins.str] | None = ...,
|
|
252
|
+
main_jar_file_uri: builtins.str = ...,
|
|
253
|
+
main_class: builtins.str = ...,
|
|
254
|
+
packages: collections.abc.Iterable[builtins.str] | None = ...,
|
|
255
|
+
repositories: collections.abc.Iterable[builtins.str] | None = ...,
|
|
256
|
+
exclude_packages: collections.abc.Iterable[builtins.str] | None = ...,
|
|
257
|
+
) -> None: ...
|
|
258
|
+
def ClearField(self, field_name: typing.Literal["archive_uris", b"archive_uris", "args", b"args", "exclude_packages", b"exclude_packages", "file_uris", b"file_uris", "jar_file_uris", b"jar_file_uris", "main_class", b"main_class", "main_jar_file_uri", b"main_jar_file_uri", "packages", b"packages", "properties", b"properties", "repositories", b"repositories"]) -> None: ...
|
|
259
|
+
|
|
260
|
+
global___SparkJob = SparkJob
|
|
261
|
+
|
|
262
|
+
@typing.final
|
|
263
|
+
class PysparkJob(google.protobuf.message.Message):
|
|
264
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
265
|
+
|
|
266
|
+
@typing.final
|
|
267
|
+
class PropertiesEntry(google.protobuf.message.Message):
|
|
268
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
269
|
+
|
|
270
|
+
KEY_FIELD_NUMBER: builtins.int
|
|
271
|
+
VALUE_FIELD_NUMBER: builtins.int
|
|
272
|
+
key: builtins.str
|
|
273
|
+
value: builtins.str
|
|
274
|
+
def __init__(
|
|
275
|
+
self,
|
|
276
|
+
*,
|
|
277
|
+
key: builtins.str = ...,
|
|
278
|
+
value: builtins.str = ...,
|
|
279
|
+
) -> None: ...
|
|
280
|
+
def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
|
|
281
|
+
|
|
282
|
+
ARGS_FIELD_NUMBER: builtins.int
|
|
283
|
+
JAR_FILE_URIS_FIELD_NUMBER: builtins.int
|
|
284
|
+
FILE_URIS_FIELD_NUMBER: builtins.int
|
|
285
|
+
ARCHIVE_URIS_FIELD_NUMBER: builtins.int
|
|
286
|
+
PROPERTIES_FIELD_NUMBER: builtins.int
|
|
287
|
+
MAIN_PYTHON_FILE_URI_FIELD_NUMBER: builtins.int
|
|
288
|
+
PYTHON_FILE_URIS_FIELD_NUMBER: builtins.int
|
|
289
|
+
PACKAGES_FIELD_NUMBER: builtins.int
|
|
290
|
+
REPOSITORIES_FIELD_NUMBER: builtins.int
|
|
291
|
+
EXCLUDE_PACKAGES_FIELD_NUMBER: builtins.int
|
|
292
|
+
main_python_file_uri: builtins.str
|
|
293
|
+
"""URI of the main Python file to use as the driver. Must be a .py file."""
|
|
294
|
+
@property
|
|
295
|
+
def args(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
296
|
+
"""Optional arguments to pass to the driver."""
|
|
297
|
+
|
|
298
|
+
@property
|
|
299
|
+
def jar_file_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
300
|
+
"""Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks."""
|
|
301
|
+
|
|
302
|
+
@property
|
|
303
|
+
def file_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
304
|
+
"""URIs of files to be copied to the working directory of Spark drivers and distributed tasks."""
|
|
305
|
+
|
|
306
|
+
@property
|
|
307
|
+
def archive_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
308
|
+
"""URIs of archives to be extracted in the working directory of Spark drivers and tasks."""
|
|
309
|
+
|
|
310
|
+
@property
|
|
311
|
+
def properties(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.str]:
|
|
312
|
+
"""A mapping of property names to values, used to configure Spark."""
|
|
313
|
+
|
|
314
|
+
@property
|
|
315
|
+
def python_file_uris(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
316
|
+
"""URIs of Python files to pass to the PySpark framework."""
|
|
317
|
+
|
|
318
|
+
@property
|
|
319
|
+
def packages(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
320
|
+
"""List of maven coordinates of jars to include on the driver and executor classpaths."""
|
|
321
|
+
|
|
322
|
+
@property
|
|
323
|
+
def repositories(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
324
|
+
"""List of additional remote repositories to search for the maven coordinates given with --packages."""
|
|
325
|
+
|
|
326
|
+
@property
|
|
327
|
+
def exclude_packages(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
328
|
+
"""List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts."""
|
|
329
|
+
|
|
330
|
+
def __init__(
|
|
331
|
+
self,
|
|
332
|
+
*,
|
|
333
|
+
args: collections.abc.Iterable[builtins.str] | None = ...,
|
|
334
|
+
jar_file_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
335
|
+
file_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
336
|
+
archive_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
337
|
+
properties: collections.abc.Mapping[builtins.str, builtins.str] | None = ...,
|
|
338
|
+
main_python_file_uri: builtins.str = ...,
|
|
339
|
+
python_file_uris: collections.abc.Iterable[builtins.str] | None = ...,
|
|
340
|
+
packages: collections.abc.Iterable[builtins.str] | None = ...,
|
|
341
|
+
repositories: collections.abc.Iterable[builtins.str] | None = ...,
|
|
342
|
+
exclude_packages: collections.abc.Iterable[builtins.str] | None = ...,
|
|
343
|
+
) -> None: ...
|
|
344
|
+
def ClearField(self, field_name: typing.Literal["archive_uris", b"archive_uris", "args", b"args", "exclude_packages", b"exclude_packages", "file_uris", b"file_uris", "jar_file_uris", b"jar_file_uris", "main_python_file_uri", b"main_python_file_uri", "packages", b"packages", "properties", b"properties", "python_file_uris", b"python_file_uris", "repositories", b"repositories"]) -> None: ...
|
|
345
|
+
|
|
346
|
+
global___PysparkJob = PysparkJob
|
|
@@ -0,0 +1,24 @@
|
|
|
1
|
+
# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT!
|
|
2
|
+
"""Client and server classes corresponding to protobuf-defined services."""
|
|
3
|
+
import grpc
|
|
4
|
+
import warnings
|
|
5
|
+
|
|
6
|
+
|
|
7
|
+
GRPC_GENERATED_VERSION = '1.70.0'
|
|
8
|
+
GRPC_VERSION = grpc.__version__
|
|
9
|
+
_version_not_supported = False
|
|
10
|
+
|
|
11
|
+
try:
|
|
12
|
+
from grpc._utilities import first_version_is_lower
|
|
13
|
+
_version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION)
|
|
14
|
+
except ImportError:
|
|
15
|
+
_version_not_supported = True
|
|
16
|
+
|
|
17
|
+
if _version_not_supported:
|
|
18
|
+
raise RuntimeError(
|
|
19
|
+
f'The grpc package installed is at version {GRPC_VERSION},'
|
|
20
|
+
+ f' but the generated code in yandex/cloud/spark/v1/job_pb2_grpc.py depends on'
|
|
21
|
+
+ f' grpcio>={GRPC_GENERATED_VERSION}.'
|
|
22
|
+
+ f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}'
|
|
23
|
+
+ f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.'
|
|
24
|
+
)
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
"""
|
|
2
|
+
@generated by mypy-protobuf. Do not edit manually!
|
|
3
|
+
isort:skip_file
|
|
4
|
+
"""
|
|
5
|
+
|
|
6
|
+
import abc
|
|
7
|
+
import collections.abc
|
|
8
|
+
import grpc
|
|
9
|
+
import grpc.aio
|
|
10
|
+
import typing
|
|
11
|
+
|
|
12
|
+
_T = typing.TypeVar("_T")
|
|
13
|
+
|
|
14
|
+
class _MaybeAsyncIterator(collections.abc.AsyncIterator[_T], collections.abc.Iterator[_T], metaclass=abc.ABCMeta): ...
|
|
15
|
+
|
|
16
|
+
class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type: ignore[misc, type-arg]
|
|
17
|
+
...
|
|
@@ -0,0 +1,97 @@
|
|
|
1
|
+
# -*- coding: utf-8 -*-
|
|
2
|
+
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
|
3
|
+
# NO CHECKED-IN PROTOBUF GENCODE
|
|
4
|
+
# source: yandex/cloud/spark/v1/job_service.proto
|
|
5
|
+
# Protobuf Python Version: 5.29.0
|
|
6
|
+
"""Generated protocol buffer code."""
|
|
7
|
+
from google.protobuf import descriptor as _descriptor
|
|
8
|
+
from google.protobuf import descriptor_pool as _descriptor_pool
|
|
9
|
+
from google.protobuf import runtime_version as _runtime_version
|
|
10
|
+
from google.protobuf import symbol_database as _symbol_database
|
|
11
|
+
from google.protobuf.internal import builder as _builder
|
|
12
|
+
_runtime_version.ValidateProtobufRuntimeVersion(
|
|
13
|
+
_runtime_version.Domain.PUBLIC,
|
|
14
|
+
5,
|
|
15
|
+
29,
|
|
16
|
+
0,
|
|
17
|
+
'',
|
|
18
|
+
'yandex/cloud/spark/v1/job_service.proto'
|
|
19
|
+
)
|
|
20
|
+
# @@protoc_insertion_point(imports)
|
|
21
|
+
|
|
22
|
+
_sym_db = _symbol_database.Default()
|
|
23
|
+
|
|
24
|
+
|
|
25
|
+
from yandex.cloud.spark.v1 import job_pb2 as yandex_dot_cloud_dot_spark_dot_v1_dot_job__pb2
|
|
26
|
+
from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operation_dot_operation__pb2
|
|
27
|
+
from yandex.cloud import validation_pb2 as yandex_dot_cloud_dot_validation__pb2
|
|
28
|
+
from yandex.cloud.api import operation_pb2 as yandex_dot_cloud_dot_api_dot_operation__pb2
|
|
29
|
+
|
|
30
|
+
|
|
31
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\'yandex/cloud/spark/v1/job_service.proto\x12\x15yandex.cloud.spark.v1\x1a\x1fyandex/cloud/spark/v1/job.proto\x1a&yandex/cloud/operation/operation.proto\x1a\x1dyandex/cloud/validation.proto\x1a yandex/cloud/api/operation.proto\"O\n\rGetJobRequest\x12 \n\ncluster_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12\x1c\n\x06job_id\x18\x02 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\"\x8d\x01\n\x0fListJobsRequest\x12 \n\ncluster_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12\x1d\n\tpage_size\x18\x02 \x01(\x03\x42\n\xfa\xc7\x31\x06<=1000\x12\x1d\n\npage_token\x18\x03 \x01(\tB\t\x8a\xc8\x31\x05<=200\x12\x1a\n\x06\x66ilter\x18\x04 \x01(\tB\n\x8a\xc8\x31\x06<=1000\"`\n\x10ListJobsResponse\x12(\n\x04jobs\x18\x01 \x03(\x0b\x32\x1a.yandex.cloud.spark.v1.Job\x12\"\n\x0fnext_page_token\x18\x02 \x01(\tB\t\x8a\xc8\x31\x05<=200\"\xe1\x01\n\x10\x43reateJobRequest\x12 \n\ncluster_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12/\n\x04name\x18\x02 \x01(\tB!\xf2\xc7\x31\x1d|[a-z][-a-z0-9]{1,61}[a-z0-9]\x12\x34\n\tspark_job\x18\x03 \x01(\x0b\x32\x1f.yandex.cloud.spark.v1.SparkJobH\x00\x12\x38\n\x0bpyspark_job\x18\x04 \x01(\x0b\x32!.yandex.cloud.spark.v1.PysparkJobH\x00\x42\n\n\x08job_spec\"O\n\x11\x43reateJobMetadata\x12 \n\ncluster_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12\x18\n\x06job_id\x18\x02 \x01(\tB\x08\x8a\xc8\x31\x04<=50\"R\n\x10\x43\x61ncelJobRequest\x12 \n\ncluster_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12\x1c\n\x06job_id\x18\x02 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\"\x94\x01\n\x11ListJobLogRequest\x12 \n\ncluster_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12\x1c\n\x06job_id\x18\x02 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=50\x12 \n\tpage_size\x18\x03 \x01(\x03\x42\r\xfa\xc7\x31\t<=1048576\x12\x1d\n\npage_token\x18\x04 \x01(\tB\t\x8a\xc8\x31\x05<=200\"I\n\x12ListJobLogResponse\x12\x0f\n\x07\x63ontent\x18\x01 \x01(\t\x12\"\n\x0fnext_page_token\x18\x02 \x01(\tB\t\x8a\xc8\x31\x05<=2002\xfc\x03\n\nJobService\x12Y\n\x04List\x12&.yandex.cloud.spark.v1.ListJobsRequest\x1a\'.yandex.cloud.spark.v1.ListJobsResponse\"\x00\x12r\n\x06\x43reate\x12\'.yandex.cloud.spark.v1.CreateJobRequest\x1a!.yandex.cloud.operation.Operation\"\x1c\xb2\xd2*\x18\n\x11\x43reateJobMetadata\x12\x03Job\x12I\n\x03Get\x12$.yandex.cloud.spark.v1.GetJobRequest\x1a\x1a.yandex.cloud.spark.v1.Job\"\x00\x12`\n\x07ListLog\x12(.yandex.cloud.spark.v1.ListJobLogRequest\x1a).yandex.cloud.spark.v1.ListJobLogResponse\"\x00\x12r\n\x06\x43\x61ncel\x12\'.yandex.cloud.spark.v1.CancelJobRequest\x1a!.yandex.cloud.operation.Operation\"\x1c\xb2\xd2*\x18\n\x11\x43reateJobMetadata\x12\x03JobB\\\n\x19yandex.cloud.api.spark.v1Z?github.com/yandex-cloud/go-genproto/yandex/cloud/spark/v1;sparkb\x06proto3')
|
|
32
|
+
|
|
33
|
+
_globals = globals()
|
|
34
|
+
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
35
|
+
_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'yandex.cloud.spark.v1.job_service_pb2', _globals)
|
|
36
|
+
if not _descriptor._USE_C_DESCRIPTORS:
|
|
37
|
+
_globals['DESCRIPTOR']._loaded_options = None
|
|
38
|
+
_globals['DESCRIPTOR']._serialized_options = b'\n\031yandex.cloud.api.spark.v1Z?github.com/yandex-cloud/go-genproto/yandex/cloud/spark/v1;spark'
|
|
39
|
+
_globals['_GETJOBREQUEST'].fields_by_name['cluster_id']._loaded_options = None
|
|
40
|
+
_globals['_GETJOBREQUEST'].fields_by_name['cluster_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
41
|
+
_globals['_GETJOBREQUEST'].fields_by_name['job_id']._loaded_options = None
|
|
42
|
+
_globals['_GETJOBREQUEST'].fields_by_name['job_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
43
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['cluster_id']._loaded_options = None
|
|
44
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['cluster_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
45
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['page_size']._loaded_options = None
|
|
46
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['page_size']._serialized_options = b'\372\3071\006<=1000'
|
|
47
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['page_token']._loaded_options = None
|
|
48
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['page_token']._serialized_options = b'\212\3101\005<=200'
|
|
49
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['filter']._loaded_options = None
|
|
50
|
+
_globals['_LISTJOBSREQUEST'].fields_by_name['filter']._serialized_options = b'\212\3101\006<=1000'
|
|
51
|
+
_globals['_LISTJOBSRESPONSE'].fields_by_name['next_page_token']._loaded_options = None
|
|
52
|
+
_globals['_LISTJOBSRESPONSE'].fields_by_name['next_page_token']._serialized_options = b'\212\3101\005<=200'
|
|
53
|
+
_globals['_CREATEJOBREQUEST'].fields_by_name['cluster_id']._loaded_options = None
|
|
54
|
+
_globals['_CREATEJOBREQUEST'].fields_by_name['cluster_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
55
|
+
_globals['_CREATEJOBREQUEST'].fields_by_name['name']._loaded_options = None
|
|
56
|
+
_globals['_CREATEJOBREQUEST'].fields_by_name['name']._serialized_options = b'\362\3071\035|[a-z][-a-z0-9]{1,61}[a-z0-9]'
|
|
57
|
+
_globals['_CREATEJOBMETADATA'].fields_by_name['cluster_id']._loaded_options = None
|
|
58
|
+
_globals['_CREATEJOBMETADATA'].fields_by_name['cluster_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
59
|
+
_globals['_CREATEJOBMETADATA'].fields_by_name['job_id']._loaded_options = None
|
|
60
|
+
_globals['_CREATEJOBMETADATA'].fields_by_name['job_id']._serialized_options = b'\212\3101\004<=50'
|
|
61
|
+
_globals['_CANCELJOBREQUEST'].fields_by_name['cluster_id']._loaded_options = None
|
|
62
|
+
_globals['_CANCELJOBREQUEST'].fields_by_name['cluster_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
63
|
+
_globals['_CANCELJOBREQUEST'].fields_by_name['job_id']._loaded_options = None
|
|
64
|
+
_globals['_CANCELJOBREQUEST'].fields_by_name['job_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
65
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['cluster_id']._loaded_options = None
|
|
66
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['cluster_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
67
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['job_id']._loaded_options = None
|
|
68
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['job_id']._serialized_options = b'\350\3071\001\212\3101\004<=50'
|
|
69
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['page_size']._loaded_options = None
|
|
70
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['page_size']._serialized_options = b'\372\3071\t<=1048576'
|
|
71
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['page_token']._loaded_options = None
|
|
72
|
+
_globals['_LISTJOBLOGREQUEST'].fields_by_name['page_token']._serialized_options = b'\212\3101\005<=200'
|
|
73
|
+
_globals['_LISTJOBLOGRESPONSE'].fields_by_name['next_page_token']._loaded_options = None
|
|
74
|
+
_globals['_LISTJOBLOGRESPONSE'].fields_by_name['next_page_token']._serialized_options = b'\212\3101\005<=200'
|
|
75
|
+
_globals['_JOBSERVICE'].methods_by_name['Create']._loaded_options = None
|
|
76
|
+
_globals['_JOBSERVICE'].methods_by_name['Create']._serialized_options = b'\262\322*\030\n\021CreateJobMetadata\022\003Job'
|
|
77
|
+
_globals['_JOBSERVICE'].methods_by_name['Cancel']._loaded_options = None
|
|
78
|
+
_globals['_JOBSERVICE'].methods_by_name['Cancel']._serialized_options = b'\262\322*\030\n\021CreateJobMetadata\022\003Job'
|
|
79
|
+
_globals['_GETJOBREQUEST']._serialized_start=204
|
|
80
|
+
_globals['_GETJOBREQUEST']._serialized_end=283
|
|
81
|
+
_globals['_LISTJOBSREQUEST']._serialized_start=286
|
|
82
|
+
_globals['_LISTJOBSREQUEST']._serialized_end=427
|
|
83
|
+
_globals['_LISTJOBSRESPONSE']._serialized_start=429
|
|
84
|
+
_globals['_LISTJOBSRESPONSE']._serialized_end=525
|
|
85
|
+
_globals['_CREATEJOBREQUEST']._serialized_start=528
|
|
86
|
+
_globals['_CREATEJOBREQUEST']._serialized_end=753
|
|
87
|
+
_globals['_CREATEJOBMETADATA']._serialized_start=755
|
|
88
|
+
_globals['_CREATEJOBMETADATA']._serialized_end=834
|
|
89
|
+
_globals['_CANCELJOBREQUEST']._serialized_start=836
|
|
90
|
+
_globals['_CANCELJOBREQUEST']._serialized_end=918
|
|
91
|
+
_globals['_LISTJOBLOGREQUEST']._serialized_start=921
|
|
92
|
+
_globals['_LISTJOBLOGREQUEST']._serialized_end=1069
|
|
93
|
+
_globals['_LISTJOBLOGRESPONSE']._serialized_start=1071
|
|
94
|
+
_globals['_LISTJOBLOGRESPONSE']._serialized_end=1144
|
|
95
|
+
_globals['_JOBSERVICE']._serialized_start=1147
|
|
96
|
+
_globals['_JOBSERVICE']._serialized_end=1655
|
|
97
|
+
# @@protoc_insertion_point(module_scope)
|