yandexcloud 0.338.0__py3-none-any.whl → 0.340.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- yandex/cloud/ai/assistants/v1/common_pb2.py +32 -27
- yandex/cloud/ai/assistants/v1/common_pb2.pyi +29 -2
- yandex/cloud/ai/dataset/v1/dataset_service_pb2.py +32 -32
- yandex/cloud/ai/dataset/v1/dataset_service_pb2.pyi +7 -1
- yandex/cloud/ai/tuning/v1/tuning_service_pb2.py +39 -3
- yandex/cloud/ai/tuning/v1/tuning_service_pb2.pyi +220 -0
- yandex/cloud/ai/tuning/v1/tuning_service_pb2_grpc.py +172 -0
- yandex/cloud/ai/tuning/v1/tuning_service_pb2_grpc.pyi +68 -0
- yandex/cloud/ai/tuning/v1/tuning_task_pb2.py +3 -3
- yandex/cloud/ai/tuning/v1/tuning_task_pb2.pyi +2 -0
- yandex/cloud/apploadbalancer/v1/load_balancer_pb2.py +50 -50
- yandex/cloud/apploadbalancer/v1/load_balancer_pb2.pyi +10 -2
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2.py +86 -66
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2.pyi +98 -2
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2_grpc.py +88 -0
- yandex/cloud/apploadbalancer/v1/load_balancer_service_pb2_grpc.pyi +40 -0
- yandex/cloud/backup/v1/policy_pb2.py +56 -52
- yandex/cloud/backup/v1/policy_pb2.pyi +36 -5
- yandex/cloud/backup/v1/resource_pb2.py +21 -19
- yandex/cloud/backup/v1/resource_pb2.pyi +32 -2
- yandex/cloud/backup/v1/resource_service_pb2.py +30 -30
- yandex/cloud/backup/v1/resource_service_pb2.pyi +5 -1
- yandex/cloud/baremetal/__init__.py +0 -0
- yandex/cloud/baremetal/v1alpha/__init__.py +0 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2.py +42 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2.pyi +109 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/configuration_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2.py +49 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2.pyi +111 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2_grpc.py +148 -0
- yandex/cloud/baremetal/v1alpha/configuration_service_pb2_grpc.pyi +80 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2.py +37 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2.pyi +33 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/dhcp_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2.py +39 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2.pyi +68 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/disk_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2.py +37 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2.pyi +36 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2.py +49 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2.pyi +87 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2_grpc.py +148 -0
- yandex/cloud/baremetal/v1alpha/hardware_pool_service_pb2_grpc.pyi +80 -0
- yandex/cloud/baremetal/v1alpha/image_pb2.py +44 -0
- yandex/cloud/baremetal/v1alpha/image_pb2.pyi +125 -0
- yandex/cloud/baremetal/v1alpha/image_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/image_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2.py +108 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2.pyi +364 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/image_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2.py +47 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2.pyi +174 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2.py +112 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2.pyi +443 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/private_subnet_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2.py +45 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2.pyi +143 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2.py +106 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2.pyi +382 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/public_subnet_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/server_pb2.py +54 -0
- yandex/cloud/baremetal/v1alpha/server_pb2.pyi +282 -0
- yandex/cloud/baremetal/v1alpha/server_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/server_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2.py +217 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2.pyi +850 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2_grpc.py +589 -0
- yandex/cloud/baremetal/v1alpha/server_service_pb2_grpc.pyi +281 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2.py +46 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2.pyi +165 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/storage_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2.py +51 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2.pyi +91 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2_grpc.py +145 -0
- yandex/cloud/baremetal/v1alpha/storage_service_pb2_grpc.pyi +70 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2.py +42 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2.pyi +77 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/vrf_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2.py +106 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2.pyi +357 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2_grpc.py +327 -0
- yandex/cloud/baremetal/v1alpha/vrf_service_pb2_grpc.pyi +170 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2.py +37 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2.pyi +29 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2_grpc.py +24 -0
- yandex/cloud/baremetal/v1alpha/zone_pb2_grpc.pyi +17 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2.py +49 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2.pyi +87 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2_grpc.py +148 -0
- yandex/cloud/baremetal/v1alpha/zone_service_pb2_grpc.pyi +80 -0
- yandex/cloud/cloudregistry/v1/registry_service_pb2.py +40 -40
- yandex/cloud/dataproc/manager/v1/job_pb2.pyi +26 -26
- yandex/cloud/dataproc/manager/v1/job_service_pb2.pyi +7 -7
- yandex/cloud/dataproc/manager/v1/job_service_pb2_grpc.py +2 -2
- yandex/cloud/dataproc/manager/v1/job_service_pb2_grpc.pyi +6 -6
- yandex/cloud/dataproc/manager/v1/manager_service_pb2.pyi +2 -2
- yandex/cloud/dataproc/manager/v1/manager_service_pb2_grpc.py +3 -3
- yandex/cloud/dataproc/manager/v1/manager_service_pb2_grpc.pyi +3 -3
- yandex/cloud/dataproc/v1/cluster_pb2.pyi +5 -5
- yandex/cloud/dataproc/v1/cluster_service_pb2.pyi +13 -13
- yandex/cloud/dataproc/v1/cluster_service_pb2_grpc.py +4 -4
- yandex/cloud/dataproc/v1/cluster_service_pb2_grpc.pyi +6 -6
- yandex/cloud/dataproc/v1/job_pb2.pyi +15 -15
- yandex/cloud/dataproc/v1/job_service_pb2.pyi +3 -3
- yandex/cloud/dataproc/v1/job_service_pb2_grpc.py +4 -4
- yandex/cloud/dataproc/v1/job_service_pb2_grpc.pyi +6 -6
- yandex/cloud/dataproc/v1/resource_preset_pb2.pyi +2 -2
- yandex/cloud/dataproc/v1/subcluster_pb2.pyi +11 -11
- yandex/cloud/dataproc/v1/subcluster_service_pb2.pyi +3 -3
- yandex/cloud/dataproc/v1/subcluster_service_pb2_grpc.py +3 -3
- yandex/cloud/dataproc/v1/subcluster_service_pb2_grpc.pyi +3 -3
- yandex/cloud/iam/v1/api_key_pb2.pyi +3 -1
- yandex/cloud/iam/v1/api_key_service_pb2.pyi +3 -1
- yandex/cloud/iam/v1/workload/oidc/federation_service_pb2.py +32 -25
- yandex/cloud/iam/v1/workload/oidc/federation_service_pb2_grpc.py +133 -0
- yandex/cloud/iam/v1/workload/oidc/federation_service_pb2_grpc.pyi +61 -0
- yandex/cloud/k8s/v1/cluster_pb2.py +40 -30
- yandex/cloud/k8s/v1/cluster_pb2.pyi +94 -2
- yandex/cloud/k8s/v1/cluster_service_pb2.py +54 -42
- yandex/cloud/k8s/v1/cluster_service_pb2.pyi +70 -4
- yandex/cloud/k8s/v1/resource_preset_pb2.py +37 -0
- yandex/cloud/k8s/v1/resource_preset_pb2.pyi +43 -0
- yandex/cloud/k8s/v1/resource_preset_pb2_grpc.py +24 -0
- yandex/cloud/k8s/v1/resource_preset_pb2_grpc.pyi +17 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2.py +53 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2.pyi +81 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2_grpc.py +146 -0
- yandex/cloud/k8s/v1/resource_preset_service_pb2_grpc.pyi +71 -0
- yandex/cloud/mdb/mysql/v1/backup_pb2.py +6 -6
- yandex/cloud/mdb/mysql/v1/backup_pb2.pyi +5 -1
- yandex/cloud/mdb/mysql/v1/config/mysql5_7_pb2.py +26 -20
- yandex/cloud/mdb/mysql/v1/config/mysql5_7_pb2.pyi +73 -2
- yandex/cloud/mdb/mysql/v1/config/mysql8_0_pb2.py +26 -20
- yandex/cloud/mdb/mysql/v1/config/mysql8_0_pb2.pyi +100 -2
- yandex/cloud/mdb/opensearch/v1/backup_pb2.py +30 -3
- yandex/cloud/mdb/opensearch/v1/backup_pb2.pyi +162 -0
- yandex/cloud/mdb/opensearch/v1/cluster_pb2.py +46 -45
- yandex/cloud/mdb/opensearch/v1/cluster_pb2.pyi +9 -2
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2.py +78 -78
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2.pyi +16 -4
- yandex/cloud/mdb/postgresql/v1/config/host10_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host10_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host11_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host11_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host12_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host12_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host13_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host13_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host14_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host14_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host15_1c_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host15_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/host16_1c_pb2.py +25 -25
- yandex/cloud/mdb/postgresql/v1/config/host16_pb2.py +25 -25
- yandex/cloud/mdb/postgresql/v1/config/host9_6_pb2.py +23 -23
- yandex/cloud/mdb/postgresql/v1/config/postgresql10_1c_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql10_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql11_1c_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql11_pb2.py +35 -35
- yandex/cloud/mdb/postgresql/v1/config/postgresql12_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql12_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql13_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql13_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql14_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql14_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql15_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql15_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql16_1c_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql16_pb2.py +37 -37
- yandex/cloud/mdb/postgresql/v1/config/postgresql9_6_pb2.py +29 -29
- yandex/cloud/monitoring/v3/unit_format_pb2.py +2 -2
- yandex/cloud/monitoring/v3/unit_format_pb2.pyi +16 -0
- yandex/cloud/searchapi/v2/gen_search_service_pb2.py +28 -22
- yandex/cloud/spark/__init__.py +0 -0
- yandex/cloud/spark/v1/__init__.py +0 -0
- yandex/cloud/spark/v1/cluster_pb2.py +100 -0
- yandex/cloud/spark/v1/cluster_pb2.pyi +444 -0
- yandex/cloud/spark/v1/cluster_pb2_grpc.py +24 -0
- yandex/cloud/spark/v1/cluster_pb2_grpc.pyi +17 -0
- yandex/cloud/spark/v1/cluster_service_pb2.py +124 -0
- yandex/cloud/spark/v1/cluster_service_pb2.pyi +333 -0
- yandex/cloud/spark/v1/cluster_service_pb2_grpc.py +366 -0
- yandex/cloud/spark/v1/cluster_service_pb2_grpc.pyi +169 -0
- yandex/cloud/spark/v1/job_pb2.py +56 -0
- yandex/cloud/spark/v1/job_pb2.pyi +346 -0
- yandex/cloud/spark/v1/job_pb2_grpc.py +24 -0
- yandex/cloud/spark/v1/job_pb2_grpc.pyi +17 -0
- yandex/cloud/spark/v1/job_service_pb2.py +97 -0
- yandex/cloud/spark/v1/job_service_pb2.pyi +224 -0
- yandex/cloud/spark/v1/job_service_pb2_grpc.py +279 -0
- yandex/cloud/spark/v1/job_service_pb2_grpc.pyi +132 -0
- yandex/cloud/spark/v1/maintenance_pb2.py +53 -0
- yandex/cloud/spark/v1/maintenance_pb2.pyi +123 -0
- yandex/cloud/spark/v1/maintenance_pb2_grpc.py +24 -0
- yandex/cloud/spark/v1/maintenance_pb2_grpc.pyi +17 -0
- yandexcloud/__init__.py +1 -1
- yandexcloud/_sdk.py +1 -0
- yandexcloud/_wrappers/__init__.py +15 -0
- yandexcloud/_wrappers/spark/__init__.py +603 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/METADATA +2 -1
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/RECORD +219 -106
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/AUTHORS +0 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/LICENSE +0 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/WHEEL +0 -0
- {yandexcloud-0.338.0.dist-info → yandexcloud-0.340.0.dist-info}/top_level.txt +0 -0
|
@@ -0,0 +1,224 @@
|
|
|
1
|
+
"""
|
|
2
|
+
@generated by mypy-protobuf. Do not edit manually!
|
|
3
|
+
isort:skip_file
|
|
4
|
+
"""
|
|
5
|
+
|
|
6
|
+
import builtins
|
|
7
|
+
import collections.abc
|
|
8
|
+
import google.protobuf.descriptor
|
|
9
|
+
import google.protobuf.internal.containers
|
|
10
|
+
import google.protobuf.message
|
|
11
|
+
import typing
|
|
12
|
+
import yandex.cloud.spark.v1.job_pb2
|
|
13
|
+
|
|
14
|
+
DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
15
|
+
|
|
16
|
+
@typing.final
|
|
17
|
+
class GetJobRequest(google.protobuf.message.Message):
|
|
18
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
19
|
+
|
|
20
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
21
|
+
JOB_ID_FIELD_NUMBER: builtins.int
|
|
22
|
+
cluster_id: builtins.str
|
|
23
|
+
"""ID of the Spark cluster."""
|
|
24
|
+
job_id: builtins.str
|
|
25
|
+
"""ID of the Spark job to return."""
|
|
26
|
+
def __init__(
|
|
27
|
+
self,
|
|
28
|
+
*,
|
|
29
|
+
cluster_id: builtins.str = ...,
|
|
30
|
+
job_id: builtins.str = ...,
|
|
31
|
+
) -> None: ...
|
|
32
|
+
def ClearField(self, field_name: typing.Literal["cluster_id", b"cluster_id", "job_id", b"job_id"]) -> None: ...
|
|
33
|
+
|
|
34
|
+
global___GetJobRequest = GetJobRequest
|
|
35
|
+
|
|
36
|
+
@typing.final
|
|
37
|
+
class ListJobsRequest(google.protobuf.message.Message):
|
|
38
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
39
|
+
|
|
40
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
41
|
+
PAGE_SIZE_FIELD_NUMBER: builtins.int
|
|
42
|
+
PAGE_TOKEN_FIELD_NUMBER: builtins.int
|
|
43
|
+
FILTER_FIELD_NUMBER: builtins.int
|
|
44
|
+
cluster_id: builtins.str
|
|
45
|
+
"""ID of the cluster to list Spark jobs of."""
|
|
46
|
+
page_size: builtins.int
|
|
47
|
+
"""The maximum number of results per page that should be returned. If the number of available
|
|
48
|
+
results is larger than `page_size`, the service returns a `next_page_token` that can be used
|
|
49
|
+
to get the next page of results in subsequent ListJobs requests.
|
|
50
|
+
Acceptable values are 0 to 1000, inclusive. Default value: 100.
|
|
51
|
+
"""
|
|
52
|
+
page_token: builtins.str
|
|
53
|
+
"""Page token. Set `page_token` to the `next_page_token` returned by a previous ListJobs
|
|
54
|
+
request to get the next page of results.
|
|
55
|
+
"""
|
|
56
|
+
filter: builtins.str
|
|
57
|
+
"""String that describes a display filter."""
|
|
58
|
+
def __init__(
|
|
59
|
+
self,
|
|
60
|
+
*,
|
|
61
|
+
cluster_id: builtins.str = ...,
|
|
62
|
+
page_size: builtins.int = ...,
|
|
63
|
+
page_token: builtins.str = ...,
|
|
64
|
+
filter: builtins.str = ...,
|
|
65
|
+
) -> None: ...
|
|
66
|
+
def ClearField(self, field_name: typing.Literal["cluster_id", b"cluster_id", "filter", b"filter", "page_size", b"page_size", "page_token", b"page_token"]) -> None: ...
|
|
67
|
+
|
|
68
|
+
global___ListJobsRequest = ListJobsRequest
|
|
69
|
+
|
|
70
|
+
@typing.final
|
|
71
|
+
class ListJobsResponse(google.protobuf.message.Message):
|
|
72
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
73
|
+
|
|
74
|
+
JOBS_FIELD_NUMBER: builtins.int
|
|
75
|
+
NEXT_PAGE_TOKEN_FIELD_NUMBER: builtins.int
|
|
76
|
+
next_page_token: builtins.str
|
|
77
|
+
"""This token allows you to get the next page of results for ListJobs requests,
|
|
78
|
+
if the number of results is larger than `page_size` specified in the request.
|
|
79
|
+
To get the next page, specify the value of `next_page_token` as a value for
|
|
80
|
+
the `page_token` parameter in the next ListClusters request. Subsequent ListClusters
|
|
81
|
+
requests will have their own `next_page_token` to continue paging through the results.
|
|
82
|
+
"""
|
|
83
|
+
@property
|
|
84
|
+
def jobs(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.spark.v1.job_pb2.Job]:
|
|
85
|
+
"""Requested list of Spark jobs."""
|
|
86
|
+
|
|
87
|
+
def __init__(
|
|
88
|
+
self,
|
|
89
|
+
*,
|
|
90
|
+
jobs: collections.abc.Iterable[yandex.cloud.spark.v1.job_pb2.Job] | None = ...,
|
|
91
|
+
next_page_token: builtins.str = ...,
|
|
92
|
+
) -> None: ...
|
|
93
|
+
def ClearField(self, field_name: typing.Literal["jobs", b"jobs", "next_page_token", b"next_page_token"]) -> None: ...
|
|
94
|
+
|
|
95
|
+
global___ListJobsResponse = ListJobsResponse
|
|
96
|
+
|
|
97
|
+
@typing.final
|
|
98
|
+
class CreateJobRequest(google.protobuf.message.Message):
|
|
99
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
100
|
+
|
|
101
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
102
|
+
NAME_FIELD_NUMBER: builtins.int
|
|
103
|
+
SPARK_JOB_FIELD_NUMBER: builtins.int
|
|
104
|
+
PYSPARK_JOB_FIELD_NUMBER: builtins.int
|
|
105
|
+
cluster_id: builtins.str
|
|
106
|
+
"""ID of the cluster to create Spark job in."""
|
|
107
|
+
name: builtins.str
|
|
108
|
+
"""Optional. Name of the job."""
|
|
109
|
+
@property
|
|
110
|
+
def spark_job(self) -> yandex.cloud.spark.v1.job_pb2.SparkJob: ...
|
|
111
|
+
@property
|
|
112
|
+
def pyspark_job(self) -> yandex.cloud.spark.v1.job_pb2.PysparkJob: ...
|
|
113
|
+
def __init__(
|
|
114
|
+
self,
|
|
115
|
+
*,
|
|
116
|
+
cluster_id: builtins.str = ...,
|
|
117
|
+
name: builtins.str = ...,
|
|
118
|
+
spark_job: yandex.cloud.spark.v1.job_pb2.SparkJob | None = ...,
|
|
119
|
+
pyspark_job: yandex.cloud.spark.v1.job_pb2.PysparkJob | None = ...,
|
|
120
|
+
) -> None: ...
|
|
121
|
+
def HasField(self, field_name: typing.Literal["job_spec", b"job_spec", "pyspark_job", b"pyspark_job", "spark_job", b"spark_job"]) -> builtins.bool: ...
|
|
122
|
+
def ClearField(self, field_name: typing.Literal["cluster_id", b"cluster_id", "job_spec", b"job_spec", "name", b"name", "pyspark_job", b"pyspark_job", "spark_job", b"spark_job"]) -> None: ...
|
|
123
|
+
def WhichOneof(self, oneof_group: typing.Literal["job_spec", b"job_spec"]) -> typing.Literal["spark_job", "pyspark_job"] | None: ...
|
|
124
|
+
|
|
125
|
+
global___CreateJobRequest = CreateJobRequest
|
|
126
|
+
|
|
127
|
+
@typing.final
|
|
128
|
+
class CreateJobMetadata(google.protobuf.message.Message):
|
|
129
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
130
|
+
|
|
131
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
132
|
+
JOB_ID_FIELD_NUMBER: builtins.int
|
|
133
|
+
cluster_id: builtins.str
|
|
134
|
+
"""ID of the Spark cluster."""
|
|
135
|
+
job_id: builtins.str
|
|
136
|
+
"""ID of the Spark job."""
|
|
137
|
+
def __init__(
|
|
138
|
+
self,
|
|
139
|
+
*,
|
|
140
|
+
cluster_id: builtins.str = ...,
|
|
141
|
+
job_id: builtins.str = ...,
|
|
142
|
+
) -> None: ...
|
|
143
|
+
def ClearField(self, field_name: typing.Literal["cluster_id", b"cluster_id", "job_id", b"job_id"]) -> None: ...
|
|
144
|
+
|
|
145
|
+
global___CreateJobMetadata = CreateJobMetadata
|
|
146
|
+
|
|
147
|
+
@typing.final
|
|
148
|
+
class CancelJobRequest(google.protobuf.message.Message):
|
|
149
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
150
|
+
|
|
151
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
152
|
+
JOB_ID_FIELD_NUMBER: builtins.int
|
|
153
|
+
cluster_id: builtins.str
|
|
154
|
+
"""ID of the Spark cluster."""
|
|
155
|
+
job_id: builtins.str
|
|
156
|
+
"""ID of the Spark job to cancel."""
|
|
157
|
+
def __init__(
|
|
158
|
+
self,
|
|
159
|
+
*,
|
|
160
|
+
cluster_id: builtins.str = ...,
|
|
161
|
+
job_id: builtins.str = ...,
|
|
162
|
+
) -> None: ...
|
|
163
|
+
def ClearField(self, field_name: typing.Literal["cluster_id", b"cluster_id", "job_id", b"job_id"]) -> None: ...
|
|
164
|
+
|
|
165
|
+
global___CancelJobRequest = CancelJobRequest
|
|
166
|
+
|
|
167
|
+
@typing.final
|
|
168
|
+
class ListJobLogRequest(google.protobuf.message.Message):
|
|
169
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
170
|
+
|
|
171
|
+
CLUSTER_ID_FIELD_NUMBER: builtins.int
|
|
172
|
+
JOB_ID_FIELD_NUMBER: builtins.int
|
|
173
|
+
PAGE_SIZE_FIELD_NUMBER: builtins.int
|
|
174
|
+
PAGE_TOKEN_FIELD_NUMBER: builtins.int
|
|
175
|
+
cluster_id: builtins.str
|
|
176
|
+
"""ID of the Spark cluster."""
|
|
177
|
+
job_id: builtins.str
|
|
178
|
+
"""ID of the Spark job to return."""
|
|
179
|
+
page_size: builtins.int
|
|
180
|
+
"""The maximum length of job output per papge that should be returned.
|
|
181
|
+
If the number of available output is larger tha `page_size`, the service returns
|
|
182
|
+
a `next_page_token` that can be used to get the next page of job output in subsequent ListLog requests.
|
|
183
|
+
Acceptable values are 0 to 1048576. Default value: 1048576.
|
|
184
|
+
"""
|
|
185
|
+
page_token: builtins.str
|
|
186
|
+
"""Page token. Set `page_token` to the `next_page_token` returned by a previous ListLog
|
|
187
|
+
request to get the next page of results.
|
|
188
|
+
"""
|
|
189
|
+
def __init__(
|
|
190
|
+
self,
|
|
191
|
+
*,
|
|
192
|
+
cluster_id: builtins.str = ...,
|
|
193
|
+
job_id: builtins.str = ...,
|
|
194
|
+
page_size: builtins.int = ...,
|
|
195
|
+
page_token: builtins.str = ...,
|
|
196
|
+
) -> None: ...
|
|
197
|
+
def ClearField(self, field_name: typing.Literal["cluster_id", b"cluster_id", "job_id", b"job_id", "page_size", b"page_size", "page_token", b"page_token"]) -> None: ...
|
|
198
|
+
|
|
199
|
+
global___ListJobLogRequest = ListJobLogRequest
|
|
200
|
+
|
|
201
|
+
@typing.final
|
|
202
|
+
class ListJobLogResponse(google.protobuf.message.Message):
|
|
203
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
204
|
+
|
|
205
|
+
CONTENT_FIELD_NUMBER: builtins.int
|
|
206
|
+
NEXT_PAGE_TOKEN_FIELD_NUMBER: builtins.int
|
|
207
|
+
content: builtins.str
|
|
208
|
+
"""Requested part of Spark Job log."""
|
|
209
|
+
next_page_token: builtins.str
|
|
210
|
+
"""This token allows you to get the next page of results for ListLog requests,
|
|
211
|
+
if the number of results is larger than `page_size` specified in the request.
|
|
212
|
+
To get the next page, specify the value of `next_page_token` as a value for
|
|
213
|
+
the `page_token` parameter in the next ListLog request. Subsequent ListLog
|
|
214
|
+
requests will have their own `next_page_token` to continue paging through the results.
|
|
215
|
+
"""
|
|
216
|
+
def __init__(
|
|
217
|
+
self,
|
|
218
|
+
*,
|
|
219
|
+
content: builtins.str = ...,
|
|
220
|
+
next_page_token: builtins.str = ...,
|
|
221
|
+
) -> None: ...
|
|
222
|
+
def ClearField(self, field_name: typing.Literal["content", b"content", "next_page_token", b"next_page_token"]) -> None: ...
|
|
223
|
+
|
|
224
|
+
global___ListJobLogResponse = ListJobLogResponse
|
|
@@ -0,0 +1,279 @@
|
|
|
1
|
+
# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT!
|
|
2
|
+
"""Client and server classes corresponding to protobuf-defined services."""
|
|
3
|
+
import grpc
|
|
4
|
+
import warnings
|
|
5
|
+
|
|
6
|
+
from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operation_dot_operation__pb2
|
|
7
|
+
from yandex.cloud.spark.v1 import job_pb2 as yandex_dot_cloud_dot_spark_dot_v1_dot_job__pb2
|
|
8
|
+
from yandex.cloud.spark.v1 import job_service_pb2 as yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2
|
|
9
|
+
|
|
10
|
+
GRPC_GENERATED_VERSION = '1.70.0'
|
|
11
|
+
GRPC_VERSION = grpc.__version__
|
|
12
|
+
_version_not_supported = False
|
|
13
|
+
|
|
14
|
+
try:
|
|
15
|
+
from grpc._utilities import first_version_is_lower
|
|
16
|
+
_version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION)
|
|
17
|
+
except ImportError:
|
|
18
|
+
_version_not_supported = True
|
|
19
|
+
|
|
20
|
+
if _version_not_supported:
|
|
21
|
+
raise RuntimeError(
|
|
22
|
+
f'The grpc package installed is at version {GRPC_VERSION},'
|
|
23
|
+
+ f' but the generated code in yandex/cloud/spark/v1/job_service_pb2_grpc.py depends on'
|
|
24
|
+
+ f' grpcio>={GRPC_GENERATED_VERSION}.'
|
|
25
|
+
+ f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}'
|
|
26
|
+
+ f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.'
|
|
27
|
+
)
|
|
28
|
+
|
|
29
|
+
|
|
30
|
+
class JobServiceStub(object):
|
|
31
|
+
"""A set of methods for managing jobs for Spark cluster.
|
|
32
|
+
"""
|
|
33
|
+
|
|
34
|
+
def __init__(self, channel):
|
|
35
|
+
"""Constructor.
|
|
36
|
+
|
|
37
|
+
Args:
|
|
38
|
+
channel: A grpc.Channel.
|
|
39
|
+
"""
|
|
40
|
+
self.List = channel.unary_unary(
|
|
41
|
+
'/yandex.cloud.spark.v1.JobService/List',
|
|
42
|
+
request_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobsRequest.SerializeToString,
|
|
43
|
+
response_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobsResponse.FromString,
|
|
44
|
+
_registered_method=True)
|
|
45
|
+
self.Create = channel.unary_unary(
|
|
46
|
+
'/yandex.cloud.spark.v1.JobService/Create',
|
|
47
|
+
request_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.CreateJobRequest.SerializeToString,
|
|
48
|
+
response_deserializer=yandex_dot_cloud_dot_operation_dot_operation__pb2.Operation.FromString,
|
|
49
|
+
_registered_method=True)
|
|
50
|
+
self.Get = channel.unary_unary(
|
|
51
|
+
'/yandex.cloud.spark.v1.JobService/Get',
|
|
52
|
+
request_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.GetJobRequest.SerializeToString,
|
|
53
|
+
response_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__pb2.Job.FromString,
|
|
54
|
+
_registered_method=True)
|
|
55
|
+
self.ListLog = channel.unary_unary(
|
|
56
|
+
'/yandex.cloud.spark.v1.JobService/ListLog',
|
|
57
|
+
request_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobLogRequest.SerializeToString,
|
|
58
|
+
response_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobLogResponse.FromString,
|
|
59
|
+
_registered_method=True)
|
|
60
|
+
self.Cancel = channel.unary_unary(
|
|
61
|
+
'/yandex.cloud.spark.v1.JobService/Cancel',
|
|
62
|
+
request_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.CancelJobRequest.SerializeToString,
|
|
63
|
+
response_deserializer=yandex_dot_cloud_dot_operation_dot_operation__pb2.Operation.FromString,
|
|
64
|
+
_registered_method=True)
|
|
65
|
+
|
|
66
|
+
|
|
67
|
+
class JobServiceServicer(object):
|
|
68
|
+
"""A set of methods for managing jobs for Spark cluster.
|
|
69
|
+
"""
|
|
70
|
+
|
|
71
|
+
def List(self, request, context):
|
|
72
|
+
"""Retrieves a list of jobs for Spark cluster.
|
|
73
|
+
"""
|
|
74
|
+
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
75
|
+
context.set_details('Method not implemented!')
|
|
76
|
+
raise NotImplementedError('Method not implemented!')
|
|
77
|
+
|
|
78
|
+
def Create(self, request, context):
|
|
79
|
+
"""Creates a job for Spark cluster.
|
|
80
|
+
"""
|
|
81
|
+
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
82
|
+
context.set_details('Method not implemented!')
|
|
83
|
+
raise NotImplementedError('Method not implemented!')
|
|
84
|
+
|
|
85
|
+
def Get(self, request, context):
|
|
86
|
+
"""Returns the specified Spark job.
|
|
87
|
+
"""
|
|
88
|
+
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
89
|
+
context.set_details('Method not implemented!')
|
|
90
|
+
raise NotImplementedError('Method not implemented!')
|
|
91
|
+
|
|
92
|
+
def ListLog(self, request, context):
|
|
93
|
+
"""Returns a log for Spark job.
|
|
94
|
+
"""
|
|
95
|
+
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
96
|
+
context.set_details('Method not implemented!')
|
|
97
|
+
raise NotImplementedError('Method not implemented!')
|
|
98
|
+
|
|
99
|
+
def Cancel(self, request, context):
|
|
100
|
+
"""Cancels the specified Spark job.
|
|
101
|
+
"""
|
|
102
|
+
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
103
|
+
context.set_details('Method not implemented!')
|
|
104
|
+
raise NotImplementedError('Method not implemented!')
|
|
105
|
+
|
|
106
|
+
|
|
107
|
+
def add_JobServiceServicer_to_server(servicer, server):
|
|
108
|
+
rpc_method_handlers = {
|
|
109
|
+
'List': grpc.unary_unary_rpc_method_handler(
|
|
110
|
+
servicer.List,
|
|
111
|
+
request_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobsRequest.FromString,
|
|
112
|
+
response_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobsResponse.SerializeToString,
|
|
113
|
+
),
|
|
114
|
+
'Create': grpc.unary_unary_rpc_method_handler(
|
|
115
|
+
servicer.Create,
|
|
116
|
+
request_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.CreateJobRequest.FromString,
|
|
117
|
+
response_serializer=yandex_dot_cloud_dot_operation_dot_operation__pb2.Operation.SerializeToString,
|
|
118
|
+
),
|
|
119
|
+
'Get': grpc.unary_unary_rpc_method_handler(
|
|
120
|
+
servicer.Get,
|
|
121
|
+
request_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.GetJobRequest.FromString,
|
|
122
|
+
response_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__pb2.Job.SerializeToString,
|
|
123
|
+
),
|
|
124
|
+
'ListLog': grpc.unary_unary_rpc_method_handler(
|
|
125
|
+
servicer.ListLog,
|
|
126
|
+
request_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobLogRequest.FromString,
|
|
127
|
+
response_serializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobLogResponse.SerializeToString,
|
|
128
|
+
),
|
|
129
|
+
'Cancel': grpc.unary_unary_rpc_method_handler(
|
|
130
|
+
servicer.Cancel,
|
|
131
|
+
request_deserializer=yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.CancelJobRequest.FromString,
|
|
132
|
+
response_serializer=yandex_dot_cloud_dot_operation_dot_operation__pb2.Operation.SerializeToString,
|
|
133
|
+
),
|
|
134
|
+
}
|
|
135
|
+
generic_handler = grpc.method_handlers_generic_handler(
|
|
136
|
+
'yandex.cloud.spark.v1.JobService', rpc_method_handlers)
|
|
137
|
+
server.add_generic_rpc_handlers((generic_handler,))
|
|
138
|
+
server.add_registered_method_handlers('yandex.cloud.spark.v1.JobService', rpc_method_handlers)
|
|
139
|
+
|
|
140
|
+
|
|
141
|
+
# This class is part of an EXPERIMENTAL API.
|
|
142
|
+
class JobService(object):
|
|
143
|
+
"""A set of methods for managing jobs for Spark cluster.
|
|
144
|
+
"""
|
|
145
|
+
|
|
146
|
+
@staticmethod
|
|
147
|
+
def List(request,
|
|
148
|
+
target,
|
|
149
|
+
options=(),
|
|
150
|
+
channel_credentials=None,
|
|
151
|
+
call_credentials=None,
|
|
152
|
+
insecure=False,
|
|
153
|
+
compression=None,
|
|
154
|
+
wait_for_ready=None,
|
|
155
|
+
timeout=None,
|
|
156
|
+
metadata=None):
|
|
157
|
+
return grpc.experimental.unary_unary(
|
|
158
|
+
request,
|
|
159
|
+
target,
|
|
160
|
+
'/yandex.cloud.spark.v1.JobService/List',
|
|
161
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobsRequest.SerializeToString,
|
|
162
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobsResponse.FromString,
|
|
163
|
+
options,
|
|
164
|
+
channel_credentials,
|
|
165
|
+
insecure,
|
|
166
|
+
call_credentials,
|
|
167
|
+
compression,
|
|
168
|
+
wait_for_ready,
|
|
169
|
+
timeout,
|
|
170
|
+
metadata,
|
|
171
|
+
_registered_method=True)
|
|
172
|
+
|
|
173
|
+
@staticmethod
|
|
174
|
+
def Create(request,
|
|
175
|
+
target,
|
|
176
|
+
options=(),
|
|
177
|
+
channel_credentials=None,
|
|
178
|
+
call_credentials=None,
|
|
179
|
+
insecure=False,
|
|
180
|
+
compression=None,
|
|
181
|
+
wait_for_ready=None,
|
|
182
|
+
timeout=None,
|
|
183
|
+
metadata=None):
|
|
184
|
+
return grpc.experimental.unary_unary(
|
|
185
|
+
request,
|
|
186
|
+
target,
|
|
187
|
+
'/yandex.cloud.spark.v1.JobService/Create',
|
|
188
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.CreateJobRequest.SerializeToString,
|
|
189
|
+
yandex_dot_cloud_dot_operation_dot_operation__pb2.Operation.FromString,
|
|
190
|
+
options,
|
|
191
|
+
channel_credentials,
|
|
192
|
+
insecure,
|
|
193
|
+
call_credentials,
|
|
194
|
+
compression,
|
|
195
|
+
wait_for_ready,
|
|
196
|
+
timeout,
|
|
197
|
+
metadata,
|
|
198
|
+
_registered_method=True)
|
|
199
|
+
|
|
200
|
+
@staticmethod
|
|
201
|
+
def Get(request,
|
|
202
|
+
target,
|
|
203
|
+
options=(),
|
|
204
|
+
channel_credentials=None,
|
|
205
|
+
call_credentials=None,
|
|
206
|
+
insecure=False,
|
|
207
|
+
compression=None,
|
|
208
|
+
wait_for_ready=None,
|
|
209
|
+
timeout=None,
|
|
210
|
+
metadata=None):
|
|
211
|
+
return grpc.experimental.unary_unary(
|
|
212
|
+
request,
|
|
213
|
+
target,
|
|
214
|
+
'/yandex.cloud.spark.v1.JobService/Get',
|
|
215
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.GetJobRequest.SerializeToString,
|
|
216
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__pb2.Job.FromString,
|
|
217
|
+
options,
|
|
218
|
+
channel_credentials,
|
|
219
|
+
insecure,
|
|
220
|
+
call_credentials,
|
|
221
|
+
compression,
|
|
222
|
+
wait_for_ready,
|
|
223
|
+
timeout,
|
|
224
|
+
metadata,
|
|
225
|
+
_registered_method=True)
|
|
226
|
+
|
|
227
|
+
@staticmethod
|
|
228
|
+
def ListLog(request,
|
|
229
|
+
target,
|
|
230
|
+
options=(),
|
|
231
|
+
channel_credentials=None,
|
|
232
|
+
call_credentials=None,
|
|
233
|
+
insecure=False,
|
|
234
|
+
compression=None,
|
|
235
|
+
wait_for_ready=None,
|
|
236
|
+
timeout=None,
|
|
237
|
+
metadata=None):
|
|
238
|
+
return grpc.experimental.unary_unary(
|
|
239
|
+
request,
|
|
240
|
+
target,
|
|
241
|
+
'/yandex.cloud.spark.v1.JobService/ListLog',
|
|
242
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobLogRequest.SerializeToString,
|
|
243
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.ListJobLogResponse.FromString,
|
|
244
|
+
options,
|
|
245
|
+
channel_credentials,
|
|
246
|
+
insecure,
|
|
247
|
+
call_credentials,
|
|
248
|
+
compression,
|
|
249
|
+
wait_for_ready,
|
|
250
|
+
timeout,
|
|
251
|
+
metadata,
|
|
252
|
+
_registered_method=True)
|
|
253
|
+
|
|
254
|
+
@staticmethod
|
|
255
|
+
def Cancel(request,
|
|
256
|
+
target,
|
|
257
|
+
options=(),
|
|
258
|
+
channel_credentials=None,
|
|
259
|
+
call_credentials=None,
|
|
260
|
+
insecure=False,
|
|
261
|
+
compression=None,
|
|
262
|
+
wait_for_ready=None,
|
|
263
|
+
timeout=None,
|
|
264
|
+
metadata=None):
|
|
265
|
+
return grpc.experimental.unary_unary(
|
|
266
|
+
request,
|
|
267
|
+
target,
|
|
268
|
+
'/yandex.cloud.spark.v1.JobService/Cancel',
|
|
269
|
+
yandex_dot_cloud_dot_spark_dot_v1_dot_job__service__pb2.CancelJobRequest.SerializeToString,
|
|
270
|
+
yandex_dot_cloud_dot_operation_dot_operation__pb2.Operation.FromString,
|
|
271
|
+
options,
|
|
272
|
+
channel_credentials,
|
|
273
|
+
insecure,
|
|
274
|
+
call_credentials,
|
|
275
|
+
compression,
|
|
276
|
+
wait_for_ready,
|
|
277
|
+
timeout,
|
|
278
|
+
metadata,
|
|
279
|
+
_registered_method=True)
|
|
@@ -0,0 +1,132 @@
|
|
|
1
|
+
"""
|
|
2
|
+
@generated by mypy-protobuf. Do not edit manually!
|
|
3
|
+
isort:skip_file
|
|
4
|
+
"""
|
|
5
|
+
|
|
6
|
+
import abc
|
|
7
|
+
import collections.abc
|
|
8
|
+
import grpc
|
|
9
|
+
import grpc.aio
|
|
10
|
+
import typing
|
|
11
|
+
import yandex.cloud.operation.operation_pb2
|
|
12
|
+
import yandex.cloud.spark.v1.job_pb2
|
|
13
|
+
import yandex.cloud.spark.v1.job_service_pb2
|
|
14
|
+
|
|
15
|
+
_T = typing.TypeVar("_T")
|
|
16
|
+
|
|
17
|
+
class _MaybeAsyncIterator(collections.abc.AsyncIterator[_T], collections.abc.Iterator[_T], metaclass=abc.ABCMeta): ...
|
|
18
|
+
|
|
19
|
+
class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type: ignore[misc, type-arg]
|
|
20
|
+
...
|
|
21
|
+
|
|
22
|
+
class JobServiceStub:
|
|
23
|
+
"""A set of methods for managing jobs for Spark cluster."""
|
|
24
|
+
|
|
25
|
+
def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
|
|
26
|
+
List: grpc.UnaryUnaryMultiCallable[
|
|
27
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobsRequest,
|
|
28
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobsResponse,
|
|
29
|
+
]
|
|
30
|
+
"""Retrieves a list of jobs for Spark cluster."""
|
|
31
|
+
|
|
32
|
+
Create: grpc.UnaryUnaryMultiCallable[
|
|
33
|
+
yandex.cloud.spark.v1.job_service_pb2.CreateJobRequest,
|
|
34
|
+
yandex.cloud.operation.operation_pb2.Operation,
|
|
35
|
+
]
|
|
36
|
+
"""Creates a job for Spark cluster."""
|
|
37
|
+
|
|
38
|
+
Get: grpc.UnaryUnaryMultiCallable[
|
|
39
|
+
yandex.cloud.spark.v1.job_service_pb2.GetJobRequest,
|
|
40
|
+
yandex.cloud.spark.v1.job_pb2.Job,
|
|
41
|
+
]
|
|
42
|
+
"""Returns the specified Spark job."""
|
|
43
|
+
|
|
44
|
+
ListLog: grpc.UnaryUnaryMultiCallable[
|
|
45
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobLogRequest,
|
|
46
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobLogResponse,
|
|
47
|
+
]
|
|
48
|
+
"""Returns a log for Spark job."""
|
|
49
|
+
|
|
50
|
+
Cancel: grpc.UnaryUnaryMultiCallable[
|
|
51
|
+
yandex.cloud.spark.v1.job_service_pb2.CancelJobRequest,
|
|
52
|
+
yandex.cloud.operation.operation_pb2.Operation,
|
|
53
|
+
]
|
|
54
|
+
"""Cancels the specified Spark job."""
|
|
55
|
+
|
|
56
|
+
class JobServiceAsyncStub:
|
|
57
|
+
"""A set of methods for managing jobs for Spark cluster."""
|
|
58
|
+
|
|
59
|
+
List: grpc.aio.UnaryUnaryMultiCallable[
|
|
60
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobsRequest,
|
|
61
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobsResponse,
|
|
62
|
+
]
|
|
63
|
+
"""Retrieves a list of jobs for Spark cluster."""
|
|
64
|
+
|
|
65
|
+
Create: grpc.aio.UnaryUnaryMultiCallable[
|
|
66
|
+
yandex.cloud.spark.v1.job_service_pb2.CreateJobRequest,
|
|
67
|
+
yandex.cloud.operation.operation_pb2.Operation,
|
|
68
|
+
]
|
|
69
|
+
"""Creates a job for Spark cluster."""
|
|
70
|
+
|
|
71
|
+
Get: grpc.aio.UnaryUnaryMultiCallable[
|
|
72
|
+
yandex.cloud.spark.v1.job_service_pb2.GetJobRequest,
|
|
73
|
+
yandex.cloud.spark.v1.job_pb2.Job,
|
|
74
|
+
]
|
|
75
|
+
"""Returns the specified Spark job."""
|
|
76
|
+
|
|
77
|
+
ListLog: grpc.aio.UnaryUnaryMultiCallable[
|
|
78
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobLogRequest,
|
|
79
|
+
yandex.cloud.spark.v1.job_service_pb2.ListJobLogResponse,
|
|
80
|
+
]
|
|
81
|
+
"""Returns a log for Spark job."""
|
|
82
|
+
|
|
83
|
+
Cancel: grpc.aio.UnaryUnaryMultiCallable[
|
|
84
|
+
yandex.cloud.spark.v1.job_service_pb2.CancelJobRequest,
|
|
85
|
+
yandex.cloud.operation.operation_pb2.Operation,
|
|
86
|
+
]
|
|
87
|
+
"""Cancels the specified Spark job."""
|
|
88
|
+
|
|
89
|
+
class JobServiceServicer(metaclass=abc.ABCMeta):
|
|
90
|
+
"""A set of methods for managing jobs for Spark cluster."""
|
|
91
|
+
|
|
92
|
+
@abc.abstractmethod
|
|
93
|
+
def List(
|
|
94
|
+
self,
|
|
95
|
+
request: yandex.cloud.spark.v1.job_service_pb2.ListJobsRequest,
|
|
96
|
+
context: _ServicerContext,
|
|
97
|
+
) -> typing.Union[yandex.cloud.spark.v1.job_service_pb2.ListJobsResponse, collections.abc.Awaitable[yandex.cloud.spark.v1.job_service_pb2.ListJobsResponse]]:
|
|
98
|
+
"""Retrieves a list of jobs for Spark cluster."""
|
|
99
|
+
|
|
100
|
+
@abc.abstractmethod
|
|
101
|
+
def Create(
|
|
102
|
+
self,
|
|
103
|
+
request: yandex.cloud.spark.v1.job_service_pb2.CreateJobRequest,
|
|
104
|
+
context: _ServicerContext,
|
|
105
|
+
) -> typing.Union[yandex.cloud.operation.operation_pb2.Operation, collections.abc.Awaitable[yandex.cloud.operation.operation_pb2.Operation]]:
|
|
106
|
+
"""Creates a job for Spark cluster."""
|
|
107
|
+
|
|
108
|
+
@abc.abstractmethod
|
|
109
|
+
def Get(
|
|
110
|
+
self,
|
|
111
|
+
request: yandex.cloud.spark.v1.job_service_pb2.GetJobRequest,
|
|
112
|
+
context: _ServicerContext,
|
|
113
|
+
) -> typing.Union[yandex.cloud.spark.v1.job_pb2.Job, collections.abc.Awaitable[yandex.cloud.spark.v1.job_pb2.Job]]:
|
|
114
|
+
"""Returns the specified Spark job."""
|
|
115
|
+
|
|
116
|
+
@abc.abstractmethod
|
|
117
|
+
def ListLog(
|
|
118
|
+
self,
|
|
119
|
+
request: yandex.cloud.spark.v1.job_service_pb2.ListJobLogRequest,
|
|
120
|
+
context: _ServicerContext,
|
|
121
|
+
) -> typing.Union[yandex.cloud.spark.v1.job_service_pb2.ListJobLogResponse, collections.abc.Awaitable[yandex.cloud.spark.v1.job_service_pb2.ListJobLogResponse]]:
|
|
122
|
+
"""Returns a log for Spark job."""
|
|
123
|
+
|
|
124
|
+
@abc.abstractmethod
|
|
125
|
+
def Cancel(
|
|
126
|
+
self,
|
|
127
|
+
request: yandex.cloud.spark.v1.job_service_pb2.CancelJobRequest,
|
|
128
|
+
context: _ServicerContext,
|
|
129
|
+
) -> typing.Union[yandex.cloud.operation.operation_pb2.Operation, collections.abc.Awaitable[yandex.cloud.operation.operation_pb2.Operation]]:
|
|
130
|
+
"""Cancels the specified Spark job."""
|
|
131
|
+
|
|
132
|
+
def add_JobServiceServicer_to_server(servicer: JobServiceServicer, server: typing.Union[grpc.Server, grpc.aio.Server]) -> None: ...
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
# -*- coding: utf-8 -*-
|
|
2
|
+
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
|
3
|
+
# NO CHECKED-IN PROTOBUF GENCODE
|
|
4
|
+
# source: yandex/cloud/spark/v1/maintenance.proto
|
|
5
|
+
# Protobuf Python Version: 5.29.0
|
|
6
|
+
"""Generated protocol buffer code."""
|
|
7
|
+
from google.protobuf import descriptor as _descriptor
|
|
8
|
+
from google.protobuf import descriptor_pool as _descriptor_pool
|
|
9
|
+
from google.protobuf import runtime_version as _runtime_version
|
|
10
|
+
from google.protobuf import symbol_database as _symbol_database
|
|
11
|
+
from google.protobuf.internal import builder as _builder
|
|
12
|
+
_runtime_version.ValidateProtobufRuntimeVersion(
|
|
13
|
+
_runtime_version.Domain.PUBLIC,
|
|
14
|
+
5,
|
|
15
|
+
29,
|
|
16
|
+
0,
|
|
17
|
+
'',
|
|
18
|
+
'yandex/cloud/spark/v1/maintenance.proto'
|
|
19
|
+
)
|
|
20
|
+
# @@protoc_insertion_point(imports)
|
|
21
|
+
|
|
22
|
+
_sym_db = _symbol_database.Default()
|
|
23
|
+
|
|
24
|
+
|
|
25
|
+
from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__pb2
|
|
26
|
+
from yandex.cloud import validation_pb2 as yandex_dot_cloud_dot_validation__pb2
|
|
27
|
+
|
|
28
|
+
|
|
29
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\'yandex/cloud/spark/v1/maintenance.proto\x12\x15yandex.cloud.spark.v1\x1a\x1fgoogle/protobuf/timestamp.proto\x1a\x1dyandex/cloud/validation.proto\"\xbc\x01\n\x11MaintenanceWindow\x12\x42\n\x07\x61nytime\x18\x01 \x01(\x0b\x32/.yandex.cloud.spark.v1.AnytimeMaintenanceWindowH\x00\x12S\n\x19weekly_maintenance_window\x18\x02 \x01(\x0b\x32..yandex.cloud.spark.v1.WeeklyMaintenanceWindowH\x00\x42\x0e\n\x06policy\x12\x04\xc0\xc1\x31\x01\"\x1a\n\x18\x41nytimeMaintenanceWindow\"\xda\x01\n\x17WeeklyMaintenanceWindow\x12\x43\n\x03\x64\x61y\x18\x01 \x01(\x0e\x32\x36.yandex.cloud.spark.v1.WeeklyMaintenanceWindow.WeekDay\x12\x16\n\x04hour\x18\x02 \x01(\x03\x42\x08\xfa\xc7\x31\x04\x31-24\"b\n\x07WeekDay\x12\x18\n\x14WEEK_DAY_UNSPECIFIED\x10\x00\x12\x07\n\x03MON\x10\x01\x12\x07\n\x03TUE\x10\x02\x12\x07\n\x03WED\x10\x03\x12\x07\n\x03THU\x10\x04\x12\x07\n\x03\x46RI\x10\x05\x12\x07\n\x03SAT\x10\x06\x12\x07\n\x03SUN\x10\x07\"\xe1\x01\n\x14MaintenanceOperation\x12\x17\n\x04info\x18\x01 \x01(\tB\t\x8a\xc8\x31\x05<=256\x12\x31\n\rdelayed_until\x18\x02 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12;\n\x17latest_maintenance_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12@\n\x1cnext_maintenance_window_time\x18\x04 \x01(\x0b\x32\x1a.google.protobuf.TimestampB\\\n\x19yandex.cloud.api.spark.v1Z?github.com/yandex-cloud/go-genproto/yandex/cloud/spark/v1;sparkb\x06proto3')
|
|
30
|
+
|
|
31
|
+
_globals = globals()
|
|
32
|
+
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
33
|
+
_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'yandex.cloud.spark.v1.maintenance_pb2', _globals)
|
|
34
|
+
if not _descriptor._USE_C_DESCRIPTORS:
|
|
35
|
+
_globals['DESCRIPTOR']._loaded_options = None
|
|
36
|
+
_globals['DESCRIPTOR']._serialized_options = b'\n\031yandex.cloud.api.spark.v1Z?github.com/yandex-cloud/go-genproto/yandex/cloud/spark/v1;spark'
|
|
37
|
+
_globals['_MAINTENANCEWINDOW'].oneofs_by_name['policy']._loaded_options = None
|
|
38
|
+
_globals['_MAINTENANCEWINDOW'].oneofs_by_name['policy']._serialized_options = b'\300\3011\001'
|
|
39
|
+
_globals['_WEEKLYMAINTENANCEWINDOW'].fields_by_name['hour']._loaded_options = None
|
|
40
|
+
_globals['_WEEKLYMAINTENANCEWINDOW'].fields_by_name['hour']._serialized_options = b'\372\3071\0041-24'
|
|
41
|
+
_globals['_MAINTENANCEOPERATION'].fields_by_name['info']._loaded_options = None
|
|
42
|
+
_globals['_MAINTENANCEOPERATION'].fields_by_name['info']._serialized_options = b'\212\3101\005<=256'
|
|
43
|
+
_globals['_MAINTENANCEWINDOW']._serialized_start=131
|
|
44
|
+
_globals['_MAINTENANCEWINDOW']._serialized_end=319
|
|
45
|
+
_globals['_ANYTIMEMAINTENANCEWINDOW']._serialized_start=321
|
|
46
|
+
_globals['_ANYTIMEMAINTENANCEWINDOW']._serialized_end=347
|
|
47
|
+
_globals['_WEEKLYMAINTENANCEWINDOW']._serialized_start=350
|
|
48
|
+
_globals['_WEEKLYMAINTENANCEWINDOW']._serialized_end=568
|
|
49
|
+
_globals['_WEEKLYMAINTENANCEWINDOW_WEEKDAY']._serialized_start=470
|
|
50
|
+
_globals['_WEEKLYMAINTENANCEWINDOW_WEEKDAY']._serialized_end=568
|
|
51
|
+
_globals['_MAINTENANCEOPERATION']._serialized_start=571
|
|
52
|
+
_globals['_MAINTENANCEOPERATION']._serialized_end=796
|
|
53
|
+
# @@protoc_insertion_point(module_scope)
|