yandexcloud 0.324.0__py3-none-any.whl → 0.325.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- yandex/cloud/ai/foundation_models/v1/embedding/embedding_service_pb2.pyi +1 -1
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_pb2.pyi +4 -2
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2.pyi +6 -3
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2_grpc.py +3 -3
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2_grpc.pyi +3 -3
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_pb2.pyi +3 -3
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2.pyi +25 -10
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.py +11 -5
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.pyi +33 -9
- yandex/cloud/ai/foundation_models/v1/text_common_pb2.pyi +3 -3
- yandex/cloud/ai/stt/v3/stt_pb2.py +5 -3
- yandex/cloud/ai/stt/v3/stt_pb2.pyi +15 -0
- yandex/cloud/ai/stt/v3/stt_service_pb2.py +10 -7
- yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.py +34 -0
- yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.pyi +18 -0
- yandex/cloud/apploadbalancer/v1/backend_group_pb2.py +26 -26
- yandex/cloud/apploadbalancer/v1/backend_group_pb2.pyi +7 -1
- yandex/cloud/backup/v1/policy_pb2.pyi +1 -1
- yandex/cloud/cdn/v1/resource_service_pb2.py +14 -14
- yandex/cloud/cdn/v1/resource_service_pb2.pyi +5 -1
- yandex/cloud/cloudregistry/v1/artifact_pb2.pyi +2 -0
- yandex/cloud/cloudregistry/v1/artifact_service_pb2.pyi +3 -3
- yandex/cloud/cloudregistry/v1/artifact_service_pb2_grpc.py +9 -6
- yandex/cloud/cloudregistry/v1/artifact_service_pb2_grpc.pyi +15 -9
- yandex/cloud/cloudregistry/v1/registry_pb2.pyi +1 -1
- yandex/cloud/cloudregistry/v1/registry_service_pb2.py +42 -38
- yandex/cloud/cloudregistry/v1/registry_service_pb2_grpc.py +2 -2
- yandex/cloud/cloudregistry/v1/registry_service_pb2_grpc.pyi +6 -6
- yandex/cloud/datasphere/v2/jobs/jobs_pb2.py +9 -5
- yandex/cloud/datasphere/v2/jobs/jobs_pb2.pyi +96 -0
- yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.py +55 -53
- yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.pyi +14 -2
- yandex/cloud/iam/v1/user_account_pb2.pyi +2 -2
- yandex/cloud/iam/v1/user_account_service_pb2_grpc.py +3 -3
- yandex/cloud/iam/v1/user_account_service_pb2_grpc.pyi +3 -3
- yandex/cloud/mdb/redis/v1/backup_pb2.py +4 -2
- yandex/cloud/mdb/redis/v1/backup_pb2.pyi +31 -1
- yandex/cloud/organizationmanager/v1/user_account_pb2.pyi +2 -2
- yandex/cloud/searchapi/__init__.py +0 -0
- yandex/cloud/searchapi/v2/__init__.py +0 -0
- yandex/cloud/searchapi/v2/search_query_pb2.py +37 -0
- yandex/cloud/searchapi/v2/search_query_pb2.pyi +100 -0
- yandex/cloud/searchapi/v2/search_query_pb2_grpc.py +4 -0
- yandex/cloud/searchapi/v2/search_query_pb2_grpc.pyi +17 -0
- yandex/cloud/searchapi/v2/search_service_pb2.py +61 -0
- yandex/cloud/searchapi/v2/search_service_pb2.pyi +218 -0
- yandex/cloud/searchapi/v2/search_service_pb2_grpc.py +70 -0
- yandex/cloud/searchapi/v2/search_service_pb2_grpc.pyi +48 -0
- yandex/cloud/serverless/containers/v1/container_pb2.py +42 -34
- yandex/cloud/serverless/containers/v1/container_pb2.pyi +52 -2
- yandex/cloud/serverless/containers/v1/container_service_pb2.py +18 -18
- yandex/cloud/serverless/containers/v1/container_service_pb2.pyi +8 -2
- yandex/cloud/serverless/eventrouter/v1/connector_pb2.py +9 -9
- yandex/cloud/serverless/eventrouter/v1/connector_pb2.pyi +8 -0
- yandex/cloud/serverless/eventrouter/v1/rule_pb2.py +31 -31
- yandex/cloud/serverless/eventrouter/v1/rule_pb2.pyi +4 -0
- yandex/cloud/serverless/workflows/v1/execution_service_pb2.py +23 -15
- yandex/cloud/serverless/workflows/v1/execution_service_pb2.pyi +32 -0
- yandex/cloud/serverless/workflows/v1/execution_service_pb2_grpc.py +34 -0
- yandex/cloud/serverless/workflows/v1/execution_service_pb2_grpc.pyi +20 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2.py +46 -40
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2.pyi +47 -2
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.py +18 -18
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.pyi +16 -3
- yandex/cloud/video/v1/channel_pb2.py +1 -1
- yandex/cloud/video/v1/channel_service_pb2.py +58 -25
- yandex/cloud/video/v1/channel_service_pb2.pyi +40 -0
- yandex/cloud/video/v1/channel_service_pb2_grpc.py +34 -0
- yandex/cloud/video/v1/channel_service_pb2_grpc.pyi +20 -0
- yandex/cloud/video/v1/episode_pb2.py +10 -8
- yandex/cloud/video/v1/episode_pb2.pyi +19 -3
- yandex/cloud/video/v1/episode_service_pb2.py +96 -43
- yandex/cloud/video/v1/episode_service_pb2.pyi +68 -6
- yandex/cloud/video/v1/episode_service_pb2_grpc.py +34 -0
- yandex/cloud/video/v1/episode_service_pb2_grpc.pyi +20 -0
- yandex/cloud/video/v1/manifest_pb2.py +3 -3
- yandex/cloud/video/v1/manifest_pb2.pyi +1 -5
- yandex/cloud/video/v1/playlist_pb2.py +29 -0
- yandex/cloud/video/v1/playlist_pb2.pyi +89 -0
- yandex/cloud/video/v1/playlist_pb2_grpc.py +4 -0
- yandex/cloud/video/v1/playlist_pb2_grpc.pyi +17 -0
- yandex/cloud/video/v1/playlist_service_pb2.py +102 -0
- yandex/cloud/video/v1/playlist_service_pb2.pyi +331 -0
- yandex/cloud/video/v1/playlist_service_pb2_grpc.py +276 -0
- yandex/cloud/video/v1/playlist_service_pb2_grpc.pyi +172 -0
- yandex/cloud/video/v1/stream_line_pb2.py +1 -1
- yandex/cloud/video/v1/stream_line_service_pb2.py +122 -55
- yandex/cloud/video/v1/stream_line_service_pb2.pyi +80 -0
- yandex/cloud/video/v1/stream_line_service_pb2_grpc.py +68 -0
- yandex/cloud/video/v1/stream_line_service_pb2_grpc.pyi +40 -0
- yandex/cloud/video/v1/stream_pb2.py +1 -1
- yandex/cloud/video/v1/stream_service_pb2.py +98 -37
- yandex/cloud/video/v1/stream_service_pb2.pyi +80 -0
- yandex/cloud/video/v1/stream_service_pb2_grpc.py +68 -0
- yandex/cloud/video/v1/stream_service_pb2_grpc.pyi +40 -0
- yandex/cloud/video/v1/subtitle_pb2.py +29 -0
- yandex/cloud/video/v1/subtitle_pb2.pyi +90 -0
- yandex/cloud/video/v1/subtitle_pb2_grpc.py +4 -0
- yandex/cloud/video/v1/subtitle_pb2_grpc.pyi +17 -0
- yandex/cloud/video/v1/subtitle_service_pb2.py +83 -0
- yandex/cloud/video/v1/subtitle_service_pb2.pyi +212 -0
- yandex/cloud/video/v1/subtitle_service_pb2_grpc.py +208 -0
- yandex/cloud/video/v1/subtitle_service_pb2_grpc.pyi +132 -0
- yandex/cloud/video/v1/thumbnail_service_pb2.py +50 -21
- yandex/cloud/video/v1/thumbnail_service_pb2.pyi +48 -0
- yandex/cloud/video/v1/thumbnail_service_pb2_grpc.py +69 -0
- yandex/cloud/video/v1/thumbnail_service_pb2_grpc.pyi +41 -0
- yandex/cloud/video/v1/video_pb2.py +18 -14
- yandex/cloud/video/v1/video_pb2.pyi +55 -3
- yandex/cloud/video/v1/video_service_pb2.py +118 -49
- yandex/cloud/video/v1/video_service_pb2.pyi +161 -6
- yandex/cloud/video/v1/video_service_pb2_grpc.py +102 -0
- yandex/cloud/video/v1/video_service_pb2_grpc.pyi +60 -0
- {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/METADATA +1 -1
- {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/RECORD +119 -93
- {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/WHEEL +1 -1
- {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/AUTHORS +0 -0
- {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/LICENSE +0 -0
- {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/top_level.txt +0 -0
|
@@ -21,7 +21,7 @@ class TextEmbeddingRequest(google.protobuf.message.Message):
|
|
|
21
21
|
MODEL_URI_FIELD_NUMBER: builtins.int
|
|
22
22
|
TEXT_FIELD_NUMBER: builtins.int
|
|
23
23
|
model_uri: builtins.str
|
|
24
|
-
"""The [
|
|
24
|
+
"""The [model URI](/docs/foundation-models/concepts/embeddings) to be used for obtaining text embeddings."""
|
|
25
25
|
text: builtins.str
|
|
26
26
|
"""The input text for which the embedding is requested."""
|
|
27
27
|
def __init__(
|
|
@@ -60,9 +60,11 @@ class ImageGenerationOptions(google.protobuf.message.Message):
|
|
|
60
60
|
SEED_FIELD_NUMBER: builtins.int
|
|
61
61
|
ASPECT_RATIO_FIELD_NUMBER: builtins.int
|
|
62
62
|
mime_type: builtins.str
|
|
63
|
-
"""MIME type of generated image format.
|
|
63
|
+
"""The [MIME type](https://en.wikipedia.org/wiki/Media_type) of generated image format.
|
|
64
|
+
For possible specifications, see [documentation](/docs/foundation-models/concepts).
|
|
65
|
+
"""
|
|
64
66
|
seed: builtins.int
|
|
65
|
-
"""Seed for image generation."""
|
|
67
|
+
"""Seed for image generation. It serves as a starting point for image generation from noise."""
|
|
66
68
|
@property
|
|
67
69
|
def aspect_ratio(self) -> global___AspectRatio:
|
|
68
70
|
"""Aspect ratio of generated image."""
|
|
@@ -15,7 +15,10 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
|
15
15
|
|
|
16
16
|
@typing.final
|
|
17
17
|
class ImageGenerationRequest(google.protobuf.message.Message):
|
|
18
|
-
"""Request for the service to generate an image.
|
|
18
|
+
"""Request for the service to generate an image.
|
|
19
|
+
|
|
20
|
+
For examples of usage, see [step-by-step guide](/docs/foundation-models/operations/yandexart/request).
|
|
21
|
+
"""
|
|
19
22
|
|
|
20
23
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
21
24
|
|
|
@@ -23,7 +26,7 @@ class ImageGenerationRequest(google.protobuf.message.Message):
|
|
|
23
26
|
MESSAGES_FIELD_NUMBER: builtins.int
|
|
24
27
|
GENERATION_OPTIONS_FIELD_NUMBER: builtins.int
|
|
25
28
|
model_uri: builtins.str
|
|
26
|
-
"""The [
|
|
29
|
+
"""The [model URI](/docs/foundation-models/concepts/yandexart/models) to be used for image generation."""
|
|
27
30
|
@property
|
|
28
31
|
def messages(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.image_generation.image_generation_pb2.Message]:
|
|
29
32
|
"""A list of messages representing the context for the image generation model."""
|
|
@@ -53,7 +56,7 @@ class ImageGenerationResponse(google.protobuf.message.Message):
|
|
|
53
56
|
IMAGE_FIELD_NUMBER: builtins.int
|
|
54
57
|
MODEL_VERSION_FIELD_NUMBER: builtins.int
|
|
55
58
|
image: builtins.bytes
|
|
56
|
-
"""The image is serialized as an array of bytes encoded in
|
|
59
|
+
"""The image is serialized as an array of bytes encoded in [Base64](https://en.wikipedia.org/wiki/Base64)."""
|
|
57
60
|
model_version: builtins.str
|
|
58
61
|
"""The model version changes with each new releases."""
|
|
59
62
|
def __init__(
|
|
@@ -7,7 +7,7 @@ from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operati
|
|
|
7
7
|
|
|
8
8
|
|
|
9
9
|
class ImageGenerationAsyncServiceStub(object):
|
|
10
|
-
"""Service for
|
|
10
|
+
"""Service for creating images based on a text description.
|
|
11
11
|
"""
|
|
12
12
|
|
|
13
13
|
def __init__(self, channel):
|
|
@@ -24,7 +24,7 @@ class ImageGenerationAsyncServiceStub(object):
|
|
|
24
24
|
|
|
25
25
|
|
|
26
26
|
class ImageGenerationAsyncServiceServicer(object):
|
|
27
|
-
"""Service for
|
|
27
|
+
"""Service for creating images based on a text description.
|
|
28
28
|
"""
|
|
29
29
|
|
|
30
30
|
def Generate(self, request, context):
|
|
@@ -50,7 +50,7 @@ def add_ImageGenerationAsyncServiceServicer_to_server(servicer, server):
|
|
|
50
50
|
|
|
51
51
|
# This class is part of an EXPERIMENTAL API.
|
|
52
52
|
class ImageGenerationAsyncService(object):
|
|
53
|
-
"""Service for
|
|
53
|
+
"""Service for creating images based on a text description.
|
|
54
54
|
"""
|
|
55
55
|
|
|
56
56
|
@staticmethod
|
|
@@ -19,7 +19,7 @@ class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type:
|
|
|
19
19
|
...
|
|
20
20
|
|
|
21
21
|
class ImageGenerationAsyncServiceStub:
|
|
22
|
-
"""Service for
|
|
22
|
+
"""Service for creating images based on a text description."""
|
|
23
23
|
|
|
24
24
|
def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
|
|
25
25
|
Generate: grpc.UnaryUnaryMultiCallable[
|
|
@@ -29,7 +29,7 @@ class ImageGenerationAsyncServiceStub:
|
|
|
29
29
|
"""A method for generating an image based on a textual description."""
|
|
30
30
|
|
|
31
31
|
class ImageGenerationAsyncServiceAsyncStub:
|
|
32
|
-
"""Service for
|
|
32
|
+
"""Service for creating images based on a text description."""
|
|
33
33
|
|
|
34
34
|
Generate: grpc.aio.UnaryUnaryMultiCallable[
|
|
35
35
|
yandex.cloud.ai.foundation_models.v1.image_generation.image_generation_service_pb2.ImageGenerationRequest,
|
|
@@ -38,7 +38,7 @@ class ImageGenerationAsyncServiceAsyncStub:
|
|
|
38
38
|
"""A method for generating an image based on a textual description."""
|
|
39
39
|
|
|
40
40
|
class ImageGenerationAsyncServiceServicer(metaclass=abc.ABCMeta):
|
|
41
|
-
"""Service for
|
|
41
|
+
"""Service for creating images based on a text description."""
|
|
42
42
|
|
|
43
43
|
@abc.abstractmethod
|
|
44
44
|
def Generate(
|
|
@@ -12,16 +12,16 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
|
12
12
|
|
|
13
13
|
@typing.final
|
|
14
14
|
class ClassificationLabel(google.protobuf.message.Message):
|
|
15
|
-
"""A pair of text
|
|
15
|
+
"""A pair of text labels and their corresponding confidence values."""
|
|
16
16
|
|
|
17
17
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
18
18
|
|
|
19
19
|
LABEL_FIELD_NUMBER: builtins.int
|
|
20
20
|
CONFIDENCE_FIELD_NUMBER: builtins.int
|
|
21
21
|
label: builtins.str
|
|
22
|
-
"""A
|
|
22
|
+
"""A class name label."""
|
|
23
23
|
confidence: builtins.float
|
|
24
|
-
"""
|
|
24
|
+
"""The probability of classifying text into a specific class."""
|
|
25
25
|
def __init__(
|
|
26
26
|
self,
|
|
27
27
|
*,
|
yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2.pyi
CHANGED
|
@@ -15,14 +15,20 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
|
15
15
|
|
|
16
16
|
@typing.final
|
|
17
17
|
class TextClassificationRequest(google.protobuf.message.Message):
|
|
18
|
-
"""Request for the service to classify text.
|
|
18
|
+
"""Request for the service to classify text with tuned model.
|
|
19
|
+
|
|
20
|
+
The names of the classes between which the model will be distributing requests must be specified during model tuning;
|
|
21
|
+
therefore, they are not provided in the request.
|
|
22
|
+
|
|
23
|
+
For examples of usage, see [step-by-step guides](/docs/operations/classifier/additionally-trained).
|
|
24
|
+
"""
|
|
19
25
|
|
|
20
26
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
21
27
|
|
|
22
28
|
MODEL_URI_FIELD_NUMBER: builtins.int
|
|
23
29
|
TEXT_FIELD_NUMBER: builtins.int
|
|
24
30
|
model_uri: builtins.str
|
|
25
|
-
"""The
|
|
31
|
+
"""The [URI](/docs/foundation-models/concepts/classifier/models) of your tuned classifier model."""
|
|
26
32
|
text: builtins.str
|
|
27
33
|
"""Text for classification."""
|
|
28
34
|
def __init__(
|
|
@@ -37,17 +43,19 @@ global___TextClassificationRequest = TextClassificationRequest
|
|
|
37
43
|
|
|
38
44
|
@typing.final
|
|
39
45
|
class TextClassificationResponse(google.protobuf.message.Message):
|
|
40
|
-
"""Response
|
|
46
|
+
"""Response with classifier predictions."""
|
|
41
47
|
|
|
42
48
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
43
49
|
|
|
44
50
|
PREDICTIONS_FIELD_NUMBER: builtins.int
|
|
45
51
|
MODEL_VERSION_FIELD_NUMBER: builtins.int
|
|
46
52
|
model_version: builtins.str
|
|
47
|
-
"""
|
|
53
|
+
"""The model version changes with each new releases."""
|
|
48
54
|
@property
|
|
49
55
|
def predictions(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationLabel]:
|
|
50
|
-
"""
|
|
56
|
+
"""The classification results with the `confidence`` values
|
|
57
|
+
for the probability of classifying the request text into each class.
|
|
58
|
+
"""
|
|
51
59
|
|
|
52
60
|
def __init__(
|
|
53
61
|
self,
|
|
@@ -61,7 +69,9 @@ global___TextClassificationResponse = TextClassificationResponse
|
|
|
61
69
|
|
|
62
70
|
@typing.final
|
|
63
71
|
class FewShotTextClassificationRequest(google.protobuf.message.Message):
|
|
64
|
-
"""Request for the service to classify text.
|
|
72
|
+
"""Request for the service to classify text.
|
|
73
|
+
For examples of usage, see [step-by-step guides](/docs/operations/classifier/readymade).
|
|
74
|
+
"""
|
|
65
75
|
|
|
66
76
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
67
77
|
|
|
@@ -71,14 +81,17 @@ class FewShotTextClassificationRequest(google.protobuf.message.Message):
|
|
|
71
81
|
TEXT_FIELD_NUMBER: builtins.int
|
|
72
82
|
SAMPLES_FIELD_NUMBER: builtins.int
|
|
73
83
|
model_uri: builtins.str
|
|
74
|
-
"""The
|
|
84
|
+
"""The [URI](/docs/foundation-models/concepts/classifier/models) of the classifier model."""
|
|
75
85
|
task_description: builtins.str
|
|
76
86
|
"""Text description of the classification task."""
|
|
77
87
|
text: builtins.str
|
|
78
88
|
"""Text for classification."""
|
|
79
89
|
@property
|
|
80
90
|
def labels(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
81
|
-
"""List of available labels for the classification result.
|
|
91
|
+
"""List of available labels for the classification result.
|
|
92
|
+
Give meaningful names to label classes: this is essential for correct classification results.
|
|
93
|
+
For example, use ``chemistry`` and ``physics`` rather than ``chm`` and ``phs`` for class names.
|
|
94
|
+
"""
|
|
82
95
|
|
|
83
96
|
@property
|
|
84
97
|
def samples(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationSample]:
|
|
@@ -106,10 +119,12 @@ class FewShotTextClassificationResponse(google.protobuf.message.Message):
|
|
|
106
119
|
PREDICTIONS_FIELD_NUMBER: builtins.int
|
|
107
120
|
MODEL_VERSION_FIELD_NUMBER: builtins.int
|
|
108
121
|
model_version: builtins.str
|
|
109
|
-
"""
|
|
122
|
+
"""The model version changes with each new releases."""
|
|
110
123
|
@property
|
|
111
124
|
def predictions(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationLabel]:
|
|
112
|
-
"""
|
|
125
|
+
"""The classification results with the `confidence`` values
|
|
126
|
+
for the probability of classifying the request text into each class.
|
|
127
|
+
"""
|
|
113
128
|
|
|
114
129
|
def __init__(
|
|
115
130
|
self,
|
yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.py
CHANGED
|
@@ -6,7 +6,7 @@ from yandex.cloud.ai.foundation_models.v1.text_classification import text_classi
|
|
|
6
6
|
|
|
7
7
|
|
|
8
8
|
class TextClassificationServiceStub(object):
|
|
9
|
-
"""Service for classifying text
|
|
9
|
+
"""Service for classifying the text requests provided in prompts.
|
|
10
10
|
"""
|
|
11
11
|
|
|
12
12
|
def __init__(self, channel):
|
|
@@ -28,18 +28,24 @@ class TextClassificationServiceStub(object):
|
|
|
28
28
|
|
|
29
29
|
|
|
30
30
|
class TextClassificationServiceServicer(object):
|
|
31
|
-
"""Service for classifying text
|
|
31
|
+
"""Service for classifying the text requests provided in prompts.
|
|
32
32
|
"""
|
|
33
33
|
|
|
34
34
|
def Classify(self, request, context):
|
|
35
|
-
"""RPC method
|
|
35
|
+
"""RPC method to classify text with tuned model.
|
|
36
|
+
|
|
37
|
+
The names of the classes between which the model will be distributing requests
|
|
38
|
+
must be specified during model tuning and are not provided in the request.
|
|
36
39
|
"""
|
|
37
40
|
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
38
41
|
context.set_details('Method not implemented!')
|
|
39
42
|
raise NotImplementedError('Method not implemented!')
|
|
40
43
|
|
|
41
44
|
def FewShotClassify(self, request, context):
|
|
42
|
-
"""RPC method for
|
|
45
|
+
"""RPC method for binary and multi-class classification.
|
|
46
|
+
|
|
47
|
+
You can provide up to 20 classes for few-shot text classification
|
|
48
|
+
with optional examples.
|
|
43
49
|
"""
|
|
44
50
|
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
45
51
|
context.set_details('Method not implemented!')
|
|
@@ -66,7 +72,7 @@ def add_TextClassificationServiceServicer_to_server(servicer, server):
|
|
|
66
72
|
|
|
67
73
|
# This class is part of an EXPERIMENTAL API.
|
|
68
74
|
class TextClassificationService(object):
|
|
69
|
-
"""Service for classifying text
|
|
75
|
+
"""Service for classifying the text requests provided in prompts.
|
|
70
76
|
"""
|
|
71
77
|
|
|
72
78
|
@staticmethod
|
yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.pyi
CHANGED
|
@@ -18,38 +18,54 @@ class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type:
|
|
|
18
18
|
...
|
|
19
19
|
|
|
20
20
|
class TextClassificationServiceStub:
|
|
21
|
-
"""Service for classifying text
|
|
21
|
+
"""Service for classifying the text requests provided in prompts."""
|
|
22
22
|
|
|
23
23
|
def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
|
|
24
24
|
Classify: grpc.UnaryUnaryMultiCallable[
|
|
25
25
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
|
|
26
26
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse,
|
|
27
27
|
]
|
|
28
|
-
"""RPC method
|
|
28
|
+
"""RPC method to classify text with tuned model.
|
|
29
|
+
|
|
30
|
+
The names of the classes between which the model will be distributing requests
|
|
31
|
+
must be specified during model tuning and are not provided in the request.
|
|
32
|
+
"""
|
|
29
33
|
|
|
30
34
|
FewShotClassify: grpc.UnaryUnaryMultiCallable[
|
|
31
35
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
|
|
32
36
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse,
|
|
33
37
|
]
|
|
34
|
-
"""RPC method for
|
|
38
|
+
"""RPC method for binary and multi-class classification.
|
|
39
|
+
|
|
40
|
+
You can provide up to 20 classes for few-shot text classification
|
|
41
|
+
with optional examples.
|
|
42
|
+
"""
|
|
35
43
|
|
|
36
44
|
class TextClassificationServiceAsyncStub:
|
|
37
|
-
"""Service for classifying text
|
|
45
|
+
"""Service for classifying the text requests provided in prompts."""
|
|
38
46
|
|
|
39
47
|
Classify: grpc.aio.UnaryUnaryMultiCallable[
|
|
40
48
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
|
|
41
49
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse,
|
|
42
50
|
]
|
|
43
|
-
"""RPC method
|
|
51
|
+
"""RPC method to classify text with tuned model.
|
|
52
|
+
|
|
53
|
+
The names of the classes between which the model will be distributing requests
|
|
54
|
+
must be specified during model tuning and are not provided in the request.
|
|
55
|
+
"""
|
|
44
56
|
|
|
45
57
|
FewShotClassify: grpc.aio.UnaryUnaryMultiCallable[
|
|
46
58
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
|
|
47
59
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse,
|
|
48
60
|
]
|
|
49
|
-
"""RPC method for
|
|
61
|
+
"""RPC method for binary and multi-class classification.
|
|
62
|
+
|
|
63
|
+
You can provide up to 20 classes for few-shot text classification
|
|
64
|
+
with optional examples.
|
|
65
|
+
"""
|
|
50
66
|
|
|
51
67
|
class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
|
|
52
|
-
"""Service for classifying text
|
|
68
|
+
"""Service for classifying the text requests provided in prompts."""
|
|
53
69
|
|
|
54
70
|
@abc.abstractmethod
|
|
55
71
|
def Classify(
|
|
@@ -57,7 +73,11 @@ class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
|
|
|
57
73
|
request: yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
|
|
58
74
|
context: _ServicerContext,
|
|
59
75
|
) -> typing.Union[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse, collections.abc.Awaitable[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse]]:
|
|
60
|
-
"""RPC method
|
|
76
|
+
"""RPC method to classify text with tuned model.
|
|
77
|
+
|
|
78
|
+
The names of the classes between which the model will be distributing requests
|
|
79
|
+
must be specified during model tuning and are not provided in the request.
|
|
80
|
+
"""
|
|
61
81
|
|
|
62
82
|
@abc.abstractmethod
|
|
63
83
|
def FewShotClassify(
|
|
@@ -65,6 +85,10 @@ class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
|
|
|
65
85
|
request: yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
|
|
66
86
|
context: _ServicerContext,
|
|
67
87
|
) -> typing.Union[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse, collections.abc.Awaitable[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse]]:
|
|
68
|
-
"""RPC method for
|
|
88
|
+
"""RPC method for binary and multi-class classification.
|
|
89
|
+
|
|
90
|
+
You can provide up to 20 classes for few-shot text classification
|
|
91
|
+
with optional examples.
|
|
92
|
+
"""
|
|
69
93
|
|
|
70
94
|
def add_TextClassificationServiceServicer_to_server(servicer: TextClassificationServiceServicer, server: typing.Union[grpc.Server, grpc.aio.Server]) -> None: ...
|
|
@@ -109,7 +109,7 @@ class ContentUsage(google.protobuf.message.Message):
|
|
|
109
109
|
input_text_tokens: builtins.int
|
|
110
110
|
"""The number of tokens in the textual part of the model input."""
|
|
111
111
|
completion_tokens: builtins.int
|
|
112
|
-
"""The
|
|
112
|
+
"""The number of tokens in the generated completion."""
|
|
113
113
|
total_tokens: builtins.int
|
|
114
114
|
"""The total number of tokens, including all input tokens and all generated tokens."""
|
|
115
115
|
def __init__(
|
|
@@ -171,10 +171,10 @@ class Alternative(google.protobuf.message.Message):
|
|
|
171
171
|
MESSAGE_FIELD_NUMBER: builtins.int
|
|
172
172
|
STATUS_FIELD_NUMBER: builtins.int
|
|
173
173
|
status: global___Alternative.AlternativeStatus.ValueType
|
|
174
|
-
"""The generation status of the alternative"""
|
|
174
|
+
"""The generation status of the alternative."""
|
|
175
175
|
@property
|
|
176
176
|
def message(self) -> global___Message:
|
|
177
|
-
"""A message
|
|
177
|
+
"""A message with the content of the alternative."""
|
|
178
178
|
|
|
179
179
|
def __init__(
|
|
180
180
|
self,
|
|
@@ -13,7 +13,7 @@ _sym_db = _symbol_database.Default()
|
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
|
|
16
|
-
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n yandex/cloud/ai/stt/v3/stt.proto\x12\x10speechkit.stt.v3\"\xe2\x03\n\x18TextNormalizationOptions\x12X\n\x12text_normalization\x18\x01 \x01(\x0e\x32<.speechkit.stt.v3.TextNormalizationOptions.TextNormalization\x12\x18\n\x10profanity_filter\x18\x02 \x01(\x08\x12\x17\n\x0fliterature_text\x18\x03 \x01(\x08\x12]\n\x15phone_formatting_mode\x18\x04 \x01(\x0e\x32>.speechkit.stt.v3.TextNormalizationOptions.PhoneFormattingMode\"x\n\x11TextNormalization\x12\"\n\x1eTEXT_NORMALIZATION_UNSPECIFIED\x10\x00\x12\x1e\n\x1aTEXT_NORMALIZATION_ENABLED\x10\x01\x12\x1f\n\x1bTEXT_NORMALIZATION_DISABLED\x10\x02\"`\n\x13PhoneFormattingMode\x12%\n!PHONE_FORMATTING_MODE_UNSPECIFIED\x10\x00\x12\"\n\x1ePHONE_FORMATTING_MODE_DISABLED\x10\x01\"\xce\x01\n\x14\x44\x65\x66\x61ultEouClassifier\x12\x43\n\x04type\x18\x01 \x01(\x0e\x32\x35.speechkit.stt.v3.DefaultEouClassifier.EouSensitivity\x12\'\n\x1fmax_pause_between_words_hint_ms\x18\x02 \x01(\x03\"H\n\x0e\x45ouSensitivity\x12\x1f\n\x1b\x45OU_SENSITIVITY_UNSPECIFIED\x10\x00\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x01\x12\x08\n\x04HIGH\x10\x02\"\x17\n\x15\x45xternalEouClassifier\"\xb2\x01\n\x14\x45ouClassifierOptions\x12\x44\n\x12\x64\x65\x66\x61ult_classifier\x18\x01 \x01(\x0b\x32&.speechkit.stt.v3.DefaultEouClassifierH\x00\x12\x46\n\x13\x65xternal_classifier\x18\x02 \x01(\x0b\x32\'.speechkit.stt.v3.ExternalEouClassifierH\x00\x42\x0c\n\nClassifier\"\xd3\x01\n\x15RecognitionClassifier\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x45\n\x08triggers\x18\x02 \x03(\x0e\x32\x33.speechkit.stt.v3.RecognitionClassifier.TriggerType\"_\n\x0bTriggerType\x12 \n\x18TRIGGER_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x10\n\x0cON_UTTERANCE\x10\x01\x12\x0c\n\x08ON_FINAL\x10\x02\x12\x0e\n\nON_PARTIAL\x10\x03\"\\\n\x1cRecognitionClassifierOptions\x12<\n\x0b\x63lassifiers\x18\x01 \x03(\x0b\x32\'.speechkit.stt.v3.RecognitionClassifier\"\x88\x01\n\x15SpeechAnalysisOptions\x12\x1f\n\x17\x65nable_speaker_analysis\x18\x01 \x01(\x08\x12$\n\x1c\x65nable_conversation_analysis\x18\x02 \x01(\x08\x12(\n descriptive_statistics_quantiles\x18\x03 \x03(\x01\"\xc7\x01\n\x08RawAudio\x12@\n\x0e\x61udio_encoding\x18\x01 \x01(\x0e\x32(.speechkit.stt.v3.RawAudio.AudioEncoding\x12\x19\n\x11sample_rate_hertz\x18\x02 \x01(\x03\x12\x1b\n\x13\x61udio_channel_count\x18\x03 \x01(\x03\"A\n\rAudioEncoding\x12\x1e\n\x1a\x41UDIO_ENCODING_UNSPECIFIED\x10\x00\x12\x10\n\x0cLINEAR16_PCM\x10\x01\"\xbf\x01\n\x0e\x43ontainerAudio\x12Q\n\x14\x63ontainer_audio_type\x18\x01 \x01(\x0e\x32\x33.speechkit.stt.v3.ContainerAudio.ContainerAudioType\"Z\n\x12\x43ontainerAudioType\x12$\n CONTAINER_AUDIO_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03WAV\x10\x01\x12\x0c\n\x08OGG_OPUS\x10\x02\x12\x07\n\x03MP3\x10\x03\"\x91\x01\n\x12\x41udioFormatOptions\x12/\n\traw_audio\x18\x01 \x01(\x0b\x32\x1a.speechkit.stt.v3.RawAudioH\x00\x12;\n\x0f\x63ontainer_audio\x18\x02 \x01(\x0b\x32 .speechkit.stt.v3.ContainerAudioH\x00\x42\r\n\x0b\x41udioFormat\"\xf7\x01\n\x1aLanguageRestrictionOptions\x12^\n\x10restriction_type\x18\x01 \x01(\x0e\x32\x44.speechkit.stt.v3.LanguageRestrictionOptions.LanguageRestrictionType\x12\x15\n\rlanguage_code\x18\x02 \x03(\t\"b\n\x17LanguageRestrictionType\x12)\n%LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tWHITELIST\x10\x01\x12\r\n\tBLACKLIST\x10\x02\"\xb2\x03\n\x17RecognitionModelOptions\x12\r\n\x05model\x18\x01 \x01(\t\x12:\n\x0c\x61udio_format\x18\x02 \x01(\x0b\x32$.speechkit.stt.v3.AudioFormatOptions\x12\x46\n\x12text_normalization\x18\x03 \x01(\x0b\x32*.speechkit.stt.v3.TextNormalizationOptions\x12J\n\x14language_restriction\x18\x04 \x01(\x0b\x32,.speechkit.stt.v3.LanguageRestrictionOptions\x12\\\n\x15\x61udio_processing_type\x18\x05 \x01(\x0e\x32=.speechkit.stt.v3.RecognitionModelOptions.AudioProcessingType\"Z\n\x13\x41udioProcessingType\x12%\n!AUDIO_PROCESSING_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tREAL_TIME\x10\x01\x12\r\n\tFULL_DATA\x10\x02\"\xde\x01\n\x16SpeakerLabelingOptions\x12R\n\x10speaker_labeling\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.SpeakerLabelingOptions.SpeakerLabeling\"p\n\x0fSpeakerLabeling\x12 \n\x1cSPEAKER_LABELING_UNSPECIFIED\x10\x00\x12\x1c\n\x18SPEAKER_LABELING_ENABLED\x10\x01\x12\x1d\n\x19SPEAKER_LABELING_DISABLED\x10\x02\"\xee\x02\n\x10StreamingOptions\x12\x44\n\x11recognition_model\x18\x01 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12>\n\x0e\x65ou_classifier\x18\x02 \x01(\x0b\x32&.speechkit.stt.v3.EouClassifierOptions\x12N\n\x16recognition_classifier\x18\x03 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x05 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptions\"\x1a\n\nAudioChunk\x12\x0c\n\x04\x64\x61ta\x18\x01 \x01(\x0c\"#\n\x0cSilenceChunk\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\x03\"\x05\n\x03\x45ou\"\xe8\x01\n\x10StreamingRequest\x12=\n\x0fsession_options\x18\x01 \x01(\x0b\x32\".speechkit.stt.v3.StreamingOptionsH\x00\x12-\n\x05\x63hunk\x18\x02 \x01(\x0b\x32\x1c.speechkit.stt.v3.AudioChunkH\x00\x12\x37\n\rsilence_chunk\x18\x03 \x01(\x0b\x32\x1e.speechkit.stt.v3.SilenceChunkH\x00\x12$\n\x03\x65ou\x18\x04 \x01(\x0b\x32\x15.speechkit.stt.v3.EouH\x00\x42\x07\n\x05\x45vent\"\xe3\x02\n\x14RecognizeFileRequest\x12\x11\n\x07\x63ontent\x18\x01 \x01(\x0cH\x00\x12\r\n\x03uri\x18\x02 \x01(\tH\x00\x12\x44\n\x11recognition_model\x18\x03 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12N\n\x16recognition_classifier\x18\x04 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x05 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x06 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptionsB\r\n\x0b\x41udioSource\"@\n\x04Word\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"@\n\x12LanguageEstimation\x12\x15\n\rlanguage_code\x18\x01 \x01(\t\x12\x13\n\x0bprobability\x18\x02 \x01(\x01\"\xbb\x01\n\x0b\x41lternative\x12%\n\x05words\x18\x01 \x03(\x0b\x32\x16.speechkit.stt.v3.Word\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x15\n\rstart_time_ms\x18\x03 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x04 \x01(\x03\x12\x12\n\nconfidence\x18\x05 \x01(\x01\x12\x37\n\tlanguages\x18\x06 \x03(\x0b\x32$.speechkit.stt.v3.LanguageEstimation\"\x1c\n\tEouUpdate\x12\x0f\n\x07time_ms\x18\x02 \x01(\x03\"a\n\x11\x41lternativeUpdate\x12\x33\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32\x1d.speechkit.stt.v3.Alternative\x12\x17\n\x0b\x63hannel_tag\x18\x02 \x01(\tB\x02\x18\x01\"\x99\x01\n\x0c\x41udioCursors\x12\x18\n\x10received_data_ms\x18\x01 \x01(\x03\x12\x15\n\rreset_time_ms\x18\x02 \x01(\x03\x12\x17\n\x0fpartial_time_ms\x18\x03 \x01(\x03\x12\x15\n\rfinal_time_ms\x18\x04 \x01(\x03\x12\x13\n\x0b\x66inal_index\x18\x05 \x01(\x03\x12\x13\n\x0b\x65ou_time_ms\x18\x06 \x01(\x03\"n\n\x0f\x46inalRefinement\x12\x13\n\x0b\x66inal_index\x18\x01 \x01(\x03\x12>\n\x0fnormalized_text\x18\x02 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x42\x06\n\x04Type\"L\n\nStatusCode\x12-\n\tcode_type\x18\x01 \x01(\x0e\x32\x1a.speechkit.stt.v3.CodeType\x12\x0f\n\x07message\x18\x02 \x01(\t\"4\n\x0bSessionUuid\x12\x0c\n\x04uuid\x18\x01 \x01(\t\x12\x17\n\x0fuser_request_id\x18\x02 \x01(\t\"K\n\x0fPhraseHighlight\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"?\n\x1aRecognitionClassifierLabel\x12\r\n\x05label\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x01\"\xa6\x01\n\x1bRecognitionClassifierResult\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x35\n\nhighlights\x18\x02 \x03(\x0b\x32!.speechkit.stt.v3.PhraseHighlight\x12<\n\x06labels\x18\x03 \x03(\x0b\x32,.speechkit.stt.v3.RecognitionClassifierLabel\"\xc7\x02\n\x1bRecognitionClassifierUpdate\x12M\n\x0bwindow_type\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.RecognitionClassifierUpdate.WindowType\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\x12H\n\x11\x63lassifier_result\x18\x04 \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierResult\"c\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x01\x12\x0e\n\nLAST_FINAL\x10\x02\x12\x10\n\x0cLAST_PARTIAL\x10\x03\"\xbb\x01\n\x15\x44\x65scriptiveStatistics\x12\x0b\n\x03min\x18\x01 \x01(\x01\x12\x0b\n\x03max\x18\x02 \x01(\x01\x12\x0c\n\x04mean\x18\x03 \x01(\x01\x12\x0b\n\x03std\x18\x04 \x01(\x01\x12\x43\n\tquantiles\x18\x05 \x03(\x0b\x32\x30.speechkit.stt.v3.DescriptiveStatistics.Quantile\x1a(\n\x08Quantile\x12\r\n\x05level\x18\x01 \x01(\x01\x12\r\n\x05value\x18\x02 \x01(\x01\"D\n\x16\x41udioSegmentBoundaries\x12\x15\n\rstart_time_ms\x18\x01 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x02 \x01(\x03\"\x87\x06\n\x0fSpeakerAnalysis\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x41\n\x0bwindow_type\x18\x02 \x01(\x0e\x32,.speechkit.stt.v3.SpeakerAnalysis.WindowType\x12\x43\n\x11speech_boundaries\x18\x03 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12\x17\n\x0ftotal_speech_ms\x18\x04 \x01(\x03\x12\x14\n\x0cspeech_ratio\x18\x05 \x01(\x01\x12\x18\n\x10total_silence_ms\x18\x06 \x01(\x03\x12\x15\n\rsilence_ratio\x18\x07 \x01(\x01\x12\x13\n\x0bwords_count\x18\x08 \x01(\x03\x12\x15\n\rletters_count\x18\t \x01(\x03\x12\x41\n\x10words_per_second\x18\n \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x43\n\x12letters_per_second\x18\x0b \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x44\n\x13words_per_utterance\x18\x0c \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x46\n\x15letters_per_utterance\x18\r \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x17\n\x0futterance_count\x18\x0e \x01(\x03\x12N\n\x1dutterance_duration_estimation\x18\x0f \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\"L\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\t\n\x05TOTAL\x10\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x02\"\x85\x06\n\x14\x43onversationAnalysis\x12I\n\x17\x63onversation_boundaries\x18\x01 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12.\n&total_simultaneous_silence_duration_ms\x18\x02 \x01(\x03\x12(\n total_simultaneous_silence_ratio\x18\x03 \x01(\x01\x12Y\n(simultaneous_silence_duration_estimation\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12-\n%total_simultaneous_speech_duration_ms\x18\x05 \x01(\x03\x12\'\n\x1ftotal_simultaneous_speech_ratio\x18\x06 \x01(\x01\x12X\n\'simultaneous_speech_duration_estimation\x18\x07 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12W\n\x12speaker_interrupts\x18\x08 \x03(\x0b\x32;.speechkit.stt.v3.ConversationAnalysis.InterruptsEvaluation\x12 \n\x18total_speech_duration_ms\x18\t \x01(\x03\x12\x1a\n\x12total_speech_ratio\x18\n \x01(\x01\x1a\xa3\x01\n\x14InterruptsEvaluation\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x18\n\x10interrupts_count\x18\x02 \x01(\x03\x12\x1e\n\x16interrupts_duration_ms\x18\x03 \x01(\x03\x12<\n\ninterrupts\x18\x04 \x03(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\"\xa5\x05\n\x11StreamingResponse\x12\x33\n\x0csession_uuid\x18\x01 \x01(\x0b\x32\x1d.speechkit.stt.v3.SessionUuid\x12\x35\n\raudio_cursors\x18\x02 \x01(\x0b\x32\x1e.speechkit.stt.v3.AudioCursors\x12\x1d\n\x15response_wall_time_ms\x18\x03 \x01(\x03\x12\x36\n\x07partial\x18\x04 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x34\n\x05\x66inal\x18\x05 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x31\n\neou_update\x18\x06 \x01(\x0b\x32\x1b.speechkit.stt.v3.EouUpdateH\x00\x12=\n\x10\x66inal_refinement\x18\x07 \x01(\x0b\x32!.speechkit.stt.v3.FinalRefinementH\x00\x12\x33\n\x0bstatus_code\x18\x08 \x01(\x0b\x32\x1c.speechkit.stt.v3.StatusCodeH\x00\x12J\n\x11\x63lassifier_update\x18\n \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierUpdateH\x00\x12=\n\x10speaker_analysis\x18\x0b \x01(\x0b\x32!.speechkit.stt.v3.SpeakerAnalysisH\x00\x12G\n\x15\x63onversation_analysis\x18\x0c \x01(\x0b\x32&.speechkit.stt.v3.ConversationAnalysisH\x00\x12\x13\n\x0b\x63hannel_tag\x18\t \x01(\tB\x07\n\x05\x45vent*O\n\x08\x43odeType\x12\x1d\n\x15\x43ODE_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x0b\n\x07WORKING\x10\x01\x12\x0b\n\x07WARNING\x10\x02\x12\n\n\x06\x43LOSED\x10\x03\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
|
|
16
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n yandex/cloud/ai/stt/v3/stt.proto\x12\x10speechkit.stt.v3\"\xe2\x03\n\x18TextNormalizationOptions\x12X\n\x12text_normalization\x18\x01 \x01(\x0e\x32<.speechkit.stt.v3.TextNormalizationOptions.TextNormalization\x12\x18\n\x10profanity_filter\x18\x02 \x01(\x08\x12\x17\n\x0fliterature_text\x18\x03 \x01(\x08\x12]\n\x15phone_formatting_mode\x18\x04 \x01(\x0e\x32>.speechkit.stt.v3.TextNormalizationOptions.PhoneFormattingMode\"x\n\x11TextNormalization\x12\"\n\x1eTEXT_NORMALIZATION_UNSPECIFIED\x10\x00\x12\x1e\n\x1aTEXT_NORMALIZATION_ENABLED\x10\x01\x12\x1f\n\x1bTEXT_NORMALIZATION_DISABLED\x10\x02\"`\n\x13PhoneFormattingMode\x12%\n!PHONE_FORMATTING_MODE_UNSPECIFIED\x10\x00\x12\"\n\x1ePHONE_FORMATTING_MODE_DISABLED\x10\x01\"\xce\x01\n\x14\x44\x65\x66\x61ultEouClassifier\x12\x43\n\x04type\x18\x01 \x01(\x0e\x32\x35.speechkit.stt.v3.DefaultEouClassifier.EouSensitivity\x12\'\n\x1fmax_pause_between_words_hint_ms\x18\x02 \x01(\x03\"H\n\x0e\x45ouSensitivity\x12\x1f\n\x1b\x45OU_SENSITIVITY_UNSPECIFIED\x10\x00\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x01\x12\x08\n\x04HIGH\x10\x02\"\x17\n\x15\x45xternalEouClassifier\"\xb2\x01\n\x14\x45ouClassifierOptions\x12\x44\n\x12\x64\x65\x66\x61ult_classifier\x18\x01 \x01(\x0b\x32&.speechkit.stt.v3.DefaultEouClassifierH\x00\x12\x46\n\x13\x65xternal_classifier\x18\x02 \x01(\x0b\x32\'.speechkit.stt.v3.ExternalEouClassifierH\x00\x42\x0c\n\nClassifier\"\xd3\x01\n\x15RecognitionClassifier\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x45\n\x08triggers\x18\x02 \x03(\x0e\x32\x33.speechkit.stt.v3.RecognitionClassifier.TriggerType\"_\n\x0bTriggerType\x12 \n\x18TRIGGER_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x10\n\x0cON_UTTERANCE\x10\x01\x12\x0c\n\x08ON_FINAL\x10\x02\x12\x0e\n\nON_PARTIAL\x10\x03\"\\\n\x1cRecognitionClassifierOptions\x12<\n\x0b\x63lassifiers\x18\x01 \x03(\x0b\x32\'.speechkit.stt.v3.RecognitionClassifier\"\x88\x01\n\x15SpeechAnalysisOptions\x12\x1f\n\x17\x65nable_speaker_analysis\x18\x01 \x01(\x08\x12$\n\x1c\x65nable_conversation_analysis\x18\x02 \x01(\x08\x12(\n descriptive_statistics_quantiles\x18\x03 \x03(\x01\"\xc7\x01\n\x08RawAudio\x12@\n\x0e\x61udio_encoding\x18\x01 \x01(\x0e\x32(.speechkit.stt.v3.RawAudio.AudioEncoding\x12\x19\n\x11sample_rate_hertz\x18\x02 \x01(\x03\x12\x1b\n\x13\x61udio_channel_count\x18\x03 \x01(\x03\"A\n\rAudioEncoding\x12\x1e\n\x1a\x41UDIO_ENCODING_UNSPECIFIED\x10\x00\x12\x10\n\x0cLINEAR16_PCM\x10\x01\"\xbf\x01\n\x0e\x43ontainerAudio\x12Q\n\x14\x63ontainer_audio_type\x18\x01 \x01(\x0e\x32\x33.speechkit.stt.v3.ContainerAudio.ContainerAudioType\"Z\n\x12\x43ontainerAudioType\x12$\n CONTAINER_AUDIO_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03WAV\x10\x01\x12\x0c\n\x08OGG_OPUS\x10\x02\x12\x07\n\x03MP3\x10\x03\"\x91\x01\n\x12\x41udioFormatOptions\x12/\n\traw_audio\x18\x01 \x01(\x0b\x32\x1a.speechkit.stt.v3.RawAudioH\x00\x12;\n\x0f\x63ontainer_audio\x18\x02 \x01(\x0b\x32 .speechkit.stt.v3.ContainerAudioH\x00\x42\r\n\x0b\x41udioFormat\"\xf7\x01\n\x1aLanguageRestrictionOptions\x12^\n\x10restriction_type\x18\x01 \x01(\x0e\x32\x44.speechkit.stt.v3.LanguageRestrictionOptions.LanguageRestrictionType\x12\x15\n\rlanguage_code\x18\x02 \x03(\t\"b\n\x17LanguageRestrictionType\x12)\n%LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tWHITELIST\x10\x01\x12\r\n\tBLACKLIST\x10\x02\"\xb2\x03\n\x17RecognitionModelOptions\x12\r\n\x05model\x18\x01 \x01(\t\x12:\n\x0c\x61udio_format\x18\x02 \x01(\x0b\x32$.speechkit.stt.v3.AudioFormatOptions\x12\x46\n\x12text_normalization\x18\x03 \x01(\x0b\x32*.speechkit.stt.v3.TextNormalizationOptions\x12J\n\x14language_restriction\x18\x04 \x01(\x0b\x32,.speechkit.stt.v3.LanguageRestrictionOptions\x12\\\n\x15\x61udio_processing_type\x18\x05 \x01(\x0e\x32=.speechkit.stt.v3.RecognitionModelOptions.AudioProcessingType\"Z\n\x13\x41udioProcessingType\x12%\n!AUDIO_PROCESSING_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tREAL_TIME\x10\x01\x12\r\n\tFULL_DATA\x10\x02\"\xde\x01\n\x16SpeakerLabelingOptions\x12R\n\x10speaker_labeling\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.SpeakerLabelingOptions.SpeakerLabeling\"p\n\x0fSpeakerLabeling\x12 \n\x1cSPEAKER_LABELING_UNSPECIFIED\x10\x00\x12\x1c\n\x18SPEAKER_LABELING_ENABLED\x10\x01\x12\x1d\n\x19SPEAKER_LABELING_DISABLED\x10\x02\"\xee\x02\n\x10StreamingOptions\x12\x44\n\x11recognition_model\x18\x01 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12>\n\x0e\x65ou_classifier\x18\x02 \x01(\x0b\x32&.speechkit.stt.v3.EouClassifierOptions\x12N\n\x16recognition_classifier\x18\x03 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x05 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptions\"\x1a\n\nAudioChunk\x12\x0c\n\x04\x64\x61ta\x18\x01 \x01(\x0c\"#\n\x0cSilenceChunk\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\x03\"\x05\n\x03\x45ou\"\xe8\x01\n\x10StreamingRequest\x12=\n\x0fsession_options\x18\x01 \x01(\x0b\x32\".speechkit.stt.v3.StreamingOptionsH\x00\x12-\n\x05\x63hunk\x18\x02 \x01(\x0b\x32\x1c.speechkit.stt.v3.AudioChunkH\x00\x12\x37\n\rsilence_chunk\x18\x03 \x01(\x0b\x32\x1e.speechkit.stt.v3.SilenceChunkH\x00\x12$\n\x03\x65ou\x18\x04 \x01(\x0b\x32\x15.speechkit.stt.v3.EouH\x00\x42\x07\n\x05\x45vent\"\xe3\x02\n\x14RecognizeFileRequest\x12\x11\n\x07\x63ontent\x18\x01 \x01(\x0cH\x00\x12\r\n\x03uri\x18\x02 \x01(\tH\x00\x12\x44\n\x11recognition_model\x18\x03 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12N\n\x16recognition_classifier\x18\x04 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x05 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x06 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptionsB\r\n\x0b\x41udioSource\"@\n\x04Word\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"@\n\x12LanguageEstimation\x12\x15\n\rlanguage_code\x18\x01 \x01(\t\x12\x13\n\x0bprobability\x18\x02 \x01(\x01\"\xbb\x01\n\x0b\x41lternative\x12%\n\x05words\x18\x01 \x03(\x0b\x32\x16.speechkit.stt.v3.Word\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x15\n\rstart_time_ms\x18\x03 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x04 \x01(\x03\x12\x12\n\nconfidence\x18\x05 \x01(\x01\x12\x37\n\tlanguages\x18\x06 \x03(\x0b\x32$.speechkit.stt.v3.LanguageEstimation\"\x1c\n\tEouUpdate\x12\x0f\n\x07time_ms\x18\x02 \x01(\x03\"a\n\x11\x41lternativeUpdate\x12\x33\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32\x1d.speechkit.stt.v3.Alternative\x12\x17\n\x0b\x63hannel_tag\x18\x02 \x01(\tB\x02\x18\x01\"\x99\x01\n\x0c\x41udioCursors\x12\x18\n\x10received_data_ms\x18\x01 \x01(\x03\x12\x15\n\rreset_time_ms\x18\x02 \x01(\x03\x12\x17\n\x0fpartial_time_ms\x18\x03 \x01(\x03\x12\x15\n\rfinal_time_ms\x18\x04 \x01(\x03\x12\x13\n\x0b\x66inal_index\x18\x05 \x01(\x03\x12\x13\n\x0b\x65ou_time_ms\x18\x06 \x01(\x03\"n\n\x0f\x46inalRefinement\x12\x13\n\x0b\x66inal_index\x18\x01 \x01(\x03\x12>\n\x0fnormalized_text\x18\x02 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x42\x06\n\x04Type\"L\n\nStatusCode\x12-\n\tcode_type\x18\x01 \x01(\x0e\x32\x1a.speechkit.stt.v3.CodeType\x12\x0f\n\x07message\x18\x02 \x01(\t\"4\n\x0bSessionUuid\x12\x0c\n\x04uuid\x18\x01 \x01(\t\x12\x17\n\x0fuser_request_id\x18\x02 \x01(\t\"K\n\x0fPhraseHighlight\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"?\n\x1aRecognitionClassifierLabel\x12\r\n\x05label\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x01\"\xa6\x01\n\x1bRecognitionClassifierResult\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x35\n\nhighlights\x18\x02 \x03(\x0b\x32!.speechkit.stt.v3.PhraseHighlight\x12<\n\x06labels\x18\x03 \x03(\x0b\x32,.speechkit.stt.v3.RecognitionClassifierLabel\"\xc7\x02\n\x1bRecognitionClassifierUpdate\x12M\n\x0bwindow_type\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.RecognitionClassifierUpdate.WindowType\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\x12H\n\x11\x63lassifier_result\x18\x04 \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierResult\"c\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x01\x12\x0e\n\nLAST_FINAL\x10\x02\x12\x10\n\x0cLAST_PARTIAL\x10\x03\"\xbb\x01\n\x15\x44\x65scriptiveStatistics\x12\x0b\n\x03min\x18\x01 \x01(\x01\x12\x0b\n\x03max\x18\x02 \x01(\x01\x12\x0c\n\x04mean\x18\x03 \x01(\x01\x12\x0b\n\x03std\x18\x04 \x01(\x01\x12\x43\n\tquantiles\x18\x05 \x03(\x0b\x32\x30.speechkit.stt.v3.DescriptiveStatistics.Quantile\x1a(\n\x08Quantile\x12\r\n\x05level\x18\x01 \x01(\x01\x12\r\n\x05value\x18\x02 \x01(\x01\"D\n\x16\x41udioSegmentBoundaries\x12\x15\n\rstart_time_ms\x18\x01 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x02 \x01(\x03\"\x87\x06\n\x0fSpeakerAnalysis\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x41\n\x0bwindow_type\x18\x02 \x01(\x0e\x32,.speechkit.stt.v3.SpeakerAnalysis.WindowType\x12\x43\n\x11speech_boundaries\x18\x03 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12\x17\n\x0ftotal_speech_ms\x18\x04 \x01(\x03\x12\x14\n\x0cspeech_ratio\x18\x05 \x01(\x01\x12\x18\n\x10total_silence_ms\x18\x06 \x01(\x03\x12\x15\n\rsilence_ratio\x18\x07 \x01(\x01\x12\x13\n\x0bwords_count\x18\x08 \x01(\x03\x12\x15\n\rletters_count\x18\t \x01(\x03\x12\x41\n\x10words_per_second\x18\n \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x43\n\x12letters_per_second\x18\x0b \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x44\n\x13words_per_utterance\x18\x0c \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x46\n\x15letters_per_utterance\x18\r \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x17\n\x0futterance_count\x18\x0e \x01(\x03\x12N\n\x1dutterance_duration_estimation\x18\x0f \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\"L\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\t\n\x05TOTAL\x10\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x02\"\x85\x06\n\x14\x43onversationAnalysis\x12I\n\x17\x63onversation_boundaries\x18\x01 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12.\n&total_simultaneous_silence_duration_ms\x18\x02 \x01(\x03\x12(\n total_simultaneous_silence_ratio\x18\x03 \x01(\x01\x12Y\n(simultaneous_silence_duration_estimation\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12-\n%total_simultaneous_speech_duration_ms\x18\x05 \x01(\x03\x12\'\n\x1ftotal_simultaneous_speech_ratio\x18\x06 \x01(\x01\x12X\n\'simultaneous_speech_duration_estimation\x18\x07 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12W\n\x12speaker_interrupts\x18\x08 \x03(\x0b\x32;.speechkit.stt.v3.ConversationAnalysis.InterruptsEvaluation\x12 \n\x18total_speech_duration_ms\x18\t \x01(\x03\x12\x1a\n\x12total_speech_ratio\x18\n \x01(\x01\x1a\xa3\x01\n\x14InterruptsEvaluation\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x18\n\x10interrupts_count\x18\x02 \x01(\x03\x12\x1e\n\x16interrupts_duration_ms\x18\x03 \x01(\x03\x12<\n\ninterrupts\x18\x04 \x03(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\"\xa5\x05\n\x11StreamingResponse\x12\x33\n\x0csession_uuid\x18\x01 \x01(\x0b\x32\x1d.speechkit.stt.v3.SessionUuid\x12\x35\n\raudio_cursors\x18\x02 \x01(\x0b\x32\x1e.speechkit.stt.v3.AudioCursors\x12\x1d\n\x15response_wall_time_ms\x18\x03 \x01(\x03\x12\x36\n\x07partial\x18\x04 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x34\n\x05\x66inal\x18\x05 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x31\n\neou_update\x18\x06 \x01(\x0b\x32\x1b.speechkit.stt.v3.EouUpdateH\x00\x12=\n\x10\x66inal_refinement\x18\x07 \x01(\x0b\x32!.speechkit.stt.v3.FinalRefinementH\x00\x12\x33\n\x0bstatus_code\x18\x08 \x01(\x0b\x32\x1c.speechkit.stt.v3.StatusCodeH\x00\x12J\n\x11\x63lassifier_update\x18\n \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierUpdateH\x00\x12=\n\x10speaker_analysis\x18\x0b \x01(\x0b\x32!.speechkit.stt.v3.SpeakerAnalysisH\x00\x12G\n\x15\x63onversation_analysis\x18\x0c \x01(\x0b\x32&.speechkit.stt.v3.ConversationAnalysisH\x00\x12\x13\n\x0b\x63hannel_tag\x18\t \x01(\tB\x07\n\x05\x45vent\"0\n\x18\x44\x65leteRecognitionRequest\x12\x14\n\x0coperation_id\x18\x01 \x01(\t*O\n\x08\x43odeType\x12\x1d\n\x15\x43ODE_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x0b\n\x07WORKING\x10\x01\x12\x0b\n\x07WARNING\x10\x02\x12\n\n\x06\x43LOSED\x10\x03\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
|
|
17
17
|
|
|
18
18
|
_globals = globals()
|
|
19
19
|
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
@@ -31,8 +31,8 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
31
31
|
_RECOGNITIONCLASSIFIERUPDATE_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._serialized_options = b'\010\001'
|
|
32
32
|
_SPEAKERANALYSIS_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._options = None
|
|
33
33
|
_SPEAKERANALYSIS_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._serialized_options = b'\010\001'
|
|
34
|
-
_globals['_CODETYPE']._serialized_start=
|
|
35
|
-
_globals['_CODETYPE']._serialized_end=
|
|
34
|
+
_globals['_CODETYPE']._serialized_start=7927
|
|
35
|
+
_globals['_CODETYPE']._serialized_end=8006
|
|
36
36
|
_globals['_TEXTNORMALIZATIONOPTIONS']._serialized_start=55
|
|
37
37
|
_globals['_TEXTNORMALIZATIONOPTIONS']._serialized_end=537
|
|
38
38
|
_globals['_TEXTNORMALIZATIONOPTIONS_TEXTNORMALIZATION']._serialized_start=319
|
|
@@ -133,4 +133,6 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
133
133
|
_globals['_CONVERSATIONANALYSIS_INTERRUPTSEVALUATION']._serialized_end=7195
|
|
134
134
|
_globals['_STREAMINGRESPONSE']._serialized_start=7198
|
|
135
135
|
_globals['_STREAMINGRESPONSE']._serialized_end=7875
|
|
136
|
+
_globals['_DELETERECOGNITIONREQUEST']._serialized_start=7877
|
|
137
|
+
_globals['_DELETERECOGNITIONREQUEST']._serialized_end=7925
|
|
136
138
|
# @@protoc_insertion_point(module_scope)
|
|
@@ -1435,3 +1435,18 @@ class StreamingResponse(google.protobuf.message.Message):
|
|
|
1435
1435
|
def WhichOneof(self, oneof_group: typing.Literal["Event", b"Event"]) -> typing.Literal["partial", "final", "eou_update", "final_refinement", "status_code", "classifier_update", "speaker_analysis", "conversation_analysis"] | None: ...
|
|
1436
1436
|
|
|
1437
1437
|
global___StreamingResponse = StreamingResponse
|
|
1438
|
+
|
|
1439
|
+
@typing.final
|
|
1440
|
+
class DeleteRecognitionRequest(google.protobuf.message.Message):
|
|
1441
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
1442
|
+
|
|
1443
|
+
OPERATION_ID_FIELD_NUMBER: builtins.int
|
|
1444
|
+
operation_id: builtins.str
|
|
1445
|
+
def __init__(
|
|
1446
|
+
self,
|
|
1447
|
+
*,
|
|
1448
|
+
operation_id: builtins.str = ...,
|
|
1449
|
+
) -> None: ...
|
|
1450
|
+
def ClearField(self, field_name: typing.Literal["operation_id", b"operation_id"]) -> None: ...
|
|
1451
|
+
|
|
1452
|
+
global___DeleteRecognitionRequest = DeleteRecognitionRequest
|
|
@@ -13,12 +13,13 @@ _sym_db = _symbol_database.Default()
|
|
|
13
13
|
|
|
14
14
|
from yandex.cloud.ai.stt.v3 import stt_pb2 as yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2
|
|
15
15
|
from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2
|
|
16
|
+
from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2
|
|
16
17
|
from yandex.cloud import validation_pb2 as yandex_dot_cloud_dot_validation__pb2
|
|
17
18
|
from yandex.cloud.api import operation_pb2 as yandex_dot_cloud_dot_api_dot_operation__pb2
|
|
18
19
|
from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operation_dot_operation__pb2
|
|
19
20
|
|
|
20
21
|
|
|
21
|
-
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n(yandex/cloud/ai/stt/v3/stt_service.proto\x12\x10speechkit.stt.v3\x1a yandex/cloud/ai/stt/v3/stt.proto\x1a\x1cgoogle/api/annotations.proto\x1a\x1dyandex/cloud/validation.proto\x1a yandex/cloud/api/operation.proto\x1a&yandex/cloud/operation/operation.proto\";\n\x15GetRecognitionRequest\x12\"\n\x0coperation_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=502q\n\nRecognizer\x12\x63\n\x12RecognizeStreaming\x12\".speechkit.stt.v3.StreamingRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x00(\x01\x30\x01\x32\
|
|
22
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n(yandex/cloud/ai/stt/v3/stt_service.proto\x12\x10speechkit.stt.v3\x1a yandex/cloud/ai/stt/v3/stt.proto\x1a\x1cgoogle/api/annotations.proto\x1a\x1bgoogle/protobuf/empty.proto\x1a\x1dyandex/cloud/validation.proto\x1a yandex/cloud/api/operation.proto\x1a&yandex/cloud/operation/operation.proto\";\n\x15GetRecognitionRequest\x12\"\n\x0coperation_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=502q\n\nRecognizer\x12\x63\n\x12RecognizeStreaming\x12\".speechkit.stt.v3.StreamingRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x00(\x01\x30\x01\x32\xaf\x03\n\x0f\x41syncRecognizer\x12\x9c\x01\n\rRecognizeFile\x12&.speechkit.stt.v3.RecognizeFileRequest\x1a!.yandex.cloud.operation.Operation\"@\xb2\xd2*\x17\x12\x15google.protobuf.Empty\x82\xd3\xe4\x93\x02\x1f\"\x1a/stt/v3/recognizeFileAsync:\x01*\x12\x80\x01\n\x0eGetRecognition\x12\'.speechkit.stt.v3.GetRecognitionRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x1e\x82\xd3\xe4\x93\x02\x18\x12\x16/stt/v3/getRecognition0\x01\x12z\n\x11\x44\x65leteRecognition\x12*.speechkit.stt.v3.DeleteRecognitionRequest\x1a\x16.google.protobuf.Empty\"!\x82\xd3\xe4\x93\x02\x1b*\x19/stt/v3/deleteRecognitionB\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
|
|
22
23
|
|
|
23
24
|
_globals = globals()
|
|
24
25
|
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
@@ -32,10 +33,12 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
32
33
|
_ASYNCRECOGNIZER.methods_by_name['RecognizeFile']._serialized_options = b'\262\322*\027\022\025google.protobuf.Empty\202\323\344\223\002\037\"\032/stt/v3/recognizeFileAsync:\001*'
|
|
33
34
|
_ASYNCRECOGNIZER.methods_by_name['GetRecognition']._options = None
|
|
34
35
|
_ASYNCRECOGNIZER.methods_by_name['GetRecognition']._serialized_options = b'\202\323\344\223\002\030\022\026/stt/v3/getRecognition'
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
_globals['
|
|
38
|
-
_globals['
|
|
39
|
-
_globals['
|
|
40
|
-
_globals['
|
|
36
|
+
_ASYNCRECOGNIZER.methods_by_name['DeleteRecognition']._options = None
|
|
37
|
+
_ASYNCRECOGNIZER.methods_by_name['DeleteRecognition']._serialized_options = b'\202\323\344\223\002\033*\031/stt/v3/deleteRecognition'
|
|
38
|
+
_globals['_GETRECOGNITIONREQUEST']._serialized_start=260
|
|
39
|
+
_globals['_GETRECOGNITIONREQUEST']._serialized_end=319
|
|
40
|
+
_globals['_RECOGNIZER']._serialized_start=321
|
|
41
|
+
_globals['_RECOGNIZER']._serialized_end=434
|
|
42
|
+
_globals['_ASYNCRECOGNIZER']._serialized_start=437
|
|
43
|
+
_globals['_ASYNCRECOGNIZER']._serialized_end=868
|
|
41
44
|
# @@protoc_insertion_point(module_scope)
|
|
@@ -2,6 +2,7 @@
|
|
|
2
2
|
"""Client and server classes corresponding to protobuf-defined services."""
|
|
3
3
|
import grpc
|
|
4
4
|
|
|
5
|
+
from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2
|
|
5
6
|
from yandex.cloud.ai.stt.v3 import stt_pb2 as yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2
|
|
6
7
|
from yandex.cloud.ai.stt.v3 import stt_service_pb2 as yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__service__pb2
|
|
7
8
|
from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operation_dot_operation__pb2
|
|
@@ -92,6 +93,11 @@ class AsyncRecognizerStub(object):
|
|
|
92
93
|
request_serializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__service__pb2.GetRecognitionRequest.SerializeToString,
|
|
93
94
|
response_deserializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.StreamingResponse.FromString,
|
|
94
95
|
)
|
|
96
|
+
self.DeleteRecognition = channel.unary_unary(
|
|
97
|
+
'/speechkit.stt.v3.AsyncRecognizer/DeleteRecognition',
|
|
98
|
+
request_serializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.DeleteRecognitionRequest.SerializeToString,
|
|
99
|
+
response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString,
|
|
100
|
+
)
|
|
95
101
|
|
|
96
102
|
|
|
97
103
|
class AsyncRecognizerServicer(object):
|
|
@@ -110,6 +116,12 @@ class AsyncRecognizerServicer(object):
|
|
|
110
116
|
context.set_details('Method not implemented!')
|
|
111
117
|
raise NotImplementedError('Method not implemented!')
|
|
112
118
|
|
|
119
|
+
def DeleteRecognition(self, request, context):
|
|
120
|
+
"""Missing associated documentation comment in .proto file."""
|
|
121
|
+
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
122
|
+
context.set_details('Method not implemented!')
|
|
123
|
+
raise NotImplementedError('Method not implemented!')
|
|
124
|
+
|
|
113
125
|
|
|
114
126
|
def add_AsyncRecognizerServicer_to_server(servicer, server):
|
|
115
127
|
rpc_method_handlers = {
|
|
@@ -123,6 +135,11 @@ def add_AsyncRecognizerServicer_to_server(servicer, server):
|
|
|
123
135
|
request_deserializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__service__pb2.GetRecognitionRequest.FromString,
|
|
124
136
|
response_serializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.StreamingResponse.SerializeToString,
|
|
125
137
|
),
|
|
138
|
+
'DeleteRecognition': grpc.unary_unary_rpc_method_handler(
|
|
139
|
+
servicer.DeleteRecognition,
|
|
140
|
+
request_deserializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.DeleteRecognitionRequest.FromString,
|
|
141
|
+
response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString,
|
|
142
|
+
),
|
|
126
143
|
}
|
|
127
144
|
generic_handler = grpc.method_handlers_generic_handler(
|
|
128
145
|
'speechkit.stt.v3.AsyncRecognizer', rpc_method_handlers)
|
|
@@ -167,3 +184,20 @@ class AsyncRecognizer(object):
|
|
|
167
184
|
yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.StreamingResponse.FromString,
|
|
168
185
|
options, channel_credentials,
|
|
169
186
|
insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
|
|
187
|
+
|
|
188
|
+
@staticmethod
|
|
189
|
+
def DeleteRecognition(request,
|
|
190
|
+
target,
|
|
191
|
+
options=(),
|
|
192
|
+
channel_credentials=None,
|
|
193
|
+
call_credentials=None,
|
|
194
|
+
insecure=False,
|
|
195
|
+
compression=None,
|
|
196
|
+
wait_for_ready=None,
|
|
197
|
+
timeout=None,
|
|
198
|
+
metadata=None):
|
|
199
|
+
return grpc.experimental.unary_unary(request, target, '/speechkit.stt.v3.AsyncRecognizer/DeleteRecognition',
|
|
200
|
+
yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.DeleteRecognitionRequest.SerializeToString,
|
|
201
|
+
google_dot_protobuf_dot_empty__pb2.Empty.FromString,
|
|
202
|
+
options, channel_credentials,
|
|
203
|
+
insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
|
|
@@ -5,6 +5,7 @@ isort:skip_file
|
|
|
5
5
|
|
|
6
6
|
import abc
|
|
7
7
|
import collections.abc
|
|
8
|
+
import google.protobuf.empty_pb2
|
|
8
9
|
import grpc
|
|
9
10
|
import grpc.aio
|
|
10
11
|
import typing
|
|
@@ -65,6 +66,11 @@ class AsyncRecognizerStub:
|
|
|
65
66
|
yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse,
|
|
66
67
|
]
|
|
67
68
|
|
|
69
|
+
DeleteRecognition: grpc.UnaryUnaryMultiCallable[
|
|
70
|
+
yandex.cloud.ai.stt.v3.stt_pb2.DeleteRecognitionRequest,
|
|
71
|
+
google.protobuf.empty_pb2.Empty,
|
|
72
|
+
]
|
|
73
|
+
|
|
68
74
|
class AsyncRecognizerAsyncStub:
|
|
69
75
|
"""A set of methods for async voice recognition."""
|
|
70
76
|
|
|
@@ -78,6 +84,11 @@ class AsyncRecognizerAsyncStub:
|
|
|
78
84
|
yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse,
|
|
79
85
|
]
|
|
80
86
|
|
|
87
|
+
DeleteRecognition: grpc.aio.UnaryUnaryMultiCallable[
|
|
88
|
+
yandex.cloud.ai.stt.v3.stt_pb2.DeleteRecognitionRequest,
|
|
89
|
+
google.protobuf.empty_pb2.Empty,
|
|
90
|
+
]
|
|
91
|
+
|
|
81
92
|
class AsyncRecognizerServicer(metaclass=abc.ABCMeta):
|
|
82
93
|
"""A set of methods for async voice recognition."""
|
|
83
94
|
|
|
@@ -95,4 +106,11 @@ class AsyncRecognizerServicer(metaclass=abc.ABCMeta):
|
|
|
95
106
|
context: _ServicerContext,
|
|
96
107
|
) -> typing.Union[collections.abc.Iterator[yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse], collections.abc.AsyncIterator[yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse]]: ...
|
|
97
108
|
|
|
109
|
+
@abc.abstractmethod
|
|
110
|
+
def DeleteRecognition(
|
|
111
|
+
self,
|
|
112
|
+
request: yandex.cloud.ai.stt.v3.stt_pb2.DeleteRecognitionRequest,
|
|
113
|
+
context: _ServicerContext,
|
|
114
|
+
) -> typing.Union[google.protobuf.empty_pb2.Empty, collections.abc.Awaitable[google.protobuf.empty_pb2.Empty]]: ...
|
|
115
|
+
|
|
98
116
|
def add_AsyncRecognizerServicer_to_server(servicer: AsyncRecognizerServicer, server: typing.Union[grpc.Server, grpc.aio.Server]) -> None: ...
|