yandexcloud 0.323.0__py3-none-any.whl → 0.325.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- yandex/cloud/ai/foundation_models/v1/embedding/embedding_service_pb2.pyi +1 -1
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_pb2.pyi +4 -2
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2.pyi +6 -3
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2_grpc.py +3 -3
- yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2_grpc.pyi +3 -3
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_pb2.pyi +3 -3
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2.pyi +25 -10
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.py +11 -5
- yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.pyi +33 -9
- yandex/cloud/ai/foundation_models/v1/text_common_pb2.py +27 -27
- yandex/cloud/ai/foundation_models/v1/text_common_pb2.pyi +18 -6
- yandex/cloud/ai/stt/v3/stt_pb2.py +5 -3
- yandex/cloud/ai/stt/v3/stt_pb2.pyi +15 -0
- yandex/cloud/ai/stt/v3/stt_service_pb2.py +10 -7
- yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.py +34 -0
- yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.pyi +18 -0
- yandex/cloud/apploadbalancer/v1/backend_group_pb2.py +26 -26
- yandex/cloud/apploadbalancer/v1/backend_group_pb2.pyi +7 -1
- yandex/cloud/backup/v1/backup_service_pb2.py +42 -41
- yandex/cloud/backup/v1/backup_service_pb2.pyi +6 -1
- yandex/cloud/backup/v1/policy_pb2.pyi +1 -1
- yandex/cloud/billing/v1/customer_service_pb2.pyi +5 -4
- yandex/cloud/cdn/v1/resource_service_pb2.py +14 -14
- yandex/cloud/cdn/v1/resource_service_pb2.pyi +5 -1
- yandex/cloud/cloudregistry/__init__.py +0 -0
- yandex/cloud/cloudregistry/v1/__init__.py +0 -0
- yandex/cloud/cloudregistry/v1/artifact_pb2.py +31 -0
- yandex/cloud/cloudregistry/v1/artifact_pb2.pyi +112 -0
- yandex/cloud/cloudregistry/v1/artifact_pb2_grpc.py +4 -0
- yandex/cloud/cloudregistry/v1/artifact_pb2_grpc.pyi +17 -0
- yandex/cloud/cloudregistry/v1/artifact_service_pb2.py +45 -0
- yandex/cloud/cloudregistry/v1/artifact_service_pb2.pyi +59 -0
- yandex/cloud/cloudregistry/v1/artifact_service_pb2_grpc.py +108 -0
- yandex/cloud/cloudregistry/v1/artifact_service_pb2_grpc.pyi +81 -0
- yandex/cloud/cloudregistry/v1/ip_permission_pb2.py +37 -0
- yandex/cloud/cloudregistry/v1/ip_permission_pb2.pyi +95 -0
- yandex/cloud/cloudregistry/v1/ip_permission_pb2_grpc.py +4 -0
- yandex/cloud/cloudregistry/v1/ip_permission_pb2_grpc.pyi +17 -0
- yandex/cloud/cloudregistry/v1/registry_pb2.py +41 -0
- yandex/cloud/cloudregistry/v1/registry_pb2.pyi +181 -0
- yandex/cloud/cloudregistry/v1/registry_pb2_grpc.py +4 -0
- yandex/cloud/cloudregistry/v1/registry_pb2_grpc.pyi +17 -0
- yandex/cloud/cloudregistry/v1/registry_service_pb2.py +157 -0
- yandex/cloud/cloudregistry/v1/registry_service_pb2.pyi +504 -0
- yandex/cloud/cloudregistry/v1/registry_service_pb2_grpc.py +451 -0
- yandex/cloud/cloudregistry/v1/registry_service_pb2_grpc.pyi +291 -0
- yandex/cloud/cloudrouter/v1/routing_instance_pb2.py +8 -10
- yandex/cloud/cloudrouter/v1/routing_instance_pb2.pyi +2 -28
- yandex/cloud/compute/v1/instance_service_pb2_grpc.py +1 -1
- yandex/cloud/compute/v1/instance_service_pb2_grpc.pyi +3 -3
- yandex/cloud/datasphere/v2/jobs/jobs_pb2.py +9 -5
- yandex/cloud/datasphere/v2/jobs/jobs_pb2.pyi +96 -0
- yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.py +55 -53
- yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.pyi +14 -2
- yandex/cloud/iam/v1/user_account_pb2.pyi +2 -2
- yandex/cloud/iam/v1/user_account_service_pb2_grpc.py +3 -3
- yandex/cloud/iam/v1/user_account_service_pb2_grpc.pyi +3 -3
- yandex/cloud/lockbox/v1/secret_service_pb2.py +56 -55
- yandex/cloud/lockbox/v1/secret_service_pb2.pyi +13 -2
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2.py +55 -45
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2.pyi +40 -0
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2_grpc.py +34 -0
- yandex/cloud/mdb/opensearch/v1/cluster_service_pb2_grpc.pyi +20 -0
- yandex/cloud/mdb/postgresql/v1/cluster_pb2.py +47 -43
- yandex/cloud/mdb/postgresql/v1/cluster_pb2.pyi +34 -6
- yandex/cloud/mdb/postgresql/v1/cluster_service_pb2.py +109 -105
- yandex/cloud/mdb/postgresql/v1/cluster_service_pb2.pyi +35 -7
- yandex/cloud/mdb/postgresql/v1/config/host17_1c_pb2.py +68 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_1c_pb2.pyi +512 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_1c_pb2_grpc.py +4 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_1c_pb2_grpc.pyi +17 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_pb2.py +68 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_pb2.pyi +512 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_pb2_grpc.py +4 -0
- yandex/cloud/mdb/postgresql/v1/config/host17_pb2_grpc.pyi +17 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_1c_pb2.py +168 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_1c_pb2.pyi +1130 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_1c_pb2_grpc.py +4 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_1c_pb2_grpc.pyi +17 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_pb2.py +168 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_pb2.pyi +1122 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_pb2_grpc.py +4 -0
- yandex/cloud/mdb/postgresql/v1/config/postgresql17_pb2_grpc.pyi +17 -0
- yandex/cloud/mdb/redis/v1/backup_pb2.py +4 -2
- yandex/cloud/mdb/redis/v1/backup_pb2.pyi +31 -1
- yandex/cloud/organizationmanager/v1/user_account_pb2.pyi +2 -2
- yandex/cloud/searchapi/__init__.py +0 -0
- yandex/cloud/searchapi/v2/__init__.py +0 -0
- yandex/cloud/searchapi/v2/search_query_pb2.py +37 -0
- yandex/cloud/searchapi/v2/search_query_pb2.pyi +100 -0
- yandex/cloud/searchapi/v2/search_query_pb2_grpc.py +4 -0
- yandex/cloud/searchapi/v2/search_query_pb2_grpc.pyi +17 -0
- yandex/cloud/searchapi/v2/search_service_pb2.py +61 -0
- yandex/cloud/searchapi/v2/search_service_pb2.pyi +218 -0
- yandex/cloud/searchapi/v2/search_service_pb2_grpc.py +70 -0
- yandex/cloud/searchapi/v2/search_service_pb2_grpc.pyi +48 -0
- yandex/cloud/serverless/containers/v1/container_pb2.py +42 -34
- yandex/cloud/serverless/containers/v1/container_pb2.pyi +52 -2
- yandex/cloud/serverless/containers/v1/container_service_pb2.py +18 -18
- yandex/cloud/serverless/containers/v1/container_service_pb2.pyi +8 -2
- yandex/cloud/serverless/eventrouter/v1/connector_pb2.py +9 -9
- yandex/cloud/serverless/eventrouter/v1/connector_pb2.pyi +8 -0
- yandex/cloud/serverless/eventrouter/v1/rule_pb2.py +31 -31
- yandex/cloud/serverless/eventrouter/v1/rule_pb2.pyi +4 -0
- yandex/cloud/serverless/workflows/v1/execution_service_pb2.py +23 -15
- yandex/cloud/serverless/workflows/v1/execution_service_pb2.pyi +32 -0
- yandex/cloud/serverless/workflows/v1/execution_service_pb2_grpc.py +34 -0
- yandex/cloud/serverless/workflows/v1/execution_service_pb2_grpc.pyi +20 -0
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2.py +46 -40
- yandex/cloud/smartwebsecurity/v1/security_profile_pb2.pyi +47 -2
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.py +18 -18
- yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.pyi +16 -3
- yandex/cloud/video/v1/channel_pb2.py +1 -1
- yandex/cloud/video/v1/channel_service_pb2.py +58 -25
- yandex/cloud/video/v1/channel_service_pb2.pyi +40 -0
- yandex/cloud/video/v1/channel_service_pb2_grpc.py +34 -0
- yandex/cloud/video/v1/channel_service_pb2_grpc.pyi +20 -0
- yandex/cloud/video/v1/episode_pb2.py +10 -8
- yandex/cloud/video/v1/episode_pb2.pyi +19 -3
- yandex/cloud/video/v1/episode_service_pb2.py +96 -43
- yandex/cloud/video/v1/episode_service_pb2.pyi +68 -6
- yandex/cloud/video/v1/episode_service_pb2_grpc.py +34 -0
- yandex/cloud/video/v1/episode_service_pb2_grpc.pyi +20 -0
- yandex/cloud/video/v1/manifest_pb2.py +3 -3
- yandex/cloud/video/v1/manifest_pb2.pyi +1 -5
- yandex/cloud/video/v1/playlist_pb2.py +29 -0
- yandex/cloud/video/v1/playlist_pb2.pyi +89 -0
- yandex/cloud/video/v1/playlist_pb2_grpc.py +4 -0
- yandex/cloud/video/v1/playlist_pb2_grpc.pyi +17 -0
- yandex/cloud/video/v1/playlist_service_pb2.py +102 -0
- yandex/cloud/video/v1/playlist_service_pb2.pyi +331 -0
- yandex/cloud/video/v1/playlist_service_pb2_grpc.py +276 -0
- yandex/cloud/video/v1/playlist_service_pb2_grpc.pyi +172 -0
- yandex/cloud/video/v1/stream_line_pb2.py +1 -1
- yandex/cloud/video/v1/stream_line_service_pb2.py +122 -55
- yandex/cloud/video/v1/stream_line_service_pb2.pyi +80 -0
- yandex/cloud/video/v1/stream_line_service_pb2_grpc.py +68 -0
- yandex/cloud/video/v1/stream_line_service_pb2_grpc.pyi +40 -0
- yandex/cloud/video/v1/stream_pb2.py +1 -1
- yandex/cloud/video/v1/stream_service_pb2.py +98 -37
- yandex/cloud/video/v1/stream_service_pb2.pyi +80 -0
- yandex/cloud/video/v1/stream_service_pb2_grpc.py +68 -0
- yandex/cloud/video/v1/stream_service_pb2_grpc.pyi +40 -0
- yandex/cloud/video/v1/subtitle_pb2.py +29 -0
- yandex/cloud/video/v1/subtitle_pb2.pyi +90 -0
- yandex/cloud/video/v1/subtitle_pb2_grpc.py +4 -0
- yandex/cloud/video/v1/subtitle_pb2_grpc.pyi +17 -0
- yandex/cloud/video/v1/subtitle_service_pb2.py +83 -0
- yandex/cloud/video/v1/subtitle_service_pb2.pyi +212 -0
- yandex/cloud/video/v1/subtitle_service_pb2_grpc.py +208 -0
- yandex/cloud/video/v1/subtitle_service_pb2_grpc.pyi +132 -0
- yandex/cloud/video/v1/thumbnail_service_pb2.py +50 -21
- yandex/cloud/video/v1/thumbnail_service_pb2.pyi +48 -0
- yandex/cloud/video/v1/thumbnail_service_pb2_grpc.py +69 -0
- yandex/cloud/video/v1/thumbnail_service_pb2_grpc.pyi +41 -0
- yandex/cloud/video/v1/video_pb2.py +18 -14
- yandex/cloud/video/v1/video_pb2.pyi +55 -3
- yandex/cloud/video/v1/video_service_pb2.py +118 -49
- yandex/cloud/video/v1/video_service_pb2.pyi +161 -6
- yandex/cloud/video/v1/video_service_pb2_grpc.py +102 -0
- yandex/cloud/video/v1/video_service_pb2_grpc.pyi +60 -0
- {yandexcloud-0.323.0.dist-info → yandexcloud-0.325.0.dist-info}/METADATA +1 -1
- {yandexcloud-0.323.0.dist-info → yandexcloud-0.325.0.dist-info}/RECORD +167 -103
- {yandexcloud-0.323.0.dist-info → yandexcloud-0.325.0.dist-info}/WHEEL +1 -1
- {yandexcloud-0.323.0.dist-info → yandexcloud-0.325.0.dist-info}/AUTHORS +0 -0
- {yandexcloud-0.323.0.dist-info → yandexcloud-0.325.0.dist-info}/LICENSE +0 -0
- {yandexcloud-0.323.0.dist-info → yandexcloud-0.325.0.dist-info}/top_level.txt +0 -0
|
@@ -21,7 +21,7 @@ class TextEmbeddingRequest(google.protobuf.message.Message):
|
|
|
21
21
|
MODEL_URI_FIELD_NUMBER: builtins.int
|
|
22
22
|
TEXT_FIELD_NUMBER: builtins.int
|
|
23
23
|
model_uri: builtins.str
|
|
24
|
-
"""The [
|
|
24
|
+
"""The [model URI](/docs/foundation-models/concepts/embeddings) to be used for obtaining text embeddings."""
|
|
25
25
|
text: builtins.str
|
|
26
26
|
"""The input text for which the embedding is requested."""
|
|
27
27
|
def __init__(
|
|
@@ -60,9 +60,11 @@ class ImageGenerationOptions(google.protobuf.message.Message):
|
|
|
60
60
|
SEED_FIELD_NUMBER: builtins.int
|
|
61
61
|
ASPECT_RATIO_FIELD_NUMBER: builtins.int
|
|
62
62
|
mime_type: builtins.str
|
|
63
|
-
"""MIME type of generated image format.
|
|
63
|
+
"""The [MIME type](https://en.wikipedia.org/wiki/Media_type) of generated image format.
|
|
64
|
+
For possible specifications, see [documentation](/docs/foundation-models/concepts).
|
|
65
|
+
"""
|
|
64
66
|
seed: builtins.int
|
|
65
|
-
"""Seed for image generation."""
|
|
67
|
+
"""Seed for image generation. It serves as a starting point for image generation from noise."""
|
|
66
68
|
@property
|
|
67
69
|
def aspect_ratio(self) -> global___AspectRatio:
|
|
68
70
|
"""Aspect ratio of generated image."""
|
|
@@ -15,7 +15,10 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
|
15
15
|
|
|
16
16
|
@typing.final
|
|
17
17
|
class ImageGenerationRequest(google.protobuf.message.Message):
|
|
18
|
-
"""Request for the service to generate an image.
|
|
18
|
+
"""Request for the service to generate an image.
|
|
19
|
+
|
|
20
|
+
For examples of usage, see [step-by-step guide](/docs/foundation-models/operations/yandexart/request).
|
|
21
|
+
"""
|
|
19
22
|
|
|
20
23
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
21
24
|
|
|
@@ -23,7 +26,7 @@ class ImageGenerationRequest(google.protobuf.message.Message):
|
|
|
23
26
|
MESSAGES_FIELD_NUMBER: builtins.int
|
|
24
27
|
GENERATION_OPTIONS_FIELD_NUMBER: builtins.int
|
|
25
28
|
model_uri: builtins.str
|
|
26
|
-
"""The [
|
|
29
|
+
"""The [model URI](/docs/foundation-models/concepts/yandexart/models) to be used for image generation."""
|
|
27
30
|
@property
|
|
28
31
|
def messages(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.image_generation.image_generation_pb2.Message]:
|
|
29
32
|
"""A list of messages representing the context for the image generation model."""
|
|
@@ -53,7 +56,7 @@ class ImageGenerationResponse(google.protobuf.message.Message):
|
|
|
53
56
|
IMAGE_FIELD_NUMBER: builtins.int
|
|
54
57
|
MODEL_VERSION_FIELD_NUMBER: builtins.int
|
|
55
58
|
image: builtins.bytes
|
|
56
|
-
"""The image is serialized as an array of bytes encoded in
|
|
59
|
+
"""The image is serialized as an array of bytes encoded in [Base64](https://en.wikipedia.org/wiki/Base64)."""
|
|
57
60
|
model_version: builtins.str
|
|
58
61
|
"""The model version changes with each new releases."""
|
|
59
62
|
def __init__(
|
|
@@ -7,7 +7,7 @@ from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operati
|
|
|
7
7
|
|
|
8
8
|
|
|
9
9
|
class ImageGenerationAsyncServiceStub(object):
|
|
10
|
-
"""Service for
|
|
10
|
+
"""Service for creating images based on a text description.
|
|
11
11
|
"""
|
|
12
12
|
|
|
13
13
|
def __init__(self, channel):
|
|
@@ -24,7 +24,7 @@ class ImageGenerationAsyncServiceStub(object):
|
|
|
24
24
|
|
|
25
25
|
|
|
26
26
|
class ImageGenerationAsyncServiceServicer(object):
|
|
27
|
-
"""Service for
|
|
27
|
+
"""Service for creating images based on a text description.
|
|
28
28
|
"""
|
|
29
29
|
|
|
30
30
|
def Generate(self, request, context):
|
|
@@ -50,7 +50,7 @@ def add_ImageGenerationAsyncServiceServicer_to_server(servicer, server):
|
|
|
50
50
|
|
|
51
51
|
# This class is part of an EXPERIMENTAL API.
|
|
52
52
|
class ImageGenerationAsyncService(object):
|
|
53
|
-
"""Service for
|
|
53
|
+
"""Service for creating images based on a text description.
|
|
54
54
|
"""
|
|
55
55
|
|
|
56
56
|
@staticmethod
|
|
@@ -19,7 +19,7 @@ class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type:
|
|
|
19
19
|
...
|
|
20
20
|
|
|
21
21
|
class ImageGenerationAsyncServiceStub:
|
|
22
|
-
"""Service for
|
|
22
|
+
"""Service for creating images based on a text description."""
|
|
23
23
|
|
|
24
24
|
def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
|
|
25
25
|
Generate: grpc.UnaryUnaryMultiCallable[
|
|
@@ -29,7 +29,7 @@ class ImageGenerationAsyncServiceStub:
|
|
|
29
29
|
"""A method for generating an image based on a textual description."""
|
|
30
30
|
|
|
31
31
|
class ImageGenerationAsyncServiceAsyncStub:
|
|
32
|
-
"""Service for
|
|
32
|
+
"""Service for creating images based on a text description."""
|
|
33
33
|
|
|
34
34
|
Generate: grpc.aio.UnaryUnaryMultiCallable[
|
|
35
35
|
yandex.cloud.ai.foundation_models.v1.image_generation.image_generation_service_pb2.ImageGenerationRequest,
|
|
@@ -38,7 +38,7 @@ class ImageGenerationAsyncServiceAsyncStub:
|
|
|
38
38
|
"""A method for generating an image based on a textual description."""
|
|
39
39
|
|
|
40
40
|
class ImageGenerationAsyncServiceServicer(metaclass=abc.ABCMeta):
|
|
41
|
-
"""Service for
|
|
41
|
+
"""Service for creating images based on a text description."""
|
|
42
42
|
|
|
43
43
|
@abc.abstractmethod
|
|
44
44
|
def Generate(
|
|
@@ -12,16 +12,16 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
|
12
12
|
|
|
13
13
|
@typing.final
|
|
14
14
|
class ClassificationLabel(google.protobuf.message.Message):
|
|
15
|
-
"""A pair of text
|
|
15
|
+
"""A pair of text labels and their corresponding confidence values."""
|
|
16
16
|
|
|
17
17
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
18
18
|
|
|
19
19
|
LABEL_FIELD_NUMBER: builtins.int
|
|
20
20
|
CONFIDENCE_FIELD_NUMBER: builtins.int
|
|
21
21
|
label: builtins.str
|
|
22
|
-
"""A
|
|
22
|
+
"""A class name label."""
|
|
23
23
|
confidence: builtins.float
|
|
24
|
-
"""
|
|
24
|
+
"""The probability of classifying text into a specific class."""
|
|
25
25
|
def __init__(
|
|
26
26
|
self,
|
|
27
27
|
*,
|
yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2.pyi
CHANGED
|
@@ -15,14 +15,20 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
|
15
15
|
|
|
16
16
|
@typing.final
|
|
17
17
|
class TextClassificationRequest(google.protobuf.message.Message):
|
|
18
|
-
"""Request for the service to classify text.
|
|
18
|
+
"""Request for the service to classify text with tuned model.
|
|
19
|
+
|
|
20
|
+
The names of the classes between which the model will be distributing requests must be specified during model tuning;
|
|
21
|
+
therefore, they are not provided in the request.
|
|
22
|
+
|
|
23
|
+
For examples of usage, see [step-by-step guides](/docs/operations/classifier/additionally-trained).
|
|
24
|
+
"""
|
|
19
25
|
|
|
20
26
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
21
27
|
|
|
22
28
|
MODEL_URI_FIELD_NUMBER: builtins.int
|
|
23
29
|
TEXT_FIELD_NUMBER: builtins.int
|
|
24
30
|
model_uri: builtins.str
|
|
25
|
-
"""The
|
|
31
|
+
"""The [URI](/docs/foundation-models/concepts/classifier/models) of your tuned classifier model."""
|
|
26
32
|
text: builtins.str
|
|
27
33
|
"""Text for classification."""
|
|
28
34
|
def __init__(
|
|
@@ -37,17 +43,19 @@ global___TextClassificationRequest = TextClassificationRequest
|
|
|
37
43
|
|
|
38
44
|
@typing.final
|
|
39
45
|
class TextClassificationResponse(google.protobuf.message.Message):
|
|
40
|
-
"""Response
|
|
46
|
+
"""Response with classifier predictions."""
|
|
41
47
|
|
|
42
48
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
43
49
|
|
|
44
50
|
PREDICTIONS_FIELD_NUMBER: builtins.int
|
|
45
51
|
MODEL_VERSION_FIELD_NUMBER: builtins.int
|
|
46
52
|
model_version: builtins.str
|
|
47
|
-
"""
|
|
53
|
+
"""The model version changes with each new releases."""
|
|
48
54
|
@property
|
|
49
55
|
def predictions(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationLabel]:
|
|
50
|
-
"""
|
|
56
|
+
"""The classification results with the `confidence`` values
|
|
57
|
+
for the probability of classifying the request text into each class.
|
|
58
|
+
"""
|
|
51
59
|
|
|
52
60
|
def __init__(
|
|
53
61
|
self,
|
|
@@ -61,7 +69,9 @@ global___TextClassificationResponse = TextClassificationResponse
|
|
|
61
69
|
|
|
62
70
|
@typing.final
|
|
63
71
|
class FewShotTextClassificationRequest(google.protobuf.message.Message):
|
|
64
|
-
"""Request for the service to classify text.
|
|
72
|
+
"""Request for the service to classify text.
|
|
73
|
+
For examples of usage, see [step-by-step guides](/docs/operations/classifier/readymade).
|
|
74
|
+
"""
|
|
65
75
|
|
|
66
76
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
67
77
|
|
|
@@ -71,14 +81,17 @@ class FewShotTextClassificationRequest(google.protobuf.message.Message):
|
|
|
71
81
|
TEXT_FIELD_NUMBER: builtins.int
|
|
72
82
|
SAMPLES_FIELD_NUMBER: builtins.int
|
|
73
83
|
model_uri: builtins.str
|
|
74
|
-
"""The
|
|
84
|
+
"""The [URI](/docs/foundation-models/concepts/classifier/models) of the classifier model."""
|
|
75
85
|
task_description: builtins.str
|
|
76
86
|
"""Text description of the classification task."""
|
|
77
87
|
text: builtins.str
|
|
78
88
|
"""Text for classification."""
|
|
79
89
|
@property
|
|
80
90
|
def labels(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
81
|
-
"""List of available labels for the classification result.
|
|
91
|
+
"""List of available labels for the classification result.
|
|
92
|
+
Give meaningful names to label classes: this is essential for correct classification results.
|
|
93
|
+
For example, use ``chemistry`` and ``physics`` rather than ``chm`` and ``phs`` for class names.
|
|
94
|
+
"""
|
|
82
95
|
|
|
83
96
|
@property
|
|
84
97
|
def samples(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationSample]:
|
|
@@ -106,10 +119,12 @@ class FewShotTextClassificationResponse(google.protobuf.message.Message):
|
|
|
106
119
|
PREDICTIONS_FIELD_NUMBER: builtins.int
|
|
107
120
|
MODEL_VERSION_FIELD_NUMBER: builtins.int
|
|
108
121
|
model_version: builtins.str
|
|
109
|
-
"""
|
|
122
|
+
"""The model version changes with each new releases."""
|
|
110
123
|
@property
|
|
111
124
|
def predictions(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationLabel]:
|
|
112
|
-
"""
|
|
125
|
+
"""The classification results with the `confidence`` values
|
|
126
|
+
for the probability of classifying the request text into each class.
|
|
127
|
+
"""
|
|
113
128
|
|
|
114
129
|
def __init__(
|
|
115
130
|
self,
|
yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.py
CHANGED
|
@@ -6,7 +6,7 @@ from yandex.cloud.ai.foundation_models.v1.text_classification import text_classi
|
|
|
6
6
|
|
|
7
7
|
|
|
8
8
|
class TextClassificationServiceStub(object):
|
|
9
|
-
"""Service for classifying text
|
|
9
|
+
"""Service for classifying the text requests provided in prompts.
|
|
10
10
|
"""
|
|
11
11
|
|
|
12
12
|
def __init__(self, channel):
|
|
@@ -28,18 +28,24 @@ class TextClassificationServiceStub(object):
|
|
|
28
28
|
|
|
29
29
|
|
|
30
30
|
class TextClassificationServiceServicer(object):
|
|
31
|
-
"""Service for classifying text
|
|
31
|
+
"""Service for classifying the text requests provided in prompts.
|
|
32
32
|
"""
|
|
33
33
|
|
|
34
34
|
def Classify(self, request, context):
|
|
35
|
-
"""RPC method
|
|
35
|
+
"""RPC method to classify text with tuned model.
|
|
36
|
+
|
|
37
|
+
The names of the classes between which the model will be distributing requests
|
|
38
|
+
must be specified during model tuning and are not provided in the request.
|
|
36
39
|
"""
|
|
37
40
|
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
38
41
|
context.set_details('Method not implemented!')
|
|
39
42
|
raise NotImplementedError('Method not implemented!')
|
|
40
43
|
|
|
41
44
|
def FewShotClassify(self, request, context):
|
|
42
|
-
"""RPC method for
|
|
45
|
+
"""RPC method for binary and multi-class classification.
|
|
46
|
+
|
|
47
|
+
You can provide up to 20 classes for few-shot text classification
|
|
48
|
+
with optional examples.
|
|
43
49
|
"""
|
|
44
50
|
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
|
|
45
51
|
context.set_details('Method not implemented!')
|
|
@@ -66,7 +72,7 @@ def add_TextClassificationServiceServicer_to_server(servicer, server):
|
|
|
66
72
|
|
|
67
73
|
# This class is part of an EXPERIMENTAL API.
|
|
68
74
|
class TextClassificationService(object):
|
|
69
|
-
"""Service for classifying text
|
|
75
|
+
"""Service for classifying the text requests provided in prompts.
|
|
70
76
|
"""
|
|
71
77
|
|
|
72
78
|
@staticmethod
|
yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.pyi
CHANGED
|
@@ -18,38 +18,54 @@ class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type:
|
|
|
18
18
|
...
|
|
19
19
|
|
|
20
20
|
class TextClassificationServiceStub:
|
|
21
|
-
"""Service for classifying text
|
|
21
|
+
"""Service for classifying the text requests provided in prompts."""
|
|
22
22
|
|
|
23
23
|
def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
|
|
24
24
|
Classify: grpc.UnaryUnaryMultiCallable[
|
|
25
25
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
|
|
26
26
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse,
|
|
27
27
|
]
|
|
28
|
-
"""RPC method
|
|
28
|
+
"""RPC method to classify text with tuned model.
|
|
29
|
+
|
|
30
|
+
The names of the classes between which the model will be distributing requests
|
|
31
|
+
must be specified during model tuning and are not provided in the request.
|
|
32
|
+
"""
|
|
29
33
|
|
|
30
34
|
FewShotClassify: grpc.UnaryUnaryMultiCallable[
|
|
31
35
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
|
|
32
36
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse,
|
|
33
37
|
]
|
|
34
|
-
"""RPC method for
|
|
38
|
+
"""RPC method for binary and multi-class classification.
|
|
39
|
+
|
|
40
|
+
You can provide up to 20 classes for few-shot text classification
|
|
41
|
+
with optional examples.
|
|
42
|
+
"""
|
|
35
43
|
|
|
36
44
|
class TextClassificationServiceAsyncStub:
|
|
37
|
-
"""Service for classifying text
|
|
45
|
+
"""Service for classifying the text requests provided in prompts."""
|
|
38
46
|
|
|
39
47
|
Classify: grpc.aio.UnaryUnaryMultiCallable[
|
|
40
48
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
|
|
41
49
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse,
|
|
42
50
|
]
|
|
43
|
-
"""RPC method
|
|
51
|
+
"""RPC method to classify text with tuned model.
|
|
52
|
+
|
|
53
|
+
The names of the classes between which the model will be distributing requests
|
|
54
|
+
must be specified during model tuning and are not provided in the request.
|
|
55
|
+
"""
|
|
44
56
|
|
|
45
57
|
FewShotClassify: grpc.aio.UnaryUnaryMultiCallable[
|
|
46
58
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
|
|
47
59
|
yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse,
|
|
48
60
|
]
|
|
49
|
-
"""RPC method for
|
|
61
|
+
"""RPC method for binary and multi-class classification.
|
|
62
|
+
|
|
63
|
+
You can provide up to 20 classes for few-shot text classification
|
|
64
|
+
with optional examples.
|
|
65
|
+
"""
|
|
50
66
|
|
|
51
67
|
class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
|
|
52
|
-
"""Service for classifying text
|
|
68
|
+
"""Service for classifying the text requests provided in prompts."""
|
|
53
69
|
|
|
54
70
|
@abc.abstractmethod
|
|
55
71
|
def Classify(
|
|
@@ -57,7 +73,11 @@ class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
|
|
|
57
73
|
request: yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
|
|
58
74
|
context: _ServicerContext,
|
|
59
75
|
) -> typing.Union[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse, collections.abc.Awaitable[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse]]:
|
|
60
|
-
"""RPC method
|
|
76
|
+
"""RPC method to classify text with tuned model.
|
|
77
|
+
|
|
78
|
+
The names of the classes between which the model will be distributing requests
|
|
79
|
+
must be specified during model tuning and are not provided in the request.
|
|
80
|
+
"""
|
|
61
81
|
|
|
62
82
|
@abc.abstractmethod
|
|
63
83
|
def FewShotClassify(
|
|
@@ -65,6 +85,10 @@ class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
|
|
|
65
85
|
request: yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
|
|
66
86
|
context: _ServicerContext,
|
|
67
87
|
) -> typing.Union[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse, collections.abc.Awaitable[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse]]:
|
|
68
|
-
"""RPC method for
|
|
88
|
+
"""RPC method for binary and multi-class classification.
|
|
89
|
+
|
|
90
|
+
You can provide up to 20 classes for few-shot text classification
|
|
91
|
+
with optional examples.
|
|
92
|
+
"""
|
|
69
93
|
|
|
70
94
|
def add_TextClassificationServiceServicer_to_server(servicer: TextClassificationServiceServicer, server: typing.Union[grpc.Server, grpc.aio.Server]) -> None: ...
|
|
@@ -15,7 +15,7 @@ from google.protobuf import struct_pb2 as google_dot_protobuf_dot_struct__pb2
|
|
|
15
15
|
from google.protobuf import wrappers_pb2 as google_dot_protobuf_dot_wrappers__pb2
|
|
16
16
|
|
|
17
17
|
|
|
18
|
-
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n6yandex/cloud/ai/foundation_models/v1/text_common.proto\x12$yandex.cloud.ai.foundation_models.v1\x1a\x1cgoogle/protobuf/struct.proto\x1a\x1egoogle/protobuf/wrappers.proto\"\x87\x01\n\x11\x43ompletionOptions\x12\x0e\n\x06stream\x18\x01 \x01(\x08\x12\x31\n\x0btemperature\x18\x02 \x01(\x0b\x32\x1c.google.protobuf.DoubleValue\x12/\n\nmax_tokens\x18\x03 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\"
|
|
18
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n6yandex/cloud/ai/foundation_models/v1/text_common.proto\x12$yandex.cloud.ai.foundation_models.v1\x1a\x1cgoogle/protobuf/struct.proto\x1a\x1egoogle/protobuf/wrappers.proto\"\x87\x01\n\x11\x43ompletionOptions\x12\x0e\n\x06stream\x18\x01 \x01(\x08\x12\x31\n\x0btemperature\x18\x02 \x01(\x0b\x32\x1c.google.protobuf.DoubleValue\x12/\n\nmax_tokens\x18\x03 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\"\xd2\x01\n\x07Message\x12\x0c\n\x04role\x18\x01 \x01(\t\x12\x0e\n\x04text\x18\x02 \x01(\tH\x00\x12L\n\x0etool_call_list\x18\x03 \x01(\x0b\x32\x32.yandex.cloud.ai.foundation_models.v1.ToolCallListH\x00\x12P\n\x10tool_result_list\x18\x04 \x01(\x0b\x32\x34.yandex.cloud.ai.foundation_models.v1.ToolResultListH\x00\x42\t\n\x07\x43ontent\"Z\n\x0c\x43ontentUsage\x12\x19\n\x11input_text_tokens\x18\x01 \x01(\x03\x12\x19\n\x11\x63ompletion_tokens\x18\x02 \x01(\x03\x12\x14\n\x0ctotal_tokens\x18\x03 \x01(\x03\"\x8c\x03\n\x0b\x41lternative\x12>\n\x07message\x18\x01 \x01(\x0b\x32-.yandex.cloud.ai.foundation_models.v1.Message\x12S\n\x06status\x18\x02 \x01(\x0e\x32\x43.yandex.cloud.ai.foundation_models.v1.Alternative.AlternativeStatus\"\xe7\x01\n\x11\x41lternativeStatus\x12\"\n\x1e\x41LTERNATIVE_STATUS_UNSPECIFIED\x10\x00\x12\x1e\n\x1a\x41LTERNATIVE_STATUS_PARTIAL\x10\x01\x12&\n\"ALTERNATIVE_STATUS_TRUNCATED_FINAL\x10\x02\x12\x1c\n\x18\x41LTERNATIVE_STATUS_FINAL\x10\x03\x12%\n!ALTERNATIVE_STATUS_CONTENT_FILTER\x10\x04\x12!\n\x1d\x41LTERNATIVE_STATUS_TOOL_CALLS\x10\x05\"2\n\x05Token\x12\n\n\x02id\x18\x01 \x01(\x03\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x0f\n\x07special\x18\x03 \x01(\x08\"Z\n\x04Tool\x12\x46\n\x08\x66unction\x18\x01 \x01(\x0b\x32\x32.yandex.cloud.ai.foundation_models.v1.FunctionToolH\x00\x42\n\n\x08ToolType\"^\n\x0c\x46unctionTool\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x02 \x01(\t\x12+\n\nparameters\x18\x03 \x01(\x0b\x32\x17.google.protobuf.Struct\"g\n\x08ToolCall\x12K\n\rfunction_call\x18\x01 \x01(\x0b\x32\x32.yandex.cloud.ai.foundation_models.v1.FunctionCallH\x00\x42\x0e\n\x0cToolCallType\"H\n\x0c\x46unctionCall\x12\x0c\n\x04name\x18\x01 \x01(\t\x12*\n\targuments\x18\x02 \x01(\x0b\x32\x17.google.protobuf.Struct\"R\n\x0cToolCallList\x12\x42\n\ntool_calls\x18\x01 \x03(\x0b\x32..yandex.cloud.ai.foundation_models.v1.ToolCall\"o\n\nToolResult\x12O\n\x0f\x66unction_result\x18\x01 \x01(\x0b\x32\x34.yandex.cloud.ai.foundation_models.v1.FunctionResultH\x00\x42\x10\n\x0eToolResultType\"@\n\x0e\x46unctionResult\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x11\n\x07\x63ontent\x18\x02 \x01(\tH\x00\x42\r\n\x0b\x43ontentType\"X\n\x0eToolResultList\x12\x46\n\x0ctool_results\x18\x01 \x03(\x0b\x32\x30.yandex.cloud.ai.foundation_models.v1.ToolResultB\x86\x01\n(yandex.cloud.api.ai.foundation_models.v1ZZgithub.com/yandex-cloud/go-genproto/yandex/cloud/ai/foundation_models/v1;foundation_modelsb\x06proto3')
|
|
19
19
|
|
|
20
20
|
_globals = globals()
|
|
21
21
|
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
@@ -25,30 +25,30 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
25
25
|
DESCRIPTOR._serialized_options = b'\n(yandex.cloud.api.ai.foundation_models.v1ZZgithub.com/yandex-cloud/go-genproto/yandex/cloud/ai/foundation_models/v1;foundation_models'
|
|
26
26
|
_globals['_COMPLETIONOPTIONS']._serialized_start=159
|
|
27
27
|
_globals['_COMPLETIONOPTIONS']._serialized_end=294
|
|
28
|
-
_globals['_MESSAGE']._serialized_start=
|
|
29
|
-
_globals['_MESSAGE']._serialized_end=
|
|
30
|
-
_globals['_CONTENTUSAGE']._serialized_start=
|
|
31
|
-
_globals['_CONTENTUSAGE']._serialized_end=
|
|
32
|
-
_globals['_ALTERNATIVE']._serialized_start=
|
|
33
|
-
_globals['_ALTERNATIVE']._serialized_end=
|
|
34
|
-
_globals['_ALTERNATIVE_ALTERNATIVESTATUS']._serialized_start=
|
|
35
|
-
_globals['_ALTERNATIVE_ALTERNATIVESTATUS']._serialized_end=
|
|
36
|
-
_globals['_TOKEN']._serialized_start=
|
|
37
|
-
_globals['_TOKEN']._serialized_end=
|
|
38
|
-
_globals['_TOOL']._serialized_start=
|
|
39
|
-
_globals['_TOOL']._serialized_end=
|
|
40
|
-
_globals['_FUNCTIONTOOL']._serialized_start=
|
|
41
|
-
_globals['_FUNCTIONTOOL']._serialized_end=
|
|
42
|
-
_globals['_TOOLCALL']._serialized_start=
|
|
43
|
-
_globals['_TOOLCALL']._serialized_end=
|
|
44
|
-
_globals['_FUNCTIONCALL']._serialized_start=
|
|
45
|
-
_globals['_FUNCTIONCALL']._serialized_end=
|
|
46
|
-
_globals['_TOOLCALLLIST']._serialized_start=
|
|
47
|
-
_globals['_TOOLCALLLIST']._serialized_end=
|
|
48
|
-
_globals['_TOOLRESULT']._serialized_start=
|
|
49
|
-
_globals['_TOOLRESULT']._serialized_end=
|
|
50
|
-
_globals['_FUNCTIONRESULT']._serialized_start=
|
|
51
|
-
_globals['_FUNCTIONRESULT']._serialized_end=
|
|
52
|
-
_globals['_TOOLRESULTLIST']._serialized_start=
|
|
53
|
-
_globals['_TOOLRESULTLIST']._serialized_end=
|
|
28
|
+
_globals['_MESSAGE']._serialized_start=297
|
|
29
|
+
_globals['_MESSAGE']._serialized_end=507
|
|
30
|
+
_globals['_CONTENTUSAGE']._serialized_start=509
|
|
31
|
+
_globals['_CONTENTUSAGE']._serialized_end=599
|
|
32
|
+
_globals['_ALTERNATIVE']._serialized_start=602
|
|
33
|
+
_globals['_ALTERNATIVE']._serialized_end=998
|
|
34
|
+
_globals['_ALTERNATIVE_ALTERNATIVESTATUS']._serialized_start=767
|
|
35
|
+
_globals['_ALTERNATIVE_ALTERNATIVESTATUS']._serialized_end=998
|
|
36
|
+
_globals['_TOKEN']._serialized_start=1000
|
|
37
|
+
_globals['_TOKEN']._serialized_end=1050
|
|
38
|
+
_globals['_TOOL']._serialized_start=1052
|
|
39
|
+
_globals['_TOOL']._serialized_end=1142
|
|
40
|
+
_globals['_FUNCTIONTOOL']._serialized_start=1144
|
|
41
|
+
_globals['_FUNCTIONTOOL']._serialized_end=1238
|
|
42
|
+
_globals['_TOOLCALL']._serialized_start=1240
|
|
43
|
+
_globals['_TOOLCALL']._serialized_end=1343
|
|
44
|
+
_globals['_FUNCTIONCALL']._serialized_start=1345
|
|
45
|
+
_globals['_FUNCTIONCALL']._serialized_end=1417
|
|
46
|
+
_globals['_TOOLCALLLIST']._serialized_start=1419
|
|
47
|
+
_globals['_TOOLCALLLIST']._serialized_end=1501
|
|
48
|
+
_globals['_TOOLRESULT']._serialized_start=1503
|
|
49
|
+
_globals['_TOOLRESULT']._serialized_end=1614
|
|
50
|
+
_globals['_FUNCTIONRESULT']._serialized_start=1616
|
|
51
|
+
_globals['_FUNCTIONRESULT']._serialized_end=1680
|
|
52
|
+
_globals['_TOOLRESULTLIST']._serialized_start=1682
|
|
53
|
+
_globals['_TOOLRESULTLIST']._serialized_end=1770
|
|
54
54
|
# @@protoc_insertion_point(module_scope)
|
|
@@ -65,6 +65,8 @@ class Message(google.protobuf.message.Message):
|
|
|
65
65
|
|
|
66
66
|
ROLE_FIELD_NUMBER: builtins.int
|
|
67
67
|
TEXT_FIELD_NUMBER: builtins.int
|
|
68
|
+
TOOL_CALL_LIST_FIELD_NUMBER: builtins.int
|
|
69
|
+
TOOL_RESULT_LIST_FIELD_NUMBER: builtins.int
|
|
68
70
|
role: builtins.str
|
|
69
71
|
"""The ID of the message sender. Supported roles:
|
|
70
72
|
* `system`: Special role used to define the behaviour of the completion model.
|
|
@@ -73,15 +75,25 @@ class Message(google.protobuf.message.Message):
|
|
|
73
75
|
"""
|
|
74
76
|
text: builtins.str
|
|
75
77
|
"""Textual content of the message."""
|
|
78
|
+
@property
|
|
79
|
+
def tool_call_list(self) -> global___ToolCallList:
|
|
80
|
+
"""List of tool calls made by the model as part of the response generation."""
|
|
81
|
+
|
|
82
|
+
@property
|
|
83
|
+
def tool_result_list(self) -> global___ToolResultList:
|
|
84
|
+
"""List of tool results returned from external tools that were invoked by the model."""
|
|
85
|
+
|
|
76
86
|
def __init__(
|
|
77
87
|
self,
|
|
78
88
|
*,
|
|
79
89
|
role: builtins.str = ...,
|
|
80
90
|
text: builtins.str = ...,
|
|
91
|
+
tool_call_list: global___ToolCallList | None = ...,
|
|
92
|
+
tool_result_list: global___ToolResultList | None = ...,
|
|
81
93
|
) -> None: ...
|
|
82
|
-
def HasField(self, field_name: typing.Literal["Content", b"Content", "text", b"text"]) -> builtins.bool: ...
|
|
83
|
-
def ClearField(self, field_name: typing.Literal["Content", b"Content", "role", b"role", "text", b"text"]) -> None: ...
|
|
84
|
-
def WhichOneof(self, oneof_group: typing.Literal["Content", b"Content"]) -> typing.Literal["text"] | None: ...
|
|
94
|
+
def HasField(self, field_name: typing.Literal["Content", b"Content", "text", b"text", "tool_call_list", b"tool_call_list", "tool_result_list", b"tool_result_list"]) -> builtins.bool: ...
|
|
95
|
+
def ClearField(self, field_name: typing.Literal["Content", b"Content", "role", b"role", "text", b"text", "tool_call_list", b"tool_call_list", "tool_result_list", b"tool_result_list"]) -> None: ...
|
|
96
|
+
def WhichOneof(self, oneof_group: typing.Literal["Content", b"Content"]) -> typing.Literal["text", "tool_call_list", "tool_result_list"] | None: ...
|
|
85
97
|
|
|
86
98
|
global___Message = Message
|
|
87
99
|
|
|
@@ -97,7 +109,7 @@ class ContentUsage(google.protobuf.message.Message):
|
|
|
97
109
|
input_text_tokens: builtins.int
|
|
98
110
|
"""The number of tokens in the textual part of the model input."""
|
|
99
111
|
completion_tokens: builtins.int
|
|
100
|
-
"""The
|
|
112
|
+
"""The number of tokens in the generated completion."""
|
|
101
113
|
total_tokens: builtins.int
|
|
102
114
|
"""The total number of tokens, including all input tokens and all generated tokens."""
|
|
103
115
|
def __init__(
|
|
@@ -159,10 +171,10 @@ class Alternative(google.protobuf.message.Message):
|
|
|
159
171
|
MESSAGE_FIELD_NUMBER: builtins.int
|
|
160
172
|
STATUS_FIELD_NUMBER: builtins.int
|
|
161
173
|
status: global___Alternative.AlternativeStatus.ValueType
|
|
162
|
-
"""The generation status of the alternative"""
|
|
174
|
+
"""The generation status of the alternative."""
|
|
163
175
|
@property
|
|
164
176
|
def message(self) -> global___Message:
|
|
165
|
-
"""A message
|
|
177
|
+
"""A message with the content of the alternative."""
|
|
166
178
|
|
|
167
179
|
def __init__(
|
|
168
180
|
self,
|
|
@@ -13,7 +13,7 @@ _sym_db = _symbol_database.Default()
|
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
|
|
16
|
-
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n yandex/cloud/ai/stt/v3/stt.proto\x12\x10speechkit.stt.v3\"\xe2\x03\n\x18TextNormalizationOptions\x12X\n\x12text_normalization\x18\x01 \x01(\x0e\x32<.speechkit.stt.v3.TextNormalizationOptions.TextNormalization\x12\x18\n\x10profanity_filter\x18\x02 \x01(\x08\x12\x17\n\x0fliterature_text\x18\x03 \x01(\x08\x12]\n\x15phone_formatting_mode\x18\x04 \x01(\x0e\x32>.speechkit.stt.v3.TextNormalizationOptions.PhoneFormattingMode\"x\n\x11TextNormalization\x12\"\n\x1eTEXT_NORMALIZATION_UNSPECIFIED\x10\x00\x12\x1e\n\x1aTEXT_NORMALIZATION_ENABLED\x10\x01\x12\x1f\n\x1bTEXT_NORMALIZATION_DISABLED\x10\x02\"`\n\x13PhoneFormattingMode\x12%\n!PHONE_FORMATTING_MODE_UNSPECIFIED\x10\x00\x12\"\n\x1ePHONE_FORMATTING_MODE_DISABLED\x10\x01\"\xce\x01\n\x14\x44\x65\x66\x61ultEouClassifier\x12\x43\n\x04type\x18\x01 \x01(\x0e\x32\x35.speechkit.stt.v3.DefaultEouClassifier.EouSensitivity\x12\'\n\x1fmax_pause_between_words_hint_ms\x18\x02 \x01(\x03\"H\n\x0e\x45ouSensitivity\x12\x1f\n\x1b\x45OU_SENSITIVITY_UNSPECIFIED\x10\x00\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x01\x12\x08\n\x04HIGH\x10\x02\"\x17\n\x15\x45xternalEouClassifier\"\xb2\x01\n\x14\x45ouClassifierOptions\x12\x44\n\x12\x64\x65\x66\x61ult_classifier\x18\x01 \x01(\x0b\x32&.speechkit.stt.v3.DefaultEouClassifierH\x00\x12\x46\n\x13\x65xternal_classifier\x18\x02 \x01(\x0b\x32\'.speechkit.stt.v3.ExternalEouClassifierH\x00\x42\x0c\n\nClassifier\"\xd3\x01\n\x15RecognitionClassifier\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x45\n\x08triggers\x18\x02 \x03(\x0e\x32\x33.speechkit.stt.v3.RecognitionClassifier.TriggerType\"_\n\x0bTriggerType\x12 \n\x18TRIGGER_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x10\n\x0cON_UTTERANCE\x10\x01\x12\x0c\n\x08ON_FINAL\x10\x02\x12\x0e\n\nON_PARTIAL\x10\x03\"\\\n\x1cRecognitionClassifierOptions\x12<\n\x0b\x63lassifiers\x18\x01 \x03(\x0b\x32\'.speechkit.stt.v3.RecognitionClassifier\"\x88\x01\n\x15SpeechAnalysisOptions\x12\x1f\n\x17\x65nable_speaker_analysis\x18\x01 \x01(\x08\x12$\n\x1c\x65nable_conversation_analysis\x18\x02 \x01(\x08\x12(\n descriptive_statistics_quantiles\x18\x03 \x03(\x01\"\xc7\x01\n\x08RawAudio\x12@\n\x0e\x61udio_encoding\x18\x01 \x01(\x0e\x32(.speechkit.stt.v3.RawAudio.AudioEncoding\x12\x19\n\x11sample_rate_hertz\x18\x02 \x01(\x03\x12\x1b\n\x13\x61udio_channel_count\x18\x03 \x01(\x03\"A\n\rAudioEncoding\x12\x1e\n\x1a\x41UDIO_ENCODING_UNSPECIFIED\x10\x00\x12\x10\n\x0cLINEAR16_PCM\x10\x01\"\xbf\x01\n\x0e\x43ontainerAudio\x12Q\n\x14\x63ontainer_audio_type\x18\x01 \x01(\x0e\x32\x33.speechkit.stt.v3.ContainerAudio.ContainerAudioType\"Z\n\x12\x43ontainerAudioType\x12$\n CONTAINER_AUDIO_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03WAV\x10\x01\x12\x0c\n\x08OGG_OPUS\x10\x02\x12\x07\n\x03MP3\x10\x03\"\x91\x01\n\x12\x41udioFormatOptions\x12/\n\traw_audio\x18\x01 \x01(\x0b\x32\x1a.speechkit.stt.v3.RawAudioH\x00\x12;\n\x0f\x63ontainer_audio\x18\x02 \x01(\x0b\x32 .speechkit.stt.v3.ContainerAudioH\x00\x42\r\n\x0b\x41udioFormat\"\xf7\x01\n\x1aLanguageRestrictionOptions\x12^\n\x10restriction_type\x18\x01 \x01(\x0e\x32\x44.speechkit.stt.v3.LanguageRestrictionOptions.LanguageRestrictionType\x12\x15\n\rlanguage_code\x18\x02 \x03(\t\"b\n\x17LanguageRestrictionType\x12)\n%LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tWHITELIST\x10\x01\x12\r\n\tBLACKLIST\x10\x02\"\xb2\x03\n\x17RecognitionModelOptions\x12\r\n\x05model\x18\x01 \x01(\t\x12:\n\x0c\x61udio_format\x18\x02 \x01(\x0b\x32$.speechkit.stt.v3.AudioFormatOptions\x12\x46\n\x12text_normalization\x18\x03 \x01(\x0b\x32*.speechkit.stt.v3.TextNormalizationOptions\x12J\n\x14language_restriction\x18\x04 \x01(\x0b\x32,.speechkit.stt.v3.LanguageRestrictionOptions\x12\\\n\x15\x61udio_processing_type\x18\x05 \x01(\x0e\x32=.speechkit.stt.v3.RecognitionModelOptions.AudioProcessingType\"Z\n\x13\x41udioProcessingType\x12%\n!AUDIO_PROCESSING_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tREAL_TIME\x10\x01\x12\r\n\tFULL_DATA\x10\x02\"\xde\x01\n\x16SpeakerLabelingOptions\x12R\n\x10speaker_labeling\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.SpeakerLabelingOptions.SpeakerLabeling\"p\n\x0fSpeakerLabeling\x12 \n\x1cSPEAKER_LABELING_UNSPECIFIED\x10\x00\x12\x1c\n\x18SPEAKER_LABELING_ENABLED\x10\x01\x12\x1d\n\x19SPEAKER_LABELING_DISABLED\x10\x02\"\xee\x02\n\x10StreamingOptions\x12\x44\n\x11recognition_model\x18\x01 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12>\n\x0e\x65ou_classifier\x18\x02 \x01(\x0b\x32&.speechkit.stt.v3.EouClassifierOptions\x12N\n\x16recognition_classifier\x18\x03 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x05 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptions\"\x1a\n\nAudioChunk\x12\x0c\n\x04\x64\x61ta\x18\x01 \x01(\x0c\"#\n\x0cSilenceChunk\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\x03\"\x05\n\x03\x45ou\"\xe8\x01\n\x10StreamingRequest\x12=\n\x0fsession_options\x18\x01 \x01(\x0b\x32\".speechkit.stt.v3.StreamingOptionsH\x00\x12-\n\x05\x63hunk\x18\x02 \x01(\x0b\x32\x1c.speechkit.stt.v3.AudioChunkH\x00\x12\x37\n\rsilence_chunk\x18\x03 \x01(\x0b\x32\x1e.speechkit.stt.v3.SilenceChunkH\x00\x12$\n\x03\x65ou\x18\x04 \x01(\x0b\x32\x15.speechkit.stt.v3.EouH\x00\x42\x07\n\x05\x45vent\"\xe3\x02\n\x14RecognizeFileRequest\x12\x11\n\x07\x63ontent\x18\x01 \x01(\x0cH\x00\x12\r\n\x03uri\x18\x02 \x01(\tH\x00\x12\x44\n\x11recognition_model\x18\x03 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12N\n\x16recognition_classifier\x18\x04 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x05 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x06 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptionsB\r\n\x0b\x41udioSource\"@\n\x04Word\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"@\n\x12LanguageEstimation\x12\x15\n\rlanguage_code\x18\x01 \x01(\t\x12\x13\n\x0bprobability\x18\x02 \x01(\x01\"\xbb\x01\n\x0b\x41lternative\x12%\n\x05words\x18\x01 \x03(\x0b\x32\x16.speechkit.stt.v3.Word\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x15\n\rstart_time_ms\x18\x03 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x04 \x01(\x03\x12\x12\n\nconfidence\x18\x05 \x01(\x01\x12\x37\n\tlanguages\x18\x06 \x03(\x0b\x32$.speechkit.stt.v3.LanguageEstimation\"\x1c\n\tEouUpdate\x12\x0f\n\x07time_ms\x18\x02 \x01(\x03\"a\n\x11\x41lternativeUpdate\x12\x33\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32\x1d.speechkit.stt.v3.Alternative\x12\x17\n\x0b\x63hannel_tag\x18\x02 \x01(\tB\x02\x18\x01\"\x99\x01\n\x0c\x41udioCursors\x12\x18\n\x10received_data_ms\x18\x01 \x01(\x03\x12\x15\n\rreset_time_ms\x18\x02 \x01(\x03\x12\x17\n\x0fpartial_time_ms\x18\x03 \x01(\x03\x12\x15\n\rfinal_time_ms\x18\x04 \x01(\x03\x12\x13\n\x0b\x66inal_index\x18\x05 \x01(\x03\x12\x13\n\x0b\x65ou_time_ms\x18\x06 \x01(\x03\"n\n\x0f\x46inalRefinement\x12\x13\n\x0b\x66inal_index\x18\x01 \x01(\x03\x12>\n\x0fnormalized_text\x18\x02 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x42\x06\n\x04Type\"L\n\nStatusCode\x12-\n\tcode_type\x18\x01 \x01(\x0e\x32\x1a.speechkit.stt.v3.CodeType\x12\x0f\n\x07message\x18\x02 \x01(\t\"4\n\x0bSessionUuid\x12\x0c\n\x04uuid\x18\x01 \x01(\t\x12\x17\n\x0fuser_request_id\x18\x02 \x01(\t\"K\n\x0fPhraseHighlight\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"?\n\x1aRecognitionClassifierLabel\x12\r\n\x05label\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x01\"\xa6\x01\n\x1bRecognitionClassifierResult\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x35\n\nhighlights\x18\x02 \x03(\x0b\x32!.speechkit.stt.v3.PhraseHighlight\x12<\n\x06labels\x18\x03 \x03(\x0b\x32,.speechkit.stt.v3.RecognitionClassifierLabel\"\xc7\x02\n\x1bRecognitionClassifierUpdate\x12M\n\x0bwindow_type\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.RecognitionClassifierUpdate.WindowType\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\x12H\n\x11\x63lassifier_result\x18\x04 \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierResult\"c\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x01\x12\x0e\n\nLAST_FINAL\x10\x02\x12\x10\n\x0cLAST_PARTIAL\x10\x03\"\xbb\x01\n\x15\x44\x65scriptiveStatistics\x12\x0b\n\x03min\x18\x01 \x01(\x01\x12\x0b\n\x03max\x18\x02 \x01(\x01\x12\x0c\n\x04mean\x18\x03 \x01(\x01\x12\x0b\n\x03std\x18\x04 \x01(\x01\x12\x43\n\tquantiles\x18\x05 \x03(\x0b\x32\x30.speechkit.stt.v3.DescriptiveStatistics.Quantile\x1a(\n\x08Quantile\x12\r\n\x05level\x18\x01 \x01(\x01\x12\r\n\x05value\x18\x02 \x01(\x01\"D\n\x16\x41udioSegmentBoundaries\x12\x15\n\rstart_time_ms\x18\x01 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x02 \x01(\x03\"\x87\x06\n\x0fSpeakerAnalysis\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x41\n\x0bwindow_type\x18\x02 \x01(\x0e\x32,.speechkit.stt.v3.SpeakerAnalysis.WindowType\x12\x43\n\x11speech_boundaries\x18\x03 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12\x17\n\x0ftotal_speech_ms\x18\x04 \x01(\x03\x12\x14\n\x0cspeech_ratio\x18\x05 \x01(\x01\x12\x18\n\x10total_silence_ms\x18\x06 \x01(\x03\x12\x15\n\rsilence_ratio\x18\x07 \x01(\x01\x12\x13\n\x0bwords_count\x18\x08 \x01(\x03\x12\x15\n\rletters_count\x18\t \x01(\x03\x12\x41\n\x10words_per_second\x18\n \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x43\n\x12letters_per_second\x18\x0b \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x44\n\x13words_per_utterance\x18\x0c \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x46\n\x15letters_per_utterance\x18\r \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x17\n\x0futterance_count\x18\x0e \x01(\x03\x12N\n\x1dutterance_duration_estimation\x18\x0f \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\"L\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\t\n\x05TOTAL\x10\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x02\"\x85\x06\n\x14\x43onversationAnalysis\x12I\n\x17\x63onversation_boundaries\x18\x01 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12.\n&total_simultaneous_silence_duration_ms\x18\x02 \x01(\x03\x12(\n total_simultaneous_silence_ratio\x18\x03 \x01(\x01\x12Y\n(simultaneous_silence_duration_estimation\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12-\n%total_simultaneous_speech_duration_ms\x18\x05 \x01(\x03\x12\'\n\x1ftotal_simultaneous_speech_ratio\x18\x06 \x01(\x01\x12X\n\'simultaneous_speech_duration_estimation\x18\x07 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12W\n\x12speaker_interrupts\x18\x08 \x03(\x0b\x32;.speechkit.stt.v3.ConversationAnalysis.InterruptsEvaluation\x12 \n\x18total_speech_duration_ms\x18\t \x01(\x03\x12\x1a\n\x12total_speech_ratio\x18\n \x01(\x01\x1a\xa3\x01\n\x14InterruptsEvaluation\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x18\n\x10interrupts_count\x18\x02 \x01(\x03\x12\x1e\n\x16interrupts_duration_ms\x18\x03 \x01(\x03\x12<\n\ninterrupts\x18\x04 \x03(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\"\xa5\x05\n\x11StreamingResponse\x12\x33\n\x0csession_uuid\x18\x01 \x01(\x0b\x32\x1d.speechkit.stt.v3.SessionUuid\x12\x35\n\raudio_cursors\x18\x02 \x01(\x0b\x32\x1e.speechkit.stt.v3.AudioCursors\x12\x1d\n\x15response_wall_time_ms\x18\x03 \x01(\x03\x12\x36\n\x07partial\x18\x04 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x34\n\x05\x66inal\x18\x05 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x31\n\neou_update\x18\x06 \x01(\x0b\x32\x1b.speechkit.stt.v3.EouUpdateH\x00\x12=\n\x10\x66inal_refinement\x18\x07 \x01(\x0b\x32!.speechkit.stt.v3.FinalRefinementH\x00\x12\x33\n\x0bstatus_code\x18\x08 \x01(\x0b\x32\x1c.speechkit.stt.v3.StatusCodeH\x00\x12J\n\x11\x63lassifier_update\x18\n \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierUpdateH\x00\x12=\n\x10speaker_analysis\x18\x0b \x01(\x0b\x32!.speechkit.stt.v3.SpeakerAnalysisH\x00\x12G\n\x15\x63onversation_analysis\x18\x0c \x01(\x0b\x32&.speechkit.stt.v3.ConversationAnalysisH\x00\x12\x13\n\x0b\x63hannel_tag\x18\t \x01(\tB\x07\n\x05\x45vent*O\n\x08\x43odeType\x12\x1d\n\x15\x43ODE_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x0b\n\x07WORKING\x10\x01\x12\x0b\n\x07WARNING\x10\x02\x12\n\n\x06\x43LOSED\x10\x03\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
|
|
16
|
+
DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n yandex/cloud/ai/stt/v3/stt.proto\x12\x10speechkit.stt.v3\"\xe2\x03\n\x18TextNormalizationOptions\x12X\n\x12text_normalization\x18\x01 \x01(\x0e\x32<.speechkit.stt.v3.TextNormalizationOptions.TextNormalization\x12\x18\n\x10profanity_filter\x18\x02 \x01(\x08\x12\x17\n\x0fliterature_text\x18\x03 \x01(\x08\x12]\n\x15phone_formatting_mode\x18\x04 \x01(\x0e\x32>.speechkit.stt.v3.TextNormalizationOptions.PhoneFormattingMode\"x\n\x11TextNormalization\x12\"\n\x1eTEXT_NORMALIZATION_UNSPECIFIED\x10\x00\x12\x1e\n\x1aTEXT_NORMALIZATION_ENABLED\x10\x01\x12\x1f\n\x1bTEXT_NORMALIZATION_DISABLED\x10\x02\"`\n\x13PhoneFormattingMode\x12%\n!PHONE_FORMATTING_MODE_UNSPECIFIED\x10\x00\x12\"\n\x1ePHONE_FORMATTING_MODE_DISABLED\x10\x01\"\xce\x01\n\x14\x44\x65\x66\x61ultEouClassifier\x12\x43\n\x04type\x18\x01 \x01(\x0e\x32\x35.speechkit.stt.v3.DefaultEouClassifier.EouSensitivity\x12\'\n\x1fmax_pause_between_words_hint_ms\x18\x02 \x01(\x03\"H\n\x0e\x45ouSensitivity\x12\x1f\n\x1b\x45OU_SENSITIVITY_UNSPECIFIED\x10\x00\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x01\x12\x08\n\x04HIGH\x10\x02\"\x17\n\x15\x45xternalEouClassifier\"\xb2\x01\n\x14\x45ouClassifierOptions\x12\x44\n\x12\x64\x65\x66\x61ult_classifier\x18\x01 \x01(\x0b\x32&.speechkit.stt.v3.DefaultEouClassifierH\x00\x12\x46\n\x13\x65xternal_classifier\x18\x02 \x01(\x0b\x32\'.speechkit.stt.v3.ExternalEouClassifierH\x00\x42\x0c\n\nClassifier\"\xd3\x01\n\x15RecognitionClassifier\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x45\n\x08triggers\x18\x02 \x03(\x0e\x32\x33.speechkit.stt.v3.RecognitionClassifier.TriggerType\"_\n\x0bTriggerType\x12 \n\x18TRIGGER_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x10\n\x0cON_UTTERANCE\x10\x01\x12\x0c\n\x08ON_FINAL\x10\x02\x12\x0e\n\nON_PARTIAL\x10\x03\"\\\n\x1cRecognitionClassifierOptions\x12<\n\x0b\x63lassifiers\x18\x01 \x03(\x0b\x32\'.speechkit.stt.v3.RecognitionClassifier\"\x88\x01\n\x15SpeechAnalysisOptions\x12\x1f\n\x17\x65nable_speaker_analysis\x18\x01 \x01(\x08\x12$\n\x1c\x65nable_conversation_analysis\x18\x02 \x01(\x08\x12(\n descriptive_statistics_quantiles\x18\x03 \x03(\x01\"\xc7\x01\n\x08RawAudio\x12@\n\x0e\x61udio_encoding\x18\x01 \x01(\x0e\x32(.speechkit.stt.v3.RawAudio.AudioEncoding\x12\x19\n\x11sample_rate_hertz\x18\x02 \x01(\x03\x12\x1b\n\x13\x61udio_channel_count\x18\x03 \x01(\x03\"A\n\rAudioEncoding\x12\x1e\n\x1a\x41UDIO_ENCODING_UNSPECIFIED\x10\x00\x12\x10\n\x0cLINEAR16_PCM\x10\x01\"\xbf\x01\n\x0e\x43ontainerAudio\x12Q\n\x14\x63ontainer_audio_type\x18\x01 \x01(\x0e\x32\x33.speechkit.stt.v3.ContainerAudio.ContainerAudioType\"Z\n\x12\x43ontainerAudioType\x12$\n CONTAINER_AUDIO_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03WAV\x10\x01\x12\x0c\n\x08OGG_OPUS\x10\x02\x12\x07\n\x03MP3\x10\x03\"\x91\x01\n\x12\x41udioFormatOptions\x12/\n\traw_audio\x18\x01 \x01(\x0b\x32\x1a.speechkit.stt.v3.RawAudioH\x00\x12;\n\x0f\x63ontainer_audio\x18\x02 \x01(\x0b\x32 .speechkit.stt.v3.ContainerAudioH\x00\x42\r\n\x0b\x41udioFormat\"\xf7\x01\n\x1aLanguageRestrictionOptions\x12^\n\x10restriction_type\x18\x01 \x01(\x0e\x32\x44.speechkit.stt.v3.LanguageRestrictionOptions.LanguageRestrictionType\x12\x15\n\rlanguage_code\x18\x02 \x03(\t\"b\n\x17LanguageRestrictionType\x12)\n%LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tWHITELIST\x10\x01\x12\r\n\tBLACKLIST\x10\x02\"\xb2\x03\n\x17RecognitionModelOptions\x12\r\n\x05model\x18\x01 \x01(\t\x12:\n\x0c\x61udio_format\x18\x02 \x01(\x0b\x32$.speechkit.stt.v3.AudioFormatOptions\x12\x46\n\x12text_normalization\x18\x03 \x01(\x0b\x32*.speechkit.stt.v3.TextNormalizationOptions\x12J\n\x14language_restriction\x18\x04 \x01(\x0b\x32,.speechkit.stt.v3.LanguageRestrictionOptions\x12\\\n\x15\x61udio_processing_type\x18\x05 \x01(\x0e\x32=.speechkit.stt.v3.RecognitionModelOptions.AudioProcessingType\"Z\n\x13\x41udioProcessingType\x12%\n!AUDIO_PROCESSING_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tREAL_TIME\x10\x01\x12\r\n\tFULL_DATA\x10\x02\"\xde\x01\n\x16SpeakerLabelingOptions\x12R\n\x10speaker_labeling\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.SpeakerLabelingOptions.SpeakerLabeling\"p\n\x0fSpeakerLabeling\x12 \n\x1cSPEAKER_LABELING_UNSPECIFIED\x10\x00\x12\x1c\n\x18SPEAKER_LABELING_ENABLED\x10\x01\x12\x1d\n\x19SPEAKER_LABELING_DISABLED\x10\x02\"\xee\x02\n\x10StreamingOptions\x12\x44\n\x11recognition_model\x18\x01 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12>\n\x0e\x65ou_classifier\x18\x02 \x01(\x0b\x32&.speechkit.stt.v3.EouClassifierOptions\x12N\n\x16recognition_classifier\x18\x03 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x05 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptions\"\x1a\n\nAudioChunk\x12\x0c\n\x04\x64\x61ta\x18\x01 \x01(\x0c\"#\n\x0cSilenceChunk\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\x03\"\x05\n\x03\x45ou\"\xe8\x01\n\x10StreamingRequest\x12=\n\x0fsession_options\x18\x01 \x01(\x0b\x32\".speechkit.stt.v3.StreamingOptionsH\x00\x12-\n\x05\x63hunk\x18\x02 \x01(\x0b\x32\x1c.speechkit.stt.v3.AudioChunkH\x00\x12\x37\n\rsilence_chunk\x18\x03 \x01(\x0b\x32\x1e.speechkit.stt.v3.SilenceChunkH\x00\x12$\n\x03\x65ou\x18\x04 \x01(\x0b\x32\x15.speechkit.stt.v3.EouH\x00\x42\x07\n\x05\x45vent\"\xe3\x02\n\x14RecognizeFileRequest\x12\x11\n\x07\x63ontent\x18\x01 \x01(\x0cH\x00\x12\r\n\x03uri\x18\x02 \x01(\tH\x00\x12\x44\n\x11recognition_model\x18\x03 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12N\n\x16recognition_classifier\x18\x04 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x05 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x06 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptionsB\r\n\x0b\x41udioSource\"@\n\x04Word\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"@\n\x12LanguageEstimation\x12\x15\n\rlanguage_code\x18\x01 \x01(\t\x12\x13\n\x0bprobability\x18\x02 \x01(\x01\"\xbb\x01\n\x0b\x41lternative\x12%\n\x05words\x18\x01 \x03(\x0b\x32\x16.speechkit.stt.v3.Word\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x15\n\rstart_time_ms\x18\x03 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x04 \x01(\x03\x12\x12\n\nconfidence\x18\x05 \x01(\x01\x12\x37\n\tlanguages\x18\x06 \x03(\x0b\x32$.speechkit.stt.v3.LanguageEstimation\"\x1c\n\tEouUpdate\x12\x0f\n\x07time_ms\x18\x02 \x01(\x03\"a\n\x11\x41lternativeUpdate\x12\x33\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32\x1d.speechkit.stt.v3.Alternative\x12\x17\n\x0b\x63hannel_tag\x18\x02 \x01(\tB\x02\x18\x01\"\x99\x01\n\x0c\x41udioCursors\x12\x18\n\x10received_data_ms\x18\x01 \x01(\x03\x12\x15\n\rreset_time_ms\x18\x02 \x01(\x03\x12\x17\n\x0fpartial_time_ms\x18\x03 \x01(\x03\x12\x15\n\rfinal_time_ms\x18\x04 \x01(\x03\x12\x13\n\x0b\x66inal_index\x18\x05 \x01(\x03\x12\x13\n\x0b\x65ou_time_ms\x18\x06 \x01(\x03\"n\n\x0f\x46inalRefinement\x12\x13\n\x0b\x66inal_index\x18\x01 \x01(\x03\x12>\n\x0fnormalized_text\x18\x02 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x42\x06\n\x04Type\"L\n\nStatusCode\x12-\n\tcode_type\x18\x01 \x01(\x0e\x32\x1a.speechkit.stt.v3.CodeType\x12\x0f\n\x07message\x18\x02 \x01(\t\"4\n\x0bSessionUuid\x12\x0c\n\x04uuid\x18\x01 \x01(\t\x12\x17\n\x0fuser_request_id\x18\x02 \x01(\t\"K\n\x0fPhraseHighlight\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"?\n\x1aRecognitionClassifierLabel\x12\r\n\x05label\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x01\"\xa6\x01\n\x1bRecognitionClassifierResult\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x35\n\nhighlights\x18\x02 \x03(\x0b\x32!.speechkit.stt.v3.PhraseHighlight\x12<\n\x06labels\x18\x03 \x03(\x0b\x32,.speechkit.stt.v3.RecognitionClassifierLabel\"\xc7\x02\n\x1bRecognitionClassifierUpdate\x12M\n\x0bwindow_type\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.RecognitionClassifierUpdate.WindowType\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\x12H\n\x11\x63lassifier_result\x18\x04 \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierResult\"c\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x01\x12\x0e\n\nLAST_FINAL\x10\x02\x12\x10\n\x0cLAST_PARTIAL\x10\x03\"\xbb\x01\n\x15\x44\x65scriptiveStatistics\x12\x0b\n\x03min\x18\x01 \x01(\x01\x12\x0b\n\x03max\x18\x02 \x01(\x01\x12\x0c\n\x04mean\x18\x03 \x01(\x01\x12\x0b\n\x03std\x18\x04 \x01(\x01\x12\x43\n\tquantiles\x18\x05 \x03(\x0b\x32\x30.speechkit.stt.v3.DescriptiveStatistics.Quantile\x1a(\n\x08Quantile\x12\r\n\x05level\x18\x01 \x01(\x01\x12\r\n\x05value\x18\x02 \x01(\x01\"D\n\x16\x41udioSegmentBoundaries\x12\x15\n\rstart_time_ms\x18\x01 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x02 \x01(\x03\"\x87\x06\n\x0fSpeakerAnalysis\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x41\n\x0bwindow_type\x18\x02 \x01(\x0e\x32,.speechkit.stt.v3.SpeakerAnalysis.WindowType\x12\x43\n\x11speech_boundaries\x18\x03 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12\x17\n\x0ftotal_speech_ms\x18\x04 \x01(\x03\x12\x14\n\x0cspeech_ratio\x18\x05 \x01(\x01\x12\x18\n\x10total_silence_ms\x18\x06 \x01(\x03\x12\x15\n\rsilence_ratio\x18\x07 \x01(\x01\x12\x13\n\x0bwords_count\x18\x08 \x01(\x03\x12\x15\n\rletters_count\x18\t \x01(\x03\x12\x41\n\x10words_per_second\x18\n \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x43\n\x12letters_per_second\x18\x0b \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x44\n\x13words_per_utterance\x18\x0c \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x46\n\x15letters_per_utterance\x18\r \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x17\n\x0futterance_count\x18\x0e \x01(\x03\x12N\n\x1dutterance_duration_estimation\x18\x0f \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\"L\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\t\n\x05TOTAL\x10\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x02\"\x85\x06\n\x14\x43onversationAnalysis\x12I\n\x17\x63onversation_boundaries\x18\x01 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12.\n&total_simultaneous_silence_duration_ms\x18\x02 \x01(\x03\x12(\n total_simultaneous_silence_ratio\x18\x03 \x01(\x01\x12Y\n(simultaneous_silence_duration_estimation\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12-\n%total_simultaneous_speech_duration_ms\x18\x05 \x01(\x03\x12\'\n\x1ftotal_simultaneous_speech_ratio\x18\x06 \x01(\x01\x12X\n\'simultaneous_speech_duration_estimation\x18\x07 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12W\n\x12speaker_interrupts\x18\x08 \x03(\x0b\x32;.speechkit.stt.v3.ConversationAnalysis.InterruptsEvaluation\x12 \n\x18total_speech_duration_ms\x18\t \x01(\x03\x12\x1a\n\x12total_speech_ratio\x18\n \x01(\x01\x1a\xa3\x01\n\x14InterruptsEvaluation\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x18\n\x10interrupts_count\x18\x02 \x01(\x03\x12\x1e\n\x16interrupts_duration_ms\x18\x03 \x01(\x03\x12<\n\ninterrupts\x18\x04 \x03(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\"\xa5\x05\n\x11StreamingResponse\x12\x33\n\x0csession_uuid\x18\x01 \x01(\x0b\x32\x1d.speechkit.stt.v3.SessionUuid\x12\x35\n\raudio_cursors\x18\x02 \x01(\x0b\x32\x1e.speechkit.stt.v3.AudioCursors\x12\x1d\n\x15response_wall_time_ms\x18\x03 \x01(\x03\x12\x36\n\x07partial\x18\x04 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x34\n\x05\x66inal\x18\x05 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x31\n\neou_update\x18\x06 \x01(\x0b\x32\x1b.speechkit.stt.v3.EouUpdateH\x00\x12=\n\x10\x66inal_refinement\x18\x07 \x01(\x0b\x32!.speechkit.stt.v3.FinalRefinementH\x00\x12\x33\n\x0bstatus_code\x18\x08 \x01(\x0b\x32\x1c.speechkit.stt.v3.StatusCodeH\x00\x12J\n\x11\x63lassifier_update\x18\n \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierUpdateH\x00\x12=\n\x10speaker_analysis\x18\x0b \x01(\x0b\x32!.speechkit.stt.v3.SpeakerAnalysisH\x00\x12G\n\x15\x63onversation_analysis\x18\x0c \x01(\x0b\x32&.speechkit.stt.v3.ConversationAnalysisH\x00\x12\x13\n\x0b\x63hannel_tag\x18\t \x01(\tB\x07\n\x05\x45vent\"0\n\x18\x44\x65leteRecognitionRequest\x12\x14\n\x0coperation_id\x18\x01 \x01(\t*O\n\x08\x43odeType\x12\x1d\n\x15\x43ODE_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x0b\n\x07WORKING\x10\x01\x12\x0b\n\x07WARNING\x10\x02\x12\n\n\x06\x43LOSED\x10\x03\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
|
|
17
17
|
|
|
18
18
|
_globals = globals()
|
|
19
19
|
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
|
|
@@ -31,8 +31,8 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
31
31
|
_RECOGNITIONCLASSIFIERUPDATE_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._serialized_options = b'\010\001'
|
|
32
32
|
_SPEAKERANALYSIS_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._options = None
|
|
33
33
|
_SPEAKERANALYSIS_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._serialized_options = b'\010\001'
|
|
34
|
-
_globals['_CODETYPE']._serialized_start=
|
|
35
|
-
_globals['_CODETYPE']._serialized_end=
|
|
34
|
+
_globals['_CODETYPE']._serialized_start=7927
|
|
35
|
+
_globals['_CODETYPE']._serialized_end=8006
|
|
36
36
|
_globals['_TEXTNORMALIZATIONOPTIONS']._serialized_start=55
|
|
37
37
|
_globals['_TEXTNORMALIZATIONOPTIONS']._serialized_end=537
|
|
38
38
|
_globals['_TEXTNORMALIZATIONOPTIONS_TEXTNORMALIZATION']._serialized_start=319
|
|
@@ -133,4 +133,6 @@ if _descriptor._USE_C_DESCRIPTORS == False:
|
|
|
133
133
|
_globals['_CONVERSATIONANALYSIS_INTERRUPTSEVALUATION']._serialized_end=7195
|
|
134
134
|
_globals['_STREAMINGRESPONSE']._serialized_start=7198
|
|
135
135
|
_globals['_STREAMINGRESPONSE']._serialized_end=7875
|
|
136
|
+
_globals['_DELETERECOGNITIONREQUEST']._serialized_start=7877
|
|
137
|
+
_globals['_DELETERECOGNITIONREQUEST']._serialized_end=7925
|
|
136
138
|
# @@protoc_insertion_point(module_scope)
|
|
@@ -1435,3 +1435,18 @@ class StreamingResponse(google.protobuf.message.Message):
|
|
|
1435
1435
|
def WhichOneof(self, oneof_group: typing.Literal["Event", b"Event"]) -> typing.Literal["partial", "final", "eou_update", "final_refinement", "status_code", "classifier_update", "speaker_analysis", "conversation_analysis"] | None: ...
|
|
1436
1436
|
|
|
1437
1437
|
global___StreamingResponse = StreamingResponse
|
|
1438
|
+
|
|
1439
|
+
@typing.final
|
|
1440
|
+
class DeleteRecognitionRequest(google.protobuf.message.Message):
|
|
1441
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
1442
|
+
|
|
1443
|
+
OPERATION_ID_FIELD_NUMBER: builtins.int
|
|
1444
|
+
operation_id: builtins.str
|
|
1445
|
+
def __init__(
|
|
1446
|
+
self,
|
|
1447
|
+
*,
|
|
1448
|
+
operation_id: builtins.str = ...,
|
|
1449
|
+
) -> None: ...
|
|
1450
|
+
def ClearField(self, field_name: typing.Literal["operation_id", b"operation_id"]) -> None: ...
|
|
1451
|
+
|
|
1452
|
+
global___DeleteRecognitionRequest = DeleteRecognitionRequest
|