yandexcloud 0.324.0__py3-none-any.whl → 0.325.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (119) hide show
  1. yandex/cloud/ai/foundation_models/v1/embedding/embedding_service_pb2.pyi +1 -1
  2. yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_pb2.pyi +4 -2
  3. yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2.pyi +6 -3
  4. yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2_grpc.py +3 -3
  5. yandex/cloud/ai/foundation_models/v1/image_generation/image_generation_service_pb2_grpc.pyi +3 -3
  6. yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_pb2.pyi +3 -3
  7. yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2.pyi +25 -10
  8. yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.py +11 -5
  9. yandex/cloud/ai/foundation_models/v1/text_classification/text_classification_service_pb2_grpc.pyi +33 -9
  10. yandex/cloud/ai/foundation_models/v1/text_common_pb2.pyi +3 -3
  11. yandex/cloud/ai/stt/v3/stt_pb2.py +5 -3
  12. yandex/cloud/ai/stt/v3/stt_pb2.pyi +15 -0
  13. yandex/cloud/ai/stt/v3/stt_service_pb2.py +10 -7
  14. yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.py +34 -0
  15. yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.pyi +18 -0
  16. yandex/cloud/apploadbalancer/v1/backend_group_pb2.py +26 -26
  17. yandex/cloud/apploadbalancer/v1/backend_group_pb2.pyi +7 -1
  18. yandex/cloud/backup/v1/policy_pb2.pyi +1 -1
  19. yandex/cloud/cdn/v1/resource_service_pb2.py +14 -14
  20. yandex/cloud/cdn/v1/resource_service_pb2.pyi +5 -1
  21. yandex/cloud/cloudregistry/v1/artifact_pb2.pyi +2 -0
  22. yandex/cloud/cloudregistry/v1/artifact_service_pb2.pyi +3 -3
  23. yandex/cloud/cloudregistry/v1/artifact_service_pb2_grpc.py +9 -6
  24. yandex/cloud/cloudregistry/v1/artifact_service_pb2_grpc.pyi +15 -9
  25. yandex/cloud/cloudregistry/v1/registry_pb2.pyi +1 -1
  26. yandex/cloud/cloudregistry/v1/registry_service_pb2.py +42 -38
  27. yandex/cloud/cloudregistry/v1/registry_service_pb2_grpc.py +2 -2
  28. yandex/cloud/cloudregistry/v1/registry_service_pb2_grpc.pyi +6 -6
  29. yandex/cloud/datasphere/v2/jobs/jobs_pb2.py +9 -5
  30. yandex/cloud/datasphere/v2/jobs/jobs_pb2.pyi +96 -0
  31. yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.py +55 -53
  32. yandex/cloud/datasphere/v2/jobs/project_job_service_pb2.pyi +14 -2
  33. yandex/cloud/iam/v1/user_account_pb2.pyi +2 -2
  34. yandex/cloud/iam/v1/user_account_service_pb2_grpc.py +3 -3
  35. yandex/cloud/iam/v1/user_account_service_pb2_grpc.pyi +3 -3
  36. yandex/cloud/mdb/redis/v1/backup_pb2.py +4 -2
  37. yandex/cloud/mdb/redis/v1/backup_pb2.pyi +31 -1
  38. yandex/cloud/organizationmanager/v1/user_account_pb2.pyi +2 -2
  39. yandex/cloud/searchapi/__init__.py +0 -0
  40. yandex/cloud/searchapi/v2/__init__.py +0 -0
  41. yandex/cloud/searchapi/v2/search_query_pb2.py +37 -0
  42. yandex/cloud/searchapi/v2/search_query_pb2.pyi +100 -0
  43. yandex/cloud/searchapi/v2/search_query_pb2_grpc.py +4 -0
  44. yandex/cloud/searchapi/v2/search_query_pb2_grpc.pyi +17 -0
  45. yandex/cloud/searchapi/v2/search_service_pb2.py +61 -0
  46. yandex/cloud/searchapi/v2/search_service_pb2.pyi +218 -0
  47. yandex/cloud/searchapi/v2/search_service_pb2_grpc.py +70 -0
  48. yandex/cloud/searchapi/v2/search_service_pb2_grpc.pyi +48 -0
  49. yandex/cloud/serverless/containers/v1/container_pb2.py +42 -34
  50. yandex/cloud/serverless/containers/v1/container_pb2.pyi +52 -2
  51. yandex/cloud/serverless/containers/v1/container_service_pb2.py +18 -18
  52. yandex/cloud/serverless/containers/v1/container_service_pb2.pyi +8 -2
  53. yandex/cloud/serverless/eventrouter/v1/connector_pb2.py +9 -9
  54. yandex/cloud/serverless/eventrouter/v1/connector_pb2.pyi +8 -0
  55. yandex/cloud/serverless/eventrouter/v1/rule_pb2.py +31 -31
  56. yandex/cloud/serverless/eventrouter/v1/rule_pb2.pyi +4 -0
  57. yandex/cloud/serverless/workflows/v1/execution_service_pb2.py +23 -15
  58. yandex/cloud/serverless/workflows/v1/execution_service_pb2.pyi +32 -0
  59. yandex/cloud/serverless/workflows/v1/execution_service_pb2_grpc.py +34 -0
  60. yandex/cloud/serverless/workflows/v1/execution_service_pb2_grpc.pyi +20 -0
  61. yandex/cloud/smartwebsecurity/v1/security_profile_pb2.py +46 -40
  62. yandex/cloud/smartwebsecurity/v1/security_profile_pb2.pyi +47 -2
  63. yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.py +18 -18
  64. yandex/cloud/smartwebsecurity/v1/security_profile_service_pb2.pyi +16 -3
  65. yandex/cloud/video/v1/channel_pb2.py +1 -1
  66. yandex/cloud/video/v1/channel_service_pb2.py +58 -25
  67. yandex/cloud/video/v1/channel_service_pb2.pyi +40 -0
  68. yandex/cloud/video/v1/channel_service_pb2_grpc.py +34 -0
  69. yandex/cloud/video/v1/channel_service_pb2_grpc.pyi +20 -0
  70. yandex/cloud/video/v1/episode_pb2.py +10 -8
  71. yandex/cloud/video/v1/episode_pb2.pyi +19 -3
  72. yandex/cloud/video/v1/episode_service_pb2.py +96 -43
  73. yandex/cloud/video/v1/episode_service_pb2.pyi +68 -6
  74. yandex/cloud/video/v1/episode_service_pb2_grpc.py +34 -0
  75. yandex/cloud/video/v1/episode_service_pb2_grpc.pyi +20 -0
  76. yandex/cloud/video/v1/manifest_pb2.py +3 -3
  77. yandex/cloud/video/v1/manifest_pb2.pyi +1 -5
  78. yandex/cloud/video/v1/playlist_pb2.py +29 -0
  79. yandex/cloud/video/v1/playlist_pb2.pyi +89 -0
  80. yandex/cloud/video/v1/playlist_pb2_grpc.py +4 -0
  81. yandex/cloud/video/v1/playlist_pb2_grpc.pyi +17 -0
  82. yandex/cloud/video/v1/playlist_service_pb2.py +102 -0
  83. yandex/cloud/video/v1/playlist_service_pb2.pyi +331 -0
  84. yandex/cloud/video/v1/playlist_service_pb2_grpc.py +276 -0
  85. yandex/cloud/video/v1/playlist_service_pb2_grpc.pyi +172 -0
  86. yandex/cloud/video/v1/stream_line_pb2.py +1 -1
  87. yandex/cloud/video/v1/stream_line_service_pb2.py +122 -55
  88. yandex/cloud/video/v1/stream_line_service_pb2.pyi +80 -0
  89. yandex/cloud/video/v1/stream_line_service_pb2_grpc.py +68 -0
  90. yandex/cloud/video/v1/stream_line_service_pb2_grpc.pyi +40 -0
  91. yandex/cloud/video/v1/stream_pb2.py +1 -1
  92. yandex/cloud/video/v1/stream_service_pb2.py +98 -37
  93. yandex/cloud/video/v1/stream_service_pb2.pyi +80 -0
  94. yandex/cloud/video/v1/stream_service_pb2_grpc.py +68 -0
  95. yandex/cloud/video/v1/stream_service_pb2_grpc.pyi +40 -0
  96. yandex/cloud/video/v1/subtitle_pb2.py +29 -0
  97. yandex/cloud/video/v1/subtitle_pb2.pyi +90 -0
  98. yandex/cloud/video/v1/subtitle_pb2_grpc.py +4 -0
  99. yandex/cloud/video/v1/subtitle_pb2_grpc.pyi +17 -0
  100. yandex/cloud/video/v1/subtitle_service_pb2.py +83 -0
  101. yandex/cloud/video/v1/subtitle_service_pb2.pyi +212 -0
  102. yandex/cloud/video/v1/subtitle_service_pb2_grpc.py +208 -0
  103. yandex/cloud/video/v1/subtitle_service_pb2_grpc.pyi +132 -0
  104. yandex/cloud/video/v1/thumbnail_service_pb2.py +50 -21
  105. yandex/cloud/video/v1/thumbnail_service_pb2.pyi +48 -0
  106. yandex/cloud/video/v1/thumbnail_service_pb2_grpc.py +69 -0
  107. yandex/cloud/video/v1/thumbnail_service_pb2_grpc.pyi +41 -0
  108. yandex/cloud/video/v1/video_pb2.py +18 -14
  109. yandex/cloud/video/v1/video_pb2.pyi +55 -3
  110. yandex/cloud/video/v1/video_service_pb2.py +118 -49
  111. yandex/cloud/video/v1/video_service_pb2.pyi +161 -6
  112. yandex/cloud/video/v1/video_service_pb2_grpc.py +102 -0
  113. yandex/cloud/video/v1/video_service_pb2_grpc.pyi +60 -0
  114. {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/METADATA +1 -1
  115. {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/RECORD +119 -93
  116. {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/WHEEL +1 -1
  117. {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/AUTHORS +0 -0
  118. {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/LICENSE +0 -0
  119. {yandexcloud-0.324.0.dist-info → yandexcloud-0.325.0.dist-info}/top_level.txt +0 -0
@@ -21,7 +21,7 @@ class TextEmbeddingRequest(google.protobuf.message.Message):
21
21
  MODEL_URI_FIELD_NUMBER: builtins.int
22
22
  TEXT_FIELD_NUMBER: builtins.int
23
23
  model_uri: builtins.str
24
- """The [ID of the model](/docs/foundation-models/concepts/embeddings) to be used for obtaining text embeddings."""
24
+ """The [model URI](/docs/foundation-models/concepts/embeddings) to be used for obtaining text embeddings."""
25
25
  text: builtins.str
26
26
  """The input text for which the embedding is requested."""
27
27
  def __init__(
@@ -60,9 +60,11 @@ class ImageGenerationOptions(google.protobuf.message.Message):
60
60
  SEED_FIELD_NUMBER: builtins.int
61
61
  ASPECT_RATIO_FIELD_NUMBER: builtins.int
62
62
  mime_type: builtins.str
63
- """MIME type of generated image format."""
63
+ """The [MIME type](https://en.wikipedia.org/wiki/Media_type) of generated image format.
64
+ For possible specifications, see [documentation](/docs/foundation-models/concepts).
65
+ """
64
66
  seed: builtins.int
65
- """Seed for image generation."""
67
+ """Seed for image generation. It serves as a starting point for image generation from noise."""
66
68
  @property
67
69
  def aspect_ratio(self) -> global___AspectRatio:
68
70
  """Aspect ratio of generated image."""
@@ -15,7 +15,10 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
15
15
 
16
16
  @typing.final
17
17
  class ImageGenerationRequest(google.protobuf.message.Message):
18
- """Request for the service to generate an image."""
18
+ """Request for the service to generate an image.
19
+
20
+ For examples of usage, see [step-by-step guide](/docs/foundation-models/operations/yandexart/request).
21
+ """
19
22
 
20
23
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
21
24
 
@@ -23,7 +26,7 @@ class ImageGenerationRequest(google.protobuf.message.Message):
23
26
  MESSAGES_FIELD_NUMBER: builtins.int
24
27
  GENERATION_OPTIONS_FIELD_NUMBER: builtins.int
25
28
  model_uri: builtins.str
26
- """The [ID of the model](/docs/foundation-models/concepts/yandexart/models) to be used for image generation."""
29
+ """The [model URI](/docs/foundation-models/concepts/yandexart/models) to be used for image generation."""
27
30
  @property
28
31
  def messages(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.image_generation.image_generation_pb2.Message]:
29
32
  """A list of messages representing the context for the image generation model."""
@@ -53,7 +56,7 @@ class ImageGenerationResponse(google.protobuf.message.Message):
53
56
  IMAGE_FIELD_NUMBER: builtins.int
54
57
  MODEL_VERSION_FIELD_NUMBER: builtins.int
55
58
  image: builtins.bytes
56
- """The image is serialized as an array of bytes encoded in base64."""
59
+ """The image is serialized as an array of bytes encoded in [Base64](https://en.wikipedia.org/wiki/Base64)."""
57
60
  model_version: builtins.str
58
61
  """The model version changes with each new releases."""
59
62
  def __init__(
@@ -7,7 +7,7 @@ from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operati
7
7
 
8
8
 
9
9
  class ImageGenerationAsyncServiceStub(object):
10
- """Service for obtaining images from input data.
10
+ """Service for creating images based on a text description.
11
11
  """
12
12
 
13
13
  def __init__(self, channel):
@@ -24,7 +24,7 @@ class ImageGenerationAsyncServiceStub(object):
24
24
 
25
25
 
26
26
  class ImageGenerationAsyncServiceServicer(object):
27
- """Service for obtaining images from input data.
27
+ """Service for creating images based on a text description.
28
28
  """
29
29
 
30
30
  def Generate(self, request, context):
@@ -50,7 +50,7 @@ def add_ImageGenerationAsyncServiceServicer_to_server(servicer, server):
50
50
 
51
51
  # This class is part of an EXPERIMENTAL API.
52
52
  class ImageGenerationAsyncService(object):
53
- """Service for obtaining images from input data.
53
+ """Service for creating images based on a text description.
54
54
  """
55
55
 
56
56
  @staticmethod
@@ -19,7 +19,7 @@ class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type:
19
19
  ...
20
20
 
21
21
  class ImageGenerationAsyncServiceStub:
22
- """Service for obtaining images from input data."""
22
+ """Service for creating images based on a text description."""
23
23
 
24
24
  def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
25
25
  Generate: grpc.UnaryUnaryMultiCallable[
@@ -29,7 +29,7 @@ class ImageGenerationAsyncServiceStub:
29
29
  """A method for generating an image based on a textual description."""
30
30
 
31
31
  class ImageGenerationAsyncServiceAsyncStub:
32
- """Service for obtaining images from input data."""
32
+ """Service for creating images based on a text description."""
33
33
 
34
34
  Generate: grpc.aio.UnaryUnaryMultiCallable[
35
35
  yandex.cloud.ai.foundation_models.v1.image_generation.image_generation_service_pb2.ImageGenerationRequest,
@@ -38,7 +38,7 @@ class ImageGenerationAsyncServiceAsyncStub:
38
38
  """A method for generating an image based on a textual description."""
39
39
 
40
40
  class ImageGenerationAsyncServiceServicer(metaclass=abc.ABCMeta):
41
- """Service for obtaining images from input data."""
41
+ """Service for creating images based on a text description."""
42
42
 
43
43
  @abc.abstractmethod
44
44
  def Generate(
@@ -12,16 +12,16 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
12
12
 
13
13
  @typing.final
14
14
  class ClassificationLabel(google.protobuf.message.Message):
15
- """A pair of text label and corresponding confidence used in classification problems."""
15
+ """A pair of text labels and their corresponding confidence values."""
16
16
 
17
17
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
18
18
 
19
19
  LABEL_FIELD_NUMBER: builtins.int
20
20
  CONFIDENCE_FIELD_NUMBER: builtins.int
21
21
  label: builtins.str
22
- """A label with a class name."""
22
+ """A class name label."""
23
23
  confidence: builtins.float
24
- """Confidence of item's belonging to a class."""
24
+ """The probability of classifying text into a specific class."""
25
25
  def __init__(
26
26
  self,
27
27
  *,
@@ -15,14 +15,20 @@ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
15
15
 
16
16
  @typing.final
17
17
  class TextClassificationRequest(google.protobuf.message.Message):
18
- """Request for the service to classify text."""
18
+ """Request for the service to classify text with tuned model.
19
+
20
+ The names of the classes between which the model will be distributing requests must be specified during model tuning;
21
+ therefore, they are not provided in the request.
22
+
23
+ For examples of usage, see [step-by-step guides](/docs/operations/classifier/additionally-trained).
24
+ """
19
25
 
20
26
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
21
27
 
22
28
  MODEL_URI_FIELD_NUMBER: builtins.int
23
29
  TEXT_FIELD_NUMBER: builtins.int
24
30
  model_uri: builtins.str
25
- """The identifier of the classification model."""
31
+ """The [URI](/docs/foundation-models/concepts/classifier/models) of your tuned classifier model."""
26
32
  text: builtins.str
27
33
  """Text for classification."""
28
34
  def __init__(
@@ -37,17 +43,19 @@ global___TextClassificationRequest = TextClassificationRequest
37
43
 
38
44
  @typing.final
39
45
  class TextClassificationResponse(google.protobuf.message.Message):
40
- """Response containing classifier predictions."""
46
+ """Response with classifier predictions."""
41
47
 
42
48
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
43
49
 
44
50
  PREDICTIONS_FIELD_NUMBER: builtins.int
45
51
  MODEL_VERSION_FIELD_NUMBER: builtins.int
46
52
  model_version: builtins.str
47
- """Model version (changes with model releases)."""
53
+ """The model version changes with each new releases."""
48
54
  @property
49
55
  def predictions(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationLabel]:
50
- """Result of classification - a list of label-confidence pairs."""
56
+ """The classification results with the `confidence`` values
57
+ for the probability of classifying the request text into each class.
58
+ """
51
59
 
52
60
  def __init__(
53
61
  self,
@@ -61,7 +69,9 @@ global___TextClassificationResponse = TextClassificationResponse
61
69
 
62
70
  @typing.final
63
71
  class FewShotTextClassificationRequest(google.protobuf.message.Message):
64
- """Request for the service to classify text."""
72
+ """Request for the service to classify text.
73
+ For examples of usage, see [step-by-step guides](/docs/operations/classifier/readymade).
74
+ """
65
75
 
66
76
  DESCRIPTOR: google.protobuf.descriptor.Descriptor
67
77
 
@@ -71,14 +81,17 @@ class FewShotTextClassificationRequest(google.protobuf.message.Message):
71
81
  TEXT_FIELD_NUMBER: builtins.int
72
82
  SAMPLES_FIELD_NUMBER: builtins.int
73
83
  model_uri: builtins.str
74
- """The identifier of the classification model."""
84
+ """The [URI](/docs/foundation-models/concepts/classifier/models) of the classifier model."""
75
85
  task_description: builtins.str
76
86
  """Text description of the classification task."""
77
87
  text: builtins.str
78
88
  """Text for classification."""
79
89
  @property
80
90
  def labels(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
81
- """List of available labels for the classification result."""
91
+ """List of available labels for the classification result.
92
+ Give meaningful names to label classes: this is essential for correct classification results.
93
+ For example, use ``chemistry`` and ``physics`` rather than ``chm`` and ``phs`` for class names.
94
+ """
82
95
 
83
96
  @property
84
97
  def samples(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationSample]:
@@ -106,10 +119,12 @@ class FewShotTextClassificationResponse(google.protobuf.message.Message):
106
119
  PREDICTIONS_FIELD_NUMBER: builtins.int
107
120
  MODEL_VERSION_FIELD_NUMBER: builtins.int
108
121
  model_version: builtins.str
109
- """Model version (changes with model releases)."""
122
+ """The model version changes with each new releases."""
110
123
  @property
111
124
  def predictions(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_pb2.ClassificationLabel]:
112
- """Result of classification - a list of label-confidence pairs."""
125
+ """The classification results with the `confidence`` values
126
+ for the probability of classifying the request text into each class.
127
+ """
113
128
 
114
129
  def __init__(
115
130
  self,
@@ -6,7 +6,7 @@ from yandex.cloud.ai.foundation_models.v1.text_classification import text_classi
6
6
 
7
7
 
8
8
  class TextClassificationServiceStub(object):
9
- """Service for classifying text from input text.
9
+ """Service for classifying the text requests provided in prompts.
10
10
  """
11
11
 
12
12
  def __init__(self, channel):
@@ -28,18 +28,24 @@ class TextClassificationServiceStub(object):
28
28
 
29
29
 
30
30
  class TextClassificationServiceServicer(object):
31
- """Service for classifying text from input text.
31
+ """Service for classifying the text requests provided in prompts.
32
32
  """
33
33
 
34
34
  def Classify(self, request, context):
35
- """RPC method for text classification.
35
+ """RPC method to classify text with tuned model.
36
+
37
+ The names of the classes between which the model will be distributing requests
38
+ must be specified during model tuning and are not provided in the request.
36
39
  """
37
40
  context.set_code(grpc.StatusCode.UNIMPLEMENTED)
38
41
  context.set_details('Method not implemented!')
39
42
  raise NotImplementedError('Method not implemented!')
40
43
 
41
44
  def FewShotClassify(self, request, context):
42
- """RPC method for few-shot text classification.
45
+ """RPC method for binary and multi-class classification.
46
+
47
+ You can provide up to 20 classes for few-shot text classification
48
+ with optional examples.
43
49
  """
44
50
  context.set_code(grpc.StatusCode.UNIMPLEMENTED)
45
51
  context.set_details('Method not implemented!')
@@ -66,7 +72,7 @@ def add_TextClassificationServiceServicer_to_server(servicer, server):
66
72
 
67
73
  # This class is part of an EXPERIMENTAL API.
68
74
  class TextClassificationService(object):
69
- """Service for classifying text from input text.
75
+ """Service for classifying the text requests provided in prompts.
70
76
  """
71
77
 
72
78
  @staticmethod
@@ -18,38 +18,54 @@ class _ServicerContext(grpc.ServicerContext, grpc.aio.ServicerContext): # type:
18
18
  ...
19
19
 
20
20
  class TextClassificationServiceStub:
21
- """Service for classifying text from input text."""
21
+ """Service for classifying the text requests provided in prompts."""
22
22
 
23
23
  def __init__(self, channel: typing.Union[grpc.Channel, grpc.aio.Channel]) -> None: ...
24
24
  Classify: grpc.UnaryUnaryMultiCallable[
25
25
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
26
26
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse,
27
27
  ]
28
- """RPC method for text classification."""
28
+ """RPC method to classify text with tuned model.
29
+
30
+ The names of the classes between which the model will be distributing requests
31
+ must be specified during model tuning and are not provided in the request.
32
+ """
29
33
 
30
34
  FewShotClassify: grpc.UnaryUnaryMultiCallable[
31
35
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
32
36
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse,
33
37
  ]
34
- """RPC method for few-shot text classification."""
38
+ """RPC method for binary and multi-class classification.
39
+
40
+ You can provide up to 20 classes for few-shot text classification
41
+ with optional examples.
42
+ """
35
43
 
36
44
  class TextClassificationServiceAsyncStub:
37
- """Service for classifying text from input text."""
45
+ """Service for classifying the text requests provided in prompts."""
38
46
 
39
47
  Classify: grpc.aio.UnaryUnaryMultiCallable[
40
48
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
41
49
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse,
42
50
  ]
43
- """RPC method for text classification."""
51
+ """RPC method to classify text with tuned model.
52
+
53
+ The names of the classes between which the model will be distributing requests
54
+ must be specified during model tuning and are not provided in the request.
55
+ """
44
56
 
45
57
  FewShotClassify: grpc.aio.UnaryUnaryMultiCallable[
46
58
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
47
59
  yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse,
48
60
  ]
49
- """RPC method for few-shot text classification."""
61
+ """RPC method for binary and multi-class classification.
62
+
63
+ You can provide up to 20 classes for few-shot text classification
64
+ with optional examples.
65
+ """
50
66
 
51
67
  class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
52
- """Service for classifying text from input text."""
68
+ """Service for classifying the text requests provided in prompts."""
53
69
 
54
70
  @abc.abstractmethod
55
71
  def Classify(
@@ -57,7 +73,11 @@ class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
57
73
  request: yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationRequest,
58
74
  context: _ServicerContext,
59
75
  ) -> typing.Union[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse, collections.abc.Awaitable[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.TextClassificationResponse]]:
60
- """RPC method for text classification."""
76
+ """RPC method to classify text with tuned model.
77
+
78
+ The names of the classes between which the model will be distributing requests
79
+ must be specified during model tuning and are not provided in the request.
80
+ """
61
81
 
62
82
  @abc.abstractmethod
63
83
  def FewShotClassify(
@@ -65,6 +85,10 @@ class TextClassificationServiceServicer(metaclass=abc.ABCMeta):
65
85
  request: yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationRequest,
66
86
  context: _ServicerContext,
67
87
  ) -> typing.Union[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse, collections.abc.Awaitable[yandex.cloud.ai.foundation_models.v1.text_classification.text_classification_service_pb2.FewShotTextClassificationResponse]]:
68
- """RPC method for few-shot text classification."""
88
+ """RPC method for binary and multi-class classification.
89
+
90
+ You can provide up to 20 classes for few-shot text classification
91
+ with optional examples.
92
+ """
69
93
 
70
94
  def add_TextClassificationServiceServicer_to_server(servicer: TextClassificationServiceServicer, server: typing.Union[grpc.Server, grpc.aio.Server]) -> None: ...
@@ -109,7 +109,7 @@ class ContentUsage(google.protobuf.message.Message):
109
109
  input_text_tokens: builtins.int
110
110
  """The number of tokens in the textual part of the model input."""
111
111
  completion_tokens: builtins.int
112
- """The total number of tokens in the generated completions."""
112
+ """The number of tokens in the generated completion."""
113
113
  total_tokens: builtins.int
114
114
  """The total number of tokens, including all input tokens and all generated tokens."""
115
115
  def __init__(
@@ -171,10 +171,10 @@ class Alternative(google.protobuf.message.Message):
171
171
  MESSAGE_FIELD_NUMBER: builtins.int
172
172
  STATUS_FIELD_NUMBER: builtins.int
173
173
  status: global___Alternative.AlternativeStatus.ValueType
174
- """The generation status of the alternative"""
174
+ """The generation status of the alternative."""
175
175
  @property
176
176
  def message(self) -> global___Message:
177
- """A message containing the content of the alternative."""
177
+ """A message with the content of the alternative."""
178
178
 
179
179
  def __init__(
180
180
  self,
@@ -13,7 +13,7 @@ _sym_db = _symbol_database.Default()
13
13
 
14
14
 
15
15
 
16
- DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n yandex/cloud/ai/stt/v3/stt.proto\x12\x10speechkit.stt.v3\"\xe2\x03\n\x18TextNormalizationOptions\x12X\n\x12text_normalization\x18\x01 \x01(\x0e\x32<.speechkit.stt.v3.TextNormalizationOptions.TextNormalization\x12\x18\n\x10profanity_filter\x18\x02 \x01(\x08\x12\x17\n\x0fliterature_text\x18\x03 \x01(\x08\x12]\n\x15phone_formatting_mode\x18\x04 \x01(\x0e\x32>.speechkit.stt.v3.TextNormalizationOptions.PhoneFormattingMode\"x\n\x11TextNormalization\x12\"\n\x1eTEXT_NORMALIZATION_UNSPECIFIED\x10\x00\x12\x1e\n\x1aTEXT_NORMALIZATION_ENABLED\x10\x01\x12\x1f\n\x1bTEXT_NORMALIZATION_DISABLED\x10\x02\"`\n\x13PhoneFormattingMode\x12%\n!PHONE_FORMATTING_MODE_UNSPECIFIED\x10\x00\x12\"\n\x1ePHONE_FORMATTING_MODE_DISABLED\x10\x01\"\xce\x01\n\x14\x44\x65\x66\x61ultEouClassifier\x12\x43\n\x04type\x18\x01 \x01(\x0e\x32\x35.speechkit.stt.v3.DefaultEouClassifier.EouSensitivity\x12\'\n\x1fmax_pause_between_words_hint_ms\x18\x02 \x01(\x03\"H\n\x0e\x45ouSensitivity\x12\x1f\n\x1b\x45OU_SENSITIVITY_UNSPECIFIED\x10\x00\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x01\x12\x08\n\x04HIGH\x10\x02\"\x17\n\x15\x45xternalEouClassifier\"\xb2\x01\n\x14\x45ouClassifierOptions\x12\x44\n\x12\x64\x65\x66\x61ult_classifier\x18\x01 \x01(\x0b\x32&.speechkit.stt.v3.DefaultEouClassifierH\x00\x12\x46\n\x13\x65xternal_classifier\x18\x02 \x01(\x0b\x32\'.speechkit.stt.v3.ExternalEouClassifierH\x00\x42\x0c\n\nClassifier\"\xd3\x01\n\x15RecognitionClassifier\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x45\n\x08triggers\x18\x02 \x03(\x0e\x32\x33.speechkit.stt.v3.RecognitionClassifier.TriggerType\"_\n\x0bTriggerType\x12 \n\x18TRIGGER_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x10\n\x0cON_UTTERANCE\x10\x01\x12\x0c\n\x08ON_FINAL\x10\x02\x12\x0e\n\nON_PARTIAL\x10\x03\"\\\n\x1cRecognitionClassifierOptions\x12<\n\x0b\x63lassifiers\x18\x01 \x03(\x0b\x32\'.speechkit.stt.v3.RecognitionClassifier\"\x88\x01\n\x15SpeechAnalysisOptions\x12\x1f\n\x17\x65nable_speaker_analysis\x18\x01 \x01(\x08\x12$\n\x1c\x65nable_conversation_analysis\x18\x02 \x01(\x08\x12(\n descriptive_statistics_quantiles\x18\x03 \x03(\x01\"\xc7\x01\n\x08RawAudio\x12@\n\x0e\x61udio_encoding\x18\x01 \x01(\x0e\x32(.speechkit.stt.v3.RawAudio.AudioEncoding\x12\x19\n\x11sample_rate_hertz\x18\x02 \x01(\x03\x12\x1b\n\x13\x61udio_channel_count\x18\x03 \x01(\x03\"A\n\rAudioEncoding\x12\x1e\n\x1a\x41UDIO_ENCODING_UNSPECIFIED\x10\x00\x12\x10\n\x0cLINEAR16_PCM\x10\x01\"\xbf\x01\n\x0e\x43ontainerAudio\x12Q\n\x14\x63ontainer_audio_type\x18\x01 \x01(\x0e\x32\x33.speechkit.stt.v3.ContainerAudio.ContainerAudioType\"Z\n\x12\x43ontainerAudioType\x12$\n CONTAINER_AUDIO_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03WAV\x10\x01\x12\x0c\n\x08OGG_OPUS\x10\x02\x12\x07\n\x03MP3\x10\x03\"\x91\x01\n\x12\x41udioFormatOptions\x12/\n\traw_audio\x18\x01 \x01(\x0b\x32\x1a.speechkit.stt.v3.RawAudioH\x00\x12;\n\x0f\x63ontainer_audio\x18\x02 \x01(\x0b\x32 .speechkit.stt.v3.ContainerAudioH\x00\x42\r\n\x0b\x41udioFormat\"\xf7\x01\n\x1aLanguageRestrictionOptions\x12^\n\x10restriction_type\x18\x01 \x01(\x0e\x32\x44.speechkit.stt.v3.LanguageRestrictionOptions.LanguageRestrictionType\x12\x15\n\rlanguage_code\x18\x02 \x03(\t\"b\n\x17LanguageRestrictionType\x12)\n%LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tWHITELIST\x10\x01\x12\r\n\tBLACKLIST\x10\x02\"\xb2\x03\n\x17RecognitionModelOptions\x12\r\n\x05model\x18\x01 \x01(\t\x12:\n\x0c\x61udio_format\x18\x02 \x01(\x0b\x32$.speechkit.stt.v3.AudioFormatOptions\x12\x46\n\x12text_normalization\x18\x03 \x01(\x0b\x32*.speechkit.stt.v3.TextNormalizationOptions\x12J\n\x14language_restriction\x18\x04 \x01(\x0b\x32,.speechkit.stt.v3.LanguageRestrictionOptions\x12\\\n\x15\x61udio_processing_type\x18\x05 \x01(\x0e\x32=.speechkit.stt.v3.RecognitionModelOptions.AudioProcessingType\"Z\n\x13\x41udioProcessingType\x12%\n!AUDIO_PROCESSING_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tREAL_TIME\x10\x01\x12\r\n\tFULL_DATA\x10\x02\"\xde\x01\n\x16SpeakerLabelingOptions\x12R\n\x10speaker_labeling\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.SpeakerLabelingOptions.SpeakerLabeling\"p\n\x0fSpeakerLabeling\x12 \n\x1cSPEAKER_LABELING_UNSPECIFIED\x10\x00\x12\x1c\n\x18SPEAKER_LABELING_ENABLED\x10\x01\x12\x1d\n\x19SPEAKER_LABELING_DISABLED\x10\x02\"\xee\x02\n\x10StreamingOptions\x12\x44\n\x11recognition_model\x18\x01 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12>\n\x0e\x65ou_classifier\x18\x02 \x01(\x0b\x32&.speechkit.stt.v3.EouClassifierOptions\x12N\n\x16recognition_classifier\x18\x03 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x05 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptions\"\x1a\n\nAudioChunk\x12\x0c\n\x04\x64\x61ta\x18\x01 \x01(\x0c\"#\n\x0cSilenceChunk\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\x03\"\x05\n\x03\x45ou\"\xe8\x01\n\x10StreamingRequest\x12=\n\x0fsession_options\x18\x01 \x01(\x0b\x32\".speechkit.stt.v3.StreamingOptionsH\x00\x12-\n\x05\x63hunk\x18\x02 \x01(\x0b\x32\x1c.speechkit.stt.v3.AudioChunkH\x00\x12\x37\n\rsilence_chunk\x18\x03 \x01(\x0b\x32\x1e.speechkit.stt.v3.SilenceChunkH\x00\x12$\n\x03\x65ou\x18\x04 \x01(\x0b\x32\x15.speechkit.stt.v3.EouH\x00\x42\x07\n\x05\x45vent\"\xe3\x02\n\x14RecognizeFileRequest\x12\x11\n\x07\x63ontent\x18\x01 \x01(\x0cH\x00\x12\r\n\x03uri\x18\x02 \x01(\tH\x00\x12\x44\n\x11recognition_model\x18\x03 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12N\n\x16recognition_classifier\x18\x04 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x05 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x06 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptionsB\r\n\x0b\x41udioSource\"@\n\x04Word\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"@\n\x12LanguageEstimation\x12\x15\n\rlanguage_code\x18\x01 \x01(\t\x12\x13\n\x0bprobability\x18\x02 \x01(\x01\"\xbb\x01\n\x0b\x41lternative\x12%\n\x05words\x18\x01 \x03(\x0b\x32\x16.speechkit.stt.v3.Word\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x15\n\rstart_time_ms\x18\x03 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x04 \x01(\x03\x12\x12\n\nconfidence\x18\x05 \x01(\x01\x12\x37\n\tlanguages\x18\x06 \x03(\x0b\x32$.speechkit.stt.v3.LanguageEstimation\"\x1c\n\tEouUpdate\x12\x0f\n\x07time_ms\x18\x02 \x01(\x03\"a\n\x11\x41lternativeUpdate\x12\x33\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32\x1d.speechkit.stt.v3.Alternative\x12\x17\n\x0b\x63hannel_tag\x18\x02 \x01(\tB\x02\x18\x01\"\x99\x01\n\x0c\x41udioCursors\x12\x18\n\x10received_data_ms\x18\x01 \x01(\x03\x12\x15\n\rreset_time_ms\x18\x02 \x01(\x03\x12\x17\n\x0fpartial_time_ms\x18\x03 \x01(\x03\x12\x15\n\rfinal_time_ms\x18\x04 \x01(\x03\x12\x13\n\x0b\x66inal_index\x18\x05 \x01(\x03\x12\x13\n\x0b\x65ou_time_ms\x18\x06 \x01(\x03\"n\n\x0f\x46inalRefinement\x12\x13\n\x0b\x66inal_index\x18\x01 \x01(\x03\x12>\n\x0fnormalized_text\x18\x02 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x42\x06\n\x04Type\"L\n\nStatusCode\x12-\n\tcode_type\x18\x01 \x01(\x0e\x32\x1a.speechkit.stt.v3.CodeType\x12\x0f\n\x07message\x18\x02 \x01(\t\"4\n\x0bSessionUuid\x12\x0c\n\x04uuid\x18\x01 \x01(\t\x12\x17\n\x0fuser_request_id\x18\x02 \x01(\t\"K\n\x0fPhraseHighlight\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"?\n\x1aRecognitionClassifierLabel\x12\r\n\x05label\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x01\"\xa6\x01\n\x1bRecognitionClassifierResult\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x35\n\nhighlights\x18\x02 \x03(\x0b\x32!.speechkit.stt.v3.PhraseHighlight\x12<\n\x06labels\x18\x03 \x03(\x0b\x32,.speechkit.stt.v3.RecognitionClassifierLabel\"\xc7\x02\n\x1bRecognitionClassifierUpdate\x12M\n\x0bwindow_type\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.RecognitionClassifierUpdate.WindowType\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\x12H\n\x11\x63lassifier_result\x18\x04 \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierResult\"c\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x01\x12\x0e\n\nLAST_FINAL\x10\x02\x12\x10\n\x0cLAST_PARTIAL\x10\x03\"\xbb\x01\n\x15\x44\x65scriptiveStatistics\x12\x0b\n\x03min\x18\x01 \x01(\x01\x12\x0b\n\x03max\x18\x02 \x01(\x01\x12\x0c\n\x04mean\x18\x03 \x01(\x01\x12\x0b\n\x03std\x18\x04 \x01(\x01\x12\x43\n\tquantiles\x18\x05 \x03(\x0b\x32\x30.speechkit.stt.v3.DescriptiveStatistics.Quantile\x1a(\n\x08Quantile\x12\r\n\x05level\x18\x01 \x01(\x01\x12\r\n\x05value\x18\x02 \x01(\x01\"D\n\x16\x41udioSegmentBoundaries\x12\x15\n\rstart_time_ms\x18\x01 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x02 \x01(\x03\"\x87\x06\n\x0fSpeakerAnalysis\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x41\n\x0bwindow_type\x18\x02 \x01(\x0e\x32,.speechkit.stt.v3.SpeakerAnalysis.WindowType\x12\x43\n\x11speech_boundaries\x18\x03 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12\x17\n\x0ftotal_speech_ms\x18\x04 \x01(\x03\x12\x14\n\x0cspeech_ratio\x18\x05 \x01(\x01\x12\x18\n\x10total_silence_ms\x18\x06 \x01(\x03\x12\x15\n\rsilence_ratio\x18\x07 \x01(\x01\x12\x13\n\x0bwords_count\x18\x08 \x01(\x03\x12\x15\n\rletters_count\x18\t \x01(\x03\x12\x41\n\x10words_per_second\x18\n \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x43\n\x12letters_per_second\x18\x0b \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x44\n\x13words_per_utterance\x18\x0c \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x46\n\x15letters_per_utterance\x18\r \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x17\n\x0futterance_count\x18\x0e \x01(\x03\x12N\n\x1dutterance_duration_estimation\x18\x0f \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\"L\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\t\n\x05TOTAL\x10\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x02\"\x85\x06\n\x14\x43onversationAnalysis\x12I\n\x17\x63onversation_boundaries\x18\x01 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12.\n&total_simultaneous_silence_duration_ms\x18\x02 \x01(\x03\x12(\n total_simultaneous_silence_ratio\x18\x03 \x01(\x01\x12Y\n(simultaneous_silence_duration_estimation\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12-\n%total_simultaneous_speech_duration_ms\x18\x05 \x01(\x03\x12\'\n\x1ftotal_simultaneous_speech_ratio\x18\x06 \x01(\x01\x12X\n\'simultaneous_speech_duration_estimation\x18\x07 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12W\n\x12speaker_interrupts\x18\x08 \x03(\x0b\x32;.speechkit.stt.v3.ConversationAnalysis.InterruptsEvaluation\x12 \n\x18total_speech_duration_ms\x18\t \x01(\x03\x12\x1a\n\x12total_speech_ratio\x18\n \x01(\x01\x1a\xa3\x01\n\x14InterruptsEvaluation\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x18\n\x10interrupts_count\x18\x02 \x01(\x03\x12\x1e\n\x16interrupts_duration_ms\x18\x03 \x01(\x03\x12<\n\ninterrupts\x18\x04 \x03(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\"\xa5\x05\n\x11StreamingResponse\x12\x33\n\x0csession_uuid\x18\x01 \x01(\x0b\x32\x1d.speechkit.stt.v3.SessionUuid\x12\x35\n\raudio_cursors\x18\x02 \x01(\x0b\x32\x1e.speechkit.stt.v3.AudioCursors\x12\x1d\n\x15response_wall_time_ms\x18\x03 \x01(\x03\x12\x36\n\x07partial\x18\x04 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x34\n\x05\x66inal\x18\x05 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x31\n\neou_update\x18\x06 \x01(\x0b\x32\x1b.speechkit.stt.v3.EouUpdateH\x00\x12=\n\x10\x66inal_refinement\x18\x07 \x01(\x0b\x32!.speechkit.stt.v3.FinalRefinementH\x00\x12\x33\n\x0bstatus_code\x18\x08 \x01(\x0b\x32\x1c.speechkit.stt.v3.StatusCodeH\x00\x12J\n\x11\x63lassifier_update\x18\n \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierUpdateH\x00\x12=\n\x10speaker_analysis\x18\x0b \x01(\x0b\x32!.speechkit.stt.v3.SpeakerAnalysisH\x00\x12G\n\x15\x63onversation_analysis\x18\x0c \x01(\x0b\x32&.speechkit.stt.v3.ConversationAnalysisH\x00\x12\x13\n\x0b\x63hannel_tag\x18\t \x01(\tB\x07\n\x05\x45vent*O\n\x08\x43odeType\x12\x1d\n\x15\x43ODE_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x0b\n\x07WORKING\x10\x01\x12\x0b\n\x07WARNING\x10\x02\x12\n\n\x06\x43LOSED\x10\x03\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
16
+ DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n yandex/cloud/ai/stt/v3/stt.proto\x12\x10speechkit.stt.v3\"\xe2\x03\n\x18TextNormalizationOptions\x12X\n\x12text_normalization\x18\x01 \x01(\x0e\x32<.speechkit.stt.v3.TextNormalizationOptions.TextNormalization\x12\x18\n\x10profanity_filter\x18\x02 \x01(\x08\x12\x17\n\x0fliterature_text\x18\x03 \x01(\x08\x12]\n\x15phone_formatting_mode\x18\x04 \x01(\x0e\x32>.speechkit.stt.v3.TextNormalizationOptions.PhoneFormattingMode\"x\n\x11TextNormalization\x12\"\n\x1eTEXT_NORMALIZATION_UNSPECIFIED\x10\x00\x12\x1e\n\x1aTEXT_NORMALIZATION_ENABLED\x10\x01\x12\x1f\n\x1bTEXT_NORMALIZATION_DISABLED\x10\x02\"`\n\x13PhoneFormattingMode\x12%\n!PHONE_FORMATTING_MODE_UNSPECIFIED\x10\x00\x12\"\n\x1ePHONE_FORMATTING_MODE_DISABLED\x10\x01\"\xce\x01\n\x14\x44\x65\x66\x61ultEouClassifier\x12\x43\n\x04type\x18\x01 \x01(\x0e\x32\x35.speechkit.stt.v3.DefaultEouClassifier.EouSensitivity\x12\'\n\x1fmax_pause_between_words_hint_ms\x18\x02 \x01(\x03\"H\n\x0e\x45ouSensitivity\x12\x1f\n\x1b\x45OU_SENSITIVITY_UNSPECIFIED\x10\x00\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x01\x12\x08\n\x04HIGH\x10\x02\"\x17\n\x15\x45xternalEouClassifier\"\xb2\x01\n\x14\x45ouClassifierOptions\x12\x44\n\x12\x64\x65\x66\x61ult_classifier\x18\x01 \x01(\x0b\x32&.speechkit.stt.v3.DefaultEouClassifierH\x00\x12\x46\n\x13\x65xternal_classifier\x18\x02 \x01(\x0b\x32\'.speechkit.stt.v3.ExternalEouClassifierH\x00\x42\x0c\n\nClassifier\"\xd3\x01\n\x15RecognitionClassifier\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x45\n\x08triggers\x18\x02 \x03(\x0e\x32\x33.speechkit.stt.v3.RecognitionClassifier.TriggerType\"_\n\x0bTriggerType\x12 \n\x18TRIGGER_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x10\n\x0cON_UTTERANCE\x10\x01\x12\x0c\n\x08ON_FINAL\x10\x02\x12\x0e\n\nON_PARTIAL\x10\x03\"\\\n\x1cRecognitionClassifierOptions\x12<\n\x0b\x63lassifiers\x18\x01 \x03(\x0b\x32\'.speechkit.stt.v3.RecognitionClassifier\"\x88\x01\n\x15SpeechAnalysisOptions\x12\x1f\n\x17\x65nable_speaker_analysis\x18\x01 \x01(\x08\x12$\n\x1c\x65nable_conversation_analysis\x18\x02 \x01(\x08\x12(\n descriptive_statistics_quantiles\x18\x03 \x03(\x01\"\xc7\x01\n\x08RawAudio\x12@\n\x0e\x61udio_encoding\x18\x01 \x01(\x0e\x32(.speechkit.stt.v3.RawAudio.AudioEncoding\x12\x19\n\x11sample_rate_hertz\x18\x02 \x01(\x03\x12\x1b\n\x13\x61udio_channel_count\x18\x03 \x01(\x03\"A\n\rAudioEncoding\x12\x1e\n\x1a\x41UDIO_ENCODING_UNSPECIFIED\x10\x00\x12\x10\n\x0cLINEAR16_PCM\x10\x01\"\xbf\x01\n\x0e\x43ontainerAudio\x12Q\n\x14\x63ontainer_audio_type\x18\x01 \x01(\x0e\x32\x33.speechkit.stt.v3.ContainerAudio.ContainerAudioType\"Z\n\x12\x43ontainerAudioType\x12$\n CONTAINER_AUDIO_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03WAV\x10\x01\x12\x0c\n\x08OGG_OPUS\x10\x02\x12\x07\n\x03MP3\x10\x03\"\x91\x01\n\x12\x41udioFormatOptions\x12/\n\traw_audio\x18\x01 \x01(\x0b\x32\x1a.speechkit.stt.v3.RawAudioH\x00\x12;\n\x0f\x63ontainer_audio\x18\x02 \x01(\x0b\x32 .speechkit.stt.v3.ContainerAudioH\x00\x42\r\n\x0b\x41udioFormat\"\xf7\x01\n\x1aLanguageRestrictionOptions\x12^\n\x10restriction_type\x18\x01 \x01(\x0e\x32\x44.speechkit.stt.v3.LanguageRestrictionOptions.LanguageRestrictionType\x12\x15\n\rlanguage_code\x18\x02 \x03(\t\"b\n\x17LanguageRestrictionType\x12)\n%LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tWHITELIST\x10\x01\x12\r\n\tBLACKLIST\x10\x02\"\xb2\x03\n\x17RecognitionModelOptions\x12\r\n\x05model\x18\x01 \x01(\t\x12:\n\x0c\x61udio_format\x18\x02 \x01(\x0b\x32$.speechkit.stt.v3.AudioFormatOptions\x12\x46\n\x12text_normalization\x18\x03 \x01(\x0b\x32*.speechkit.stt.v3.TextNormalizationOptions\x12J\n\x14language_restriction\x18\x04 \x01(\x0b\x32,.speechkit.stt.v3.LanguageRestrictionOptions\x12\\\n\x15\x61udio_processing_type\x18\x05 \x01(\x0e\x32=.speechkit.stt.v3.RecognitionModelOptions.AudioProcessingType\"Z\n\x13\x41udioProcessingType\x12%\n!AUDIO_PROCESSING_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tREAL_TIME\x10\x01\x12\r\n\tFULL_DATA\x10\x02\"\xde\x01\n\x16SpeakerLabelingOptions\x12R\n\x10speaker_labeling\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.SpeakerLabelingOptions.SpeakerLabeling\"p\n\x0fSpeakerLabeling\x12 \n\x1cSPEAKER_LABELING_UNSPECIFIED\x10\x00\x12\x1c\n\x18SPEAKER_LABELING_ENABLED\x10\x01\x12\x1d\n\x19SPEAKER_LABELING_DISABLED\x10\x02\"\xee\x02\n\x10StreamingOptions\x12\x44\n\x11recognition_model\x18\x01 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12>\n\x0e\x65ou_classifier\x18\x02 \x01(\x0b\x32&.speechkit.stt.v3.EouClassifierOptions\x12N\n\x16recognition_classifier\x18\x03 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x05 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptions\"\x1a\n\nAudioChunk\x12\x0c\n\x04\x64\x61ta\x18\x01 \x01(\x0c\"#\n\x0cSilenceChunk\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\x03\"\x05\n\x03\x45ou\"\xe8\x01\n\x10StreamingRequest\x12=\n\x0fsession_options\x18\x01 \x01(\x0b\x32\".speechkit.stt.v3.StreamingOptionsH\x00\x12-\n\x05\x63hunk\x18\x02 \x01(\x0b\x32\x1c.speechkit.stt.v3.AudioChunkH\x00\x12\x37\n\rsilence_chunk\x18\x03 \x01(\x0b\x32\x1e.speechkit.stt.v3.SilenceChunkH\x00\x12$\n\x03\x65ou\x18\x04 \x01(\x0b\x32\x15.speechkit.stt.v3.EouH\x00\x42\x07\n\x05\x45vent\"\xe3\x02\n\x14RecognizeFileRequest\x12\x11\n\x07\x63ontent\x18\x01 \x01(\x0cH\x00\x12\r\n\x03uri\x18\x02 \x01(\tH\x00\x12\x44\n\x11recognition_model\x18\x03 \x01(\x0b\x32).speechkit.stt.v3.RecognitionModelOptions\x12N\n\x16recognition_classifier\x18\x04 \x01(\x0b\x32..speechkit.stt.v3.RecognitionClassifierOptions\x12@\n\x0fspeech_analysis\x18\x05 \x01(\x0b\x32\'.speechkit.stt.v3.SpeechAnalysisOptions\x12\x42\n\x10speaker_labeling\x18\x06 \x01(\x0b\x32(.speechkit.stt.v3.SpeakerLabelingOptionsB\r\n\x0b\x41udioSource\"@\n\x04Word\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"@\n\x12LanguageEstimation\x12\x15\n\rlanguage_code\x18\x01 \x01(\t\x12\x13\n\x0bprobability\x18\x02 \x01(\x01\"\xbb\x01\n\x0b\x41lternative\x12%\n\x05words\x18\x01 \x03(\x0b\x32\x16.speechkit.stt.v3.Word\x12\x0c\n\x04text\x18\x02 \x01(\t\x12\x15\n\rstart_time_ms\x18\x03 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x04 \x01(\x03\x12\x12\n\nconfidence\x18\x05 \x01(\x01\x12\x37\n\tlanguages\x18\x06 \x03(\x0b\x32$.speechkit.stt.v3.LanguageEstimation\"\x1c\n\tEouUpdate\x12\x0f\n\x07time_ms\x18\x02 \x01(\x03\"a\n\x11\x41lternativeUpdate\x12\x33\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32\x1d.speechkit.stt.v3.Alternative\x12\x17\n\x0b\x63hannel_tag\x18\x02 \x01(\tB\x02\x18\x01\"\x99\x01\n\x0c\x41udioCursors\x12\x18\n\x10received_data_ms\x18\x01 \x01(\x03\x12\x15\n\rreset_time_ms\x18\x02 \x01(\x03\x12\x17\n\x0fpartial_time_ms\x18\x03 \x01(\x03\x12\x15\n\rfinal_time_ms\x18\x04 \x01(\x03\x12\x13\n\x0b\x66inal_index\x18\x05 \x01(\x03\x12\x13\n\x0b\x65ou_time_ms\x18\x06 \x01(\x03\"n\n\x0f\x46inalRefinement\x12\x13\n\x0b\x66inal_index\x18\x01 \x01(\x03\x12>\n\x0fnormalized_text\x18\x02 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x42\x06\n\x04Type\"L\n\nStatusCode\x12-\n\tcode_type\x18\x01 \x01(\x0e\x32\x1a.speechkit.stt.v3.CodeType\x12\x0f\n\x07message\x18\x02 \x01(\t\"4\n\x0bSessionUuid\x12\x0c\n\x04uuid\x18\x01 \x01(\t\x12\x17\n\x0fuser_request_id\x18\x02 \x01(\t\"K\n\x0fPhraseHighlight\x12\x0c\n\x04text\x18\x01 \x01(\t\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\"?\n\x1aRecognitionClassifierLabel\x12\r\n\x05label\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x01\"\xa6\x01\n\x1bRecognitionClassifierResult\x12\x12\n\nclassifier\x18\x01 \x01(\t\x12\x35\n\nhighlights\x18\x02 \x03(\x0b\x32!.speechkit.stt.v3.PhraseHighlight\x12<\n\x06labels\x18\x03 \x03(\x0b\x32,.speechkit.stt.v3.RecognitionClassifierLabel\"\xc7\x02\n\x1bRecognitionClassifierUpdate\x12M\n\x0bwindow_type\x18\x01 \x01(\x0e\x32\x38.speechkit.stt.v3.RecognitionClassifierUpdate.WindowType\x12\x15\n\rstart_time_ms\x18\x02 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x03 \x01(\x03\x12H\n\x11\x63lassifier_result\x18\x04 \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierResult\"c\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x01\x12\x0e\n\nLAST_FINAL\x10\x02\x12\x10\n\x0cLAST_PARTIAL\x10\x03\"\xbb\x01\n\x15\x44\x65scriptiveStatistics\x12\x0b\n\x03min\x18\x01 \x01(\x01\x12\x0b\n\x03max\x18\x02 \x01(\x01\x12\x0c\n\x04mean\x18\x03 \x01(\x01\x12\x0b\n\x03std\x18\x04 \x01(\x01\x12\x43\n\tquantiles\x18\x05 \x03(\x0b\x32\x30.speechkit.stt.v3.DescriptiveStatistics.Quantile\x1a(\n\x08Quantile\x12\r\n\x05level\x18\x01 \x01(\x01\x12\r\n\x05value\x18\x02 \x01(\x01\"D\n\x16\x41udioSegmentBoundaries\x12\x15\n\rstart_time_ms\x18\x01 \x01(\x03\x12\x13\n\x0b\x65nd_time_ms\x18\x02 \x01(\x03\"\x87\x06\n\x0fSpeakerAnalysis\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x41\n\x0bwindow_type\x18\x02 \x01(\x0e\x32,.speechkit.stt.v3.SpeakerAnalysis.WindowType\x12\x43\n\x11speech_boundaries\x18\x03 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12\x17\n\x0ftotal_speech_ms\x18\x04 \x01(\x03\x12\x14\n\x0cspeech_ratio\x18\x05 \x01(\x01\x12\x18\n\x10total_silence_ms\x18\x06 \x01(\x03\x12\x15\n\rsilence_ratio\x18\x07 \x01(\x01\x12\x13\n\x0bwords_count\x18\x08 \x01(\x03\x12\x15\n\rletters_count\x18\t \x01(\x03\x12\x41\n\x10words_per_second\x18\n \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x43\n\x12letters_per_second\x18\x0b \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x44\n\x13words_per_utterance\x18\x0c \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x46\n\x15letters_per_utterance\x18\r \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12\x17\n\x0futterance_count\x18\x0e \x01(\x03\x12N\n\x1dutterance_duration_estimation\x18\x0f \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\"L\n\nWindowType\x12\x1f\n\x17WINDOW_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\t\n\x05TOTAL\x10\x01\x12\x12\n\x0eLAST_UTTERANCE\x10\x02\"\x85\x06\n\x14\x43onversationAnalysis\x12I\n\x17\x63onversation_boundaries\x18\x01 \x01(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\x12.\n&total_simultaneous_silence_duration_ms\x18\x02 \x01(\x03\x12(\n total_simultaneous_silence_ratio\x18\x03 \x01(\x01\x12Y\n(simultaneous_silence_duration_estimation\x18\x04 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12-\n%total_simultaneous_speech_duration_ms\x18\x05 \x01(\x03\x12\'\n\x1ftotal_simultaneous_speech_ratio\x18\x06 \x01(\x01\x12X\n\'simultaneous_speech_duration_estimation\x18\x07 \x01(\x0b\x32\'.speechkit.stt.v3.DescriptiveStatistics\x12W\n\x12speaker_interrupts\x18\x08 \x03(\x0b\x32;.speechkit.stt.v3.ConversationAnalysis.InterruptsEvaluation\x12 \n\x18total_speech_duration_ms\x18\t \x01(\x03\x12\x1a\n\x12total_speech_ratio\x18\n \x01(\x01\x1a\xa3\x01\n\x14InterruptsEvaluation\x12\x13\n\x0bspeaker_tag\x18\x01 \x01(\t\x12\x18\n\x10interrupts_count\x18\x02 \x01(\x03\x12\x1e\n\x16interrupts_duration_ms\x18\x03 \x01(\x03\x12<\n\ninterrupts\x18\x04 \x03(\x0b\x32(.speechkit.stt.v3.AudioSegmentBoundaries\"\xa5\x05\n\x11StreamingResponse\x12\x33\n\x0csession_uuid\x18\x01 \x01(\x0b\x32\x1d.speechkit.stt.v3.SessionUuid\x12\x35\n\raudio_cursors\x18\x02 \x01(\x0b\x32\x1e.speechkit.stt.v3.AudioCursors\x12\x1d\n\x15response_wall_time_ms\x18\x03 \x01(\x03\x12\x36\n\x07partial\x18\x04 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x34\n\x05\x66inal\x18\x05 \x01(\x0b\x32#.speechkit.stt.v3.AlternativeUpdateH\x00\x12\x31\n\neou_update\x18\x06 \x01(\x0b\x32\x1b.speechkit.stt.v3.EouUpdateH\x00\x12=\n\x10\x66inal_refinement\x18\x07 \x01(\x0b\x32!.speechkit.stt.v3.FinalRefinementH\x00\x12\x33\n\x0bstatus_code\x18\x08 \x01(\x0b\x32\x1c.speechkit.stt.v3.StatusCodeH\x00\x12J\n\x11\x63lassifier_update\x18\n \x01(\x0b\x32-.speechkit.stt.v3.RecognitionClassifierUpdateH\x00\x12=\n\x10speaker_analysis\x18\x0b \x01(\x0b\x32!.speechkit.stt.v3.SpeakerAnalysisH\x00\x12G\n\x15\x63onversation_analysis\x18\x0c \x01(\x0b\x32&.speechkit.stt.v3.ConversationAnalysisH\x00\x12\x13\n\x0b\x63hannel_tag\x18\t \x01(\tB\x07\n\x05\x45vent\"0\n\x18\x44\x65leteRecognitionRequest\x12\x14\n\x0coperation_id\x18\x01 \x01(\t*O\n\x08\x43odeType\x12\x1d\n\x15\x43ODE_TYPE_UNSPECIFIED\x10\x00\x1a\x02\x08\x01\x12\x0b\n\x07WORKING\x10\x01\x12\x0b\n\x07WARNING\x10\x02\x12\n\n\x06\x43LOSED\x10\x03\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
17
17
 
18
18
  _globals = globals()
19
19
  _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
@@ -31,8 +31,8 @@ if _descriptor._USE_C_DESCRIPTORS == False:
31
31
  _RECOGNITIONCLASSIFIERUPDATE_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._serialized_options = b'\010\001'
32
32
  _SPEAKERANALYSIS_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._options = None
33
33
  _SPEAKERANALYSIS_WINDOWTYPE.values_by_name["WINDOW_TYPE_UNSPECIFIED"]._serialized_options = b'\010\001'
34
- _globals['_CODETYPE']._serialized_start=7877
35
- _globals['_CODETYPE']._serialized_end=7956
34
+ _globals['_CODETYPE']._serialized_start=7927
35
+ _globals['_CODETYPE']._serialized_end=8006
36
36
  _globals['_TEXTNORMALIZATIONOPTIONS']._serialized_start=55
37
37
  _globals['_TEXTNORMALIZATIONOPTIONS']._serialized_end=537
38
38
  _globals['_TEXTNORMALIZATIONOPTIONS_TEXTNORMALIZATION']._serialized_start=319
@@ -133,4 +133,6 @@ if _descriptor._USE_C_DESCRIPTORS == False:
133
133
  _globals['_CONVERSATIONANALYSIS_INTERRUPTSEVALUATION']._serialized_end=7195
134
134
  _globals['_STREAMINGRESPONSE']._serialized_start=7198
135
135
  _globals['_STREAMINGRESPONSE']._serialized_end=7875
136
+ _globals['_DELETERECOGNITIONREQUEST']._serialized_start=7877
137
+ _globals['_DELETERECOGNITIONREQUEST']._serialized_end=7925
136
138
  # @@protoc_insertion_point(module_scope)
@@ -1435,3 +1435,18 @@ class StreamingResponse(google.protobuf.message.Message):
1435
1435
  def WhichOneof(self, oneof_group: typing.Literal["Event", b"Event"]) -> typing.Literal["partial", "final", "eou_update", "final_refinement", "status_code", "classifier_update", "speaker_analysis", "conversation_analysis"] | None: ...
1436
1436
 
1437
1437
  global___StreamingResponse = StreamingResponse
1438
+
1439
+ @typing.final
1440
+ class DeleteRecognitionRequest(google.protobuf.message.Message):
1441
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1442
+
1443
+ OPERATION_ID_FIELD_NUMBER: builtins.int
1444
+ operation_id: builtins.str
1445
+ def __init__(
1446
+ self,
1447
+ *,
1448
+ operation_id: builtins.str = ...,
1449
+ ) -> None: ...
1450
+ def ClearField(self, field_name: typing.Literal["operation_id", b"operation_id"]) -> None: ...
1451
+
1452
+ global___DeleteRecognitionRequest = DeleteRecognitionRequest
@@ -13,12 +13,13 @@ _sym_db = _symbol_database.Default()
13
13
 
14
14
  from yandex.cloud.ai.stt.v3 import stt_pb2 as yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2
15
15
  from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2
16
+ from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2
16
17
  from yandex.cloud import validation_pb2 as yandex_dot_cloud_dot_validation__pb2
17
18
  from yandex.cloud.api import operation_pb2 as yandex_dot_cloud_dot_api_dot_operation__pb2
18
19
  from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operation_dot_operation__pb2
19
20
 
20
21
 
21
- DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n(yandex/cloud/ai/stt/v3/stt_service.proto\x12\x10speechkit.stt.v3\x1a yandex/cloud/ai/stt/v3/stt.proto\x1a\x1cgoogle/api/annotations.proto\x1a\x1dyandex/cloud/validation.proto\x1a yandex/cloud/api/operation.proto\x1a&yandex/cloud/operation/operation.proto\";\n\x15GetRecognitionRequest\x12\"\n\x0coperation_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=502q\n\nRecognizer\x12\x63\n\x12RecognizeStreaming\x12\".speechkit.stt.v3.StreamingRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x00(\x01\x30\x01\x32\xb3\x02\n\x0f\x41syncRecognizer\x12\x9c\x01\n\rRecognizeFile\x12&.speechkit.stt.v3.RecognizeFileRequest\x1a!.yandex.cloud.operation.Operation\"@\xb2\xd2*\x17\x12\x15google.protobuf.Empty\x82\xd3\xe4\x93\x02\x1f\"\x1a/stt/v3/recognizeFileAsync:\x01*\x12\x80\x01\n\x0eGetRecognition\x12\'.speechkit.stt.v3.GetRecognitionRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x1e\x82\xd3\xe4\x93\x02\x18\x12\x16/stt/v3/getRecognition0\x01\x42\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
22
+ DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n(yandex/cloud/ai/stt/v3/stt_service.proto\x12\x10speechkit.stt.v3\x1a yandex/cloud/ai/stt/v3/stt.proto\x1a\x1cgoogle/api/annotations.proto\x1a\x1bgoogle/protobuf/empty.proto\x1a\x1dyandex/cloud/validation.proto\x1a yandex/cloud/api/operation.proto\x1a&yandex/cloud/operation/operation.proto\";\n\x15GetRecognitionRequest\x12\"\n\x0coperation_id\x18\x01 \x01(\tB\x0c\xe8\xc7\x31\x01\x8a\xc8\x31\x04<=502q\n\nRecognizer\x12\x63\n\x12RecognizeStreaming\x12\".speechkit.stt.v3.StreamingRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x00(\x01\x30\x01\x32\xaf\x03\n\x0f\x41syncRecognizer\x12\x9c\x01\n\rRecognizeFile\x12&.speechkit.stt.v3.RecognizeFileRequest\x1a!.yandex.cloud.operation.Operation\"@\xb2\xd2*\x17\x12\x15google.protobuf.Empty\x82\xd3\xe4\x93\x02\x1f\"\x1a/stt/v3/recognizeFileAsync:\x01*\x12\x80\x01\n\x0eGetRecognition\x12\'.speechkit.stt.v3.GetRecognitionRequest\x1a#.speechkit.stt.v3.StreamingResponse\"\x1e\x82\xd3\xe4\x93\x02\x18\x12\x16/stt/v3/getRecognition0\x01\x12z\n\x11\x44\x65leteRecognition\x12*.speechkit.stt.v3.DeleteRecognitionRequest\x1a\x16.google.protobuf.Empty\"!\x82\xd3\xe4\x93\x02\x1b*\x19/stt/v3/deleteRecognitionB\\\n\x1ayandex.cloud.api.ai.stt.v3Z>github.com/yandex-cloud/go-genproto/yandex/cloud/ai/stt/v3;sttb\x06proto3')
22
23
 
23
24
  _globals = globals()
24
25
  _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
@@ -32,10 +33,12 @@ if _descriptor._USE_C_DESCRIPTORS == False:
32
33
  _ASYNCRECOGNIZER.methods_by_name['RecognizeFile']._serialized_options = b'\262\322*\027\022\025google.protobuf.Empty\202\323\344\223\002\037\"\032/stt/v3/recognizeFileAsync:\001*'
33
34
  _ASYNCRECOGNIZER.methods_by_name['GetRecognition']._options = None
34
35
  _ASYNCRECOGNIZER.methods_by_name['GetRecognition']._serialized_options = b'\202\323\344\223\002\030\022\026/stt/v3/getRecognition'
35
- _globals['_GETRECOGNITIONREQUEST']._serialized_start=231
36
- _globals['_GETRECOGNITIONREQUEST']._serialized_end=290
37
- _globals['_RECOGNIZER']._serialized_start=292
38
- _globals['_RECOGNIZER']._serialized_end=405
39
- _globals['_ASYNCRECOGNIZER']._serialized_start=408
40
- _globals['_ASYNCRECOGNIZER']._serialized_end=715
36
+ _ASYNCRECOGNIZER.methods_by_name['DeleteRecognition']._options = None
37
+ _ASYNCRECOGNIZER.methods_by_name['DeleteRecognition']._serialized_options = b'\202\323\344\223\002\033*\031/stt/v3/deleteRecognition'
38
+ _globals['_GETRECOGNITIONREQUEST']._serialized_start=260
39
+ _globals['_GETRECOGNITIONREQUEST']._serialized_end=319
40
+ _globals['_RECOGNIZER']._serialized_start=321
41
+ _globals['_RECOGNIZER']._serialized_end=434
42
+ _globals['_ASYNCRECOGNIZER']._serialized_start=437
43
+ _globals['_ASYNCRECOGNIZER']._serialized_end=868
41
44
  # @@protoc_insertion_point(module_scope)
@@ -2,6 +2,7 @@
2
2
  """Client and server classes corresponding to protobuf-defined services."""
3
3
  import grpc
4
4
 
5
+ from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2
5
6
  from yandex.cloud.ai.stt.v3 import stt_pb2 as yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2
6
7
  from yandex.cloud.ai.stt.v3 import stt_service_pb2 as yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__service__pb2
7
8
  from yandex.cloud.operation import operation_pb2 as yandex_dot_cloud_dot_operation_dot_operation__pb2
@@ -92,6 +93,11 @@ class AsyncRecognizerStub(object):
92
93
  request_serializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__service__pb2.GetRecognitionRequest.SerializeToString,
93
94
  response_deserializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.StreamingResponse.FromString,
94
95
  )
96
+ self.DeleteRecognition = channel.unary_unary(
97
+ '/speechkit.stt.v3.AsyncRecognizer/DeleteRecognition',
98
+ request_serializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.DeleteRecognitionRequest.SerializeToString,
99
+ response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString,
100
+ )
95
101
 
96
102
 
97
103
  class AsyncRecognizerServicer(object):
@@ -110,6 +116,12 @@ class AsyncRecognizerServicer(object):
110
116
  context.set_details('Method not implemented!')
111
117
  raise NotImplementedError('Method not implemented!')
112
118
 
119
+ def DeleteRecognition(self, request, context):
120
+ """Missing associated documentation comment in .proto file."""
121
+ context.set_code(grpc.StatusCode.UNIMPLEMENTED)
122
+ context.set_details('Method not implemented!')
123
+ raise NotImplementedError('Method not implemented!')
124
+
113
125
 
114
126
  def add_AsyncRecognizerServicer_to_server(servicer, server):
115
127
  rpc_method_handlers = {
@@ -123,6 +135,11 @@ def add_AsyncRecognizerServicer_to_server(servicer, server):
123
135
  request_deserializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__service__pb2.GetRecognitionRequest.FromString,
124
136
  response_serializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.StreamingResponse.SerializeToString,
125
137
  ),
138
+ 'DeleteRecognition': grpc.unary_unary_rpc_method_handler(
139
+ servicer.DeleteRecognition,
140
+ request_deserializer=yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.DeleteRecognitionRequest.FromString,
141
+ response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString,
142
+ ),
126
143
  }
127
144
  generic_handler = grpc.method_handlers_generic_handler(
128
145
  'speechkit.stt.v3.AsyncRecognizer', rpc_method_handlers)
@@ -167,3 +184,20 @@ class AsyncRecognizer(object):
167
184
  yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.StreamingResponse.FromString,
168
185
  options, channel_credentials,
169
186
  insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
187
+
188
+ @staticmethod
189
+ def DeleteRecognition(request,
190
+ target,
191
+ options=(),
192
+ channel_credentials=None,
193
+ call_credentials=None,
194
+ insecure=False,
195
+ compression=None,
196
+ wait_for_ready=None,
197
+ timeout=None,
198
+ metadata=None):
199
+ return grpc.experimental.unary_unary(request, target, '/speechkit.stt.v3.AsyncRecognizer/DeleteRecognition',
200
+ yandex_dot_cloud_dot_ai_dot_stt_dot_v3_dot_stt__pb2.DeleteRecognitionRequest.SerializeToString,
201
+ google_dot_protobuf_dot_empty__pb2.Empty.FromString,
202
+ options, channel_credentials,
203
+ insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
@@ -5,6 +5,7 @@ isort:skip_file
5
5
 
6
6
  import abc
7
7
  import collections.abc
8
+ import google.protobuf.empty_pb2
8
9
  import grpc
9
10
  import grpc.aio
10
11
  import typing
@@ -65,6 +66,11 @@ class AsyncRecognizerStub:
65
66
  yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse,
66
67
  ]
67
68
 
69
+ DeleteRecognition: grpc.UnaryUnaryMultiCallable[
70
+ yandex.cloud.ai.stt.v3.stt_pb2.DeleteRecognitionRequest,
71
+ google.protobuf.empty_pb2.Empty,
72
+ ]
73
+
68
74
  class AsyncRecognizerAsyncStub:
69
75
  """A set of methods for async voice recognition."""
70
76
 
@@ -78,6 +84,11 @@ class AsyncRecognizerAsyncStub:
78
84
  yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse,
79
85
  ]
80
86
 
87
+ DeleteRecognition: grpc.aio.UnaryUnaryMultiCallable[
88
+ yandex.cloud.ai.stt.v3.stt_pb2.DeleteRecognitionRequest,
89
+ google.protobuf.empty_pb2.Empty,
90
+ ]
91
+
81
92
  class AsyncRecognizerServicer(metaclass=abc.ABCMeta):
82
93
  """A set of methods for async voice recognition."""
83
94
 
@@ -95,4 +106,11 @@ class AsyncRecognizerServicer(metaclass=abc.ABCMeta):
95
106
  context: _ServicerContext,
96
107
  ) -> typing.Union[collections.abc.Iterator[yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse], collections.abc.AsyncIterator[yandex.cloud.ai.stt.v3.stt_pb2.StreamingResponse]]: ...
97
108
 
109
+ @abc.abstractmethod
110
+ def DeleteRecognition(
111
+ self,
112
+ request: yandex.cloud.ai.stt.v3.stt_pb2.DeleteRecognitionRequest,
113
+ context: _ServicerContext,
114
+ ) -> typing.Union[google.protobuf.empty_pb2.Empty, collections.abc.Awaitable[google.protobuf.empty_pb2.Empty]]: ...
115
+
98
116
  def add_AsyncRecognizerServicer_to_server(servicer: AsyncRecognizerServicer, server: typing.Union[grpc.Server, grpc.aio.Server]) -> None: ...