yandexcloud 0.357.0__py3-none-any.whl → 0.358.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of yandexcloud might be problematic. Click here for more details.
- yandex/cloud/ai/foundation_models/v1/text_common_pb2.pyi +1 -1
- yandex/cloud/ai/stt/v3/stt_pb2.pyi +138 -126
- yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.py +13 -10
- yandex/cloud/ai/stt/v3/stt_service_pb2_grpc.pyi +21 -12
- yandex/cloud/ai/tts/v3/tts_pb2.pyi +1 -1
- yandex/cloud/backup/v1/provider_service_pb2.py +11 -11
- yandex/cloud/backup/v1/provider_service_pb2.pyi +10 -2
- yandex/cloud/backup/v1/resource_service_pb2.py +42 -35
- yandex/cloud/backup/v1/resource_service_pb2.pyi +44 -0
- yandex/cloud/backup/v1/resource_service_pb2_grpc.py +44 -0
- yandex/cloud/backup/v1/resource_service_pb2_grpc.pyi +20 -0
- yandex/cloud/cdn/v1/origin_group_pb2.py +4 -2
- yandex/cloud/cdn/v1/origin_group_pb2.pyi +30 -2
- yandex/cloud/cdn/v1/origin_group_service_pb2.pyi +9 -1
- yandex/cloud/cdn/v1/origin_pb2.pyi +3 -3
- yandex/cloud/cdn/v1/origin_service_pb2.pyi +11 -3
- yandex/cloud/cdn/v1/provider_service_pb2_grpc.py +1 -1
- yandex/cloud/cdn/v1/provider_service_pb2_grpc.pyi +3 -3
- yandex/cloud/cdn/v1/resource_pb2.pyi +3 -3
- yandex/cloud/cdn/v1/resource_service_pb2.pyi +9 -1
- yandex/cloud/cdn/v1/resource_service_pb2_grpc.py +1 -2
- yandex/cloud/cdn/v1/resource_service_pb2_grpc.pyi +3 -9
- yandex/cloud/connectionmanager/v1/clickhouse_pb2.py +10 -8
- yandex/cloud/connectionmanager/v1/clickhouse_pb2.pyi +25 -4
- yandex/cloud/connectionmanager/v1/connection_pb2.py +12 -11
- yandex/cloud/connectionmanager/v1/connection_pb2.pyi +10 -3
- yandex/cloud/connectionmanager/v1/kafka_pb2.py +51 -0
- yandex/cloud/connectionmanager/v1/kafka_pb2.pyi +181 -0
- yandex/cloud/connectionmanager/v1/kafka_pb2_grpc.py +24 -0
- yandex/cloud/connectionmanager/v1/kafka_pb2_grpc.pyi +17 -0
- yandex/cloud/loadbalancer/v1/network_load_balancer_pb2.py +8 -8
- yandex/cloud/loadbalancer/v1/network_load_balancer_pb2.pyi +5 -1
- yandex/cloud/mdb/clickhouse/v1/config/clickhouse_pb2.pyi +2 -0
- yandex/cloud/organizationmanager/v1/group_service_pb2.py +47 -47
- yandexcloud/__init__.py +1 -1
- {yandexcloud-0.357.0.dist-info → yandexcloud-0.358.0.dist-info}/METADATA +1 -1
- {yandexcloud-0.357.0.dist-info → yandexcloud-0.358.0.dist-info}/RECORD +41 -37
- {yandexcloud-0.357.0.dist-info → yandexcloud-0.358.0.dist-info}/AUTHORS +0 -0
- {yandexcloud-0.357.0.dist-info → yandexcloud-0.358.0.dist-info}/LICENSE +0 -0
- {yandexcloud-0.357.0.dist-info → yandexcloud-0.358.0.dist-info}/WHEEL +0 -0
- {yandexcloud-0.357.0.dist-info → yandexcloud-0.358.0.dist-info}/top_level.txt +0 -0
|
@@ -30,7 +30,7 @@ class _CodeTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumT
|
|
|
30
30
|
WORKING: _CodeType.ValueType # 1
|
|
31
31
|
"""All good."""
|
|
32
32
|
WARNING: _CodeType.ValueType # 2
|
|
33
|
-
"""For example, if speech is sent not in real
|
|
33
|
+
"""For example, if speech is sent not in real-time or context is unknown and we've made fallback."""
|
|
34
34
|
CLOSED: _CodeType.ValueType # 3
|
|
35
35
|
"""After session was closed."""
|
|
36
36
|
|
|
@@ -40,14 +40,16 @@ CODE_TYPE_UNSPECIFIED: CodeType.ValueType # 0
|
|
|
40
40
|
WORKING: CodeType.ValueType # 1
|
|
41
41
|
"""All good."""
|
|
42
42
|
WARNING: CodeType.ValueType # 2
|
|
43
|
-
"""For example, if speech is sent not in real
|
|
43
|
+
"""For example, if speech is sent not in real-time or context is unknown and we've made fallback."""
|
|
44
44
|
CLOSED: CodeType.ValueType # 3
|
|
45
45
|
"""After session was closed."""
|
|
46
46
|
global___CodeType = CodeType
|
|
47
47
|
|
|
48
48
|
@typing.final
|
|
49
49
|
class TextNormalizationOptions(google.protobuf.message.Message):
|
|
50
|
-
"""Options
|
|
50
|
+
"""Options for post-processing text results. The normalization levels depend on the settings and the language.
|
|
51
|
+
For detailed information, see [documentation](/docs/speechkit/stt/normalization).
|
|
52
|
+
"""
|
|
51
53
|
|
|
52
54
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
53
55
|
|
|
@@ -59,18 +61,18 @@ class TextNormalizationOptions(google.protobuf.message.Message):
|
|
|
59
61
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
60
62
|
TEXT_NORMALIZATION_UNSPECIFIED: TextNormalizationOptions._TextNormalization.ValueType # 0
|
|
61
63
|
TEXT_NORMALIZATION_ENABLED: TextNormalizationOptions._TextNormalization.ValueType # 1
|
|
62
|
-
"""Enable
|
|
64
|
+
"""Enable converting numbers, dates and time from text to numeric format."""
|
|
63
65
|
TEXT_NORMALIZATION_DISABLED: TextNormalizationOptions._TextNormalization.ValueType # 2
|
|
64
|
-
"""Disable normalization"""
|
|
66
|
+
"""Disable all normalization. Default value."""
|
|
65
67
|
|
|
66
68
|
class TextNormalization(_TextNormalization, metaclass=_TextNormalizationEnumTypeWrapper):
|
|
67
|
-
"""
|
|
69
|
+
"""Base-level normalization."""
|
|
68
70
|
|
|
69
71
|
TEXT_NORMALIZATION_UNSPECIFIED: TextNormalizationOptions.TextNormalization.ValueType # 0
|
|
70
72
|
TEXT_NORMALIZATION_ENABLED: TextNormalizationOptions.TextNormalization.ValueType # 1
|
|
71
|
-
"""Enable
|
|
73
|
+
"""Enable converting numbers, dates and time from text to numeric format."""
|
|
72
74
|
TEXT_NORMALIZATION_DISABLED: TextNormalizationOptions.TextNormalization.ValueType # 2
|
|
73
|
-
"""Disable normalization"""
|
|
75
|
+
"""Disable all normalization. Default value."""
|
|
74
76
|
|
|
75
77
|
class _PhoneFormattingMode:
|
|
76
78
|
ValueType = typing.NewType("ValueType", builtins.int)
|
|
@@ -122,19 +124,26 @@ class DefaultEouClassifier(google.protobuf.message.Message):
|
|
|
122
124
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
123
125
|
EOU_SENSITIVITY_UNSPECIFIED: DefaultEouClassifier._EouSensitivity.ValueType # 0
|
|
124
126
|
DEFAULT: DefaultEouClassifier._EouSensitivity.ValueType # 1
|
|
127
|
+
"""Default and more conservative EOU detector."""
|
|
125
128
|
HIGH: DefaultEouClassifier._EouSensitivity.ValueType # 2
|
|
129
|
+
"""A high-sensitive and fast EOU detector, which may produce more false positives."""
|
|
126
130
|
|
|
127
131
|
class EouSensitivity(_EouSensitivity, metaclass=_EouSensitivityEnumTypeWrapper): ...
|
|
128
132
|
EOU_SENSITIVITY_UNSPECIFIED: DefaultEouClassifier.EouSensitivity.ValueType # 0
|
|
129
133
|
DEFAULT: DefaultEouClassifier.EouSensitivity.ValueType # 1
|
|
134
|
+
"""Default and more conservative EOU detector."""
|
|
130
135
|
HIGH: DefaultEouClassifier.EouSensitivity.ValueType # 2
|
|
136
|
+
"""A high-sensitive and fast EOU detector, which may produce more false positives."""
|
|
131
137
|
|
|
132
138
|
TYPE_FIELD_NUMBER: builtins.int
|
|
133
139
|
MAX_PAUSE_BETWEEN_WORDS_HINT_MS_FIELD_NUMBER: builtins.int
|
|
134
140
|
type: global___DefaultEouClassifier.EouSensitivity.ValueType
|
|
135
|
-
"""EOU sensitivity.
|
|
141
|
+
"""EOU sensitivity."""
|
|
136
142
|
max_pause_between_words_hint_ms: builtins.int
|
|
137
|
-
"""Hint for max pause between words.
|
|
143
|
+
"""Hint for max pause between words. SpeechKit EOU detector could use this information to adjust the speed of the EOU detection.
|
|
144
|
+
For example, a long pause between words will help distinguish between the end of utterance from slow speech like `One <long pause> two <long pause> three`.
|
|
145
|
+
A short pause can be helpful if the speaker is speaking quickly and does not emphasize pauses between sentences.
|
|
146
|
+
"""
|
|
138
147
|
def __init__(
|
|
139
148
|
self,
|
|
140
149
|
*,
|
|
@@ -147,7 +156,7 @@ global___DefaultEouClassifier = DefaultEouClassifier
|
|
|
147
156
|
|
|
148
157
|
@typing.final
|
|
149
158
|
class ExternalEouClassifier(google.protobuf.message.Message):
|
|
150
|
-
"""Use EOU provided by user"""
|
|
159
|
+
"""Use EOU provided by user."""
|
|
151
160
|
|
|
152
161
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
153
162
|
|
|
@@ -165,11 +174,11 @@ class EouClassifierOptions(google.protobuf.message.Message):
|
|
|
165
174
|
EXTERNAL_CLASSIFIER_FIELD_NUMBER: builtins.int
|
|
166
175
|
@property
|
|
167
176
|
def default_classifier(self) -> global___DefaultEouClassifier:
|
|
168
|
-
"""EOU classifier provided by SpeechKit.
|
|
177
|
+
"""Default EOU classifier provided by SpeechKit."""
|
|
169
178
|
|
|
170
179
|
@property
|
|
171
180
|
def external_classifier(self) -> global___ExternalEouClassifier:
|
|
172
|
-
"""EOU
|
|
181
|
+
"""EOU classifier enforced by external messages from user."""
|
|
173
182
|
|
|
174
183
|
def __init__(
|
|
175
184
|
self,
|
|
@@ -195,22 +204,22 @@ class RecognitionClassifier(google.protobuf.message.Message):
|
|
|
195
204
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
196
205
|
TRIGGER_TYPE_UNSPECIFIED: RecognitionClassifier._TriggerType.ValueType # 0
|
|
197
206
|
ON_UTTERANCE: RecognitionClassifier._TriggerType.ValueType # 1
|
|
198
|
-
"""Apply classifier to utterance responses"""
|
|
207
|
+
"""Apply classifier to utterance responses."""
|
|
199
208
|
ON_FINAL: RecognitionClassifier._TriggerType.ValueType # 2
|
|
200
|
-
"""Apply classifier to final responses"""
|
|
209
|
+
"""Apply classifier to final responses."""
|
|
201
210
|
ON_PARTIAL: RecognitionClassifier._TriggerType.ValueType # 3
|
|
202
|
-
"""Apply classifier to partial responses"""
|
|
211
|
+
"""Apply classifier to partial responses."""
|
|
203
212
|
|
|
204
213
|
class TriggerType(_TriggerType, metaclass=_TriggerTypeEnumTypeWrapper):
|
|
205
214
|
"""Type of recognition classifier trigger."""
|
|
206
215
|
|
|
207
216
|
TRIGGER_TYPE_UNSPECIFIED: RecognitionClassifier.TriggerType.ValueType # 0
|
|
208
217
|
ON_UTTERANCE: RecognitionClassifier.TriggerType.ValueType # 1
|
|
209
|
-
"""Apply classifier to utterance responses"""
|
|
218
|
+
"""Apply classifier to utterance responses."""
|
|
210
219
|
ON_FINAL: RecognitionClassifier.TriggerType.ValueType # 2
|
|
211
|
-
"""Apply classifier to final responses"""
|
|
220
|
+
"""Apply classifier to final responses."""
|
|
212
221
|
ON_PARTIAL: RecognitionClassifier.TriggerType.ValueType # 3
|
|
213
|
-
"""Apply classifier to partial responses"""
|
|
222
|
+
"""Apply classifier to partial responses."""
|
|
214
223
|
|
|
215
224
|
CLASSIFIER_FIELD_NUMBER: builtins.int
|
|
216
225
|
TRIGGERS_FIELD_NUMBER: builtins.int
|
|
@@ -218,7 +227,7 @@ class RecognitionClassifier(google.protobuf.message.Message):
|
|
|
218
227
|
"""Classifier name"""
|
|
219
228
|
@property
|
|
220
229
|
def triggers(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[global___RecognitionClassifier.TriggerType.ValueType]:
|
|
221
|
-
"""Describes the types of responses to which the classification results will come"""
|
|
230
|
+
"""Describes the types of responses to which the classification results will come. Classification responses will follow the responses of the specified types."""
|
|
222
231
|
|
|
223
232
|
def __init__(
|
|
224
233
|
self,
|
|
@@ -237,7 +246,7 @@ class RecognitionClassifierOptions(google.protobuf.message.Message):
|
|
|
237
246
|
CLASSIFIERS_FIELD_NUMBER: builtins.int
|
|
238
247
|
@property
|
|
239
248
|
def classifiers(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___RecognitionClassifier]:
|
|
240
|
-
"""List of classifiers to use"""
|
|
249
|
+
"""List of classifiers to use. For detailed information and usage example, see [documentation](/docs/speechkit/stt/analysis)."""
|
|
241
250
|
|
|
242
251
|
def __init__(
|
|
243
252
|
self,
|
|
@@ -299,9 +308,9 @@ class RawAudio(google.protobuf.message.Message):
|
|
|
299
308
|
SAMPLE_RATE_HERTZ_FIELD_NUMBER: builtins.int
|
|
300
309
|
AUDIO_CHANNEL_COUNT_FIELD_NUMBER: builtins.int
|
|
301
310
|
audio_encoding: global___RawAudio.AudioEncoding.ValueType
|
|
302
|
-
"""Type of audio encoding"""
|
|
311
|
+
"""Type of audio encoding."""
|
|
303
312
|
sample_rate_hertz: builtins.int
|
|
304
|
-
"""PCM sample rate"""
|
|
313
|
+
"""PCM sample rate."""
|
|
305
314
|
audio_channel_count: builtins.int
|
|
306
315
|
"""PCM channel count. Currently only single channel audio is supported in real-time recognition."""
|
|
307
316
|
def __init__(
|
|
@@ -366,7 +375,7 @@ class AudioFormatOptions(google.protobuf.message.Message):
|
|
|
366
375
|
CONTAINER_AUDIO_FIELD_NUMBER: builtins.int
|
|
367
376
|
@property
|
|
368
377
|
def raw_audio(self) -> global___RawAudio:
|
|
369
|
-
"""
|
|
378
|
+
"""RAW audio without container."""
|
|
370
379
|
|
|
371
380
|
@property
|
|
372
381
|
def container_audio(self) -> global___ContainerAudio:
|
|
@@ -386,7 +395,7 @@ global___AudioFormatOptions = AudioFormatOptions
|
|
|
386
395
|
|
|
387
396
|
@typing.final
|
|
388
397
|
class LanguageRestrictionOptions(google.protobuf.message.Message):
|
|
389
|
-
"""Type of restriction for the list of languages expected in the incoming
|
|
398
|
+
"""Type of restriction for the list of languages expected in the incoming audio."""
|
|
390
399
|
|
|
391
400
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
392
401
|
|
|
@@ -398,24 +407,27 @@ class LanguageRestrictionOptions(google.protobuf.message.Message):
|
|
|
398
407
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
399
408
|
LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED: LanguageRestrictionOptions._LanguageRestrictionType.ValueType # 0
|
|
400
409
|
WHITELIST: LanguageRestrictionOptions._LanguageRestrictionType.ValueType # 1
|
|
401
|
-
"""The
|
|
410
|
+
"""The list of most possible languages in the incoming audio."""
|
|
402
411
|
BLACKLIST: LanguageRestrictionOptions._LanguageRestrictionType.ValueType # 2
|
|
403
|
-
"""The
|
|
412
|
+
"""The list of languages that are likely not to be included in the incoming audio."""
|
|
404
413
|
|
|
405
414
|
class LanguageRestrictionType(_LanguageRestrictionType, metaclass=_LanguageRestrictionTypeEnumTypeWrapper): ...
|
|
406
415
|
LANGUAGE_RESTRICTION_TYPE_UNSPECIFIED: LanguageRestrictionOptions.LanguageRestrictionType.ValueType # 0
|
|
407
416
|
WHITELIST: LanguageRestrictionOptions.LanguageRestrictionType.ValueType # 1
|
|
408
|
-
"""The
|
|
417
|
+
"""The list of most possible languages in the incoming audio."""
|
|
409
418
|
BLACKLIST: LanguageRestrictionOptions.LanguageRestrictionType.ValueType # 2
|
|
410
|
-
"""The
|
|
419
|
+
"""The list of languages that are likely not to be included in the incoming audio."""
|
|
411
420
|
|
|
412
421
|
RESTRICTION_TYPE_FIELD_NUMBER: builtins.int
|
|
413
422
|
LANGUAGE_CODE_FIELD_NUMBER: builtins.int
|
|
414
423
|
restriction_type: global___LanguageRestrictionOptions.LanguageRestrictionType.ValueType
|
|
415
|
-
"""Language restriction type
|
|
424
|
+
"""Language restriction type.
|
|
425
|
+
All of these restrictions are used by the model as guidelines, not as strict rules.
|
|
426
|
+
The language is recognized for each sentence. If a sentence has phrases in different languages, all of them will be transcribed in the most probable language.
|
|
427
|
+
"""
|
|
416
428
|
@property
|
|
417
429
|
def language_code(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
418
|
-
"""The list of language codes to restrict recognition in the case of an auto model"""
|
|
430
|
+
"""The list of [language codes](/docs/speechkit/stt/models) to restrict recognition in the case of an auto model."""
|
|
419
431
|
|
|
420
432
|
def __init__(
|
|
421
433
|
self,
|
|
@@ -460,9 +472,9 @@ class SummarizationProperty(google.protobuf.message.Message):
|
|
|
460
472
|
instruction: builtins.str
|
|
461
473
|
"""Summarization instruction for model."""
|
|
462
474
|
json_object: builtins.bool
|
|
463
|
-
"""When set to true, the model will
|
|
464
|
-
Be sure to
|
|
465
|
-
Otherwise, it may
|
|
475
|
+
"""When set to true, the model will return a valid JSON object.
|
|
476
|
+
Be sure to ask the model explicitly for JSON.
|
|
477
|
+
Otherwise, it may produce excessive whitespace and run indefinitely until it reaches the token limit.
|
|
466
478
|
"""
|
|
467
479
|
@property
|
|
468
480
|
def json_schema(self) -> global___JsonSchema:
|
|
@@ -535,16 +547,16 @@ class RecognitionModelOptions(google.protobuf.message.Message):
|
|
|
535
547
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
536
548
|
AUDIO_PROCESSING_TYPE_UNSPECIFIED: RecognitionModelOptions._AudioProcessingType.ValueType # 0
|
|
537
549
|
REAL_TIME: RecognitionModelOptions._AudioProcessingType.ValueType # 1
|
|
538
|
-
"""Process audio in mode optimized for real-time recognition, i.e. send partials and final responses as soon as possible"""
|
|
550
|
+
"""Process audio in mode optimized for real-time recognition, i.e. send partials and final responses as soon as possible."""
|
|
539
551
|
FULL_DATA: RecognitionModelOptions._AudioProcessingType.ValueType # 2
|
|
540
|
-
"""Process audio after all data was received"""
|
|
552
|
+
"""Process audio after all data was received."""
|
|
541
553
|
|
|
542
554
|
class AudioProcessingType(_AudioProcessingType, metaclass=_AudioProcessingTypeEnumTypeWrapper): ...
|
|
543
555
|
AUDIO_PROCESSING_TYPE_UNSPECIFIED: RecognitionModelOptions.AudioProcessingType.ValueType # 0
|
|
544
556
|
REAL_TIME: RecognitionModelOptions.AudioProcessingType.ValueType # 1
|
|
545
|
-
"""Process audio in mode optimized for real-time recognition, i.e. send partials and final responses as soon as possible"""
|
|
557
|
+
"""Process audio in mode optimized for real-time recognition, i.e. send partials and final responses as soon as possible."""
|
|
546
558
|
FULL_DATA: RecognitionModelOptions.AudioProcessingType.ValueType # 2
|
|
547
|
-
"""Process audio after all data was received"""
|
|
559
|
+
"""Process audio after all data was received."""
|
|
548
560
|
|
|
549
561
|
MODEL_FIELD_NUMBER: builtins.int
|
|
550
562
|
AUDIO_FORMAT_FIELD_NUMBER: builtins.int
|
|
@@ -552,11 +564,15 @@ class RecognitionModelOptions(google.protobuf.message.Message):
|
|
|
552
564
|
LANGUAGE_RESTRICTION_FIELD_NUMBER: builtins.int
|
|
553
565
|
AUDIO_PROCESSING_TYPE_FIELD_NUMBER: builtins.int
|
|
554
566
|
model: builtins.str
|
|
555
|
-
"""Sets the recognition model for the cloud version of SpeechKit.
|
|
567
|
+
"""Sets the recognition model for the cloud version of SpeechKit.
|
|
568
|
+
For `Recognizer.RecognizeStreaming`, possible values are `general`, `general:rc`, `general:deprecated`.
|
|
569
|
+
For `AsyncRecognizer.RecognizeFile`, possible values are `general`, `general:rc`, `general:deprecated`, `deferred-general`, `deferred-general:rc`, and `deferred-general:deprecated`.
|
|
556
570
|
The model is ignored for SpeechKit Hybrid.
|
|
557
571
|
"""
|
|
558
572
|
audio_processing_type: global___RecognitionModelOptions.AudioProcessingType.ValueType
|
|
559
|
-
"""
|
|
573
|
+
"""For `Recognizer.RecognizeStreaming`, defines the audio data processing mode. Default is `REAL_TIME`.
|
|
574
|
+
For `AsyncRecognizer.RecognizeFile`, this field is ignored.
|
|
575
|
+
"""
|
|
560
576
|
@property
|
|
561
577
|
def audio_format(self) -> global___AudioFormatOptions:
|
|
562
578
|
"""Specified input audio."""
|
|
@@ -595,20 +611,20 @@ class SpeakerLabelingOptions(google.protobuf.message.Message):
|
|
|
595
611
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
596
612
|
SPEAKER_LABELING_UNSPECIFIED: SpeakerLabelingOptions._SpeakerLabeling.ValueType # 0
|
|
597
613
|
SPEAKER_LABELING_ENABLED: SpeakerLabelingOptions._SpeakerLabeling.ValueType # 1
|
|
598
|
-
"""Enable speaker labeling"""
|
|
614
|
+
"""Enable speaker labeling."""
|
|
599
615
|
SPEAKER_LABELING_DISABLED: SpeakerLabelingOptions._SpeakerLabeling.ValueType # 2
|
|
600
|
-
"""Disable speaker labeling"""
|
|
616
|
+
"""Disable speaker labeling. Default value."""
|
|
601
617
|
|
|
602
618
|
class SpeakerLabeling(_SpeakerLabeling, metaclass=_SpeakerLabelingEnumTypeWrapper): ...
|
|
603
619
|
SPEAKER_LABELING_UNSPECIFIED: SpeakerLabelingOptions.SpeakerLabeling.ValueType # 0
|
|
604
620
|
SPEAKER_LABELING_ENABLED: SpeakerLabelingOptions.SpeakerLabeling.ValueType # 1
|
|
605
|
-
"""Enable speaker labeling"""
|
|
621
|
+
"""Enable speaker labeling."""
|
|
606
622
|
SPEAKER_LABELING_DISABLED: SpeakerLabelingOptions.SpeakerLabeling.ValueType # 2
|
|
607
|
-
"""Disable speaker labeling"""
|
|
623
|
+
"""Disable speaker labeling. Default value."""
|
|
608
624
|
|
|
609
625
|
SPEAKER_LABELING_FIELD_NUMBER: builtins.int
|
|
610
626
|
speaker_labeling: global___SpeakerLabelingOptions.SpeakerLabeling.ValueType
|
|
611
|
-
"""Specifies the execution of speaker labeling.
|
|
627
|
+
"""Specifies the execution of speaker labeling."""
|
|
612
628
|
def __init__(
|
|
613
629
|
self,
|
|
614
630
|
*,
|
|
@@ -634,7 +650,7 @@ class StreamingOptions(google.protobuf.message.Message):
|
|
|
634
650
|
|
|
635
651
|
@property
|
|
636
652
|
def eou_classifier(self) -> global___EouClassifierOptions:
|
|
637
|
-
"""Configuration for end of utterance detection model."""
|
|
653
|
+
"""Configuration for an end of utterance detection model."""
|
|
638
654
|
|
|
639
655
|
@property
|
|
640
656
|
def recognition_classifier(self) -> global___RecognitionClassifierOptions:
|
|
@@ -646,11 +662,11 @@ class StreamingOptions(google.protobuf.message.Message):
|
|
|
646
662
|
|
|
647
663
|
@property
|
|
648
664
|
def speaker_labeling(self) -> global___SpeakerLabelingOptions:
|
|
649
|
-
"""Configuration for speaker labeling"""
|
|
665
|
+
"""Configuration for speaker labeling."""
|
|
650
666
|
|
|
651
667
|
@property
|
|
652
668
|
def summarization(self) -> global___SummarizationOptions:
|
|
653
|
-
"""Summarization options"""
|
|
669
|
+
"""Summarization options."""
|
|
654
670
|
|
|
655
671
|
def __init__(
|
|
656
672
|
self,
|
|
@@ -705,7 +721,7 @@ global___SilenceChunk = SilenceChunk
|
|
|
705
721
|
|
|
706
722
|
@typing.final
|
|
707
723
|
class Eou(google.protobuf.message.Message):
|
|
708
|
-
"""Force EOU"""
|
|
724
|
+
"""Force EOU."""
|
|
709
725
|
|
|
710
726
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
711
727
|
|
|
@@ -717,10 +733,8 @@ global___Eou = Eou
|
|
|
717
733
|
|
|
718
734
|
@typing.final
|
|
719
735
|
class StreamingRequest(google.protobuf.message.Message):
|
|
720
|
-
"""Streaming audio request
|
|
721
|
-
Events are control messages from user.
|
|
722
|
-
First message should be session options.
|
|
723
|
-
The next messages are audio data chunks or control messages.
|
|
736
|
+
"""Streaming audio request.
|
|
737
|
+
Events are control messages from user. First message should be session options. The next messages are audio data chunks or control messages.
|
|
724
738
|
"""
|
|
725
739
|
|
|
726
740
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
@@ -773,7 +787,7 @@ class RecognizeFileRequest(google.protobuf.message.Message):
|
|
|
773
787
|
content: builtins.bytes
|
|
774
788
|
"""Bytes with data"""
|
|
775
789
|
uri: builtins.str
|
|
776
|
-
"""S3 data
|
|
790
|
+
"""S3 data URL"""
|
|
777
791
|
@property
|
|
778
792
|
def recognition_model(self) -> global___RecognitionModelOptions:
|
|
779
793
|
"""Configuration for speech recognition model."""
|
|
@@ -962,18 +976,18 @@ class AudioCursors(google.protobuf.message.Message):
|
|
|
962
976
|
reset_time_ms: builtins.int
|
|
963
977
|
"""Input stream reset data."""
|
|
964
978
|
partial_time_ms: builtins.int
|
|
965
|
-
"""How much audio was processed. This time includes trimming silences as well.
|
|
966
|
-
to update recognition results (includes silence as well).
|
|
979
|
+
"""How much audio was processed. This time includes trimming silences as well.
|
|
980
|
+
This cursor is moved after server received enough data to update recognition results (includes silence as well).
|
|
967
981
|
"""
|
|
968
982
|
final_time_ms: builtins.int
|
|
969
|
-
"""Time of last final. This cursor is moved when server decides that recognition from start of audio until final_time_ms will not change anymore
|
|
970
|
-
usually this
|
|
983
|
+
"""Time of last final. This cursor is moved when server decides that recognition from start of audio until `final_time_ms` will not change anymore
|
|
984
|
+
usually this event is followed by EOU detection. This behavior could change in future.
|
|
971
985
|
"""
|
|
972
986
|
final_index: builtins.int
|
|
973
987
|
"""This is index of last final server send. Incremented after each new final."""
|
|
974
988
|
eou_time_ms: builtins.int
|
|
975
989
|
"""Estimated time of EOU. Cursor is updated after each new EOU is sent.
|
|
976
|
-
For external classifier this equals to received_data_ms at the moment EOU event arrives.
|
|
990
|
+
For external classifier this equals to [received_data_ms] at the moment EOU event arrives.
|
|
977
991
|
For internal classifier this is estimation of time. The time is not exact and has the same guarantees as word timings.
|
|
978
992
|
"""
|
|
979
993
|
def __init__(
|
|
@@ -1018,7 +1032,7 @@ global___FinalRefinement = FinalRefinement
|
|
|
1018
1032
|
|
|
1019
1033
|
@typing.final
|
|
1020
1034
|
class StatusCode(google.protobuf.message.Message):
|
|
1021
|
-
"""Status message"""
|
|
1035
|
+
"""Status message."""
|
|
1022
1036
|
|
|
1023
1037
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
1024
1038
|
|
|
@@ -1068,11 +1082,11 @@ class PhraseHighlight(google.protobuf.message.Message):
|
|
|
1068
1082
|
START_TIME_MS_FIELD_NUMBER: builtins.int
|
|
1069
1083
|
END_TIME_MS_FIELD_NUMBER: builtins.int
|
|
1070
1084
|
text: builtins.str
|
|
1071
|
-
"""Text transcription of the highlighted audio segment"""
|
|
1085
|
+
"""Text transcription of the highlighted audio segment."""
|
|
1072
1086
|
start_time_ms: builtins.int
|
|
1073
|
-
"""Start time of the highlighted audio segment"""
|
|
1087
|
+
"""Start time of the highlighted audio segment."""
|
|
1074
1088
|
end_time_ms: builtins.int
|
|
1075
|
-
"""End time of the highlighted audio segment"""
|
|
1089
|
+
"""End time of the highlighted audio segment."""
|
|
1076
1090
|
def __init__(
|
|
1077
1091
|
self,
|
|
1078
1092
|
*,
|
|
@@ -1091,9 +1105,9 @@ class RecognitionClassifierLabel(google.protobuf.message.Message):
|
|
|
1091
1105
|
LABEL_FIELD_NUMBER: builtins.int
|
|
1092
1106
|
CONFIDENCE_FIELD_NUMBER: builtins.int
|
|
1093
1107
|
label: builtins.str
|
|
1094
|
-
"""The label of the class predicted by the classifier"""
|
|
1108
|
+
"""The label of the class predicted by the classifier."""
|
|
1095
1109
|
confidence: builtins.float
|
|
1096
|
-
"""The prediction confidence"""
|
|
1110
|
+
"""The prediction confidence."""
|
|
1097
1111
|
def __init__(
|
|
1098
1112
|
self,
|
|
1099
1113
|
*,
|
|
@@ -1112,14 +1126,14 @@ class RecognitionClassifierResult(google.protobuf.message.Message):
|
|
|
1112
1126
|
HIGHLIGHTS_FIELD_NUMBER: builtins.int
|
|
1113
1127
|
LABELS_FIELD_NUMBER: builtins.int
|
|
1114
1128
|
classifier: builtins.str
|
|
1115
|
-
"""Name of the triggered classifier"""
|
|
1129
|
+
"""Name of the triggered classifier."""
|
|
1116
1130
|
@property
|
|
1117
1131
|
def highlights(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___PhraseHighlight]:
|
|
1118
|
-
"""List of highlights, i.e. parts of phrase that determine the result of the classification"""
|
|
1132
|
+
"""List of highlights, i.e. parts of phrase that determine the result of the classification."""
|
|
1119
1133
|
|
|
1120
1134
|
@property
|
|
1121
1135
|
def labels(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___RecognitionClassifierLabel]:
|
|
1122
|
-
"""Classifier predictions"""
|
|
1136
|
+
"""Classifier predictions."""
|
|
1123
1137
|
|
|
1124
1138
|
def __init__(
|
|
1125
1139
|
self,
|
|
@@ -1144,34 +1158,34 @@ class RecognitionClassifierUpdate(google.protobuf.message.Message):
|
|
|
1144
1158
|
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
1145
1159
|
WINDOW_TYPE_UNSPECIFIED: RecognitionClassifierUpdate._WindowType.ValueType # 0
|
|
1146
1160
|
LAST_UTTERANCE: RecognitionClassifierUpdate._WindowType.ValueType # 1
|
|
1147
|
-
"""The result of applying the classifier to the last utterance response"""
|
|
1161
|
+
"""The result of applying the classifier to the last utterance response."""
|
|
1148
1162
|
LAST_FINAL: RecognitionClassifierUpdate._WindowType.ValueType # 2
|
|
1149
|
-
"""The result of applying the classifier to the last final response"""
|
|
1163
|
+
"""The result of applying the classifier to the last final response."""
|
|
1150
1164
|
LAST_PARTIAL: RecognitionClassifierUpdate._WindowType.ValueType # 3
|
|
1151
|
-
"""The result of applying the classifier to the last partial response"""
|
|
1165
|
+
"""The result of applying the classifier to the last partial response."""
|
|
1152
1166
|
|
|
1153
1167
|
class WindowType(_WindowType, metaclass=_WindowTypeEnumTypeWrapper): ...
|
|
1154
1168
|
WINDOW_TYPE_UNSPECIFIED: RecognitionClassifierUpdate.WindowType.ValueType # 0
|
|
1155
1169
|
LAST_UTTERANCE: RecognitionClassifierUpdate.WindowType.ValueType # 1
|
|
1156
|
-
"""The result of applying the classifier to the last utterance response"""
|
|
1170
|
+
"""The result of applying the classifier to the last utterance response."""
|
|
1157
1171
|
LAST_FINAL: RecognitionClassifierUpdate.WindowType.ValueType # 2
|
|
1158
|
-
"""The result of applying the classifier to the last final response"""
|
|
1172
|
+
"""The result of applying the classifier to the last final response."""
|
|
1159
1173
|
LAST_PARTIAL: RecognitionClassifierUpdate.WindowType.ValueType # 3
|
|
1160
|
-
"""The result of applying the classifier to the last partial response"""
|
|
1174
|
+
"""The result of applying the classifier to the last partial response."""
|
|
1161
1175
|
|
|
1162
1176
|
WINDOW_TYPE_FIELD_NUMBER: builtins.int
|
|
1163
1177
|
START_TIME_MS_FIELD_NUMBER: builtins.int
|
|
1164
1178
|
END_TIME_MS_FIELD_NUMBER: builtins.int
|
|
1165
1179
|
CLASSIFIER_RESULT_FIELD_NUMBER: builtins.int
|
|
1166
1180
|
window_type: global___RecognitionClassifierUpdate.WindowType.ValueType
|
|
1167
|
-
"""Response window type"""
|
|
1181
|
+
"""Response window type."""
|
|
1168
1182
|
start_time_ms: builtins.int
|
|
1169
|
-
"""Start time of the audio segment used for classification"""
|
|
1183
|
+
"""Start time of the audio segment used for classification."""
|
|
1170
1184
|
end_time_ms: builtins.int
|
|
1171
|
-
"""End time of the audio segment used for classification"""
|
|
1185
|
+
"""End time of the audio segment used for classification."""
|
|
1172
1186
|
@property
|
|
1173
1187
|
def classifier_result(self) -> global___RecognitionClassifierResult:
|
|
1174
|
-
"""Result for dictionary-based classifier"""
|
|
1188
|
+
"""Result for dictionary-based classifier."""
|
|
1175
1189
|
|
|
1176
1190
|
def __init__(
|
|
1177
1191
|
self,
|
|
@@ -1197,9 +1211,9 @@ class DescriptiveStatistics(google.protobuf.message.Message):
|
|
|
1197
1211
|
LEVEL_FIELD_NUMBER: builtins.int
|
|
1198
1212
|
VALUE_FIELD_NUMBER: builtins.int
|
|
1199
1213
|
level: builtins.float
|
|
1200
|
-
"""Quantile level in range (0, 1)"""
|
|
1214
|
+
"""Quantile level in range (0, 1)."""
|
|
1201
1215
|
value: builtins.float
|
|
1202
|
-
"""Quantile value"""
|
|
1216
|
+
"""Quantile value."""
|
|
1203
1217
|
def __init__(
|
|
1204
1218
|
self,
|
|
1205
1219
|
*,
|
|
@@ -1214,16 +1228,16 @@ class DescriptiveStatistics(google.protobuf.message.Message):
|
|
|
1214
1228
|
STD_FIELD_NUMBER: builtins.int
|
|
1215
1229
|
QUANTILES_FIELD_NUMBER: builtins.int
|
|
1216
1230
|
min: builtins.float
|
|
1217
|
-
"""Minimum observed value"""
|
|
1231
|
+
"""Minimum observed value."""
|
|
1218
1232
|
max: builtins.float
|
|
1219
|
-
"""Maximum observed value"""
|
|
1233
|
+
"""Maximum observed value."""
|
|
1220
1234
|
mean: builtins.float
|
|
1221
|
-
"""Estimated mean of distribution"""
|
|
1235
|
+
"""Estimated mean of distribution."""
|
|
1222
1236
|
std: builtins.float
|
|
1223
|
-
"""Estimated standard deviation of distribution"""
|
|
1237
|
+
"""Estimated standard deviation of distribution."""
|
|
1224
1238
|
@property
|
|
1225
1239
|
def quantiles(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___DescriptiveStatistics.Quantile]:
|
|
1226
|
-
"""List of evaluated quantiles"""
|
|
1240
|
+
"""List of evaluated quantiles."""
|
|
1227
1241
|
|
|
1228
1242
|
def __init__(
|
|
1229
1243
|
self,
|
|
@@ -1245,9 +1259,9 @@ class AudioSegmentBoundaries(google.protobuf.message.Message):
|
|
|
1245
1259
|
START_TIME_MS_FIELD_NUMBER: builtins.int
|
|
1246
1260
|
END_TIME_MS_FIELD_NUMBER: builtins.int
|
|
1247
1261
|
start_time_ms: builtins.int
|
|
1248
|
-
"""Audio segment start time"""
|
|
1262
|
+
"""Audio segment start time."""
|
|
1249
1263
|
end_time_ms: builtins.int
|
|
1250
|
-
"""Audio segment end time"""
|
|
1264
|
+
"""Audio segment end time."""
|
|
1251
1265
|
def __init__(
|
|
1252
1266
|
self,
|
|
1253
1267
|
*,
|
|
@@ -1297,42 +1311,42 @@ class SpeakerAnalysis(google.protobuf.message.Message):
|
|
|
1297
1311
|
UTTERANCE_COUNT_FIELD_NUMBER: builtins.int
|
|
1298
1312
|
UTTERANCE_DURATION_ESTIMATION_FIELD_NUMBER: builtins.int
|
|
1299
1313
|
speaker_tag: builtins.str
|
|
1300
|
-
"""Speaker tag"""
|
|
1314
|
+
"""Speaker tag."""
|
|
1301
1315
|
window_type: global___SpeakerAnalysis.WindowType.ValueType
|
|
1302
|
-
"""Response window type"""
|
|
1316
|
+
"""Response window type."""
|
|
1303
1317
|
total_speech_ms: builtins.int
|
|
1304
|
-
"""Total speech duration"""
|
|
1318
|
+
"""Total speech duration."""
|
|
1305
1319
|
speech_ratio: builtins.float
|
|
1306
|
-
"""Speech ratio within audio segment"""
|
|
1320
|
+
"""Speech ratio within audio segment."""
|
|
1307
1321
|
total_silence_ms: builtins.int
|
|
1308
|
-
"""Total silence
|
|
1322
|
+
"""Total duration of silence."""
|
|
1309
1323
|
silence_ratio: builtins.float
|
|
1310
|
-
"""Silence ratio within audio segment"""
|
|
1324
|
+
"""Silence ratio within audio segment."""
|
|
1311
1325
|
words_count: builtins.int
|
|
1312
|
-
"""Number of words in recognized speech"""
|
|
1326
|
+
"""Number of words in recognized speech."""
|
|
1313
1327
|
letters_count: builtins.int
|
|
1314
|
-
"""Number of letters in recognized speech"""
|
|
1328
|
+
"""Number of letters in recognized speech."""
|
|
1315
1329
|
utterance_count: builtins.int
|
|
1316
1330
|
"""Number of utterances"""
|
|
1317
1331
|
@property
|
|
1318
1332
|
def speech_boundaries(self) -> global___AudioSegmentBoundaries:
|
|
1319
|
-
"""Audio segment boundaries"""
|
|
1333
|
+
"""Audio segment boundaries."""
|
|
1320
1334
|
|
|
1321
1335
|
@property
|
|
1322
1336
|
def words_per_second(self) -> global___DescriptiveStatistics:
|
|
1323
|
-
"""Descriptive statistics for words per second distribution"""
|
|
1337
|
+
"""Descriptive statistics for words per second distribution."""
|
|
1324
1338
|
|
|
1325
1339
|
@property
|
|
1326
1340
|
def letters_per_second(self) -> global___DescriptiveStatistics:
|
|
1327
|
-
"""Descriptive statistics for letters per second distribution"""
|
|
1341
|
+
"""Descriptive statistics for letters per second distribution."""
|
|
1328
1342
|
|
|
1329
1343
|
@property
|
|
1330
1344
|
def words_per_utterance(self) -> global___DescriptiveStatistics:
|
|
1331
|
-
"""Descriptive statistics for words per utterance distribution"""
|
|
1345
|
+
"""Descriptive statistics for words per utterance distribution."""
|
|
1332
1346
|
|
|
1333
1347
|
@property
|
|
1334
1348
|
def letters_per_utterance(self) -> global___DescriptiveStatistics:
|
|
1335
|
-
"""Descriptive statistics for letters per utterance distribution"""
|
|
1349
|
+
"""Descriptive statistics for letters per utterance distribution."""
|
|
1336
1350
|
|
|
1337
1351
|
@property
|
|
1338
1352
|
def utterance_duration_estimation(self) -> global___DescriptiveStatistics:
|
|
@@ -1375,14 +1389,14 @@ class ConversationAnalysis(google.protobuf.message.Message):
|
|
|
1375
1389
|
INTERRUPTS_DURATION_MS_FIELD_NUMBER: builtins.int
|
|
1376
1390
|
INTERRUPTS_FIELD_NUMBER: builtins.int
|
|
1377
1391
|
speaker_tag: builtins.str
|
|
1378
|
-
"""Speaker tag"""
|
|
1392
|
+
"""Speaker tag."""
|
|
1379
1393
|
interrupts_count: builtins.int
|
|
1380
|
-
"""Number of interrupts made by the speaker"""
|
|
1394
|
+
"""Number of interrupts made by the speaker."""
|
|
1381
1395
|
interrupts_duration_ms: builtins.int
|
|
1382
|
-
"""Total duration of all interrupts"""
|
|
1396
|
+
"""Total duration of all interrupts."""
|
|
1383
1397
|
@property
|
|
1384
1398
|
def interrupts(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___AudioSegmentBoundaries]:
|
|
1385
|
-
"""Boundaries for every interrupt"""
|
|
1399
|
+
"""Boundaries for every interrupt."""
|
|
1386
1400
|
|
|
1387
1401
|
def __init__(
|
|
1388
1402
|
self,
|
|
@@ -1405,32 +1419,32 @@ class ConversationAnalysis(google.protobuf.message.Message):
|
|
|
1405
1419
|
TOTAL_SPEECH_DURATION_MS_FIELD_NUMBER: builtins.int
|
|
1406
1420
|
TOTAL_SPEECH_RATIO_FIELD_NUMBER: builtins.int
|
|
1407
1421
|
total_simultaneous_silence_duration_ms: builtins.int
|
|
1408
|
-
"""Total simultaneous silence duration"""
|
|
1422
|
+
"""Total simultaneous silence duration."""
|
|
1409
1423
|
total_simultaneous_silence_ratio: builtins.float
|
|
1410
|
-
"""Simultaneous silence ratio within audio segment"""
|
|
1424
|
+
"""Simultaneous silence ratio within audio segment."""
|
|
1411
1425
|
total_simultaneous_speech_duration_ms: builtins.int
|
|
1412
|
-
"""Total simultaneous speech duration"""
|
|
1426
|
+
"""Total simultaneous speech duration."""
|
|
1413
1427
|
total_simultaneous_speech_ratio: builtins.float
|
|
1414
|
-
"""Simultaneous speech ratio within audio segment"""
|
|
1428
|
+
"""Simultaneous speech ratio within audio segment."""
|
|
1415
1429
|
total_speech_duration_ms: builtins.int
|
|
1416
|
-
"""Total speech duration, including both simultaneous and separate speech"""
|
|
1430
|
+
"""Total speech duration, including both simultaneous and separate speech."""
|
|
1417
1431
|
total_speech_ratio: builtins.float
|
|
1418
|
-
"""Total speech ratio within audio segment"""
|
|
1432
|
+
"""Total speech ratio within audio segment."""
|
|
1419
1433
|
@property
|
|
1420
1434
|
def conversation_boundaries(self) -> global___AudioSegmentBoundaries:
|
|
1421
|
-
"""Audio segment boundaries"""
|
|
1435
|
+
"""Audio segment boundaries."""
|
|
1422
1436
|
|
|
1423
1437
|
@property
|
|
1424
1438
|
def simultaneous_silence_duration_estimation(self) -> global___DescriptiveStatistics:
|
|
1425
|
-
"""Descriptive statistics for simultaneous silence duration distribution"""
|
|
1439
|
+
"""Descriptive statistics for simultaneous silence duration distribution."""
|
|
1426
1440
|
|
|
1427
1441
|
@property
|
|
1428
1442
|
def simultaneous_speech_duration_estimation(self) -> global___DescriptiveStatistics:
|
|
1429
|
-
"""Descriptive statistics for simultaneous speech duration distribution"""
|
|
1443
|
+
"""Descriptive statistics for simultaneous speech duration distribution."""
|
|
1430
1444
|
|
|
1431
1445
|
@property
|
|
1432
1446
|
def speaker_interrupts(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ConversationAnalysis.InterruptsEvaluation]:
|
|
1433
|
-
"""Interrupts description for every speaker"""
|
|
1447
|
+
"""Interrupts description for every speaker."""
|
|
1434
1448
|
|
|
1435
1449
|
def __init__(
|
|
1436
1450
|
self,
|
|
@@ -1505,9 +1519,7 @@ global___Summarization = Summarization
|
|
|
1505
1519
|
@typing.final
|
|
1506
1520
|
class StreamingResponse(google.protobuf.message.Message):
|
|
1507
1521
|
"""Responses from server.
|
|
1508
|
-
Each response contains session
|
|
1509
|
-
AudioCursors
|
|
1510
|
-
plus specific event
|
|
1522
|
+
Each response contains session UUID, AudioCursors, and specific event.
|
|
1511
1523
|
"""
|
|
1512
1524
|
|
|
1513
1525
|
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
@@ -1526,12 +1538,12 @@ class StreamingResponse(google.protobuf.message.Message):
|
|
|
1526
1538
|
SUMMARIZATION_FIELD_NUMBER: builtins.int
|
|
1527
1539
|
CHANNEL_TAG_FIELD_NUMBER: builtins.int
|
|
1528
1540
|
response_wall_time_ms: builtins.int
|
|
1529
|
-
"""Wall clock on server side. This is time when server wrote results to stream"""
|
|
1541
|
+
"""Wall clock on server side. This is time when server wrote results to stream."""
|
|
1530
1542
|
channel_tag: builtins.str
|
|
1531
1543
|
"""Tag for distinguish audio channels."""
|
|
1532
1544
|
@property
|
|
1533
1545
|
def session_uuid(self) -> global___SessionUuid:
|
|
1534
|
-
"""Session identifier"""
|
|
1546
|
+
"""Session identifier."""
|
|
1535
1547
|
|
|
1536
1548
|
@property
|
|
1537
1549
|
def audio_cursors(self) -> global___AudioCursors:
|
|
@@ -1539,13 +1551,13 @@ class StreamingResponse(google.protobuf.message.Message):
|
|
|
1539
1551
|
|
|
1540
1552
|
@property
|
|
1541
1553
|
def partial(self) -> global___AlternativeUpdate:
|
|
1542
|
-
"""Partial results, server will send them regularly after enough audio data was received from user.
|
|
1543
|
-
from final_time_ms to partial_time_ms
|
|
1554
|
+
"""Partial results, server will send them regularly after enough audio data was received from user.
|
|
1555
|
+
This is the current text estimation from `final_time_ms` to `partial_time_ms`. Could change after new data will arrive.
|
|
1544
1556
|
"""
|
|
1545
1557
|
|
|
1546
1558
|
@property
|
|
1547
1559
|
def final(self) -> global___AlternativeUpdate:
|
|
1548
|
-
"""Final results, the recognition is now fixed until final_time_ms
|
|
1560
|
+
"""Final results, the recognition is now fixed until `final_time_ms`. For now, final is sent only if the EOU event was triggered. This behavior could be changed in future releases."""
|
|
1549
1561
|
|
|
1550
1562
|
@property
|
|
1551
1563
|
def eou_update(self) -> global___EouUpdate:
|
|
@@ -1565,19 +1577,19 @@ class StreamingResponse(google.protobuf.message.Message):
|
|
|
1565
1577
|
|
|
1566
1578
|
@property
|
|
1567
1579
|
def classifier_update(self) -> global___RecognitionClassifierUpdate:
|
|
1568
|
-
"""Result of the triggered classifier"""
|
|
1580
|
+
"""Result of the triggered classifier."""
|
|
1569
1581
|
|
|
1570
1582
|
@property
|
|
1571
1583
|
def speaker_analysis(self) -> global___SpeakerAnalysis:
|
|
1572
|
-
"""Speech statistics for every speaker"""
|
|
1584
|
+
"""Speech statistics for every speaker."""
|
|
1573
1585
|
|
|
1574
1586
|
@property
|
|
1575
1587
|
def conversation_analysis(self) -> global___ConversationAnalysis:
|
|
1576
|
-
"""Conversation statistics"""
|
|
1588
|
+
"""Conversation statistics."""
|
|
1577
1589
|
|
|
1578
1590
|
@property
|
|
1579
1591
|
def summarization(self) -> global___Summarization:
|
|
1580
|
-
"""Summary"""
|
|
1592
|
+
"""Summary."""
|
|
1581
1593
|
|
|
1582
1594
|
def __init__(
|
|
1583
1595
|
self,
|