sarvamai 0.1.5a13__tar.gz → 0.1.6__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/PKG-INFO +1 -1
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/pyproject.toml +1 -1
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/__init__.py +15 -4
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/client_wrapper.py +2 -2
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/audio_data.py +3 -2
- sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/__init__.py +17 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/client.py +10 -8
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/raw_client.py +10 -8
- sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/types/__init__.py +15 -0
- sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_high_vad_sensitivity.py +5 -0
- sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_vad_signals.py +5 -0
- sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/__init__.py +15 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/client.py +12 -8
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/raw_client.py +12 -8
- sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/types/__init__.py +13 -0
- sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/types/speech_to_text_translate_streaming_high_vad_sensitivity.py +5 -0
- sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/types/speech_to_text_translate_streaming_vad_signals.py +5 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text_to_speech/client.py +6 -14
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text_to_speech/raw_client.py +6 -14
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/__init__.py +0 -2
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/audio_data.py +1 -2
- sarvamai-0.1.6/src/sarvamai/types/text_to_speech_model.py +5 -0
- sarvamai-0.1.6/src/sarvamai/types/text_to_speech_speaker.py +7 -0
- sarvamai-0.1.5a13/src/sarvamai/speech_to_text_streaming/__init__.py +0 -7
- sarvamai-0.1.5a13/src/sarvamai/speech_to_text_streaming/types/__init__.py +0 -8
- sarvamai-0.1.5a13/src/sarvamai/speech_to_text_translate_streaming/__init__.py +0 -7
- sarvamai-0.1.5a13/src/sarvamai/speech_to_text_translate_streaming/types/__init__.py +0 -7
- sarvamai-0.1.5a13/src/sarvamai/types/audio_data_encoding.py +0 -5
- sarvamai-0.1.5a13/src/sarvamai/types/text_to_speech_model.py +0 -5
- sarvamai-0.1.5a13/src/sarvamai/types/text_to_speech_speaker.py +0 -28
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/README.md +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/chat/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/chat/client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/chat/raw_client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/api_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/datetime_utils.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/events.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/file.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/force_multipart.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/http_client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/http_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/jsonable_encoder.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/pydantic_utilities.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/query_encoder.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/remove_none_from_dict.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/request_options.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/serialization.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/environment.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/bad_request_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/forbidden_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/internal_server_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/service_unavailable_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/too_many_requests_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/unprocessable_entity_error.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/play.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/py.typed +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/audio_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_assistant_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_system_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_user_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_response_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/choice.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/completion_usage.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/config_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/create_chat_completion_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/diarized_entry.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/diarized_transcript.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/error_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/error_details.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/error_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/events_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/language_identification_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_response_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_streaming_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_transcription_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_response_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_streaming_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_transcription_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/stop_configuration.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/text_to_speech_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/timestamps_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/transcription_metrics.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/translation_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/transliteration_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text/client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text/raw_client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/socket_client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_language_code.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/socket_client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/types/speech_to_text_translate_streaming_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text/client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text/raw_client.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text_to_speech/__init__.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/audio_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_assistant_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_system_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_user_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_response_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/choice.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/completion_usage.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/config_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/create_chat_completion_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/diarized_entry.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/diarized_transcript.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_code.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_details.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_message.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/events_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/finish_reason.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/format.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/language_identification_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/numerals_format.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/reasoning_effort.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/response_type.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/role.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/sarvam_model_ids.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_sample_rate.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_response_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_streaming_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_transcription_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_response_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_streaming_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_transcription_data.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/spoken_form_numerals_format.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/stop_configuration.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/text_to_speech_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/text_to_speech_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/timestamps_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transcription_metrics.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_mode.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_model.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_source_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_speaker_gender.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_target_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translation_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translatiterate_target_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transliterate_mode.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transliterate_source_language.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transliteration_response.py +0 -0
- {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/version.py +0 -0
|
@@ -4,7 +4,6 @@
|
|
|
4
4
|
|
|
5
5
|
from .types import (
|
|
6
6
|
AudioData,
|
|
7
|
-
AudioDataEncoding,
|
|
8
7
|
AudioMessage,
|
|
9
8
|
ChatCompletionRequestAssistantMessage,
|
|
10
9
|
ChatCompletionRequestMessage,
|
|
@@ -113,14 +112,22 @@ from .requests import (
|
|
|
113
112
|
TranslationResponseParams,
|
|
114
113
|
TransliterationResponseParams,
|
|
115
114
|
)
|
|
116
|
-
from .speech_to_text_streaming import
|
|
117
|
-
|
|
115
|
+
from .speech_to_text_streaming import (
|
|
116
|
+
SpeechToTextStreamingHighVadSensitivity,
|
|
117
|
+
SpeechToTextStreamingLanguageCode,
|
|
118
|
+
SpeechToTextStreamingModel,
|
|
119
|
+
SpeechToTextStreamingVadSignals,
|
|
120
|
+
)
|
|
121
|
+
from .speech_to_text_translate_streaming import (
|
|
122
|
+
SpeechToTextTranslateStreamingHighVadSensitivity,
|
|
123
|
+
SpeechToTextTranslateStreamingModel,
|
|
124
|
+
SpeechToTextTranslateStreamingVadSignals,
|
|
125
|
+
)
|
|
118
126
|
from .version import __version__
|
|
119
127
|
|
|
120
128
|
__all__ = [
|
|
121
129
|
"AsyncSarvamAI",
|
|
122
130
|
"AudioData",
|
|
123
|
-
"AudioDataEncoding",
|
|
124
131
|
"AudioDataParams",
|
|
125
132
|
"AudioMessage",
|
|
126
133
|
"AudioMessageParams",
|
|
@@ -183,10 +190,12 @@ __all__ = [
|
|
|
183
190
|
"SpeechToTextResponseData",
|
|
184
191
|
"SpeechToTextResponseDataParams",
|
|
185
192
|
"SpeechToTextResponseParams",
|
|
193
|
+
"SpeechToTextStreamingHighVadSensitivity",
|
|
186
194
|
"SpeechToTextStreamingLanguageCode",
|
|
187
195
|
"SpeechToTextStreamingModel",
|
|
188
196
|
"SpeechToTextStreamingResponse",
|
|
189
197
|
"SpeechToTextStreamingResponseParams",
|
|
198
|
+
"SpeechToTextStreamingVadSignals",
|
|
190
199
|
"SpeechToTextTranscriptionData",
|
|
191
200
|
"SpeechToTextTranscriptionDataParams",
|
|
192
201
|
"SpeechToTextTranslateLanguage",
|
|
@@ -195,9 +204,11 @@ __all__ = [
|
|
|
195
204
|
"SpeechToTextTranslateResponseData",
|
|
196
205
|
"SpeechToTextTranslateResponseDataParams",
|
|
197
206
|
"SpeechToTextTranslateResponseParams",
|
|
207
|
+
"SpeechToTextTranslateStreamingHighVadSensitivity",
|
|
198
208
|
"SpeechToTextTranslateStreamingModel",
|
|
199
209
|
"SpeechToTextTranslateStreamingResponse",
|
|
200
210
|
"SpeechToTextTranslateStreamingResponseParams",
|
|
211
|
+
"SpeechToTextTranslateStreamingVadSignals",
|
|
201
212
|
"SpeechToTextTranslateTranscriptionData",
|
|
202
213
|
"SpeechToTextTranslateTranscriptionDataParams",
|
|
203
214
|
"SpokenFormNumeralsFormat",
|
|
@@ -17,10 +17,10 @@ class BaseClientWrapper:
|
|
|
17
17
|
|
|
18
18
|
def get_headers(self) -> typing.Dict[str, str]:
|
|
19
19
|
headers: typing.Dict[str, str] = {
|
|
20
|
-
"User-Agent": "sarvamai/0.1.
|
|
20
|
+
"User-Agent": "sarvamai/0.1.6",
|
|
21
21
|
"X-Fern-Language": "Python",
|
|
22
22
|
"X-Fern-SDK-Name": "sarvamai",
|
|
23
|
-
"X-Fern-SDK-Version": "0.1.
|
|
23
|
+
"X-Fern-SDK-Version": "0.1.6",
|
|
24
24
|
}
|
|
25
25
|
headers["api-subscription-key"] = self.api_subscription_key
|
|
26
26
|
return headers
|
|
@@ -1,7 +1,8 @@
|
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
|
2
2
|
|
|
3
|
+
import typing
|
|
4
|
+
|
|
3
5
|
import typing_extensions
|
|
4
|
-
from ..types.audio_data_encoding import AudioDataEncoding
|
|
5
6
|
|
|
6
7
|
|
|
7
8
|
class AudioDataParams(typing_extensions.TypedDict):
|
|
@@ -15,7 +16,7 @@ class AudioDataParams(typing_extensions.TypedDict):
|
|
|
15
16
|
Audio sample rate in Hz (16kHz preferred, 8kHz least preferred)
|
|
16
17
|
"""
|
|
17
18
|
|
|
18
|
-
encoding:
|
|
19
|
+
encoding: typing.Literal["audio/wav"]
|
|
19
20
|
"""
|
|
20
21
|
Audio encoding format
|
|
21
22
|
"""
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
# isort: skip_file
|
|
4
|
+
|
|
5
|
+
from .types import (
|
|
6
|
+
SpeechToTextStreamingHighVadSensitivity,
|
|
7
|
+
SpeechToTextStreamingLanguageCode,
|
|
8
|
+
SpeechToTextStreamingModel,
|
|
9
|
+
SpeechToTextStreamingVadSignals,
|
|
10
|
+
)
|
|
11
|
+
|
|
12
|
+
__all__ = [
|
|
13
|
+
"SpeechToTextStreamingHighVadSensitivity",
|
|
14
|
+
"SpeechToTextStreamingLanguageCode",
|
|
15
|
+
"SpeechToTextStreamingModel",
|
|
16
|
+
"SpeechToTextStreamingVadSignals",
|
|
17
|
+
]
|
|
@@ -11,8 +11,10 @@ from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
|
|
11
11
|
from ..core.request_options import RequestOptions
|
|
12
12
|
from .raw_client import AsyncRawSpeechToTextStreamingClient, RawSpeechToTextStreamingClient
|
|
13
13
|
from .socket_client import AsyncSpeechToTextStreamingSocketClient, SpeechToTextStreamingSocketClient
|
|
14
|
+
from .types.speech_to_text_streaming_high_vad_sensitivity import SpeechToTextStreamingHighVadSensitivity
|
|
14
15
|
from .types.speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
|
|
15
16
|
from .types.speech_to_text_streaming_model import SpeechToTextStreamingModel
|
|
17
|
+
from .types.speech_to_text_streaming_vad_signals import SpeechToTextStreamingVadSignals
|
|
16
18
|
|
|
17
19
|
|
|
18
20
|
class SpeechToTextStreamingClient:
|
|
@@ -36,8 +38,8 @@ class SpeechToTextStreamingClient:
|
|
|
36
38
|
*,
|
|
37
39
|
language_code: SpeechToTextStreamingLanguageCode,
|
|
38
40
|
model: typing.Optional[SpeechToTextStreamingModel] = None,
|
|
39
|
-
high_vad_sensitivity: typing.Optional[
|
|
40
|
-
vad_signals: typing.Optional[
|
|
41
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
|
|
42
|
+
vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
|
|
41
43
|
api_subscription_key: typing.Optional[str] = None,
|
|
42
44
|
request_options: typing.Optional[RequestOptions] = None,
|
|
43
45
|
) -> typing.Iterator[SpeechToTextStreamingSocketClient]:
|
|
@@ -52,10 +54,10 @@ class SpeechToTextStreamingClient:
|
|
|
52
54
|
model : typing.Optional[SpeechToTextStreamingModel]
|
|
53
55
|
Speech to text model to use
|
|
54
56
|
|
|
55
|
-
high_vad_sensitivity : typing.Optional[
|
|
57
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
|
|
56
58
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
57
59
|
|
|
58
|
-
vad_signals : typing.Optional[
|
|
60
|
+
vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
|
|
59
61
|
Enable VAD signals in response
|
|
60
62
|
|
|
61
63
|
api_subscription_key : typing.Optional[str]
|
|
@@ -123,8 +125,8 @@ class AsyncSpeechToTextStreamingClient:
|
|
|
123
125
|
*,
|
|
124
126
|
language_code: SpeechToTextStreamingLanguageCode,
|
|
125
127
|
model: typing.Optional[SpeechToTextStreamingModel] = None,
|
|
126
|
-
high_vad_sensitivity: typing.Optional[
|
|
127
|
-
vad_signals: typing.Optional[
|
|
128
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
|
|
129
|
+
vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
|
|
128
130
|
api_subscription_key: typing.Optional[str] = None,
|
|
129
131
|
request_options: typing.Optional[RequestOptions] = None,
|
|
130
132
|
) -> typing.AsyncIterator[AsyncSpeechToTextStreamingSocketClient]:
|
|
@@ -139,10 +141,10 @@ class AsyncSpeechToTextStreamingClient:
|
|
|
139
141
|
model : typing.Optional[SpeechToTextStreamingModel]
|
|
140
142
|
Speech to text model to use
|
|
141
143
|
|
|
142
|
-
high_vad_sensitivity : typing.Optional[
|
|
144
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
|
|
143
145
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
144
146
|
|
|
145
|
-
vad_signals : typing.Optional[
|
|
147
|
+
vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
|
|
146
148
|
Enable VAD signals in response
|
|
147
149
|
|
|
148
150
|
api_subscription_key : typing.Optional[str]
|
|
@@ -10,8 +10,10 @@ from ..core.api_error import ApiError
|
|
|
10
10
|
from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
|
11
11
|
from ..core.request_options import RequestOptions
|
|
12
12
|
from .socket_client import AsyncSpeechToTextStreamingSocketClient, SpeechToTextStreamingSocketClient
|
|
13
|
+
from .types.speech_to_text_streaming_high_vad_sensitivity import SpeechToTextStreamingHighVadSensitivity
|
|
13
14
|
from .types.speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
|
|
14
15
|
from .types.speech_to_text_streaming_model import SpeechToTextStreamingModel
|
|
16
|
+
from .types.speech_to_text_streaming_vad_signals import SpeechToTextStreamingVadSignals
|
|
15
17
|
|
|
16
18
|
|
|
17
19
|
class RawSpeechToTextStreamingClient:
|
|
@@ -24,8 +26,8 @@ class RawSpeechToTextStreamingClient:
|
|
|
24
26
|
*,
|
|
25
27
|
language_code: SpeechToTextStreamingLanguageCode,
|
|
26
28
|
model: typing.Optional[SpeechToTextStreamingModel] = None,
|
|
27
|
-
high_vad_sensitivity: typing.Optional[
|
|
28
|
-
vad_signals: typing.Optional[
|
|
29
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
|
|
30
|
+
vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
|
|
29
31
|
api_subscription_key: typing.Optional[str] = None,
|
|
30
32
|
request_options: typing.Optional[RequestOptions] = None,
|
|
31
33
|
) -> typing.Iterator[SpeechToTextStreamingSocketClient]:
|
|
@@ -40,10 +42,10 @@ class RawSpeechToTextStreamingClient:
|
|
|
40
42
|
model : typing.Optional[SpeechToTextStreamingModel]
|
|
41
43
|
Speech to text model to use
|
|
42
44
|
|
|
43
|
-
high_vad_sensitivity : typing.Optional[
|
|
45
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
|
|
44
46
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
45
47
|
|
|
46
|
-
vad_signals : typing.Optional[
|
|
48
|
+
vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
|
|
47
49
|
Enable VAD signals in response
|
|
48
50
|
|
|
49
51
|
api_subscription_key : typing.Optional[str]
|
|
@@ -100,8 +102,8 @@ class AsyncRawSpeechToTextStreamingClient:
|
|
|
100
102
|
*,
|
|
101
103
|
language_code: SpeechToTextStreamingLanguageCode,
|
|
102
104
|
model: typing.Optional[SpeechToTextStreamingModel] = None,
|
|
103
|
-
high_vad_sensitivity: typing.Optional[
|
|
104
|
-
vad_signals: typing.Optional[
|
|
105
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
|
|
106
|
+
vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
|
|
105
107
|
api_subscription_key: typing.Optional[str] = None,
|
|
106
108
|
request_options: typing.Optional[RequestOptions] = None,
|
|
107
109
|
) -> typing.AsyncIterator[AsyncSpeechToTextStreamingSocketClient]:
|
|
@@ -116,10 +118,10 @@ class AsyncRawSpeechToTextStreamingClient:
|
|
|
116
118
|
model : typing.Optional[SpeechToTextStreamingModel]
|
|
117
119
|
Speech to text model to use
|
|
118
120
|
|
|
119
|
-
high_vad_sensitivity : typing.Optional[
|
|
121
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
|
|
120
122
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
121
123
|
|
|
122
|
-
vad_signals : typing.Optional[
|
|
124
|
+
vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
|
|
123
125
|
Enable VAD signals in response
|
|
124
126
|
|
|
125
127
|
api_subscription_key : typing.Optional[str]
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
# isort: skip_file
|
|
4
|
+
|
|
5
|
+
from .speech_to_text_streaming_high_vad_sensitivity import SpeechToTextStreamingHighVadSensitivity
|
|
6
|
+
from .speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
|
|
7
|
+
from .speech_to_text_streaming_model import SpeechToTextStreamingModel
|
|
8
|
+
from .speech_to_text_streaming_vad_signals import SpeechToTextStreamingVadSignals
|
|
9
|
+
|
|
10
|
+
__all__ = [
|
|
11
|
+
"SpeechToTextStreamingHighVadSensitivity",
|
|
12
|
+
"SpeechToTextStreamingLanguageCode",
|
|
13
|
+
"SpeechToTextStreamingModel",
|
|
14
|
+
"SpeechToTextStreamingVadSignals",
|
|
15
|
+
]
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
# isort: skip_file
|
|
4
|
+
|
|
5
|
+
from .types import (
|
|
6
|
+
SpeechToTextTranslateStreamingHighVadSensitivity,
|
|
7
|
+
SpeechToTextTranslateStreamingModel,
|
|
8
|
+
SpeechToTextTranslateStreamingVadSignals,
|
|
9
|
+
)
|
|
10
|
+
|
|
11
|
+
__all__ = [
|
|
12
|
+
"SpeechToTextTranslateStreamingHighVadSensitivity",
|
|
13
|
+
"SpeechToTextTranslateStreamingModel",
|
|
14
|
+
"SpeechToTextTranslateStreamingVadSignals",
|
|
15
|
+
]
|
{sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/client.py
RENAMED
|
@@ -11,7 +11,11 @@ from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
|
|
11
11
|
from ..core.request_options import RequestOptions
|
|
12
12
|
from .raw_client import AsyncRawSpeechToTextTranslateStreamingClient, RawSpeechToTextTranslateStreamingClient
|
|
13
13
|
from .socket_client import AsyncSpeechToTextTranslateStreamingSocketClient, SpeechToTextTranslateStreamingSocketClient
|
|
14
|
+
from .types.speech_to_text_translate_streaming_high_vad_sensitivity import (
|
|
15
|
+
SpeechToTextTranslateStreamingHighVadSensitivity,
|
|
16
|
+
)
|
|
14
17
|
from .types.speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
|
|
18
|
+
from .types.speech_to_text_translate_streaming_vad_signals import SpeechToTextTranslateStreamingVadSignals
|
|
15
19
|
|
|
16
20
|
|
|
17
21
|
class SpeechToTextTranslateStreamingClient:
|
|
@@ -34,8 +38,8 @@ class SpeechToTextTranslateStreamingClient:
|
|
|
34
38
|
self,
|
|
35
39
|
*,
|
|
36
40
|
model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
|
|
37
|
-
high_vad_sensitivity: typing.Optional[
|
|
38
|
-
vad_signals: typing.Optional[
|
|
41
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
|
|
42
|
+
vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
|
|
39
43
|
api_subscription_key: typing.Optional[str] = None,
|
|
40
44
|
request_options: typing.Optional[RequestOptions] = None,
|
|
41
45
|
) -> typing.Iterator[SpeechToTextTranslateStreamingSocketClient]:
|
|
@@ -47,10 +51,10 @@ class SpeechToTextTranslateStreamingClient:
|
|
|
47
51
|
model : typing.Optional[SpeechToTextTranslateStreamingModel]
|
|
48
52
|
Speech to text model to use (defaults to "saaras:v2.5" if not specified)
|
|
49
53
|
|
|
50
|
-
high_vad_sensitivity : typing.Optional[
|
|
54
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
|
|
51
55
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
52
56
|
|
|
53
|
-
vad_signals : typing.Optional[
|
|
57
|
+
vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
|
|
54
58
|
Enable VAD signals in response
|
|
55
59
|
|
|
56
60
|
api_subscription_key : typing.Optional[str]
|
|
@@ -115,8 +119,8 @@ class AsyncSpeechToTextTranslateStreamingClient:
|
|
|
115
119
|
self,
|
|
116
120
|
*,
|
|
117
121
|
model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
|
|
118
|
-
high_vad_sensitivity: typing.Optional[
|
|
119
|
-
vad_signals: typing.Optional[
|
|
122
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
|
|
123
|
+
vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
|
|
120
124
|
api_subscription_key: typing.Optional[str] = None,
|
|
121
125
|
request_options: typing.Optional[RequestOptions] = None,
|
|
122
126
|
) -> typing.AsyncIterator[AsyncSpeechToTextTranslateStreamingSocketClient]:
|
|
@@ -128,10 +132,10 @@ class AsyncSpeechToTextTranslateStreamingClient:
|
|
|
128
132
|
model : typing.Optional[SpeechToTextTranslateStreamingModel]
|
|
129
133
|
Speech to text model to use (defaults to "saaras:v2.5" if not specified)
|
|
130
134
|
|
|
131
|
-
high_vad_sensitivity : typing.Optional[
|
|
135
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
|
|
132
136
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
133
137
|
|
|
134
|
-
vad_signals : typing.Optional[
|
|
138
|
+
vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
|
|
135
139
|
Enable VAD signals in response
|
|
136
140
|
|
|
137
141
|
api_subscription_key : typing.Optional[str]
|
{sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/raw_client.py
RENAMED
|
@@ -10,7 +10,11 @@ from ..core.api_error import ApiError
|
|
|
10
10
|
from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
|
11
11
|
from ..core.request_options import RequestOptions
|
|
12
12
|
from .socket_client import AsyncSpeechToTextTranslateStreamingSocketClient, SpeechToTextTranslateStreamingSocketClient
|
|
13
|
+
from .types.speech_to_text_translate_streaming_high_vad_sensitivity import (
|
|
14
|
+
SpeechToTextTranslateStreamingHighVadSensitivity,
|
|
15
|
+
)
|
|
13
16
|
from .types.speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
|
|
17
|
+
from .types.speech_to_text_translate_streaming_vad_signals import SpeechToTextTranslateStreamingVadSignals
|
|
14
18
|
|
|
15
19
|
|
|
16
20
|
class RawSpeechToTextTranslateStreamingClient:
|
|
@@ -22,8 +26,8 @@ class RawSpeechToTextTranslateStreamingClient:
|
|
|
22
26
|
self,
|
|
23
27
|
*,
|
|
24
28
|
model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
|
|
25
|
-
high_vad_sensitivity: typing.Optional[
|
|
26
|
-
vad_signals: typing.Optional[
|
|
29
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
|
|
30
|
+
vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
|
|
27
31
|
api_subscription_key: typing.Optional[str] = None,
|
|
28
32
|
request_options: typing.Optional[RequestOptions] = None,
|
|
29
33
|
) -> typing.Iterator[SpeechToTextTranslateStreamingSocketClient]:
|
|
@@ -35,10 +39,10 @@ class RawSpeechToTextTranslateStreamingClient:
|
|
|
35
39
|
model : typing.Optional[SpeechToTextTranslateStreamingModel]
|
|
36
40
|
Speech to text model to use (defaults to "saaras:v2.5" if not specified)
|
|
37
41
|
|
|
38
|
-
high_vad_sensitivity : typing.Optional[
|
|
42
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
|
|
39
43
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
40
44
|
|
|
41
|
-
vad_signals : typing.Optional[
|
|
45
|
+
vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
|
|
42
46
|
Enable VAD signals in response
|
|
43
47
|
|
|
44
48
|
api_subscription_key : typing.Optional[str]
|
|
@@ -92,8 +96,8 @@ class AsyncRawSpeechToTextTranslateStreamingClient:
|
|
|
92
96
|
self,
|
|
93
97
|
*,
|
|
94
98
|
model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
|
|
95
|
-
high_vad_sensitivity: typing.Optional[
|
|
96
|
-
vad_signals: typing.Optional[
|
|
99
|
+
high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
|
|
100
|
+
vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
|
|
97
101
|
api_subscription_key: typing.Optional[str] = None,
|
|
98
102
|
request_options: typing.Optional[RequestOptions] = None,
|
|
99
103
|
) -> typing.AsyncIterator[AsyncSpeechToTextTranslateStreamingSocketClient]:
|
|
@@ -105,10 +109,10 @@ class AsyncRawSpeechToTextTranslateStreamingClient:
|
|
|
105
109
|
model : typing.Optional[SpeechToTextTranslateStreamingModel]
|
|
106
110
|
Speech to text model to use (defaults to "saaras:v2.5" if not specified)
|
|
107
111
|
|
|
108
|
-
high_vad_sensitivity : typing.Optional[
|
|
112
|
+
high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
|
|
109
113
|
Enable high VAD (Voice Activity Detection) sensitivity
|
|
110
114
|
|
|
111
|
-
vad_signals : typing.Optional[
|
|
115
|
+
vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
|
|
112
116
|
Enable VAD signals in response
|
|
113
117
|
|
|
114
118
|
api_subscription_key : typing.Optional[str]
|
|
@@ -0,0 +1,13 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
# isort: skip_file
|
|
4
|
+
|
|
5
|
+
from .speech_to_text_translate_streaming_high_vad_sensitivity import SpeechToTextTranslateStreamingHighVadSensitivity
|
|
6
|
+
from .speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
|
|
7
|
+
from .speech_to_text_translate_streaming_vad_signals import SpeechToTextTranslateStreamingVadSignals
|
|
8
|
+
|
|
9
|
+
__all__ = [
|
|
10
|
+
"SpeechToTextTranslateStreamingHighVadSensitivity",
|
|
11
|
+
"SpeechToTextTranslateStreamingModel",
|
|
12
|
+
"SpeechToTextTranslateStreamingVadSignals",
|
|
13
|
+
]
|
|
@@ -58,13 +58,9 @@ class TextToSpeechClient:
|
|
|
58
58
|
speaker : typing.Optional[TextToSpeechSpeaker]
|
|
59
59
|
The speaker voice to be used for the output audio.
|
|
60
60
|
|
|
61
|
-
**Default:**
|
|
62
|
-
|
|
63
|
-
**Model Compatibility (Speakers compatible with respective models):**
|
|
64
|
-
- **bulbul:v1:**
|
|
65
|
-
- Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
|
|
66
|
-
- Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
|
|
61
|
+
**Default:** Anushka
|
|
67
62
|
|
|
63
|
+
**Model Compatibility (Speakers compatible with respective model):**
|
|
68
64
|
- **bulbul:v2:**
|
|
69
65
|
- Female: Anushka, Manisha, Vidya, Arya
|
|
70
66
|
- Male: Abhilash, Karun, Hitesh
|
|
@@ -87,7 +83,7 @@ class TextToSpeechClient:
|
|
|
87
83
|
Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
|
|
88
84
|
|
|
89
85
|
model : typing.Optional[TextToSpeechModel]
|
|
90
|
-
Specifies the model to use for text-to-speech conversion. Default is bulbul:
|
|
86
|
+
Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
|
|
91
87
|
|
|
92
88
|
request_options : typing.Optional[RequestOptions]
|
|
93
89
|
Request-specific configuration.
|
|
@@ -167,13 +163,9 @@ class AsyncTextToSpeechClient:
|
|
|
167
163
|
speaker : typing.Optional[TextToSpeechSpeaker]
|
|
168
164
|
The speaker voice to be used for the output audio.
|
|
169
165
|
|
|
170
|
-
**Default:**
|
|
171
|
-
|
|
172
|
-
**Model Compatibility (Speakers compatible with respective models):**
|
|
173
|
-
- **bulbul:v1:**
|
|
174
|
-
- Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
|
|
175
|
-
- Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
|
|
166
|
+
**Default:** Anushka
|
|
176
167
|
|
|
168
|
+
**Model Compatibility (Speakers compatible with respective model):**
|
|
177
169
|
- **bulbul:v2:**
|
|
178
170
|
- Female: Anushka, Manisha, Vidya, Arya
|
|
179
171
|
- Male: Abhilash, Karun, Hitesh
|
|
@@ -196,7 +188,7 @@ class AsyncTextToSpeechClient:
|
|
|
196
188
|
Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
|
|
197
189
|
|
|
198
190
|
model : typing.Optional[TextToSpeechModel]
|
|
199
|
-
Specifies the model to use for text-to-speech conversion. Default is bulbul:
|
|
191
|
+
Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
|
|
200
192
|
|
|
201
193
|
request_options : typing.Optional[RequestOptions]
|
|
202
194
|
Request-specific configuration.
|
|
@@ -55,13 +55,9 @@ class RawTextToSpeechClient:
|
|
|
55
55
|
speaker : typing.Optional[TextToSpeechSpeaker]
|
|
56
56
|
The speaker voice to be used for the output audio.
|
|
57
57
|
|
|
58
|
-
**Default:**
|
|
59
|
-
|
|
60
|
-
**Model Compatibility (Speakers compatible with respective models):**
|
|
61
|
-
- **bulbul:v1:**
|
|
62
|
-
- Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
|
|
63
|
-
- Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
|
|
58
|
+
**Default:** Anushka
|
|
64
59
|
|
|
60
|
+
**Model Compatibility (Speakers compatible with respective model):**
|
|
65
61
|
- **bulbul:v2:**
|
|
66
62
|
- Female: Anushka, Manisha, Vidya, Arya
|
|
67
63
|
- Male: Abhilash, Karun, Hitesh
|
|
@@ -84,7 +80,7 @@ class RawTextToSpeechClient:
|
|
|
84
80
|
Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
|
|
85
81
|
|
|
86
82
|
model : typing.Optional[TextToSpeechModel]
|
|
87
|
-
Specifies the model to use for text-to-speech conversion. Default is bulbul:
|
|
83
|
+
Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
|
|
88
84
|
|
|
89
85
|
request_options : typing.Optional[RequestOptions]
|
|
90
86
|
Request-specific configuration.
|
|
@@ -218,13 +214,9 @@ class AsyncRawTextToSpeechClient:
|
|
|
218
214
|
speaker : typing.Optional[TextToSpeechSpeaker]
|
|
219
215
|
The speaker voice to be used for the output audio.
|
|
220
216
|
|
|
221
|
-
**Default:**
|
|
222
|
-
|
|
223
|
-
**Model Compatibility (Speakers compatible with respective models):**
|
|
224
|
-
- **bulbul:v1:**
|
|
225
|
-
- Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
|
|
226
|
-
- Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
|
|
217
|
+
**Default:** Anushka
|
|
227
218
|
|
|
219
|
+
**Model Compatibility (Speakers compatible with respective model):**
|
|
228
220
|
- **bulbul:v2:**
|
|
229
221
|
- Female: Anushka, Manisha, Vidya, Arya
|
|
230
222
|
- Male: Abhilash, Karun, Hitesh
|
|
@@ -247,7 +239,7 @@ class AsyncRawTextToSpeechClient:
|
|
|
247
239
|
Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
|
|
248
240
|
|
|
249
241
|
model : typing.Optional[TextToSpeechModel]
|
|
250
|
-
Specifies the model to use for text-to-speech conversion. Default is bulbul:
|
|
242
|
+
Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
|
|
251
243
|
|
|
252
244
|
request_options : typing.Optional[RequestOptions]
|
|
253
245
|
Request-specific configuration.
|
|
@@ -3,7 +3,6 @@
|
|
|
3
3
|
# isort: skip_file
|
|
4
4
|
|
|
5
5
|
from .audio_data import AudioData
|
|
6
|
-
from .audio_data_encoding import AudioDataEncoding
|
|
7
6
|
from .audio_message import AudioMessage
|
|
8
7
|
from .chat_completion_request_assistant_message import ChatCompletionRequestAssistantMessage
|
|
9
8
|
from .chat_completion_request_message import (
|
|
@@ -68,7 +67,6 @@ from .transliteration_response import TransliterationResponse
|
|
|
68
67
|
|
|
69
68
|
__all__ = [
|
|
70
69
|
"AudioData",
|
|
71
|
-
"AudioDataEncoding",
|
|
72
70
|
"AudioMessage",
|
|
73
71
|
"ChatCompletionRequestAssistantMessage",
|
|
74
72
|
"ChatCompletionRequestMessage",
|
|
@@ -4,7 +4,6 @@ import typing
|
|
|
4
4
|
|
|
5
5
|
import pydantic
|
|
6
6
|
from ..core.pydantic_utilities import IS_PYDANTIC_V2, UniversalBaseModel
|
|
7
|
-
from .audio_data_encoding import AudioDataEncoding
|
|
8
7
|
|
|
9
8
|
|
|
10
9
|
class AudioData(UniversalBaseModel):
|
|
@@ -18,7 +17,7 @@ class AudioData(UniversalBaseModel):
|
|
|
18
17
|
Audio sample rate in Hz (16kHz preferred, 8kHz least preferred)
|
|
19
18
|
"""
|
|
20
19
|
|
|
21
|
-
encoding:
|
|
20
|
+
encoding: typing.Literal["audio/wav"] = pydantic.Field(default="audio/wav")
|
|
22
21
|
"""
|
|
23
22
|
Audio encoding format
|
|
24
23
|
"""
|
|
@@ -1,8 +0,0 @@
|
|
|
1
|
-
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
-
|
|
3
|
-
# isort: skip_file
|
|
4
|
-
|
|
5
|
-
from .speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
|
|
6
|
-
from .speech_to_text_streaming_model import SpeechToTextStreamingModel
|
|
7
|
-
|
|
8
|
-
__all__ = ["SpeechToTextStreamingLanguageCode", "SpeechToTextStreamingModel"]
|