sarvamai 0.1.5a13__tar.gz → 0.1.6__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (158) hide show
  1. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/PKG-INFO +1 -1
  2. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/pyproject.toml +1 -1
  3. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/__init__.py +15 -4
  4. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/client_wrapper.py +2 -2
  5. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/audio_data.py +3 -2
  6. sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/__init__.py +17 -0
  7. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/client.py +10 -8
  8. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/raw_client.py +10 -8
  9. sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/types/__init__.py +15 -0
  10. sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_high_vad_sensitivity.py +5 -0
  11. sarvamai-0.1.6/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_vad_signals.py +5 -0
  12. sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/__init__.py +15 -0
  13. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/client.py +12 -8
  14. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/raw_client.py +12 -8
  15. sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/types/__init__.py +13 -0
  16. sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/types/speech_to_text_translate_streaming_high_vad_sensitivity.py +5 -0
  17. sarvamai-0.1.6/src/sarvamai/speech_to_text_translate_streaming/types/speech_to_text_translate_streaming_vad_signals.py +5 -0
  18. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text_to_speech/client.py +6 -14
  19. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text_to_speech/raw_client.py +6 -14
  20. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/__init__.py +0 -2
  21. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/audio_data.py +1 -2
  22. sarvamai-0.1.6/src/sarvamai/types/text_to_speech_model.py +5 -0
  23. sarvamai-0.1.6/src/sarvamai/types/text_to_speech_speaker.py +7 -0
  24. sarvamai-0.1.5a13/src/sarvamai/speech_to_text_streaming/__init__.py +0 -7
  25. sarvamai-0.1.5a13/src/sarvamai/speech_to_text_streaming/types/__init__.py +0 -8
  26. sarvamai-0.1.5a13/src/sarvamai/speech_to_text_translate_streaming/__init__.py +0 -7
  27. sarvamai-0.1.5a13/src/sarvamai/speech_to_text_translate_streaming/types/__init__.py +0 -7
  28. sarvamai-0.1.5a13/src/sarvamai/types/audio_data_encoding.py +0 -5
  29. sarvamai-0.1.5a13/src/sarvamai/types/text_to_speech_model.py +0 -5
  30. sarvamai-0.1.5a13/src/sarvamai/types/text_to_speech_speaker.py +0 -28
  31. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/README.md +0 -0
  32. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/chat/__init__.py +0 -0
  33. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/chat/client.py +0 -0
  34. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/chat/raw_client.py +0 -0
  35. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/client.py +0 -0
  36. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/__init__.py +0 -0
  37. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/api_error.py +0 -0
  38. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/datetime_utils.py +0 -0
  39. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/events.py +0 -0
  40. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/file.py +0 -0
  41. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/force_multipart.py +0 -0
  42. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/http_client.py +0 -0
  43. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/http_response.py +0 -0
  44. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/jsonable_encoder.py +0 -0
  45. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/pydantic_utilities.py +0 -0
  46. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/query_encoder.py +0 -0
  47. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/remove_none_from_dict.py +0 -0
  48. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/request_options.py +0 -0
  49. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/core/serialization.py +0 -0
  50. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/environment.py +0 -0
  51. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/__init__.py +0 -0
  52. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/bad_request_error.py +0 -0
  53. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/forbidden_error.py +0 -0
  54. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/internal_server_error.py +0 -0
  55. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/service_unavailable_error.py +0 -0
  56. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/too_many_requests_error.py +0 -0
  57. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/errors/unprocessable_entity_error.py +0 -0
  58. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/play.py +0 -0
  59. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/py.typed +0 -0
  60. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/__init__.py +0 -0
  61. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/audio_message.py +0 -0
  62. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_assistant_message.py +0 -0
  63. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_message.py +0 -0
  64. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_system_message.py +0 -0
  65. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_request_user_message.py +0 -0
  66. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/chat_completion_response_message.py +0 -0
  67. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/choice.py +0 -0
  68. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/completion_usage.py +0 -0
  69. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/config_message.py +0 -0
  70. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/create_chat_completion_response.py +0 -0
  71. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/diarized_entry.py +0 -0
  72. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/diarized_transcript.py +0 -0
  73. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/error_data.py +0 -0
  74. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/error_details.py +0 -0
  75. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/error_message.py +0 -0
  76. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/events_data.py +0 -0
  77. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/language_identification_response.py +0 -0
  78. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_response.py +0 -0
  79. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_response_data.py +0 -0
  80. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_streaming_response.py +0 -0
  81. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_transcription_data.py +0 -0
  82. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_response.py +0 -0
  83. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_response_data.py +0 -0
  84. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_streaming_response.py +0 -0
  85. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/speech_to_text_translate_transcription_data.py +0 -0
  86. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/stop_configuration.py +0 -0
  87. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/text_to_speech_response.py +0 -0
  88. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/timestamps_model.py +0 -0
  89. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/transcription_metrics.py +0 -0
  90. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/translation_response.py +0 -0
  91. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/requests/transliteration_response.py +0 -0
  92. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text/__init__.py +0 -0
  93. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text/client.py +0 -0
  94. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text/raw_client.py +0 -0
  95. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/socket_client.py +0 -0
  96. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_language_code.py +0 -0
  97. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_streaming/types/speech_to_text_streaming_model.py +0 -0
  98. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/socket_client.py +0 -0
  99. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/speech_to_text_translate_streaming/types/speech_to_text_translate_streaming_model.py +0 -0
  100. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text/__init__.py +0 -0
  101. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text/client.py +0 -0
  102. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text/raw_client.py +0 -0
  103. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/text_to_speech/__init__.py +0 -0
  104. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/audio_message.py +0 -0
  105. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_assistant_message.py +0 -0
  106. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_message.py +0 -0
  107. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_system_message.py +0 -0
  108. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_request_user_message.py +0 -0
  109. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/chat_completion_response_message.py +0 -0
  110. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/choice.py +0 -0
  111. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/completion_usage.py +0 -0
  112. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/config_message.py +0 -0
  113. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/create_chat_completion_response.py +0 -0
  114. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/diarized_entry.py +0 -0
  115. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/diarized_transcript.py +0 -0
  116. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_code.py +0 -0
  117. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_data.py +0 -0
  118. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_details.py +0 -0
  119. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/error_message.py +0 -0
  120. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/events_data.py +0 -0
  121. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/finish_reason.py +0 -0
  122. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/format.py +0 -0
  123. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/language_identification_response.py +0 -0
  124. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/numerals_format.py +0 -0
  125. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/reasoning_effort.py +0 -0
  126. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/response_type.py +0 -0
  127. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/role.py +0 -0
  128. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/sarvam_model_ids.py +0 -0
  129. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_sample_rate.py +0 -0
  130. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_language.py +0 -0
  131. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_model.py +0 -0
  132. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_response.py +0 -0
  133. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_response_data.py +0 -0
  134. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_streaming_response.py +0 -0
  135. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_transcription_data.py +0 -0
  136. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_language.py +0 -0
  137. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_model.py +0 -0
  138. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_response.py +0 -0
  139. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_response_data.py +0 -0
  140. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_streaming_response.py +0 -0
  141. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/speech_to_text_translate_transcription_data.py +0 -0
  142. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/spoken_form_numerals_format.py +0 -0
  143. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/stop_configuration.py +0 -0
  144. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/text_to_speech_language.py +0 -0
  145. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/text_to_speech_response.py +0 -0
  146. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/timestamps_model.py +0 -0
  147. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transcription_metrics.py +0 -0
  148. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_mode.py +0 -0
  149. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_model.py +0 -0
  150. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_source_language.py +0 -0
  151. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_speaker_gender.py +0 -0
  152. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translate_target_language.py +0 -0
  153. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translation_response.py +0 -0
  154. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/translatiterate_target_language.py +0 -0
  155. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transliterate_mode.py +0 -0
  156. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transliterate_source_language.py +0 -0
  157. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/types/transliteration_response.py +0 -0
  158. {sarvamai-0.1.5a13 → sarvamai-0.1.6}/src/sarvamai/version.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: sarvamai
3
- Version: 0.1.5a13
3
+ Version: 0.1.6
4
4
  Summary:
5
5
  Requires-Python: >=3.8,<4.0
6
6
  Classifier: Intended Audience :: Developers
@@ -3,7 +3,7 @@ name = "sarvamai"
3
3
 
4
4
  [tool.poetry]
5
5
  name = "sarvamai"
6
- version = "0.1.5a13"
6
+ version = "0.1.6"
7
7
  description = ""
8
8
  readme = "README.md"
9
9
  authors = []
@@ -4,7 +4,6 @@
4
4
 
5
5
  from .types import (
6
6
  AudioData,
7
- AudioDataEncoding,
8
7
  AudioMessage,
9
8
  ChatCompletionRequestAssistantMessage,
10
9
  ChatCompletionRequestMessage,
@@ -113,14 +112,22 @@ from .requests import (
113
112
  TranslationResponseParams,
114
113
  TransliterationResponseParams,
115
114
  )
116
- from .speech_to_text_streaming import SpeechToTextStreamingLanguageCode, SpeechToTextStreamingModel
117
- from .speech_to_text_translate_streaming import SpeechToTextTranslateStreamingModel
115
+ from .speech_to_text_streaming import (
116
+ SpeechToTextStreamingHighVadSensitivity,
117
+ SpeechToTextStreamingLanguageCode,
118
+ SpeechToTextStreamingModel,
119
+ SpeechToTextStreamingVadSignals,
120
+ )
121
+ from .speech_to_text_translate_streaming import (
122
+ SpeechToTextTranslateStreamingHighVadSensitivity,
123
+ SpeechToTextTranslateStreamingModel,
124
+ SpeechToTextTranslateStreamingVadSignals,
125
+ )
118
126
  from .version import __version__
119
127
 
120
128
  __all__ = [
121
129
  "AsyncSarvamAI",
122
130
  "AudioData",
123
- "AudioDataEncoding",
124
131
  "AudioDataParams",
125
132
  "AudioMessage",
126
133
  "AudioMessageParams",
@@ -183,10 +190,12 @@ __all__ = [
183
190
  "SpeechToTextResponseData",
184
191
  "SpeechToTextResponseDataParams",
185
192
  "SpeechToTextResponseParams",
193
+ "SpeechToTextStreamingHighVadSensitivity",
186
194
  "SpeechToTextStreamingLanguageCode",
187
195
  "SpeechToTextStreamingModel",
188
196
  "SpeechToTextStreamingResponse",
189
197
  "SpeechToTextStreamingResponseParams",
198
+ "SpeechToTextStreamingVadSignals",
190
199
  "SpeechToTextTranscriptionData",
191
200
  "SpeechToTextTranscriptionDataParams",
192
201
  "SpeechToTextTranslateLanguage",
@@ -195,9 +204,11 @@ __all__ = [
195
204
  "SpeechToTextTranslateResponseData",
196
205
  "SpeechToTextTranslateResponseDataParams",
197
206
  "SpeechToTextTranslateResponseParams",
207
+ "SpeechToTextTranslateStreamingHighVadSensitivity",
198
208
  "SpeechToTextTranslateStreamingModel",
199
209
  "SpeechToTextTranslateStreamingResponse",
200
210
  "SpeechToTextTranslateStreamingResponseParams",
211
+ "SpeechToTextTranslateStreamingVadSignals",
201
212
  "SpeechToTextTranslateTranscriptionData",
202
213
  "SpeechToTextTranslateTranscriptionDataParams",
203
214
  "SpokenFormNumeralsFormat",
@@ -17,10 +17,10 @@ class BaseClientWrapper:
17
17
 
18
18
  def get_headers(self) -> typing.Dict[str, str]:
19
19
  headers: typing.Dict[str, str] = {
20
- "User-Agent": "sarvamai/0.1.5a13",
20
+ "User-Agent": "sarvamai/0.1.6",
21
21
  "X-Fern-Language": "Python",
22
22
  "X-Fern-SDK-Name": "sarvamai",
23
- "X-Fern-SDK-Version": "0.1.5a13",
23
+ "X-Fern-SDK-Version": "0.1.6",
24
24
  }
25
25
  headers["api-subscription-key"] = self.api_subscription_key
26
26
  return headers
@@ -1,7 +1,8 @@
1
1
  # This file was auto-generated by Fern from our API Definition.
2
2
 
3
+ import typing
4
+
3
5
  import typing_extensions
4
- from ..types.audio_data_encoding import AudioDataEncoding
5
6
 
6
7
 
7
8
  class AudioDataParams(typing_extensions.TypedDict):
@@ -15,7 +16,7 @@ class AudioDataParams(typing_extensions.TypedDict):
15
16
  Audio sample rate in Hz (16kHz preferred, 8kHz least preferred)
16
17
  """
17
18
 
18
- encoding: AudioDataEncoding
19
+ encoding: typing.Literal["audio/wav"]
19
20
  """
20
21
  Audio encoding format
21
22
  """
@@ -0,0 +1,17 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ # isort: skip_file
4
+
5
+ from .types import (
6
+ SpeechToTextStreamingHighVadSensitivity,
7
+ SpeechToTextStreamingLanguageCode,
8
+ SpeechToTextStreamingModel,
9
+ SpeechToTextStreamingVadSignals,
10
+ )
11
+
12
+ __all__ = [
13
+ "SpeechToTextStreamingHighVadSensitivity",
14
+ "SpeechToTextStreamingLanguageCode",
15
+ "SpeechToTextStreamingModel",
16
+ "SpeechToTextStreamingVadSignals",
17
+ ]
@@ -11,8 +11,10 @@ from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
11
11
  from ..core.request_options import RequestOptions
12
12
  from .raw_client import AsyncRawSpeechToTextStreamingClient, RawSpeechToTextStreamingClient
13
13
  from .socket_client import AsyncSpeechToTextStreamingSocketClient, SpeechToTextStreamingSocketClient
14
+ from .types.speech_to_text_streaming_high_vad_sensitivity import SpeechToTextStreamingHighVadSensitivity
14
15
  from .types.speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
15
16
  from .types.speech_to_text_streaming_model import SpeechToTextStreamingModel
17
+ from .types.speech_to_text_streaming_vad_signals import SpeechToTextStreamingVadSignals
16
18
 
17
19
 
18
20
  class SpeechToTextStreamingClient:
@@ -36,8 +38,8 @@ class SpeechToTextStreamingClient:
36
38
  *,
37
39
  language_code: SpeechToTextStreamingLanguageCode,
38
40
  model: typing.Optional[SpeechToTextStreamingModel] = None,
39
- high_vad_sensitivity: typing.Optional[str] = None,
40
- vad_signals: typing.Optional[str] = None,
41
+ high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
42
+ vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
41
43
  api_subscription_key: typing.Optional[str] = None,
42
44
  request_options: typing.Optional[RequestOptions] = None,
43
45
  ) -> typing.Iterator[SpeechToTextStreamingSocketClient]:
@@ -52,10 +54,10 @@ class SpeechToTextStreamingClient:
52
54
  model : typing.Optional[SpeechToTextStreamingModel]
53
55
  Speech to text model to use
54
56
 
55
- high_vad_sensitivity : typing.Optional[str]
57
+ high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
56
58
  Enable high VAD (Voice Activity Detection) sensitivity
57
59
 
58
- vad_signals : typing.Optional[str]
60
+ vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
59
61
  Enable VAD signals in response
60
62
 
61
63
  api_subscription_key : typing.Optional[str]
@@ -123,8 +125,8 @@ class AsyncSpeechToTextStreamingClient:
123
125
  *,
124
126
  language_code: SpeechToTextStreamingLanguageCode,
125
127
  model: typing.Optional[SpeechToTextStreamingModel] = None,
126
- high_vad_sensitivity: typing.Optional[str] = None,
127
- vad_signals: typing.Optional[str] = None,
128
+ high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
129
+ vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
128
130
  api_subscription_key: typing.Optional[str] = None,
129
131
  request_options: typing.Optional[RequestOptions] = None,
130
132
  ) -> typing.AsyncIterator[AsyncSpeechToTextStreamingSocketClient]:
@@ -139,10 +141,10 @@ class AsyncSpeechToTextStreamingClient:
139
141
  model : typing.Optional[SpeechToTextStreamingModel]
140
142
  Speech to text model to use
141
143
 
142
- high_vad_sensitivity : typing.Optional[str]
144
+ high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
143
145
  Enable high VAD (Voice Activity Detection) sensitivity
144
146
 
145
- vad_signals : typing.Optional[str]
147
+ vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
146
148
  Enable VAD signals in response
147
149
 
148
150
  api_subscription_key : typing.Optional[str]
@@ -10,8 +10,10 @@ from ..core.api_error import ApiError
10
10
  from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
11
11
  from ..core.request_options import RequestOptions
12
12
  from .socket_client import AsyncSpeechToTextStreamingSocketClient, SpeechToTextStreamingSocketClient
13
+ from .types.speech_to_text_streaming_high_vad_sensitivity import SpeechToTextStreamingHighVadSensitivity
13
14
  from .types.speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
14
15
  from .types.speech_to_text_streaming_model import SpeechToTextStreamingModel
16
+ from .types.speech_to_text_streaming_vad_signals import SpeechToTextStreamingVadSignals
15
17
 
16
18
 
17
19
  class RawSpeechToTextStreamingClient:
@@ -24,8 +26,8 @@ class RawSpeechToTextStreamingClient:
24
26
  *,
25
27
  language_code: SpeechToTextStreamingLanguageCode,
26
28
  model: typing.Optional[SpeechToTextStreamingModel] = None,
27
- high_vad_sensitivity: typing.Optional[str] = None,
28
- vad_signals: typing.Optional[str] = None,
29
+ high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
30
+ vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
29
31
  api_subscription_key: typing.Optional[str] = None,
30
32
  request_options: typing.Optional[RequestOptions] = None,
31
33
  ) -> typing.Iterator[SpeechToTextStreamingSocketClient]:
@@ -40,10 +42,10 @@ class RawSpeechToTextStreamingClient:
40
42
  model : typing.Optional[SpeechToTextStreamingModel]
41
43
  Speech to text model to use
42
44
 
43
- high_vad_sensitivity : typing.Optional[str]
45
+ high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
44
46
  Enable high VAD (Voice Activity Detection) sensitivity
45
47
 
46
- vad_signals : typing.Optional[str]
48
+ vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
47
49
  Enable VAD signals in response
48
50
 
49
51
  api_subscription_key : typing.Optional[str]
@@ -100,8 +102,8 @@ class AsyncRawSpeechToTextStreamingClient:
100
102
  *,
101
103
  language_code: SpeechToTextStreamingLanguageCode,
102
104
  model: typing.Optional[SpeechToTextStreamingModel] = None,
103
- high_vad_sensitivity: typing.Optional[str] = None,
104
- vad_signals: typing.Optional[str] = None,
105
+ high_vad_sensitivity: typing.Optional[SpeechToTextStreamingHighVadSensitivity] = None,
106
+ vad_signals: typing.Optional[SpeechToTextStreamingVadSignals] = None,
105
107
  api_subscription_key: typing.Optional[str] = None,
106
108
  request_options: typing.Optional[RequestOptions] = None,
107
109
  ) -> typing.AsyncIterator[AsyncSpeechToTextStreamingSocketClient]:
@@ -116,10 +118,10 @@ class AsyncRawSpeechToTextStreamingClient:
116
118
  model : typing.Optional[SpeechToTextStreamingModel]
117
119
  Speech to text model to use
118
120
 
119
- high_vad_sensitivity : typing.Optional[str]
121
+ high_vad_sensitivity : typing.Optional[SpeechToTextStreamingHighVadSensitivity]
120
122
  Enable high VAD (Voice Activity Detection) sensitivity
121
123
 
122
- vad_signals : typing.Optional[str]
124
+ vad_signals : typing.Optional[SpeechToTextStreamingVadSignals]
123
125
  Enable VAD signals in response
124
126
 
125
127
  api_subscription_key : typing.Optional[str]
@@ -0,0 +1,15 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ # isort: skip_file
4
+
5
+ from .speech_to_text_streaming_high_vad_sensitivity import SpeechToTextStreamingHighVadSensitivity
6
+ from .speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
7
+ from .speech_to_text_streaming_model import SpeechToTextStreamingModel
8
+ from .speech_to_text_streaming_vad_signals import SpeechToTextStreamingVadSignals
9
+
10
+ __all__ = [
11
+ "SpeechToTextStreamingHighVadSensitivity",
12
+ "SpeechToTextStreamingLanguageCode",
13
+ "SpeechToTextStreamingModel",
14
+ "SpeechToTextStreamingVadSignals",
15
+ ]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ SpeechToTextStreamingHighVadSensitivity = typing.Union[typing.Literal["true", "false"], typing.Any]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ SpeechToTextStreamingVadSignals = typing.Union[typing.Literal["true", "false"], typing.Any]
@@ -0,0 +1,15 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ # isort: skip_file
4
+
5
+ from .types import (
6
+ SpeechToTextTranslateStreamingHighVadSensitivity,
7
+ SpeechToTextTranslateStreamingModel,
8
+ SpeechToTextTranslateStreamingVadSignals,
9
+ )
10
+
11
+ __all__ = [
12
+ "SpeechToTextTranslateStreamingHighVadSensitivity",
13
+ "SpeechToTextTranslateStreamingModel",
14
+ "SpeechToTextTranslateStreamingVadSignals",
15
+ ]
@@ -11,7 +11,11 @@ from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
11
11
  from ..core.request_options import RequestOptions
12
12
  from .raw_client import AsyncRawSpeechToTextTranslateStreamingClient, RawSpeechToTextTranslateStreamingClient
13
13
  from .socket_client import AsyncSpeechToTextTranslateStreamingSocketClient, SpeechToTextTranslateStreamingSocketClient
14
+ from .types.speech_to_text_translate_streaming_high_vad_sensitivity import (
15
+ SpeechToTextTranslateStreamingHighVadSensitivity,
16
+ )
14
17
  from .types.speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
18
+ from .types.speech_to_text_translate_streaming_vad_signals import SpeechToTextTranslateStreamingVadSignals
15
19
 
16
20
 
17
21
  class SpeechToTextTranslateStreamingClient:
@@ -34,8 +38,8 @@ class SpeechToTextTranslateStreamingClient:
34
38
  self,
35
39
  *,
36
40
  model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
37
- high_vad_sensitivity: typing.Optional[str] = None,
38
- vad_signals: typing.Optional[str] = None,
41
+ high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
42
+ vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
39
43
  api_subscription_key: typing.Optional[str] = None,
40
44
  request_options: typing.Optional[RequestOptions] = None,
41
45
  ) -> typing.Iterator[SpeechToTextTranslateStreamingSocketClient]:
@@ -47,10 +51,10 @@ class SpeechToTextTranslateStreamingClient:
47
51
  model : typing.Optional[SpeechToTextTranslateStreamingModel]
48
52
  Speech to text model to use (defaults to "saaras:v2.5" if not specified)
49
53
 
50
- high_vad_sensitivity : typing.Optional[str]
54
+ high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
51
55
  Enable high VAD (Voice Activity Detection) sensitivity
52
56
 
53
- vad_signals : typing.Optional[str]
57
+ vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
54
58
  Enable VAD signals in response
55
59
 
56
60
  api_subscription_key : typing.Optional[str]
@@ -115,8 +119,8 @@ class AsyncSpeechToTextTranslateStreamingClient:
115
119
  self,
116
120
  *,
117
121
  model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
118
- high_vad_sensitivity: typing.Optional[str] = None,
119
- vad_signals: typing.Optional[str] = None,
122
+ high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
123
+ vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
120
124
  api_subscription_key: typing.Optional[str] = None,
121
125
  request_options: typing.Optional[RequestOptions] = None,
122
126
  ) -> typing.AsyncIterator[AsyncSpeechToTextTranslateStreamingSocketClient]:
@@ -128,10 +132,10 @@ class AsyncSpeechToTextTranslateStreamingClient:
128
132
  model : typing.Optional[SpeechToTextTranslateStreamingModel]
129
133
  Speech to text model to use (defaults to "saaras:v2.5" if not specified)
130
134
 
131
- high_vad_sensitivity : typing.Optional[str]
135
+ high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
132
136
  Enable high VAD (Voice Activity Detection) sensitivity
133
137
 
134
- vad_signals : typing.Optional[str]
138
+ vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
135
139
  Enable VAD signals in response
136
140
 
137
141
  api_subscription_key : typing.Optional[str]
@@ -10,7 +10,11 @@ from ..core.api_error import ApiError
10
10
  from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
11
11
  from ..core.request_options import RequestOptions
12
12
  from .socket_client import AsyncSpeechToTextTranslateStreamingSocketClient, SpeechToTextTranslateStreamingSocketClient
13
+ from .types.speech_to_text_translate_streaming_high_vad_sensitivity import (
14
+ SpeechToTextTranslateStreamingHighVadSensitivity,
15
+ )
13
16
  from .types.speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
17
+ from .types.speech_to_text_translate_streaming_vad_signals import SpeechToTextTranslateStreamingVadSignals
14
18
 
15
19
 
16
20
  class RawSpeechToTextTranslateStreamingClient:
@@ -22,8 +26,8 @@ class RawSpeechToTextTranslateStreamingClient:
22
26
  self,
23
27
  *,
24
28
  model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
25
- high_vad_sensitivity: typing.Optional[str] = None,
26
- vad_signals: typing.Optional[str] = None,
29
+ high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
30
+ vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
27
31
  api_subscription_key: typing.Optional[str] = None,
28
32
  request_options: typing.Optional[RequestOptions] = None,
29
33
  ) -> typing.Iterator[SpeechToTextTranslateStreamingSocketClient]:
@@ -35,10 +39,10 @@ class RawSpeechToTextTranslateStreamingClient:
35
39
  model : typing.Optional[SpeechToTextTranslateStreamingModel]
36
40
  Speech to text model to use (defaults to "saaras:v2.5" if not specified)
37
41
 
38
- high_vad_sensitivity : typing.Optional[str]
42
+ high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
39
43
  Enable high VAD (Voice Activity Detection) sensitivity
40
44
 
41
- vad_signals : typing.Optional[str]
45
+ vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
42
46
  Enable VAD signals in response
43
47
 
44
48
  api_subscription_key : typing.Optional[str]
@@ -92,8 +96,8 @@ class AsyncRawSpeechToTextTranslateStreamingClient:
92
96
  self,
93
97
  *,
94
98
  model: typing.Optional[SpeechToTextTranslateStreamingModel] = None,
95
- high_vad_sensitivity: typing.Optional[str] = None,
96
- vad_signals: typing.Optional[str] = None,
99
+ high_vad_sensitivity: typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity] = None,
100
+ vad_signals: typing.Optional[SpeechToTextTranslateStreamingVadSignals] = None,
97
101
  api_subscription_key: typing.Optional[str] = None,
98
102
  request_options: typing.Optional[RequestOptions] = None,
99
103
  ) -> typing.AsyncIterator[AsyncSpeechToTextTranslateStreamingSocketClient]:
@@ -105,10 +109,10 @@ class AsyncRawSpeechToTextTranslateStreamingClient:
105
109
  model : typing.Optional[SpeechToTextTranslateStreamingModel]
106
110
  Speech to text model to use (defaults to "saaras:v2.5" if not specified)
107
111
 
108
- high_vad_sensitivity : typing.Optional[str]
112
+ high_vad_sensitivity : typing.Optional[SpeechToTextTranslateStreamingHighVadSensitivity]
109
113
  Enable high VAD (Voice Activity Detection) sensitivity
110
114
 
111
- vad_signals : typing.Optional[str]
115
+ vad_signals : typing.Optional[SpeechToTextTranslateStreamingVadSignals]
112
116
  Enable VAD signals in response
113
117
 
114
118
  api_subscription_key : typing.Optional[str]
@@ -0,0 +1,13 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ # isort: skip_file
4
+
5
+ from .speech_to_text_translate_streaming_high_vad_sensitivity import SpeechToTextTranslateStreamingHighVadSensitivity
6
+ from .speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
7
+ from .speech_to_text_translate_streaming_vad_signals import SpeechToTextTranslateStreamingVadSignals
8
+
9
+ __all__ = [
10
+ "SpeechToTextTranslateStreamingHighVadSensitivity",
11
+ "SpeechToTextTranslateStreamingModel",
12
+ "SpeechToTextTranslateStreamingVadSignals",
13
+ ]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ SpeechToTextTranslateStreamingHighVadSensitivity = typing.Union[typing.Literal["true", "false"], typing.Any]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ SpeechToTextTranslateStreamingVadSignals = typing.Union[typing.Literal["true", "false"], typing.Any]
@@ -58,13 +58,9 @@ class TextToSpeechClient:
58
58
  speaker : typing.Optional[TextToSpeechSpeaker]
59
59
  The speaker voice to be used for the output audio.
60
60
 
61
- **Default:** Meera
62
-
63
- **Model Compatibility (Speakers compatible with respective models):**
64
- - **bulbul:v1:**
65
- - Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
66
- - Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
61
+ **Default:** Anushka
67
62
 
63
+ **Model Compatibility (Speakers compatible with respective model):**
68
64
  - **bulbul:v2:**
69
65
  - Female: Anushka, Manisha, Vidya, Arya
70
66
  - Male: Abhilash, Karun, Hitesh
@@ -87,7 +83,7 @@ class TextToSpeechClient:
87
83
  Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
88
84
 
89
85
  model : typing.Optional[TextToSpeechModel]
90
- Specifies the model to use for text-to-speech conversion. Default is bulbul:v1.
86
+ Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
91
87
 
92
88
  request_options : typing.Optional[RequestOptions]
93
89
  Request-specific configuration.
@@ -167,13 +163,9 @@ class AsyncTextToSpeechClient:
167
163
  speaker : typing.Optional[TextToSpeechSpeaker]
168
164
  The speaker voice to be used for the output audio.
169
165
 
170
- **Default:** Meera
171
-
172
- **Model Compatibility (Speakers compatible with respective models):**
173
- - **bulbul:v1:**
174
- - Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
175
- - Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
166
+ **Default:** Anushka
176
167
 
168
+ **Model Compatibility (Speakers compatible with respective model):**
177
169
  - **bulbul:v2:**
178
170
  - Female: Anushka, Manisha, Vidya, Arya
179
171
  - Male: Abhilash, Karun, Hitesh
@@ -196,7 +188,7 @@ class AsyncTextToSpeechClient:
196
188
  Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
197
189
 
198
190
  model : typing.Optional[TextToSpeechModel]
199
- Specifies the model to use for text-to-speech conversion. Default is bulbul:v1.
191
+ Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
200
192
 
201
193
  request_options : typing.Optional[RequestOptions]
202
194
  Request-specific configuration.
@@ -55,13 +55,9 @@ class RawTextToSpeechClient:
55
55
  speaker : typing.Optional[TextToSpeechSpeaker]
56
56
  The speaker voice to be used for the output audio.
57
57
 
58
- **Default:** Meera
59
-
60
- **Model Compatibility (Speakers compatible with respective models):**
61
- - **bulbul:v1:**
62
- - Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
63
- - Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
58
+ **Default:** Anushka
64
59
 
60
+ **Model Compatibility (Speakers compatible with respective model):**
65
61
  - **bulbul:v2:**
66
62
  - Female: Anushka, Manisha, Vidya, Arya
67
63
  - Male: Abhilash, Karun, Hitesh
@@ -84,7 +80,7 @@ class RawTextToSpeechClient:
84
80
  Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
85
81
 
86
82
  model : typing.Optional[TextToSpeechModel]
87
- Specifies the model to use for text-to-speech conversion. Default is bulbul:v1.
83
+ Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
88
84
 
89
85
  request_options : typing.Optional[RequestOptions]
90
86
  Request-specific configuration.
@@ -218,13 +214,9 @@ class AsyncRawTextToSpeechClient:
218
214
  speaker : typing.Optional[TextToSpeechSpeaker]
219
215
  The speaker voice to be used for the output audio.
220
216
 
221
- **Default:** Meera
222
-
223
- **Model Compatibility (Speakers compatible with respective models):**
224
- - **bulbul:v1:**
225
- - Female: Diya, Maya, Meera, Pavithra, Maitreyi, Misha
226
- - Male: Amol, Arjun, Amartya, Arvind, Neel, Vian
217
+ **Default:** Anushka
227
218
 
219
+ **Model Compatibility (Speakers compatible with respective model):**
228
220
  - **bulbul:v2:**
229
221
  - Female: Anushka, Manisha, Vidya, Arya
230
222
  - Male: Abhilash, Karun, Hitesh
@@ -247,7 +239,7 @@ class AsyncRawTextToSpeechClient:
247
239
  Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
248
240
 
249
241
  model : typing.Optional[TextToSpeechModel]
250
- Specifies the model to use for text-to-speech conversion. Default is bulbul:v1.
242
+ Specifies the model to use for text-to-speech conversion. Default is bulbul:v2.
251
243
 
252
244
  request_options : typing.Optional[RequestOptions]
253
245
  Request-specific configuration.
@@ -3,7 +3,6 @@
3
3
  # isort: skip_file
4
4
 
5
5
  from .audio_data import AudioData
6
- from .audio_data_encoding import AudioDataEncoding
7
6
  from .audio_message import AudioMessage
8
7
  from .chat_completion_request_assistant_message import ChatCompletionRequestAssistantMessage
9
8
  from .chat_completion_request_message import (
@@ -68,7 +67,6 @@ from .transliteration_response import TransliterationResponse
68
67
 
69
68
  __all__ = [
70
69
  "AudioData",
71
- "AudioDataEncoding",
72
70
  "AudioMessage",
73
71
  "ChatCompletionRequestAssistantMessage",
74
72
  "ChatCompletionRequestMessage",
@@ -4,7 +4,6 @@ import typing
4
4
 
5
5
  import pydantic
6
6
  from ..core.pydantic_utilities import IS_PYDANTIC_V2, UniversalBaseModel
7
- from .audio_data_encoding import AudioDataEncoding
8
7
 
9
8
 
10
9
  class AudioData(UniversalBaseModel):
@@ -18,7 +17,7 @@ class AudioData(UniversalBaseModel):
18
17
  Audio sample rate in Hz (16kHz preferred, 8kHz least preferred)
19
18
  """
20
19
 
21
- encoding: AudioDataEncoding = pydantic.Field()
20
+ encoding: typing.Literal["audio/wav"] = pydantic.Field(default="audio/wav")
22
21
  """
23
22
  Audio encoding format
24
23
  """
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ TextToSpeechModel = typing.Literal["bulbul:v2"]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ TextToSpeechSpeaker = typing.Union[
6
+ typing.Literal["anushka", "abhilash", "manisha", "vidya", "arya", "karun", "hitesh"], typing.Any
7
+ ]
@@ -1,7 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- # isort: skip_file
4
-
5
- from .types import SpeechToTextStreamingLanguageCode, SpeechToTextStreamingModel
6
-
7
- __all__ = ["SpeechToTextStreamingLanguageCode", "SpeechToTextStreamingModel"]
@@ -1,8 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- # isort: skip_file
4
-
5
- from .speech_to_text_streaming_language_code import SpeechToTextStreamingLanguageCode
6
- from .speech_to_text_streaming_model import SpeechToTextStreamingModel
7
-
8
- __all__ = ["SpeechToTextStreamingLanguageCode", "SpeechToTextStreamingModel"]
@@ -1,7 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- # isort: skip_file
4
-
5
- from .types import SpeechToTextTranslateStreamingModel
6
-
7
- __all__ = ["SpeechToTextTranslateStreamingModel"]
@@ -1,7 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- # isort: skip_file
4
-
5
- from .speech_to_text_translate_streaming_model import SpeechToTextTranslateStreamingModel
6
-
7
- __all__ = ["SpeechToTextTranslateStreamingModel"]
@@ -1,5 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- import typing
4
-
5
- AudioDataEncoding = typing.Union[typing.Literal["audio/wav", "audio/pcm", "audio/flac", "audio/mp3"], typing.Any]
@@ -1,5 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- import typing
4
-
5
- TextToSpeechModel = typing.Union[typing.Literal["bulbul:v1", "bulbul:v2"], typing.Any]