sarvamai 0.1.17a0__py3-none-any.whl → 0.1.18__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- sarvamai/__init__.py +2 -3
- sarvamai/core/client_wrapper.py +2 -2
- sarvamai/speech_to_text/__init__.py +0 -3
- sarvamai/speech_to_text/client.py +9 -10
- sarvamai/speech_to_text/raw_client.py +9 -10
- sarvamai/types/__init__.py +2 -0
- sarvamai/{speech_to_text/types/speech_to_text_translate_request_input_audio_codec.py → types/input_audio_codec.py} +1 -1
- {sarvamai-0.1.17a0.dist-info → sarvamai-0.1.18.dist-info}/METADATA +1 -1
- {sarvamai-0.1.17a0.dist-info → sarvamai-0.1.18.dist-info}/RECORD +10 -12
- sarvamai/speech_to_text/types/__init__.py +0 -8
- sarvamai/speech_to_text/types/speech_to_text_transcribe_request_input_audio_codec.py +0 -33
- {sarvamai-0.1.17a0.dist-info → sarvamai-0.1.18.dist-info}/WHEEL +0 -0
sarvamai/__init__.py
CHANGED
|
@@ -44,6 +44,7 @@ from .types import (
|
|
|
44
44
|
FilesUploadResponse,
|
|
45
45
|
FinishReason,
|
|
46
46
|
FlushSignal,
|
|
47
|
+
InputAudioCodec,
|
|
47
48
|
JobState,
|
|
48
49
|
JobStatusV1Response,
|
|
49
50
|
LanguageIdentificationResponse,
|
|
@@ -174,7 +175,6 @@ from .requests import (
|
|
|
174
175
|
TranslationResponseParams,
|
|
175
176
|
TransliterationResponseParams,
|
|
176
177
|
)
|
|
177
|
-
from .speech_to_text import SpeechToTextTranscribeRequestInputAudioCodec, SpeechToTextTranslateRequestInputAudioCodec
|
|
178
178
|
from .speech_to_text_streaming import (
|
|
179
179
|
SpeechToTextStreamingHighVadSensitivity,
|
|
180
180
|
SpeechToTextStreamingLanguageCode,
|
|
@@ -267,6 +267,7 @@ __all__ = [
|
|
|
267
267
|
"FlushSignal",
|
|
268
268
|
"FlushSignalParams",
|
|
269
269
|
"ForbiddenError",
|
|
270
|
+
"InputAudioCodec",
|
|
270
271
|
"InternalServerError",
|
|
271
272
|
"JobState",
|
|
272
273
|
"JobStatusV1Response",
|
|
@@ -302,14 +303,12 @@ __all__ = [
|
|
|
302
303
|
"SpeechToTextStreamingResponse",
|
|
303
304
|
"SpeechToTextStreamingResponseParams",
|
|
304
305
|
"SpeechToTextStreamingVadSignals",
|
|
305
|
-
"SpeechToTextTranscribeRequestInputAudioCodec",
|
|
306
306
|
"SpeechToTextTranscriptionData",
|
|
307
307
|
"SpeechToTextTranscriptionDataParams",
|
|
308
308
|
"SpeechToTextTranslateJobParameters",
|
|
309
309
|
"SpeechToTextTranslateJobParametersParams",
|
|
310
310
|
"SpeechToTextTranslateLanguage",
|
|
311
311
|
"SpeechToTextTranslateModel",
|
|
312
|
-
"SpeechToTextTranslateRequestInputAudioCodec",
|
|
313
312
|
"SpeechToTextTranslateResponse",
|
|
314
313
|
"SpeechToTextTranslateResponseData",
|
|
315
314
|
"SpeechToTextTranslateResponseDataParams",
|
sarvamai/core/client_wrapper.py
CHANGED
|
@@ -23,10 +23,10 @@ class BaseClientWrapper:
|
|
|
23
23
|
|
|
24
24
|
def get_headers(self) -> typing.Dict[str, str]:
|
|
25
25
|
headers: typing.Dict[str, str] = {
|
|
26
|
-
"User-Agent": "sarvamai/0.1.
|
|
26
|
+
"User-Agent": "sarvamai/0.1.18",
|
|
27
27
|
"X-Fern-Language": "Python",
|
|
28
28
|
"X-Fern-SDK-Name": "sarvamai",
|
|
29
|
-
"X-Fern-SDK-Version": "0.1.
|
|
29
|
+
"X-Fern-SDK-Version": "0.1.18",
|
|
30
30
|
**(self.get_custom_headers() or {}),
|
|
31
31
|
}
|
|
32
32
|
headers["api-subscription-key"] = self.api_subscription_key
|
|
@@ -5,14 +5,13 @@ import typing
|
|
|
5
5
|
from .. import core
|
|
6
6
|
from ..core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
|
7
7
|
from ..core.request_options import RequestOptions
|
|
8
|
+
from ..types.input_audio_codec import InputAudioCodec
|
|
8
9
|
from ..types.speech_to_text_language import SpeechToTextLanguage
|
|
9
10
|
from ..types.speech_to_text_model import SpeechToTextModel
|
|
10
11
|
from ..types.speech_to_text_response import SpeechToTextResponse
|
|
11
12
|
from ..types.speech_to_text_translate_model import SpeechToTextTranslateModel
|
|
12
13
|
from ..types.speech_to_text_translate_response import SpeechToTextTranslateResponse
|
|
13
14
|
from .raw_client import AsyncRawSpeechToTextClient, RawSpeechToTextClient
|
|
14
|
-
from .types.speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
|
|
15
|
-
from .types.speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
|
|
16
15
|
|
|
17
16
|
# this is used as the default value for optional parameters
|
|
18
17
|
OMIT = typing.cast(typing.Any, ...)
|
|
@@ -39,7 +38,7 @@ class SpeechToTextClient:
|
|
|
39
38
|
file: core.File,
|
|
40
39
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
41
40
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
42
|
-
input_audio_codec: typing.Optional[
|
|
41
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
43
42
|
request_options: typing.Optional[RequestOptions] = None,
|
|
44
43
|
) -> SpeechToTextResponse:
|
|
45
44
|
"""
|
|
@@ -71,7 +70,7 @@ class SpeechToTextClient:
|
|
|
71
70
|
For the `saarika:v2.5` model, it is optional.
|
|
72
71
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
73
72
|
|
|
74
|
-
input_audio_codec : typing.Optional[
|
|
73
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
75
74
|
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
76
75
|
|
|
77
76
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -106,7 +105,7 @@ class SpeechToTextClient:
|
|
|
106
105
|
file: core.File,
|
|
107
106
|
prompt: typing.Optional[str] = OMIT,
|
|
108
107
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
109
|
-
input_audio_codec: typing.Optional[
|
|
108
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
110
109
|
request_options: typing.Optional[RequestOptions] = None,
|
|
111
110
|
) -> SpeechToTextTranslateResponse:
|
|
112
111
|
"""
|
|
@@ -135,7 +134,7 @@ class SpeechToTextClient:
|
|
|
135
134
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
136
135
|
Model to be used for converting speech to text in target language
|
|
137
136
|
|
|
138
|
-
input_audio_codec : typing.Optional[
|
|
137
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
139
138
|
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
140
139
|
|
|
141
140
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -182,7 +181,7 @@ class AsyncSpeechToTextClient:
|
|
|
182
181
|
file: core.File,
|
|
183
182
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
184
183
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
185
|
-
input_audio_codec: typing.Optional[
|
|
184
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
186
185
|
request_options: typing.Optional[RequestOptions] = None,
|
|
187
186
|
) -> SpeechToTextResponse:
|
|
188
187
|
"""
|
|
@@ -214,7 +213,7 @@ class AsyncSpeechToTextClient:
|
|
|
214
213
|
For the `saarika:v2.5` model, it is optional.
|
|
215
214
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
216
215
|
|
|
217
|
-
input_audio_codec : typing.Optional[
|
|
216
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
218
217
|
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
219
218
|
|
|
220
219
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -257,7 +256,7 @@ class AsyncSpeechToTextClient:
|
|
|
257
256
|
file: core.File,
|
|
258
257
|
prompt: typing.Optional[str] = OMIT,
|
|
259
258
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
260
|
-
input_audio_codec: typing.Optional[
|
|
259
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
261
260
|
request_options: typing.Optional[RequestOptions] = None,
|
|
262
261
|
) -> SpeechToTextTranslateResponse:
|
|
263
262
|
"""
|
|
@@ -286,7 +285,7 @@ class AsyncSpeechToTextClient:
|
|
|
286
285
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
287
286
|
Model to be used for converting speech to text in target language
|
|
288
287
|
|
|
289
|
-
input_audio_codec : typing.Optional[
|
|
288
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
290
289
|
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
291
290
|
|
|
292
291
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -15,13 +15,12 @@ from ..errors.internal_server_error import InternalServerError
|
|
|
15
15
|
from ..errors.service_unavailable_error import ServiceUnavailableError
|
|
16
16
|
from ..errors.too_many_requests_error import TooManyRequestsError
|
|
17
17
|
from ..errors.unprocessable_entity_error import UnprocessableEntityError
|
|
18
|
+
from ..types.input_audio_codec import InputAudioCodec
|
|
18
19
|
from ..types.speech_to_text_language import SpeechToTextLanguage
|
|
19
20
|
from ..types.speech_to_text_model import SpeechToTextModel
|
|
20
21
|
from ..types.speech_to_text_response import SpeechToTextResponse
|
|
21
22
|
from ..types.speech_to_text_translate_model import SpeechToTextTranslateModel
|
|
22
23
|
from ..types.speech_to_text_translate_response import SpeechToTextTranslateResponse
|
|
23
|
-
from .types.speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
|
|
24
|
-
from .types.speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
|
|
25
24
|
|
|
26
25
|
# this is used as the default value for optional parameters
|
|
27
26
|
OMIT = typing.cast(typing.Any, ...)
|
|
@@ -37,7 +36,7 @@ class RawSpeechToTextClient:
|
|
|
37
36
|
file: core.File,
|
|
38
37
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
39
38
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
40
|
-
input_audio_codec: typing.Optional[
|
|
39
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
41
40
|
request_options: typing.Optional[RequestOptions] = None,
|
|
42
41
|
) -> HttpResponse[SpeechToTextResponse]:
|
|
43
42
|
"""
|
|
@@ -69,7 +68,7 @@ class RawSpeechToTextClient:
|
|
|
69
68
|
For the `saarika:v2.5` model, it is optional.
|
|
70
69
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
71
70
|
|
|
72
|
-
input_audio_codec : typing.Optional[
|
|
71
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
73
72
|
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
74
73
|
|
|
75
74
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -183,7 +182,7 @@ class RawSpeechToTextClient:
|
|
|
183
182
|
file: core.File,
|
|
184
183
|
prompt: typing.Optional[str] = OMIT,
|
|
185
184
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
186
|
-
input_audio_codec: typing.Optional[
|
|
185
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
187
186
|
request_options: typing.Optional[RequestOptions] = None,
|
|
188
187
|
) -> HttpResponse[SpeechToTextTranslateResponse]:
|
|
189
188
|
"""
|
|
@@ -212,7 +211,7 @@ class RawSpeechToTextClient:
|
|
|
212
211
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
213
212
|
Model to be used for converting speech to text in target language
|
|
214
213
|
|
|
215
|
-
input_audio_codec : typing.Optional[
|
|
214
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
216
215
|
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
217
216
|
|
|
218
217
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -331,7 +330,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
331
330
|
file: core.File,
|
|
332
331
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
333
332
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
334
|
-
input_audio_codec: typing.Optional[
|
|
333
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
335
334
|
request_options: typing.Optional[RequestOptions] = None,
|
|
336
335
|
) -> AsyncHttpResponse[SpeechToTextResponse]:
|
|
337
336
|
"""
|
|
@@ -363,7 +362,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
363
362
|
For the `saarika:v2.5` model, it is optional.
|
|
364
363
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
365
364
|
|
|
366
|
-
input_audio_codec : typing.Optional[
|
|
365
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
367
366
|
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
368
367
|
|
|
369
368
|
request_options : typing.Optional[RequestOptions]
|
|
@@ -477,7 +476,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
477
476
|
file: core.File,
|
|
478
477
|
prompt: typing.Optional[str] = OMIT,
|
|
479
478
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
480
|
-
input_audio_codec: typing.Optional[
|
|
479
|
+
input_audio_codec: typing.Optional[InputAudioCodec] = OMIT,
|
|
481
480
|
request_options: typing.Optional[RequestOptions] = None,
|
|
482
481
|
) -> AsyncHttpResponse[SpeechToTextTranslateResponse]:
|
|
483
482
|
"""
|
|
@@ -506,7 +505,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
506
505
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
507
506
|
Model to be used for converting speech to text in target language
|
|
508
507
|
|
|
509
|
-
input_audio_codec : typing.Optional[
|
|
508
|
+
input_audio_codec : typing.Optional[InputAudioCodec]
|
|
510
509
|
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
511
510
|
|
|
512
511
|
request_options : typing.Optional[RequestOptions]
|
sarvamai/types/__init__.py
CHANGED
|
@@ -45,6 +45,7 @@ from .files_request import FilesRequest
|
|
|
45
45
|
from .files_upload_response import FilesUploadResponse
|
|
46
46
|
from .finish_reason import FinishReason
|
|
47
47
|
from .flush_signal import FlushSignal
|
|
48
|
+
from .input_audio_codec import InputAudioCodec
|
|
48
49
|
from .job_state import JobState
|
|
49
50
|
from .job_status_v_1_response import JobStatusV1Response
|
|
50
51
|
from .language_identification_response import LanguageIdentificationResponse
|
|
@@ -137,6 +138,7 @@ __all__ = [
|
|
|
137
138
|
"FilesUploadResponse",
|
|
138
139
|
"FinishReason",
|
|
139
140
|
"FlushSignal",
|
|
141
|
+
"InputAudioCodec",
|
|
140
142
|
"JobState",
|
|
141
143
|
"JobStatusV1Response",
|
|
142
144
|
"LanguageIdentificationResponse",
|
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
sarvamai/__init__.py,sha256=
|
|
1
|
+
sarvamai/__init__.py,sha256=5j3UOszpw8O8ihXP2HwhHPgwCGnzPu5-8w9FDV51k1s,10694
|
|
2
2
|
sarvamai/chat/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
|
|
3
3
|
sarvamai/chat/client.py,sha256=xOSj83Gr6Q7eY2qUeATiuXYQqBqWqSCQlIEopK5fKus,11022
|
|
4
4
|
sarvamai/chat/raw_client.py,sha256=A2kRuZcVWlJhyYCD7YKgqNkZEp3cYa1731KhRkhirU0,17885
|
|
5
5
|
sarvamai/client.py,sha256=J30X_os1lPf8Wml0KDFEf6p8VGHhgF_lf3nw1T2D3qo,8207
|
|
6
6
|
sarvamai/core/__init__.py,sha256=YE2CtXeASe1RAbaI39twKWYKCuT4tW5is9HWHhJjR_g,1653
|
|
7
7
|
sarvamai/core/api_error.py,sha256=44vPoTyWN59gonCIZMdzw7M1uspygiLnr3GNFOoVL2Q,614
|
|
8
|
-
sarvamai/core/client_wrapper.py,sha256=
|
|
8
|
+
sarvamai/core/client_wrapper.py,sha256=FlSYn4WDn_xmsib3MOFWOsbYdumr8MJCHcYR4MHwGpI,2566
|
|
9
9
|
sarvamai/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
|
|
10
10
|
sarvamai/core/events.py,sha256=HvKBdSoYcFetk7cgNXb7FxuY-FtY8NtUhZIN7mGVx8U,1159
|
|
11
11
|
sarvamai/core/file.py,sha256=d4NNbX8XvXP32z8KpK2Xovv33nFfruIrpz0QWxlgpZk,2663
|
|
@@ -83,12 +83,9 @@ sarvamai/requests/timestamps_model.py,sha256=LkaYC5poipZe4r5-7Tp_Y058RXYWXwocn46
|
|
|
83
83
|
sarvamai/requests/transcription_metrics.py,sha256=FDclX2Z9Z3azrDXxtZW8xbkxxWMZQXEbaYMMheAAAdk,321
|
|
84
84
|
sarvamai/requests/translation_response.py,sha256=8iwQeZB1purHY757bIQI-n9QeVRBItaAVcBJ_la-k1Y,414
|
|
85
85
|
sarvamai/requests/transliteration_response.py,sha256=KqRkqnegLmt7LjdVxjRePX6RoqaLm64KFGZ6q7mXyfw,426
|
|
86
|
-
sarvamai/speech_to_text/__init__.py,sha256=
|
|
87
|
-
sarvamai/speech_to_text/client.py,sha256=
|
|
88
|
-
sarvamai/speech_to_text/raw_client.py,sha256=
|
|
89
|
-
sarvamai/speech_to_text/types/__init__.py,sha256=HIUDEN0mwbyXM15kMPzNydaxB9FEY_Fjzj5RsMFyef4,410
|
|
90
|
-
sarvamai/speech_to_text/types/speech_to_text_transcribe_request_input_audio_codec.py,sha256=FRi5ofYQFd4TmDbAqZ0Dx2mL0F7FRovwJGn1OI8N4q0,592
|
|
91
|
-
sarvamai/speech_to_text/types/speech_to_text_translate_request_input_audio_codec.py,sha256=q5ZnWwicyFtdOX9EMDUkHipSIIAMstCVYyjesminEWM,591
|
|
86
|
+
sarvamai/speech_to_text/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
|
|
87
|
+
sarvamai/speech_to_text/client.py,sha256=JZ-3ZenTTcJkSAIb7Hkj8zBS4r2TV_jlynl4LjxOnZg,12393
|
|
88
|
+
sarvamai/speech_to_text/raw_client.py,sha256=oCGHyVtVBpXGRSWtAvR_r2j3tEumc9VEDjOYqMVxv5w,26677
|
|
92
89
|
sarvamai/speech_to_text_job/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
|
|
93
90
|
sarvamai/speech_to_text_job/client.py,sha256=WSGBJxYcNxl77Zd1X6VVWjg4zshqecXf6WCyhfLXVlI,18007
|
|
94
91
|
sarvamai/speech_to_text_job/job.py,sha256=K8HOmwrYd6l82-MZfWDBmNkZeeERyg9YOihnFfvl-Js,15021
|
|
@@ -124,7 +121,7 @@ sarvamai/text_to_speech_streaming/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7
|
|
|
124
121
|
sarvamai/text_to_speech_streaming/client.py,sha256=tcYTUmWGseGQkLOsJdO4S3eZrFstjaLCPi5OcKflfus,6306
|
|
125
122
|
sarvamai/text_to_speech_streaming/raw_client.py,sha256=VdL5HYpUotIm5HaOqTwRkyZqoV0BuzwCtAfcB0g1y10,5487
|
|
126
123
|
sarvamai/text_to_speech_streaming/socket_client.py,sha256=NEcijnvjuNcWfzqpBi-xWsXVkL0NPq6EGAkEjnaq9hw,13909
|
|
127
|
-
sarvamai/types/__init__.py,sha256=
|
|
124
|
+
sarvamai/types/__init__.py,sha256=ilRhyVqS8gsWX_73IclMel-MzGS_zL5jk-7VrsdlmDU,7862
|
|
128
125
|
sarvamai/types/audio_data.py,sha256=rgOukLkLNJ_HBBVE2g5dfEL2CWjRoGiMvCtpq0qTB1Y,829
|
|
129
126
|
sarvamai/types/audio_message.py,sha256=sB4EgkWkWJzipYXobkmM9AYZTTZtCpg_ySKssUeznUE,560
|
|
130
127
|
sarvamai/types/audio_output.py,sha256=Eq-YUZa1mSDwt7bax2c4Vv2gBlyM_JBJWzHhTAhFSko,621
|
|
@@ -163,6 +160,7 @@ sarvamai/types/files_request.py,sha256=Jh8xPjoOTjY7DOE2EieoRqtkWkYxz9j-BP8TvWxuR
|
|
|
163
160
|
sarvamai/types/files_upload_response.py,sha256=wRntZyh1-LGpo4-x_986Nv2A9rv9asDx93pqoQxNpTY,804
|
|
164
161
|
sarvamai/types/finish_reason.py,sha256=PBWtBNkX4FMaODmlUehpF6qLB5uH_zR-Mw3M4uhIB6U,209
|
|
165
162
|
sarvamai/types/flush_signal.py,sha256=N7MJWb658KoxRpFN9cIbyQGY45zZcg8YCou3E1v--9o,759
|
|
163
|
+
sarvamai/types/input_audio_codec.py,sha256=P3rz6lg-T34E7psQ1DX4e2yvPxVJDDsMVn4TSdEpVEY,563
|
|
166
164
|
sarvamai/types/job_state.py,sha256=H6Zph2mIcjsd3upEDt1VzIEORkEpnIDs0kH8BvIyrow,189
|
|
167
165
|
sarvamai/types/job_status_v_1_response.py,sha256=jusn-3XLCk8vCnu3q3GGSzz_yYgKB5PY_01Q32-1aJk,1605
|
|
168
166
|
sarvamai/types/language_identification_response.py,sha256=jG4ZQ6KQHCiEDqC51OniOwiRdW14Fbz22bbTsUDp_kc,1483
|
|
@@ -213,6 +211,6 @@ sarvamai/types/transliterate_mode.py,sha256=1jSEMlGcoLkWuk12TgoOpSgwifa4rThGKZ1h
|
|
|
213
211
|
sarvamai/types/transliterate_source_language.py,sha256=bSY9wJszF0sg-Cgg6F-YcWC8ly1mIlj9rqa15-jBtx8,283
|
|
214
212
|
sarvamai/types/transliteration_response.py,sha256=yt-lzTbDeJ_ZL4I8kQa6oESxA9ebeJJY7LfFHpdEsmM,815
|
|
215
213
|
sarvamai/version.py,sha256=Qkp3Ee9YH-O9RTix90e0i7iNrFAGN-QDt2AFwGA4n8k,75
|
|
216
|
-
sarvamai-0.1.
|
|
217
|
-
sarvamai-0.1.
|
|
218
|
-
sarvamai-0.1.
|
|
214
|
+
sarvamai-0.1.18.dist-info/METADATA,sha256=rLo999GWURJJolBIN9QEUmGz8wPxM5eKlBVJIV2PkJY,26751
|
|
215
|
+
sarvamai-0.1.18.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
|
|
216
|
+
sarvamai-0.1.18.dist-info/RECORD,,
|
|
@@ -1,8 +0,0 @@
|
|
|
1
|
-
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
-
|
|
3
|
-
# isort: skip_file
|
|
4
|
-
|
|
5
|
-
from .speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
|
|
6
|
-
from .speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
|
|
7
|
-
|
|
8
|
-
__all__ = ["SpeechToTextTranscribeRequestInputAudioCodec", "SpeechToTextTranslateRequestInputAudioCodec"]
|
|
@@ -1,33 +0,0 @@
|
|
|
1
|
-
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
-
|
|
3
|
-
import typing
|
|
4
|
-
|
|
5
|
-
SpeechToTextTranscribeRequestInputAudioCodec = typing.Union[
|
|
6
|
-
typing.Literal[
|
|
7
|
-
"wav",
|
|
8
|
-
"x-wav",
|
|
9
|
-
"wave",
|
|
10
|
-
"mp3",
|
|
11
|
-
"mpeg",
|
|
12
|
-
"mpeg3",
|
|
13
|
-
"x-mp3",
|
|
14
|
-
"x-mpeg-3",
|
|
15
|
-
"aac",
|
|
16
|
-
"x-aac",
|
|
17
|
-
"aiff",
|
|
18
|
-
"x-aiff",
|
|
19
|
-
"ogg",
|
|
20
|
-
"opus",
|
|
21
|
-
"flac",
|
|
22
|
-
"x-flac",
|
|
23
|
-
"mp4",
|
|
24
|
-
"x-m4a",
|
|
25
|
-
"amr",
|
|
26
|
-
"x-ms-wma",
|
|
27
|
-
"webm",
|
|
28
|
-
"pcm_s16le",
|
|
29
|
-
"pcm_l16",
|
|
30
|
-
"pcm_raw",
|
|
31
|
-
],
|
|
32
|
-
typing.Any,
|
|
33
|
-
]
|
|
File without changes
|