sarvamai 0.1.16a0__py3-none-any.whl → 0.1.17a0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- sarvamai/__init__.py +3 -0
- sarvamai/core/client_wrapper.py +2 -2
- sarvamai/speech_to_text/__init__.py +3 -0
- sarvamai/speech_to_text/client.py +34 -6
- sarvamai/speech_to_text/raw_client.py +24 -2
- sarvamai/speech_to_text/types/__init__.py +8 -0
- sarvamai/speech_to_text/types/speech_to_text_transcribe_request_input_audio_codec.py +33 -0
- sarvamai/speech_to_text/types/speech_to_text_translate_request_input_audio_codec.py +33 -0
- {sarvamai-0.1.16a0.dist-info → sarvamai-0.1.17a0.dist-info}/METADATA +1 -1
- {sarvamai-0.1.16a0.dist-info → sarvamai-0.1.17a0.dist-info}/RECORD +11 -8
- {sarvamai-0.1.16a0.dist-info → sarvamai-0.1.17a0.dist-info}/WHEEL +0 -0
sarvamai/__init__.py
CHANGED
|
@@ -174,6 +174,7 @@ from .requests import (
|
|
|
174
174
|
TranslationResponseParams,
|
|
175
175
|
TransliterationResponseParams,
|
|
176
176
|
)
|
|
177
|
+
from .speech_to_text import SpeechToTextTranscribeRequestInputAudioCodec, SpeechToTextTranslateRequestInputAudioCodec
|
|
177
178
|
from .speech_to_text_streaming import (
|
|
178
179
|
SpeechToTextStreamingHighVadSensitivity,
|
|
179
180
|
SpeechToTextStreamingLanguageCode,
|
|
@@ -301,12 +302,14 @@ __all__ = [
|
|
|
301
302
|
"SpeechToTextStreamingResponse",
|
|
302
303
|
"SpeechToTextStreamingResponseParams",
|
|
303
304
|
"SpeechToTextStreamingVadSignals",
|
|
305
|
+
"SpeechToTextTranscribeRequestInputAudioCodec",
|
|
304
306
|
"SpeechToTextTranscriptionData",
|
|
305
307
|
"SpeechToTextTranscriptionDataParams",
|
|
306
308
|
"SpeechToTextTranslateJobParameters",
|
|
307
309
|
"SpeechToTextTranslateJobParametersParams",
|
|
308
310
|
"SpeechToTextTranslateLanguage",
|
|
309
311
|
"SpeechToTextTranslateModel",
|
|
312
|
+
"SpeechToTextTranslateRequestInputAudioCodec",
|
|
310
313
|
"SpeechToTextTranslateResponse",
|
|
311
314
|
"SpeechToTextTranslateResponseData",
|
|
312
315
|
"SpeechToTextTranslateResponseDataParams",
|
sarvamai/core/client_wrapper.py
CHANGED
|
@@ -23,10 +23,10 @@ class BaseClientWrapper:
|
|
|
23
23
|
|
|
24
24
|
def get_headers(self) -> typing.Dict[str, str]:
|
|
25
25
|
headers: typing.Dict[str, str] = {
|
|
26
|
-
"User-Agent": "sarvamai/0.1.
|
|
26
|
+
"User-Agent": "sarvamai/0.1.17a0",
|
|
27
27
|
"X-Fern-Language": "Python",
|
|
28
28
|
"X-Fern-SDK-Name": "sarvamai",
|
|
29
|
-
"X-Fern-SDK-Version": "0.1.
|
|
29
|
+
"X-Fern-SDK-Version": "0.1.17a0",
|
|
30
30
|
**(self.get_custom_headers() or {}),
|
|
31
31
|
}
|
|
32
32
|
headers["api-subscription-key"] = self.api_subscription_key
|
|
@@ -11,6 +11,8 @@ from ..types.speech_to_text_response import SpeechToTextResponse
|
|
|
11
11
|
from ..types.speech_to_text_translate_model import SpeechToTextTranslateModel
|
|
12
12
|
from ..types.speech_to_text_translate_response import SpeechToTextTranslateResponse
|
|
13
13
|
from .raw_client import AsyncRawSpeechToTextClient, RawSpeechToTextClient
|
|
14
|
+
from .types.speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
|
|
15
|
+
from .types.speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
|
|
14
16
|
|
|
15
17
|
# this is used as the default value for optional parameters
|
|
16
18
|
OMIT = typing.cast(typing.Any, ...)
|
|
@@ -37,6 +39,7 @@ class SpeechToTextClient:
|
|
|
37
39
|
file: core.File,
|
|
38
40
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
39
41
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
42
|
+
input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
|
|
40
43
|
request_options: typing.Optional[RequestOptions] = None,
|
|
41
44
|
) -> SpeechToTextResponse:
|
|
42
45
|
"""
|
|
@@ -68,6 +71,9 @@ class SpeechToTextClient:
|
|
|
68
71
|
For the `saarika:v2.5` model, it is optional.
|
|
69
72
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
70
73
|
|
|
74
|
+
input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
|
|
75
|
+
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
76
|
+
|
|
71
77
|
request_options : typing.Optional[RequestOptions]
|
|
72
78
|
Request-specific configuration.
|
|
73
79
|
|
|
@@ -86,7 +92,11 @@ class SpeechToTextClient:
|
|
|
86
92
|
client.speech_to_text.transcribe()
|
|
87
93
|
"""
|
|
88
94
|
_response = self._raw_client.transcribe(
|
|
89
|
-
file=file,
|
|
95
|
+
file=file,
|
|
96
|
+
model=model,
|
|
97
|
+
language_code=language_code,
|
|
98
|
+
input_audio_codec=input_audio_codec,
|
|
99
|
+
request_options=request_options,
|
|
90
100
|
)
|
|
91
101
|
return _response.data
|
|
92
102
|
|
|
@@ -96,6 +106,7 @@ class SpeechToTextClient:
|
|
|
96
106
|
file: core.File,
|
|
97
107
|
prompt: typing.Optional[str] = OMIT,
|
|
98
108
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
109
|
+
input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
|
|
99
110
|
request_options: typing.Optional[RequestOptions] = None,
|
|
100
111
|
) -> SpeechToTextTranslateResponse:
|
|
101
112
|
"""
|
|
@@ -119,11 +130,14 @@ class SpeechToTextClient:
|
|
|
119
130
|
See core.File for more documentation
|
|
120
131
|
|
|
121
132
|
prompt : typing.Optional[str]
|
|
122
|
-
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn
|
|
133
|
+
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
|
|
123
134
|
|
|
124
135
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
125
136
|
Model to be used for converting speech to text in target language
|
|
126
137
|
|
|
138
|
+
input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
|
|
139
|
+
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
140
|
+
|
|
127
141
|
request_options : typing.Optional[RequestOptions]
|
|
128
142
|
Request-specific configuration.
|
|
129
143
|
|
|
@@ -141,7 +155,9 @@ class SpeechToTextClient:
|
|
|
141
155
|
)
|
|
142
156
|
client.speech_to_text.translate()
|
|
143
157
|
"""
|
|
144
|
-
_response = self._raw_client.translate(
|
|
158
|
+
_response = self._raw_client.translate(
|
|
159
|
+
file=file, prompt=prompt, model=model, input_audio_codec=input_audio_codec, request_options=request_options
|
|
160
|
+
)
|
|
145
161
|
return _response.data
|
|
146
162
|
|
|
147
163
|
|
|
@@ -166,6 +182,7 @@ class AsyncSpeechToTextClient:
|
|
|
166
182
|
file: core.File,
|
|
167
183
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
168
184
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
185
|
+
input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
|
|
169
186
|
request_options: typing.Optional[RequestOptions] = None,
|
|
170
187
|
) -> SpeechToTextResponse:
|
|
171
188
|
"""
|
|
@@ -197,6 +214,9 @@ class AsyncSpeechToTextClient:
|
|
|
197
214
|
For the `saarika:v2.5` model, it is optional.
|
|
198
215
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
199
216
|
|
|
217
|
+
input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
|
|
218
|
+
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
219
|
+
|
|
200
220
|
request_options : typing.Optional[RequestOptions]
|
|
201
221
|
Request-specific configuration.
|
|
202
222
|
|
|
@@ -223,7 +243,11 @@ class AsyncSpeechToTextClient:
|
|
|
223
243
|
asyncio.run(main())
|
|
224
244
|
"""
|
|
225
245
|
_response = await self._raw_client.transcribe(
|
|
226
|
-
file=file,
|
|
246
|
+
file=file,
|
|
247
|
+
model=model,
|
|
248
|
+
language_code=language_code,
|
|
249
|
+
input_audio_codec=input_audio_codec,
|
|
250
|
+
request_options=request_options,
|
|
227
251
|
)
|
|
228
252
|
return _response.data
|
|
229
253
|
|
|
@@ -233,6 +257,7 @@ class AsyncSpeechToTextClient:
|
|
|
233
257
|
file: core.File,
|
|
234
258
|
prompt: typing.Optional[str] = OMIT,
|
|
235
259
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
260
|
+
input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
|
|
236
261
|
request_options: typing.Optional[RequestOptions] = None,
|
|
237
262
|
) -> SpeechToTextTranslateResponse:
|
|
238
263
|
"""
|
|
@@ -256,11 +281,14 @@ class AsyncSpeechToTextClient:
|
|
|
256
281
|
See core.File for more documentation
|
|
257
282
|
|
|
258
283
|
prompt : typing.Optional[str]
|
|
259
|
-
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn
|
|
284
|
+
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
|
|
260
285
|
|
|
261
286
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
262
287
|
Model to be used for converting speech to text in target language
|
|
263
288
|
|
|
289
|
+
input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
|
|
290
|
+
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
291
|
+
|
|
264
292
|
request_options : typing.Optional[RequestOptions]
|
|
265
293
|
Request-specific configuration.
|
|
266
294
|
|
|
@@ -287,6 +315,6 @@ class AsyncSpeechToTextClient:
|
|
|
287
315
|
asyncio.run(main())
|
|
288
316
|
"""
|
|
289
317
|
_response = await self._raw_client.translate(
|
|
290
|
-
file=file, prompt=prompt, model=model, request_options=request_options
|
|
318
|
+
file=file, prompt=prompt, model=model, input_audio_codec=input_audio_codec, request_options=request_options
|
|
291
319
|
)
|
|
292
320
|
return _response.data
|
|
@@ -20,6 +20,8 @@ from ..types.speech_to_text_model import SpeechToTextModel
|
|
|
20
20
|
from ..types.speech_to_text_response import SpeechToTextResponse
|
|
21
21
|
from ..types.speech_to_text_translate_model import SpeechToTextTranslateModel
|
|
22
22
|
from ..types.speech_to_text_translate_response import SpeechToTextTranslateResponse
|
|
23
|
+
from .types.speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
|
|
24
|
+
from .types.speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
|
|
23
25
|
|
|
24
26
|
# this is used as the default value for optional parameters
|
|
25
27
|
OMIT = typing.cast(typing.Any, ...)
|
|
@@ -35,6 +37,7 @@ class RawSpeechToTextClient:
|
|
|
35
37
|
file: core.File,
|
|
36
38
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
37
39
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
40
|
+
input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
|
|
38
41
|
request_options: typing.Optional[RequestOptions] = None,
|
|
39
42
|
) -> HttpResponse[SpeechToTextResponse]:
|
|
40
43
|
"""
|
|
@@ -66,6 +69,9 @@ class RawSpeechToTextClient:
|
|
|
66
69
|
For the `saarika:v2.5` model, it is optional.
|
|
67
70
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
68
71
|
|
|
72
|
+
input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
|
|
73
|
+
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
74
|
+
|
|
69
75
|
request_options : typing.Optional[RequestOptions]
|
|
70
76
|
Request-specific configuration.
|
|
71
77
|
|
|
@@ -81,6 +87,7 @@ class RawSpeechToTextClient:
|
|
|
81
87
|
data={
|
|
82
88
|
"model": model,
|
|
83
89
|
"language_code": language_code,
|
|
90
|
+
"input_audio_codec": input_audio_codec,
|
|
84
91
|
},
|
|
85
92
|
files={
|
|
86
93
|
"file": file,
|
|
@@ -176,6 +183,7 @@ class RawSpeechToTextClient:
|
|
|
176
183
|
file: core.File,
|
|
177
184
|
prompt: typing.Optional[str] = OMIT,
|
|
178
185
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
186
|
+
input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
|
|
179
187
|
request_options: typing.Optional[RequestOptions] = None,
|
|
180
188
|
) -> HttpResponse[SpeechToTextTranslateResponse]:
|
|
181
189
|
"""
|
|
@@ -199,11 +207,14 @@ class RawSpeechToTextClient:
|
|
|
199
207
|
See core.File for more documentation
|
|
200
208
|
|
|
201
209
|
prompt : typing.Optional[str]
|
|
202
|
-
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn
|
|
210
|
+
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
|
|
203
211
|
|
|
204
212
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
205
213
|
Model to be used for converting speech to text in target language
|
|
206
214
|
|
|
215
|
+
input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
|
|
216
|
+
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
217
|
+
|
|
207
218
|
request_options : typing.Optional[RequestOptions]
|
|
208
219
|
Request-specific configuration.
|
|
209
220
|
|
|
@@ -219,6 +230,7 @@ class RawSpeechToTextClient:
|
|
|
219
230
|
data={
|
|
220
231
|
"prompt": prompt,
|
|
221
232
|
"model": model,
|
|
233
|
+
"input_audio_codec": input_audio_codec,
|
|
222
234
|
},
|
|
223
235
|
files={
|
|
224
236
|
"file": file,
|
|
@@ -319,6 +331,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
319
331
|
file: core.File,
|
|
320
332
|
model: typing.Optional[SpeechToTextModel] = OMIT,
|
|
321
333
|
language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
|
|
334
|
+
input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
|
|
322
335
|
request_options: typing.Optional[RequestOptions] = None,
|
|
323
336
|
) -> AsyncHttpResponse[SpeechToTextResponse]:
|
|
324
337
|
"""
|
|
@@ -350,6 +363,9 @@ class AsyncRawSpeechToTextClient:
|
|
|
350
363
|
For the `saarika:v2.5` model, it is optional.
|
|
351
364
|
`unknown`: Use this when the language is not known; the API will detect it automatically.
|
|
352
365
|
|
|
366
|
+
input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
|
|
367
|
+
Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
|
|
368
|
+
|
|
353
369
|
request_options : typing.Optional[RequestOptions]
|
|
354
370
|
Request-specific configuration.
|
|
355
371
|
|
|
@@ -365,6 +381,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
365
381
|
data={
|
|
366
382
|
"model": model,
|
|
367
383
|
"language_code": language_code,
|
|
384
|
+
"input_audio_codec": input_audio_codec,
|
|
368
385
|
},
|
|
369
386
|
files={
|
|
370
387
|
"file": file,
|
|
@@ -460,6 +477,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
460
477
|
file: core.File,
|
|
461
478
|
prompt: typing.Optional[str] = OMIT,
|
|
462
479
|
model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
|
|
480
|
+
input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
|
|
463
481
|
request_options: typing.Optional[RequestOptions] = None,
|
|
464
482
|
) -> AsyncHttpResponse[SpeechToTextTranslateResponse]:
|
|
465
483
|
"""
|
|
@@ -483,11 +501,14 @@ class AsyncRawSpeechToTextClient:
|
|
|
483
501
|
See core.File for more documentation
|
|
484
502
|
|
|
485
503
|
prompt : typing.Optional[str]
|
|
486
|
-
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn
|
|
504
|
+
Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
|
|
487
505
|
|
|
488
506
|
model : typing.Optional[SpeechToTextTranslateModel]
|
|
489
507
|
Model to be used for converting speech to text in target language
|
|
490
508
|
|
|
509
|
+
input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
|
|
510
|
+
Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
|
|
511
|
+
|
|
491
512
|
request_options : typing.Optional[RequestOptions]
|
|
492
513
|
Request-specific configuration.
|
|
493
514
|
|
|
@@ -503,6 +524,7 @@ class AsyncRawSpeechToTextClient:
|
|
|
503
524
|
data={
|
|
504
525
|
"prompt": prompt,
|
|
505
526
|
"model": model,
|
|
527
|
+
"input_audio_codec": input_audio_codec,
|
|
506
528
|
},
|
|
507
529
|
files={
|
|
508
530
|
"file": file,
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
# isort: skip_file
|
|
4
|
+
|
|
5
|
+
from .speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
|
|
6
|
+
from .speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
|
|
7
|
+
|
|
8
|
+
__all__ = ["SpeechToTextTranscribeRequestInputAudioCodec", "SpeechToTextTranslateRequestInputAudioCodec"]
|
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import typing
|
|
4
|
+
|
|
5
|
+
SpeechToTextTranscribeRequestInputAudioCodec = typing.Union[
|
|
6
|
+
typing.Literal[
|
|
7
|
+
"wav",
|
|
8
|
+
"x-wav",
|
|
9
|
+
"wave",
|
|
10
|
+
"mp3",
|
|
11
|
+
"mpeg",
|
|
12
|
+
"mpeg3",
|
|
13
|
+
"x-mp3",
|
|
14
|
+
"x-mpeg-3",
|
|
15
|
+
"aac",
|
|
16
|
+
"x-aac",
|
|
17
|
+
"aiff",
|
|
18
|
+
"x-aiff",
|
|
19
|
+
"ogg",
|
|
20
|
+
"opus",
|
|
21
|
+
"flac",
|
|
22
|
+
"x-flac",
|
|
23
|
+
"mp4",
|
|
24
|
+
"x-m4a",
|
|
25
|
+
"amr",
|
|
26
|
+
"x-ms-wma",
|
|
27
|
+
"webm",
|
|
28
|
+
"pcm_s16le",
|
|
29
|
+
"pcm_l16",
|
|
30
|
+
"pcm_raw",
|
|
31
|
+
],
|
|
32
|
+
typing.Any,
|
|
33
|
+
]
|
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import typing
|
|
4
|
+
|
|
5
|
+
SpeechToTextTranslateRequestInputAudioCodec = typing.Union[
|
|
6
|
+
typing.Literal[
|
|
7
|
+
"wav",
|
|
8
|
+
"x-wav",
|
|
9
|
+
"wave",
|
|
10
|
+
"mp3",
|
|
11
|
+
"mpeg",
|
|
12
|
+
"mpeg3",
|
|
13
|
+
"x-mp3",
|
|
14
|
+
"x-mpeg-3",
|
|
15
|
+
"aac",
|
|
16
|
+
"x-aac",
|
|
17
|
+
"aiff",
|
|
18
|
+
"x-aiff",
|
|
19
|
+
"ogg",
|
|
20
|
+
"opus",
|
|
21
|
+
"flac",
|
|
22
|
+
"x-flac",
|
|
23
|
+
"mp4",
|
|
24
|
+
"x-m4a",
|
|
25
|
+
"amr",
|
|
26
|
+
"x-ms-wma",
|
|
27
|
+
"webm",
|
|
28
|
+
"pcm_s16le",
|
|
29
|
+
"pcm_l16",
|
|
30
|
+
"pcm_raw",
|
|
31
|
+
],
|
|
32
|
+
typing.Any,
|
|
33
|
+
]
|
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
sarvamai/__init__.py,sha256=
|
|
1
|
+
sarvamai/__init__.py,sha256=_AvjD4uJaddFuA7gBmWXoGqU9JEo9mHpVq-cqk47jws,10871
|
|
2
2
|
sarvamai/chat/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
|
|
3
3
|
sarvamai/chat/client.py,sha256=xOSj83Gr6Q7eY2qUeATiuXYQqBqWqSCQlIEopK5fKus,11022
|
|
4
4
|
sarvamai/chat/raw_client.py,sha256=A2kRuZcVWlJhyYCD7YKgqNkZEp3cYa1731KhRkhirU0,17885
|
|
5
5
|
sarvamai/client.py,sha256=J30X_os1lPf8Wml0KDFEf6p8VGHhgF_lf3nw1T2D3qo,8207
|
|
6
6
|
sarvamai/core/__init__.py,sha256=YE2CtXeASe1RAbaI39twKWYKCuT4tW5is9HWHhJjR_g,1653
|
|
7
7
|
sarvamai/core/api_error.py,sha256=44vPoTyWN59gonCIZMdzw7M1uspygiLnr3GNFOoVL2Q,614
|
|
8
|
-
sarvamai/core/client_wrapper.py,sha256=
|
|
8
|
+
sarvamai/core/client_wrapper.py,sha256=OJ4aojUSt1w1Qnvo0h5xwdvbFZKBybSJqwWw9tvkbVM,2570
|
|
9
9
|
sarvamai/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
|
|
10
10
|
sarvamai/core/events.py,sha256=HvKBdSoYcFetk7cgNXb7FxuY-FtY8NtUhZIN7mGVx8U,1159
|
|
11
11
|
sarvamai/core/file.py,sha256=d4NNbX8XvXP32z8KpK2Xovv33nFfruIrpz0QWxlgpZk,2663
|
|
@@ -83,9 +83,12 @@ sarvamai/requests/timestamps_model.py,sha256=LkaYC5poipZe4r5-7Tp_Y058RXYWXwocn46
|
|
|
83
83
|
sarvamai/requests/transcription_metrics.py,sha256=FDclX2Z9Z3azrDXxtZW8xbkxxWMZQXEbaYMMheAAAdk,321
|
|
84
84
|
sarvamai/requests/translation_response.py,sha256=8iwQeZB1purHY757bIQI-n9QeVRBItaAVcBJ_la-k1Y,414
|
|
85
85
|
sarvamai/requests/transliteration_response.py,sha256=KqRkqnegLmt7LjdVxjRePX6RoqaLm64KFGZ6q7mXyfw,426
|
|
86
|
-
sarvamai/speech_to_text/__init__.py,sha256=
|
|
87
|
-
sarvamai/speech_to_text/client.py,sha256=
|
|
88
|
-
sarvamai/speech_to_text/raw_client.py,sha256=
|
|
86
|
+
sarvamai/speech_to_text/__init__.py,sha256=U3RTBb6u3QwsV3fyh-RX_HQ3AI2BD0i-Y8ec3gqTjdI,301
|
|
87
|
+
sarvamai/speech_to_text/client.py,sha256=Z_MZH-W70BbNYkax_CV3ZIhgxT0d44rnXtN3yLTMIzc,12797
|
|
88
|
+
sarvamai/speech_to_text/raw_client.py,sha256=kHU4ngENfwqrKBZq0GZwAlYONPWf3ssH3hQvJhEyEXg,27081
|
|
89
|
+
sarvamai/speech_to_text/types/__init__.py,sha256=HIUDEN0mwbyXM15kMPzNydaxB9FEY_Fjzj5RsMFyef4,410
|
|
90
|
+
sarvamai/speech_to_text/types/speech_to_text_transcribe_request_input_audio_codec.py,sha256=FRi5ofYQFd4TmDbAqZ0Dx2mL0F7FRovwJGn1OI8N4q0,592
|
|
91
|
+
sarvamai/speech_to_text/types/speech_to_text_translate_request_input_audio_codec.py,sha256=q5ZnWwicyFtdOX9EMDUkHipSIIAMstCVYyjesminEWM,591
|
|
89
92
|
sarvamai/speech_to_text_job/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
|
|
90
93
|
sarvamai/speech_to_text_job/client.py,sha256=WSGBJxYcNxl77Zd1X6VVWjg4zshqecXf6WCyhfLXVlI,18007
|
|
91
94
|
sarvamai/speech_to_text_job/job.py,sha256=K8HOmwrYd6l82-MZfWDBmNkZeeERyg9YOihnFfvl-Js,15021
|
|
@@ -210,6 +213,6 @@ sarvamai/types/transliterate_mode.py,sha256=1jSEMlGcoLkWuk12TgoOpSgwifa4rThGKZ1h
|
|
|
210
213
|
sarvamai/types/transliterate_source_language.py,sha256=bSY9wJszF0sg-Cgg6F-YcWC8ly1mIlj9rqa15-jBtx8,283
|
|
211
214
|
sarvamai/types/transliteration_response.py,sha256=yt-lzTbDeJ_ZL4I8kQa6oESxA9ebeJJY7LfFHpdEsmM,815
|
|
212
215
|
sarvamai/version.py,sha256=Qkp3Ee9YH-O9RTix90e0i7iNrFAGN-QDt2AFwGA4n8k,75
|
|
213
|
-
sarvamai-0.1.
|
|
214
|
-
sarvamai-0.1.
|
|
215
|
-
sarvamai-0.1.
|
|
216
|
+
sarvamai-0.1.17a0.dist-info/METADATA,sha256=Frl6Uw6tD1u701FFbq4j6anJgmBbbNDmZr_ZNZlZfh8,26753
|
|
217
|
+
sarvamai-0.1.17a0.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
|
|
218
|
+
sarvamai-0.1.17a0.dist-info/RECORD,,
|
|
File without changes
|