sarvamai 0.1.16a0__py3-none-any.whl → 0.1.17a0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
sarvamai/__init__.py CHANGED
@@ -174,6 +174,7 @@ from .requests import (
174
174
  TranslationResponseParams,
175
175
  TransliterationResponseParams,
176
176
  )
177
+ from .speech_to_text import SpeechToTextTranscribeRequestInputAudioCodec, SpeechToTextTranslateRequestInputAudioCodec
177
178
  from .speech_to_text_streaming import (
178
179
  SpeechToTextStreamingHighVadSensitivity,
179
180
  SpeechToTextStreamingLanguageCode,
@@ -301,12 +302,14 @@ __all__ = [
301
302
  "SpeechToTextStreamingResponse",
302
303
  "SpeechToTextStreamingResponseParams",
303
304
  "SpeechToTextStreamingVadSignals",
305
+ "SpeechToTextTranscribeRequestInputAudioCodec",
304
306
  "SpeechToTextTranscriptionData",
305
307
  "SpeechToTextTranscriptionDataParams",
306
308
  "SpeechToTextTranslateJobParameters",
307
309
  "SpeechToTextTranslateJobParametersParams",
308
310
  "SpeechToTextTranslateLanguage",
309
311
  "SpeechToTextTranslateModel",
312
+ "SpeechToTextTranslateRequestInputAudioCodec",
310
313
  "SpeechToTextTranslateResponse",
311
314
  "SpeechToTextTranslateResponseData",
312
315
  "SpeechToTextTranslateResponseDataParams",
@@ -23,10 +23,10 @@ class BaseClientWrapper:
23
23
 
24
24
  def get_headers(self) -> typing.Dict[str, str]:
25
25
  headers: typing.Dict[str, str] = {
26
- "User-Agent": "sarvamai/0.1.16a0",
26
+ "User-Agent": "sarvamai/0.1.17a0",
27
27
  "X-Fern-Language": "Python",
28
28
  "X-Fern-SDK-Name": "sarvamai",
29
- "X-Fern-SDK-Version": "0.1.16a0",
29
+ "X-Fern-SDK-Version": "0.1.17a0",
30
30
  **(self.get_custom_headers() or {}),
31
31
  }
32
32
  headers["api-subscription-key"] = self.api_subscription_key
@@ -2,3 +2,6 @@
2
2
 
3
3
  # isort: skip_file
4
4
 
5
+ from .types import SpeechToTextTranscribeRequestInputAudioCodec, SpeechToTextTranslateRequestInputAudioCodec
6
+
7
+ __all__ = ["SpeechToTextTranscribeRequestInputAudioCodec", "SpeechToTextTranslateRequestInputAudioCodec"]
@@ -11,6 +11,8 @@ from ..types.speech_to_text_response import SpeechToTextResponse
11
11
  from ..types.speech_to_text_translate_model import SpeechToTextTranslateModel
12
12
  from ..types.speech_to_text_translate_response import SpeechToTextTranslateResponse
13
13
  from .raw_client import AsyncRawSpeechToTextClient, RawSpeechToTextClient
14
+ from .types.speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
15
+ from .types.speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
14
16
 
15
17
  # this is used as the default value for optional parameters
16
18
  OMIT = typing.cast(typing.Any, ...)
@@ -37,6 +39,7 @@ class SpeechToTextClient:
37
39
  file: core.File,
38
40
  model: typing.Optional[SpeechToTextModel] = OMIT,
39
41
  language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
42
+ input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
40
43
  request_options: typing.Optional[RequestOptions] = None,
41
44
  ) -> SpeechToTextResponse:
42
45
  """
@@ -68,6 +71,9 @@ class SpeechToTextClient:
68
71
  For the `saarika:v2.5` model, it is optional.
69
72
  `unknown`: Use this when the language is not known; the API will detect it automatically.
70
73
 
74
+ input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
75
+ Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
76
+
71
77
  request_options : typing.Optional[RequestOptions]
72
78
  Request-specific configuration.
73
79
 
@@ -86,7 +92,11 @@ class SpeechToTextClient:
86
92
  client.speech_to_text.transcribe()
87
93
  """
88
94
  _response = self._raw_client.transcribe(
89
- file=file, model=model, language_code=language_code, request_options=request_options
95
+ file=file,
96
+ model=model,
97
+ language_code=language_code,
98
+ input_audio_codec=input_audio_codec,
99
+ request_options=request_options,
90
100
  )
91
101
  return _response.data
92
102
 
@@ -96,6 +106,7 @@ class SpeechToTextClient:
96
106
  file: core.File,
97
107
  prompt: typing.Optional[str] = OMIT,
98
108
  model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
109
+ input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
99
110
  request_options: typing.Optional[RequestOptions] = None,
100
111
  ) -> SpeechToTextTranslateResponse:
101
112
  """
@@ -119,11 +130,14 @@ class SpeechToTextClient:
119
130
  See core.File for more documentation
120
131
 
121
132
  prompt : typing.Optional[str]
122
- Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesnt match the prompt performance of large language models.
133
+ Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
123
134
 
124
135
  model : typing.Optional[SpeechToTextTranslateModel]
125
136
  Model to be used for converting speech to text in target language
126
137
 
138
+ input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
139
+ Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
140
+
127
141
  request_options : typing.Optional[RequestOptions]
128
142
  Request-specific configuration.
129
143
 
@@ -141,7 +155,9 @@ class SpeechToTextClient:
141
155
  )
142
156
  client.speech_to_text.translate()
143
157
  """
144
- _response = self._raw_client.translate(file=file, prompt=prompt, model=model, request_options=request_options)
158
+ _response = self._raw_client.translate(
159
+ file=file, prompt=prompt, model=model, input_audio_codec=input_audio_codec, request_options=request_options
160
+ )
145
161
  return _response.data
146
162
 
147
163
 
@@ -166,6 +182,7 @@ class AsyncSpeechToTextClient:
166
182
  file: core.File,
167
183
  model: typing.Optional[SpeechToTextModel] = OMIT,
168
184
  language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
185
+ input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
169
186
  request_options: typing.Optional[RequestOptions] = None,
170
187
  ) -> SpeechToTextResponse:
171
188
  """
@@ -197,6 +214,9 @@ class AsyncSpeechToTextClient:
197
214
  For the `saarika:v2.5` model, it is optional.
198
215
  `unknown`: Use this when the language is not known; the API will detect it automatically.
199
216
 
217
+ input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
218
+ Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
219
+
200
220
  request_options : typing.Optional[RequestOptions]
201
221
  Request-specific configuration.
202
222
 
@@ -223,7 +243,11 @@ class AsyncSpeechToTextClient:
223
243
  asyncio.run(main())
224
244
  """
225
245
  _response = await self._raw_client.transcribe(
226
- file=file, model=model, language_code=language_code, request_options=request_options
246
+ file=file,
247
+ model=model,
248
+ language_code=language_code,
249
+ input_audio_codec=input_audio_codec,
250
+ request_options=request_options,
227
251
  )
228
252
  return _response.data
229
253
 
@@ -233,6 +257,7 @@ class AsyncSpeechToTextClient:
233
257
  file: core.File,
234
258
  prompt: typing.Optional[str] = OMIT,
235
259
  model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
260
+ input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
236
261
  request_options: typing.Optional[RequestOptions] = None,
237
262
  ) -> SpeechToTextTranslateResponse:
238
263
  """
@@ -256,11 +281,14 @@ class AsyncSpeechToTextClient:
256
281
  See core.File for more documentation
257
282
 
258
283
  prompt : typing.Optional[str]
259
- Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesnt match the prompt performance of large language models.
284
+ Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
260
285
 
261
286
  model : typing.Optional[SpeechToTextTranslateModel]
262
287
  Model to be used for converting speech to text in target language
263
288
 
289
+ input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
290
+ Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
291
+
264
292
  request_options : typing.Optional[RequestOptions]
265
293
  Request-specific configuration.
266
294
 
@@ -287,6 +315,6 @@ class AsyncSpeechToTextClient:
287
315
  asyncio.run(main())
288
316
  """
289
317
  _response = await self._raw_client.translate(
290
- file=file, prompt=prompt, model=model, request_options=request_options
318
+ file=file, prompt=prompt, model=model, input_audio_codec=input_audio_codec, request_options=request_options
291
319
  )
292
320
  return _response.data
@@ -20,6 +20,8 @@ from ..types.speech_to_text_model import SpeechToTextModel
20
20
  from ..types.speech_to_text_response import SpeechToTextResponse
21
21
  from ..types.speech_to_text_translate_model import SpeechToTextTranslateModel
22
22
  from ..types.speech_to_text_translate_response import SpeechToTextTranslateResponse
23
+ from .types.speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
24
+ from .types.speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
23
25
 
24
26
  # this is used as the default value for optional parameters
25
27
  OMIT = typing.cast(typing.Any, ...)
@@ -35,6 +37,7 @@ class RawSpeechToTextClient:
35
37
  file: core.File,
36
38
  model: typing.Optional[SpeechToTextModel] = OMIT,
37
39
  language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
40
+ input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
38
41
  request_options: typing.Optional[RequestOptions] = None,
39
42
  ) -> HttpResponse[SpeechToTextResponse]:
40
43
  """
@@ -66,6 +69,9 @@ class RawSpeechToTextClient:
66
69
  For the `saarika:v2.5` model, it is optional.
67
70
  `unknown`: Use this when the language is not known; the API will detect it automatically.
68
71
 
72
+ input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
73
+ Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
74
+
69
75
  request_options : typing.Optional[RequestOptions]
70
76
  Request-specific configuration.
71
77
 
@@ -81,6 +87,7 @@ class RawSpeechToTextClient:
81
87
  data={
82
88
  "model": model,
83
89
  "language_code": language_code,
90
+ "input_audio_codec": input_audio_codec,
84
91
  },
85
92
  files={
86
93
  "file": file,
@@ -176,6 +183,7 @@ class RawSpeechToTextClient:
176
183
  file: core.File,
177
184
  prompt: typing.Optional[str] = OMIT,
178
185
  model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
186
+ input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
179
187
  request_options: typing.Optional[RequestOptions] = None,
180
188
  ) -> HttpResponse[SpeechToTextTranslateResponse]:
181
189
  """
@@ -199,11 +207,14 @@ class RawSpeechToTextClient:
199
207
  See core.File for more documentation
200
208
 
201
209
  prompt : typing.Optional[str]
202
- Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesnt match the prompt performance of large language models.
210
+ Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
203
211
 
204
212
  model : typing.Optional[SpeechToTextTranslateModel]
205
213
  Model to be used for converting speech to text in target language
206
214
 
215
+ input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
216
+ Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
217
+
207
218
  request_options : typing.Optional[RequestOptions]
208
219
  Request-specific configuration.
209
220
 
@@ -219,6 +230,7 @@ class RawSpeechToTextClient:
219
230
  data={
220
231
  "prompt": prompt,
221
232
  "model": model,
233
+ "input_audio_codec": input_audio_codec,
222
234
  },
223
235
  files={
224
236
  "file": file,
@@ -319,6 +331,7 @@ class AsyncRawSpeechToTextClient:
319
331
  file: core.File,
320
332
  model: typing.Optional[SpeechToTextModel] = OMIT,
321
333
  language_code: typing.Optional[SpeechToTextLanguage] = OMIT,
334
+ input_audio_codec: typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec] = OMIT,
322
335
  request_options: typing.Optional[RequestOptions] = None,
323
336
  ) -> AsyncHttpResponse[SpeechToTextResponse]:
324
337
  """
@@ -350,6 +363,9 @@ class AsyncRawSpeechToTextClient:
350
363
  For the `saarika:v2.5` model, it is optional.
351
364
  `unknown`: Use this when the language is not known; the API will detect it automatically.
352
365
 
366
+ input_audio_codec : typing.Optional[SpeechToTextTranscribeRequestInputAudioCodec]
367
+ Input Audio codec/format of the input file. PCM files are supported only at 16kHz sample rate.
368
+
353
369
  request_options : typing.Optional[RequestOptions]
354
370
  Request-specific configuration.
355
371
 
@@ -365,6 +381,7 @@ class AsyncRawSpeechToTextClient:
365
381
  data={
366
382
  "model": model,
367
383
  "language_code": language_code,
384
+ "input_audio_codec": input_audio_codec,
368
385
  },
369
386
  files={
370
387
  "file": file,
@@ -460,6 +477,7 @@ class AsyncRawSpeechToTextClient:
460
477
  file: core.File,
461
478
  prompt: typing.Optional[str] = OMIT,
462
479
  model: typing.Optional[SpeechToTextTranslateModel] = OMIT,
480
+ input_audio_codec: typing.Optional[SpeechToTextTranslateRequestInputAudioCodec] = OMIT,
463
481
  request_options: typing.Optional[RequestOptions] = None,
464
482
  ) -> AsyncHttpResponse[SpeechToTextTranslateResponse]:
465
483
  """
@@ -483,11 +501,14 @@ class AsyncRawSpeechToTextClient:
483
501
  See core.File for more documentation
484
502
 
485
503
  prompt : typing.Optional[str]
486
- Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesnt match the prompt performance of large language models.
504
+ Conversation context can be passed as a prompt to boost model accuracy. However, the current system is at an experimentation stage and doesn't match the prompt performance of large language models.
487
505
 
488
506
  model : typing.Optional[SpeechToTextTranslateModel]
489
507
  Model to be used for converting speech to text in target language
490
508
 
509
+ input_audio_codec : typing.Optional[SpeechToTextTranslateRequestInputAudioCodec]
510
+ Audio codec/format of the input file. Our API automatically detects all codec formats, but for PCM files specifically (pcm_s16le, pcm_l16, pcm_raw), you must pass this parameter. PCM files are supported only at 16kHz sample rate.
511
+
491
512
  request_options : typing.Optional[RequestOptions]
492
513
  Request-specific configuration.
493
514
 
@@ -503,6 +524,7 @@ class AsyncRawSpeechToTextClient:
503
524
  data={
504
525
  "prompt": prompt,
505
526
  "model": model,
527
+ "input_audio_codec": input_audio_codec,
506
528
  },
507
529
  files={
508
530
  "file": file,
@@ -0,0 +1,8 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ # isort: skip_file
4
+
5
+ from .speech_to_text_transcribe_request_input_audio_codec import SpeechToTextTranscribeRequestInputAudioCodec
6
+ from .speech_to_text_translate_request_input_audio_codec import SpeechToTextTranslateRequestInputAudioCodec
7
+
8
+ __all__ = ["SpeechToTextTranscribeRequestInputAudioCodec", "SpeechToTextTranslateRequestInputAudioCodec"]
@@ -0,0 +1,33 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ SpeechToTextTranscribeRequestInputAudioCodec = typing.Union[
6
+ typing.Literal[
7
+ "wav",
8
+ "x-wav",
9
+ "wave",
10
+ "mp3",
11
+ "mpeg",
12
+ "mpeg3",
13
+ "x-mp3",
14
+ "x-mpeg-3",
15
+ "aac",
16
+ "x-aac",
17
+ "aiff",
18
+ "x-aiff",
19
+ "ogg",
20
+ "opus",
21
+ "flac",
22
+ "x-flac",
23
+ "mp4",
24
+ "x-m4a",
25
+ "amr",
26
+ "x-ms-wma",
27
+ "webm",
28
+ "pcm_s16le",
29
+ "pcm_l16",
30
+ "pcm_raw",
31
+ ],
32
+ typing.Any,
33
+ ]
@@ -0,0 +1,33 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ SpeechToTextTranslateRequestInputAudioCodec = typing.Union[
6
+ typing.Literal[
7
+ "wav",
8
+ "x-wav",
9
+ "wave",
10
+ "mp3",
11
+ "mpeg",
12
+ "mpeg3",
13
+ "x-mp3",
14
+ "x-mpeg-3",
15
+ "aac",
16
+ "x-aac",
17
+ "aiff",
18
+ "x-aiff",
19
+ "ogg",
20
+ "opus",
21
+ "flac",
22
+ "x-flac",
23
+ "mp4",
24
+ "x-m4a",
25
+ "amr",
26
+ "x-ms-wma",
27
+ "webm",
28
+ "pcm_s16le",
29
+ "pcm_l16",
30
+ "pcm_raw",
31
+ ],
32
+ typing.Any,
33
+ ]
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: sarvamai
3
- Version: 0.1.16a0
3
+ Version: 0.1.17a0
4
4
  Summary:
5
5
  Requires-Python: >=3.8,<4.0
6
6
  Classifier: Intended Audience :: Developers
@@ -1,11 +1,11 @@
1
- sarvamai/__init__.py,sha256=o9bK5DhrKYzmhZbOvKNNl-Fcif9og4eynOp47Dna4vY,10650
1
+ sarvamai/__init__.py,sha256=_AvjD4uJaddFuA7gBmWXoGqU9JEo9mHpVq-cqk47jws,10871
2
2
  sarvamai/chat/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
3
3
  sarvamai/chat/client.py,sha256=xOSj83Gr6Q7eY2qUeATiuXYQqBqWqSCQlIEopK5fKus,11022
4
4
  sarvamai/chat/raw_client.py,sha256=A2kRuZcVWlJhyYCD7YKgqNkZEp3cYa1731KhRkhirU0,17885
5
5
  sarvamai/client.py,sha256=J30X_os1lPf8Wml0KDFEf6p8VGHhgF_lf3nw1T2D3qo,8207
6
6
  sarvamai/core/__init__.py,sha256=YE2CtXeASe1RAbaI39twKWYKCuT4tW5is9HWHhJjR_g,1653
7
7
  sarvamai/core/api_error.py,sha256=44vPoTyWN59gonCIZMdzw7M1uspygiLnr3GNFOoVL2Q,614
8
- sarvamai/core/client_wrapper.py,sha256=ya3guU64Ap76A6U2WEq0HFnOz6Ds_qLLmIqk0NgoRWE,2570
8
+ sarvamai/core/client_wrapper.py,sha256=OJ4aojUSt1w1Qnvo0h5xwdvbFZKBybSJqwWw9tvkbVM,2570
9
9
  sarvamai/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
10
10
  sarvamai/core/events.py,sha256=HvKBdSoYcFetk7cgNXb7FxuY-FtY8NtUhZIN7mGVx8U,1159
11
11
  sarvamai/core/file.py,sha256=d4NNbX8XvXP32z8KpK2Xovv33nFfruIrpz0QWxlgpZk,2663
@@ -83,9 +83,12 @@ sarvamai/requests/timestamps_model.py,sha256=LkaYC5poipZe4r5-7Tp_Y058RXYWXwocn46
83
83
  sarvamai/requests/transcription_metrics.py,sha256=FDclX2Z9Z3azrDXxtZW8xbkxxWMZQXEbaYMMheAAAdk,321
84
84
  sarvamai/requests/translation_response.py,sha256=8iwQeZB1purHY757bIQI-n9QeVRBItaAVcBJ_la-k1Y,414
85
85
  sarvamai/requests/transliteration_response.py,sha256=KqRkqnegLmt7LjdVxjRePX6RoqaLm64KFGZ6q7mXyfw,426
86
- sarvamai/speech_to_text/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
87
- sarvamai/speech_to_text/client.py,sha256=v9p6bqRbA9h3f4wuylPTVDwetG7MCOlEtmrqW18KHCU,10857
88
- sarvamai/speech_to_text/raw_client.py,sha256=IhflpQX1P2kvZw8L2qbneMbRs5gtr9ye6ebvEK-hn3A,25185
86
+ sarvamai/speech_to_text/__init__.py,sha256=U3RTBb6u3QwsV3fyh-RX_HQ3AI2BD0i-Y8ec3gqTjdI,301
87
+ sarvamai/speech_to_text/client.py,sha256=Z_MZH-W70BbNYkax_CV3ZIhgxT0d44rnXtN3yLTMIzc,12797
88
+ sarvamai/speech_to_text/raw_client.py,sha256=kHU4ngENfwqrKBZq0GZwAlYONPWf3ssH3hQvJhEyEXg,27081
89
+ sarvamai/speech_to_text/types/__init__.py,sha256=HIUDEN0mwbyXM15kMPzNydaxB9FEY_Fjzj5RsMFyef4,410
90
+ sarvamai/speech_to_text/types/speech_to_text_transcribe_request_input_audio_codec.py,sha256=FRi5ofYQFd4TmDbAqZ0Dx2mL0F7FRovwJGn1OI8N4q0,592
91
+ sarvamai/speech_to_text/types/speech_to_text_translate_request_input_audio_codec.py,sha256=q5ZnWwicyFtdOX9EMDUkHipSIIAMstCVYyjesminEWM,591
89
92
  sarvamai/speech_to_text_job/__init__.py,sha256=_VhToAyIt_5axN6CLJwtxg3-CO7THa_23pbUzqhXJa4,85
90
93
  sarvamai/speech_to_text_job/client.py,sha256=WSGBJxYcNxl77Zd1X6VVWjg4zshqecXf6WCyhfLXVlI,18007
91
94
  sarvamai/speech_to_text_job/job.py,sha256=K8HOmwrYd6l82-MZfWDBmNkZeeERyg9YOihnFfvl-Js,15021
@@ -210,6 +213,6 @@ sarvamai/types/transliterate_mode.py,sha256=1jSEMlGcoLkWuk12TgoOpSgwifa4rThGKZ1h
210
213
  sarvamai/types/transliterate_source_language.py,sha256=bSY9wJszF0sg-Cgg6F-YcWC8ly1mIlj9rqa15-jBtx8,283
211
214
  sarvamai/types/transliteration_response.py,sha256=yt-lzTbDeJ_ZL4I8kQa6oESxA9ebeJJY7LfFHpdEsmM,815
212
215
  sarvamai/version.py,sha256=Qkp3Ee9YH-O9RTix90e0i7iNrFAGN-QDt2AFwGA4n8k,75
213
- sarvamai-0.1.16a0.dist-info/METADATA,sha256=kJFllmgiNTGqzAiFZ0nntIXPASSibBS3sRNTIcdaWhk,26753
214
- sarvamai-0.1.16a0.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
215
- sarvamai-0.1.16a0.dist-info/RECORD,,
216
+ sarvamai-0.1.17a0.dist-info/METADATA,sha256=Frl6Uw6tD1u701FFbq4j6anJgmBbbNDmZr_ZNZlZfh8,26753
217
+ sarvamai-0.1.17a0.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
218
+ sarvamai-0.1.17a0.dist-info/RECORD,,