smallestai 3.1.0__py3-none-any.whl → 4.0.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of smallestai might be problematic. Click here for more details.
- smallestai/__init__.py +1 -1
- smallestai/waves/__init__.py +2 -2
- smallestai/waves/async_waves_client.py +42 -69
- smallestai/waves/stream_tts.py +189 -254
- smallestai/waves/utils.py +3 -49
- smallestai/waves/waves_client.py +41 -69
- {smallestai-3.1.0.dist-info → smallestai-4.0.0.dist-info}/METADATA +2 -1
- {smallestai-3.1.0.dist-info → smallestai-4.0.0.dist-info}/RECORD +11 -11
- {smallestai-3.1.0.dist-info → smallestai-4.0.0.dist-info}/WHEEL +1 -1
- {smallestai-3.1.0.dist-info → smallestai-4.0.0.dist-info}/licenses/LICENSE +0 -0
- {smallestai-3.1.0.dist-info → smallestai-4.0.0.dist-info}/top_level.txt +0 -0
smallestai/__init__.py
CHANGED
smallestai/waves/__init__.py
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
from smallestai.waves.waves_client import WavesClient
|
|
2
2
|
from smallestai.waves.async_waves_client import AsyncWavesClient
|
|
3
|
-
from smallestai.waves.stream_tts import
|
|
3
|
+
from smallestai.waves.stream_tts import WavesStreamingTTS, TTSConfig
|
|
4
4
|
|
|
5
|
-
__all__ = ["WavesClient", "AsyncWavesClient", "
|
|
5
|
+
__all__ = ["WavesClient", "AsyncWavesClient", "WavesStreamingTTS", "TTSConfig"]
|
|
@@ -4,10 +4,10 @@ import json
|
|
|
4
4
|
import aiohttp
|
|
5
5
|
import aiofiles
|
|
6
6
|
import requests
|
|
7
|
-
from typing import Optional, Union, List
|
|
7
|
+
from typing import Optional, Union, List
|
|
8
8
|
|
|
9
9
|
from smallestai.waves.exceptions import TTSError, APIError
|
|
10
|
-
from smallestai.waves.utils import (TTSOptions, validate_input,
|
|
10
|
+
from smallestai.waves.utils import (TTSOptions, validate_input,
|
|
11
11
|
get_smallest_languages, get_smallest_models, ALLOWED_AUDIO_EXTENSIONS, API_BASE_URL)
|
|
12
12
|
|
|
13
13
|
|
|
@@ -22,7 +22,8 @@ class AsyncWavesClient:
|
|
|
22
22
|
consistency: Optional[float] = 0.5,
|
|
23
23
|
similarity: Optional[float] = 0.0,
|
|
24
24
|
enhancement: Optional[int] = 1,
|
|
25
|
-
|
|
25
|
+
language: Optional[str] = "en",
|
|
26
|
+
output_format: Optional[str] = "wav"
|
|
26
27
|
) -> None:
|
|
27
28
|
"""
|
|
28
29
|
AsyncSmallest Instance for asynchronous text-to-speech synthesis.
|
|
@@ -40,7 +41,8 @@ class AsyncWavesClient:
|
|
|
40
41
|
- consistency (float): This parameter controls word repetition and skipping. Decrease it to prevent skipped words, and increase it to prevent repetition. Only supported in `lightning-large` model. Range - [0, 1]
|
|
41
42
|
- similarity (float): This parameter controls the similarity between the synthesized audio and the reference audio. Increase it to make the speech more similar to the reference audio. Only supported in `lightning-large` model. Range - [0, 1]
|
|
42
43
|
- enhancement (int): Enhances speech quality at the cost of increased latency. Only supported in `lightning-large` model. Range - [0, 2].
|
|
43
|
-
-
|
|
44
|
+
- language (str): The language for synthesis. Default is "en".
|
|
45
|
+
- output_format (str): The output audio format. Options: "pcm", "mp3", "wav", "mulaw". Default is "pcm".
|
|
44
46
|
|
|
45
47
|
Methods:
|
|
46
48
|
- get_languages: Returns a list of available languages for synthesis.
|
|
@@ -61,11 +63,12 @@ class AsyncWavesClient:
|
|
|
61
63
|
sample_rate=sample_rate,
|
|
62
64
|
voice_id=voice_id,
|
|
63
65
|
api_key=self.api_key,
|
|
64
|
-
add_wav_header=add_wav_header,
|
|
65
66
|
speed=speed,
|
|
66
67
|
consistency=consistency,
|
|
67
68
|
similarity=similarity,
|
|
68
|
-
enhancement=enhancement
|
|
69
|
+
enhancement=enhancement,
|
|
70
|
+
language=language,
|
|
71
|
+
output_format=output_format
|
|
69
72
|
)
|
|
70
73
|
self.session = None
|
|
71
74
|
|
|
@@ -130,18 +133,14 @@ class AsyncWavesClient:
|
|
|
130
133
|
async def synthesize(
|
|
131
134
|
self,
|
|
132
135
|
text: str,
|
|
133
|
-
stream: Optional[bool] = False,
|
|
134
|
-
save_as: Optional[str] = None,
|
|
135
136
|
**kwargs
|
|
136
|
-
) -> Union[bytes
|
|
137
|
+
) -> Union[bytes]:
|
|
137
138
|
"""
|
|
138
139
|
Asynchronously synthesize speech from the provided text.
|
|
139
140
|
|
|
140
141
|
Args:
|
|
141
142
|
- text (str): The text to be converted to speech.
|
|
142
143
|
- stream (Optional[bool]): If True, returns an iterator yielding audio chunks instead of a full byte array.
|
|
143
|
-
- save_as (Optional[str]): If provided, the synthesized audio will be saved to this file path.
|
|
144
|
-
The file must have a .wav extension.
|
|
145
144
|
- kwargs: Additional optional parameters to override `__init__` options for this call.
|
|
146
145
|
|
|
147
146
|
Returns:
|
|
@@ -151,7 +150,7 @@ class AsyncWavesClient:
|
|
|
151
150
|
- Otherwise, returns the synthesized audio content as bytes.
|
|
152
151
|
|
|
153
152
|
Raises:
|
|
154
|
-
- TTSError: If the provided file name does not have a .wav extension when `save_as` is specified.
|
|
153
|
+
- TTSError: If the provided file name does not have a .wav or .mp3 extension when `save_as` is specified.
|
|
155
154
|
- APIError: If the API request fails or returns an error.
|
|
156
155
|
- ValueError: If an unexpected parameter is passed in `kwargs`.
|
|
157
156
|
"""
|
|
@@ -172,65 +171,40 @@ class AsyncWavesClient:
|
|
|
172
171
|
for key, value in kwargs.items():
|
|
173
172
|
setattr(opts, key, value)
|
|
174
173
|
|
|
175
|
-
text = preprocess_text(text)
|
|
176
174
|
validate_input(text, opts.model, opts.sample_rate, opts.speed, opts.consistency, opts.similarity, opts.enhancement)
|
|
177
175
|
|
|
178
|
-
|
|
179
|
-
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
189
|
-
"voice_id": opts.voice_id,
|
|
190
|
-
"add_wav_header": False,
|
|
191
|
-
"speed": opts.speed,
|
|
192
|
-
"model": opts.model
|
|
193
|
-
}
|
|
194
|
-
|
|
195
|
-
if opts.model == "lightning-large" or opts.model == "lightning-v2":
|
|
196
|
-
if opts.consistency is not None:
|
|
197
|
-
payload["consistency"] = opts.consistency
|
|
198
|
-
if opts.similarity is not None:
|
|
199
|
-
payload["similarity"] = opts.similarity
|
|
200
|
-
if opts.enhancement is not None:
|
|
201
|
-
payload["enhancement"] = opts.enhancement
|
|
202
|
-
|
|
203
|
-
|
|
204
|
-
headers = {
|
|
205
|
-
"Authorization": f"Bearer {self.api_key}",
|
|
206
|
-
"Content-Type": "application/json",
|
|
207
|
-
}
|
|
208
|
-
|
|
209
|
-
async with self.session.post(f"{API_BASE_URL}/{opts.model}/get_speech", json=payload, headers=headers) as res:
|
|
210
|
-
if res.status != 200:
|
|
211
|
-
raise APIError(f"Failed to synthesize speech: {await res.text()}. For more information, visit https://waves.smallest.ai/")
|
|
212
|
-
|
|
213
|
-
yield await res.read()
|
|
176
|
+
payload = {
|
|
177
|
+
"text": text,
|
|
178
|
+
"voice_id": opts.voice_id,
|
|
179
|
+
"sample_rate": opts.sample_rate,
|
|
180
|
+
"speed": opts.speed,
|
|
181
|
+
"consistency": opts.consistency,
|
|
182
|
+
"similarity": opts.similarity,
|
|
183
|
+
"enhancement": opts.enhancement,
|
|
184
|
+
"language": opts.language,
|
|
185
|
+
"output_format": opts.output_format
|
|
186
|
+
}
|
|
214
187
|
|
|
215
|
-
if
|
|
216
|
-
|
|
217
|
-
|
|
218
|
-
|
|
219
|
-
|
|
220
|
-
|
|
221
|
-
|
|
222
|
-
|
|
223
|
-
|
|
224
|
-
|
|
225
|
-
|
|
226
|
-
|
|
227
|
-
return None
|
|
228
|
-
|
|
229
|
-
if opts.add_wav_header:
|
|
230
|
-
return add_wav_header(audio_content, opts.sample_rate)
|
|
188
|
+
if opts.model == "lightning-large" or opts.model == "lightning-v2":
|
|
189
|
+
if opts.consistency is not None:
|
|
190
|
+
payload["consistency"] = opts.consistency
|
|
191
|
+
if opts.similarity is not None:
|
|
192
|
+
payload["similarity"] = opts.similarity
|
|
193
|
+
if opts.enhancement is not None:
|
|
194
|
+
payload["enhancement"] = opts.enhancement
|
|
195
|
+
|
|
196
|
+
headers = {
|
|
197
|
+
"Authorization": f"Bearer {self.api_key}",
|
|
198
|
+
"Content-Type": "application/json",
|
|
199
|
+
}
|
|
231
200
|
|
|
232
|
-
|
|
201
|
+
async with self.session.post(f"{API_BASE_URL}/{opts.model}/get_speech", json=payload, headers=headers) as res:
|
|
202
|
+
if res.status != 200:
|
|
203
|
+
raise APIError(f"Failed to synthesize speech: {await res.text()}. For more information, visit https://waves.smallest.ai/")
|
|
204
|
+
|
|
205
|
+
audio_bytes = await res.content.read()
|
|
233
206
|
|
|
207
|
+
return audio_bytes
|
|
234
208
|
finally:
|
|
235
209
|
if should_cleanup and self.session:
|
|
236
210
|
await self.session.close()
|
|
@@ -316,9 +290,8 @@ class AsyncWavesClient:
|
|
|
316
290
|
if res.status != 200:
|
|
317
291
|
raise APIError(f"Failed to delete voice: {await res.text()}. For more information, visit https://waves.smallest.ai/")
|
|
318
292
|
|
|
319
|
-
return await res.
|
|
320
|
-
|
|
293
|
+
return json.dumps(await res.json(), indent=4, ensure_ascii=False)
|
|
321
294
|
finally:
|
|
322
295
|
if should_cleanup and self.session:
|
|
323
296
|
await self.session.close()
|
|
324
|
-
self.session = None
|
|
297
|
+
self.session = None
|
smallestai/waves/stream_tts.py
CHANGED
|
@@ -1,272 +1,207 @@
|
|
|
1
|
-
import
|
|
1
|
+
import json
|
|
2
|
+
import base64
|
|
2
3
|
import time
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
from typing import
|
|
6
|
-
|
|
7
|
-
from
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
- Low latency between text generation and speech output.
|
|
32
|
-
|
|
33
|
-
Args:
|
|
34
|
-
tts_instance: The text-to-speech engine to use (Smallest or AsyncSmallest)
|
|
35
|
-
queue_timeout: How long to wait for new text (seconds, default: 1.0)
|
|
36
|
-
max_retries: Number of retry attempts for failed synthesis (default: 3)
|
|
37
|
-
verbose: Whether to log detailed metrics about TTS requests (default: False)
|
|
38
|
-
"""
|
|
39
|
-
self.tts_instance = tts_instance
|
|
40
|
-
self.tts_instance.opts.add_wav_header = False
|
|
41
|
-
self.sentence_end_regex = SENTENCE_END_REGEX
|
|
42
|
-
self.queue_timeout = queue_timeout
|
|
43
|
-
self.max_retries = max_retries
|
|
44
|
-
self.queue = Queue()
|
|
45
|
-
self.buffer_size = 250
|
|
46
|
-
self.stop_flag = False
|
|
47
|
-
self.verbose = verbose
|
|
4
|
+
import threading
|
|
5
|
+
import queue
|
|
6
|
+
from typing import Generator
|
|
7
|
+
from dataclasses import dataclass
|
|
8
|
+
from websocket import WebSocketApp
|
|
9
|
+
|
|
10
|
+
@dataclass
|
|
11
|
+
class TTSConfig:
|
|
12
|
+
voice_id: str
|
|
13
|
+
api_key: str
|
|
14
|
+
language: str = "en"
|
|
15
|
+
sample_rate: int = 24000
|
|
16
|
+
speed: float = 1.0
|
|
17
|
+
consistency: float = 0.5
|
|
18
|
+
enhancement: int = 1
|
|
19
|
+
similarity: float = 0
|
|
20
|
+
max_buffer_flush_ms: int = 0
|
|
21
|
+
|
|
22
|
+
class WavesStreamingTTS:
|
|
23
|
+
def __init__(self, config: TTSConfig):
|
|
24
|
+
self.config = config
|
|
25
|
+
self.ws_url = "wss://waves-api.smallest.ai/api/v1/lightning-v2/get_speech/stream"
|
|
26
|
+
self.ws = None
|
|
27
|
+
self.audio_queue = queue.Queue()
|
|
28
|
+
self.error_queue = queue.Queue()
|
|
29
|
+
self.is_complete = False
|
|
30
|
+
self.is_connected = False
|
|
31
|
+
self.request_id = None
|
|
48
32
|
|
|
49
|
-
|
|
50
|
-
self.
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
self.
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
async for chunk in llm_output:
|
|
71
|
-
buffer += chunk
|
|
72
|
-
|
|
73
|
-
while len(buffer) > self.buffer_size:
|
|
74
|
-
chunk_text = buffer[:self.buffer_size]
|
|
75
|
-
last_break_index = -1
|
|
76
|
-
|
|
77
|
-
# Find last sentence boundary using regex
|
|
78
|
-
for i in range(len(chunk_text) - 1, -1, -1):
|
|
79
|
-
if self.sentence_end_regex.match(chunk_text[:i + 1]):
|
|
80
|
-
last_break_index = i
|
|
81
|
-
break
|
|
82
|
-
|
|
83
|
-
if last_break_index == -1:
|
|
84
|
-
# Fallback to space if no sentence boundary found
|
|
85
|
-
last_space = chunk_text.rfind(' ')
|
|
86
|
-
if last_space != -1:
|
|
87
|
-
last_break_index = last_space
|
|
88
|
-
else:
|
|
89
|
-
last_break_index = self.buffer_size - 1
|
|
90
|
-
|
|
91
|
-
# Add chunk to queue and update buffer
|
|
92
|
-
self.queue.put(f'{buffer[:last_break_index + 1].replace("—", " ").strip()} ')
|
|
93
|
-
buffer = buffer[last_break_index + 1:].strip()
|
|
94
|
-
|
|
95
|
-
# Don't forget the remaining text
|
|
96
|
-
if buffer:
|
|
97
|
-
self.queue.put(f'{buffer.replace("—", " ").strip()} ')
|
|
98
|
-
|
|
99
|
-
self.stop_flag = True
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
def _synthesize_sync(self, sentence: str, retries: int = 0) -> Optional[bytes]:
|
|
103
|
-
"""Synchronously synthesizes a given sentence."""
|
|
104
|
-
request_start_time = time.time()
|
|
105
|
-
request_id = self.request_count + 1
|
|
33
|
+
def _get_headers(self):
|
|
34
|
+
return [f"Authorization: Bearer {self.config.api_key}"]
|
|
35
|
+
|
|
36
|
+
def _create_payload(self, text: str, continue_stream: bool = False, flush: bool = False):
|
|
37
|
+
return {
|
|
38
|
+
"voice_id": self.config.voice_id,
|
|
39
|
+
"text": text,
|
|
40
|
+
"language": self.config.language,
|
|
41
|
+
"sample_rate": self.config.sample_rate,
|
|
42
|
+
"speed": self.config.speed,
|
|
43
|
+
"consistency": self.config.consistency,
|
|
44
|
+
"similarity": self.config.similarity,
|
|
45
|
+
"enhancement": self.config.enhancement,
|
|
46
|
+
"max_buffer_flush_ms": self.config.max_buffer_flush_ms,
|
|
47
|
+
"continue": continue_stream,
|
|
48
|
+
"flush": flush
|
|
49
|
+
}
|
|
50
|
+
|
|
51
|
+
def _on_open(self, ws):
|
|
52
|
+
self.is_connected = True
|
|
106
53
|
|
|
54
|
+
def _on_message(self, ws, message):
|
|
107
55
|
try:
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
request_end_time = time.time()
|
|
56
|
+
data = json.loads(message)
|
|
57
|
+
status = data.get("status", "")
|
|
111
58
|
|
|
112
|
-
if
|
|
113
|
-
|
|
114
|
-
|
|
115
|
-
self.first_api_response_time = time.time() - self.start_time
|
|
59
|
+
if status == "error":
|
|
60
|
+
self.error_queue.put(Exception(data.get("message", "Unknown error")))
|
|
61
|
+
return
|
|
116
62
|
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
"text": sentence,
|
|
120
|
-
"start_time": request_start_time - self.start_time,
|
|
121
|
-
"end_time": request_end_time - self.start_time,
|
|
122
|
-
"duration": request_duration,
|
|
123
|
-
"char_count": len(sentence),
|
|
124
|
-
"retries": retries
|
|
125
|
-
})
|
|
63
|
+
if not self.request_id:
|
|
64
|
+
self.request_id = data.get("request_id")
|
|
126
65
|
|
|
127
|
-
|
|
128
|
-
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
133
|
-
|
|
134
|
-
|
|
135
|
-
|
|
136
|
-
|
|
66
|
+
audio_b64 = data.get("data", {}).get("audio")
|
|
67
|
+
if audio_b64:
|
|
68
|
+
self.audio_queue.put(base64.b64decode(audio_b64))
|
|
69
|
+
|
|
70
|
+
if status == "complete":
|
|
71
|
+
self.is_complete = True
|
|
72
|
+
self.audio_queue.put(None)
|
|
73
|
+
|
|
74
|
+
except Exception as e:
|
|
75
|
+
self.error_queue.put(e)
|
|
76
|
+
|
|
77
|
+
def _on_error(self, ws, error):
|
|
78
|
+
self.error_queue.put(error)
|
|
79
|
+
|
|
80
|
+
def _on_close(self, ws, *args):
|
|
81
|
+
self.is_connected = False
|
|
82
|
+
if not self.is_complete:
|
|
83
|
+
self.audio_queue.put(None)
|
|
84
|
+
|
|
85
|
+
def _connect(self):
|
|
86
|
+
if self.ws:
|
|
87
|
+
self.ws.close()
|
|
137
88
|
|
|
138
|
-
|
|
139
|
-
|
|
140
|
-
|
|
141
|
-
|
|
142
|
-
|
|
89
|
+
self.ws = WebSocketApp(
|
|
90
|
+
self.ws_url,
|
|
91
|
+
header=self._get_headers(),
|
|
92
|
+
on_open=self._on_open,
|
|
93
|
+
on_message=self._on_message,
|
|
94
|
+
on_error=self._on_error,
|
|
95
|
+
on_close=self._on_close
|
|
96
|
+
)
|
|
143
97
|
|
|
144
|
-
|
|
145
|
-
|
|
146
|
-
|
|
147
|
-
|
|
98
|
+
ws_thread = threading.Thread(target=self.ws.run_forever)
|
|
99
|
+
ws_thread.daemon = True
|
|
100
|
+
ws_thread.start()
|
|
101
|
+
|
|
102
|
+
timeout = 5.0
|
|
103
|
+
start_time = time.time()
|
|
104
|
+
while not self.is_connected and time.time() - start_time < timeout:
|
|
105
|
+
time.sleep(0.1)
|
|
148
106
|
|
|
149
|
-
|
|
150
|
-
|
|
151
|
-
|
|
152
|
-
|
|
153
|
-
|
|
154
|
-
|
|
155
|
-
|
|
156
|
-
|
|
157
|
-
|
|
158
|
-
|
|
159
|
-
|
|
160
|
-
|
|
161
|
-
|
|
162
|
-
})
|
|
107
|
+
if not self.is_connected:
|
|
108
|
+
raise Exception("Failed to connect to WebSocket")
|
|
109
|
+
|
|
110
|
+
def synthesize(self, text: str) -> Generator[bytes, None, None]:
|
|
111
|
+
self._reset_state()
|
|
112
|
+
self._connect()
|
|
113
|
+
|
|
114
|
+
payload = self._create_payload(text)
|
|
115
|
+
self.ws.send(json.dumps(payload))
|
|
116
|
+
|
|
117
|
+
while True:
|
|
118
|
+
if not self.error_queue.empty():
|
|
119
|
+
raise self.error_queue.get()
|
|
163
120
|
|
|
164
|
-
return audio_content
|
|
165
|
-
except APIError as e:
|
|
166
|
-
if retries < self.max_retries:
|
|
167
|
-
if self.verbose:
|
|
168
|
-
print(f"Retry {retries + 1}/{self.max_retries} for request: '{sentence[:30]}...'")
|
|
169
|
-
return await self._synthesize_async(sentence, retries + 1)
|
|
170
|
-
else:
|
|
171
|
-
if self.verbose:
|
|
172
|
-
print(f"Synthesis failed for sentence: {sentence} - Error: {e}. Retries Exhausted, for more information, visit https://waves.smallest.ai/")
|
|
173
|
-
return None
|
|
174
|
-
|
|
175
|
-
|
|
176
|
-
async def _run_synthesis(self) -> AsyncGenerator[bytes, None]:
|
|
177
|
-
"""
|
|
178
|
-
Continuously synthesizes sentences from the queue, yielding audio content.
|
|
179
|
-
If no sentences are in the queue, it waits until new data is available or streaming is complete.
|
|
180
|
-
"""
|
|
181
|
-
while not self.stop_flag or not self.queue.empty():
|
|
182
121
|
try:
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
if isinstance(self.tts_instance, AsyncWavesClient):
|
|
186
|
-
audio_content = await self._synthesize_async(sentence)
|
|
187
|
-
else:
|
|
188
|
-
loop = asyncio.get_running_loop()
|
|
189
|
-
audio_content = await loop.run_in_executor(None, self._synthesize_sync, sentence)
|
|
190
|
-
|
|
191
|
-
if audio_content:
|
|
192
|
-
yield audio_content
|
|
193
|
-
|
|
194
|
-
except Empty:
|
|
195
|
-
# Quick check if we should exit
|
|
196
|
-
if self.stop_flag and self.queue.empty():
|
|
122
|
+
chunk = self.audio_queue.get(timeout=1.0)
|
|
123
|
+
if chunk is None:
|
|
197
124
|
break
|
|
198
|
-
|
|
199
|
-
|
|
200
|
-
|
|
201
|
-
|
|
202
|
-
|
|
203
|
-
|
|
204
|
-
|
|
205
|
-
|
|
206
|
-
|
|
207
|
-
|
|
208
|
-
|
|
209
|
-
|
|
210
|
-
|
|
211
|
-
print(f"TEXT-TO-AUDIO STREAM METRICS")
|
|
212
|
-
print("="*100)
|
|
125
|
+
yield chunk
|
|
126
|
+
except queue.Empty:
|
|
127
|
+
if self.is_complete:
|
|
128
|
+
break
|
|
129
|
+
continue
|
|
130
|
+
|
|
131
|
+
self.ws.close()
|
|
132
|
+
|
|
133
|
+
def synthesize_streaming(self, text_stream: Generator[str, None, None],
|
|
134
|
+
continue_stream: bool = True,
|
|
135
|
+
auto_flush: bool = True) -> Generator[bytes, None, None]:
|
|
136
|
+
self._reset_state()
|
|
137
|
+
self._connect()
|
|
213
138
|
|
|
214
|
-
|
|
215
|
-
|
|
216
|
-
|
|
217
|
-
|
|
139
|
+
def send_text():
|
|
140
|
+
try:
|
|
141
|
+
for text_chunk in text_stream:
|
|
142
|
+
if text_chunk.strip():
|
|
143
|
+
payload = self._create_payload(text_chunk, continue_stream=continue_stream)
|
|
144
|
+
self.ws.send(json.dumps(payload))
|
|
145
|
+
|
|
146
|
+
if auto_flush:
|
|
147
|
+
flush_payload = self._create_payload("", flush=True)
|
|
148
|
+
self.ws.send(json.dumps(flush_payload))
|
|
149
|
+
except Exception as e:
|
|
150
|
+
self.error_queue.put(e)
|
|
218
151
|
|
|
219
|
-
|
|
220
|
-
|
|
221
|
-
|
|
222
|
-
print("\n" + header)
|
|
223
|
-
print("-" * 100)
|
|
152
|
+
sender_thread = threading.Thread(target=send_text)
|
|
153
|
+
sender_thread.daemon = True
|
|
154
|
+
sender_thread.start()
|
|
224
155
|
|
|
225
|
-
|
|
226
|
-
|
|
227
|
-
|
|
228
|
-
f"{log['id']:4} "
|
|
229
|
-
f"{log['start_time']:10.3f} "
|
|
230
|
-
f"{log['end_time']:10.3f} "
|
|
231
|
-
f"{log['duration']:12.3f} "
|
|
232
|
-
f"{log['char_count']:15} "
|
|
233
|
-
f"{log['text'][:50]}{'...' if len(log['text']) > 50 else ''}"
|
|
234
|
-
)
|
|
235
|
-
print(row)
|
|
236
|
-
|
|
237
|
-
# Print retry information if any
|
|
238
|
-
if log['retries'] > 0:
|
|
239
|
-
print(f"{'':4} {'':10} {'':10} {'':12} {'':15} Retries: {log['retries']}")
|
|
156
|
+
while True:
|
|
157
|
+
if not self.error_queue.empty():
|
|
158
|
+
raise self.error_queue.get()
|
|
240
159
|
|
|
241
|
-
|
|
242
|
-
|
|
243
|
-
|
|
244
|
-
|
|
245
|
-
|
|
246
|
-
|
|
247
|
-
|
|
248
|
-
|
|
249
|
-
|
|
250
|
-
|
|
251
|
-
|
|
252
|
-
|
|
253
|
-
|
|
254
|
-
|
|
255
|
-
|
|
256
|
-
|
|
257
|
-
|
|
258
|
-
|
|
259
|
-
|
|
260
|
-
|
|
261
|
-
|
|
262
|
-
|
|
263
|
-
|
|
264
|
-
|
|
265
|
-
|
|
266
|
-
|
|
267
|
-
|
|
268
|
-
|
|
269
|
-
llm_thread.join()
|
|
160
|
+
try:
|
|
161
|
+
chunk = self.audio_queue.get(timeout=1.0)
|
|
162
|
+
if chunk is None:
|
|
163
|
+
break
|
|
164
|
+
yield chunk
|
|
165
|
+
except queue.Empty:
|
|
166
|
+
if self.is_complete:
|
|
167
|
+
break
|
|
168
|
+
continue
|
|
169
|
+
|
|
170
|
+
self.ws.close()
|
|
171
|
+
|
|
172
|
+
def send_text_chunk(self, text: str, continue_stream: bool = True, flush: bool = False):
|
|
173
|
+
if not self.is_connected:
|
|
174
|
+
raise Exception("WebSocket not connected")
|
|
175
|
+
payload = self._create_payload(text, continue_stream=continue_stream, flush=flush)
|
|
176
|
+
self.ws.send(json.dumps(payload))
|
|
177
|
+
|
|
178
|
+
def flush_buffer(self):
|
|
179
|
+
if not self.is_connected:
|
|
180
|
+
raise Exception("WebSocket not connected")
|
|
181
|
+
payload = self._create_payload("", flush=True)
|
|
182
|
+
self.ws.send(json.dumps(payload))
|
|
183
|
+
|
|
184
|
+
def start_streaming_session(self) -> Generator[bytes, None, None]:
|
|
185
|
+
self._reset_state()
|
|
186
|
+
self._connect()
|
|
270
187
|
|
|
271
|
-
|
|
272
|
-
|
|
188
|
+
while True:
|
|
189
|
+
if not self.error_queue.empty():
|
|
190
|
+
raise self.error_queue.get()
|
|
191
|
+
|
|
192
|
+
try:
|
|
193
|
+
chunk = self.audio_queue.get(timeout=0.1)
|
|
194
|
+
if chunk is None:
|
|
195
|
+
break
|
|
196
|
+
yield chunk
|
|
197
|
+
except queue.Empty:
|
|
198
|
+
if self.is_complete:
|
|
199
|
+
break
|
|
200
|
+
continue
|
|
201
|
+
|
|
202
|
+
def _reset_state(self):
|
|
203
|
+
self.audio_queue = queue.Queue()
|
|
204
|
+
self.error_queue = queue.Queue()
|
|
205
|
+
self.is_complete = False
|
|
206
|
+
self.is_connected = False
|
|
207
|
+
self.request_id = None
|
smallestai/waves/utils.py
CHANGED
|
@@ -1,8 +1,5 @@
|
|
|
1
|
-
import re
|
|
2
|
-
import io
|
|
3
1
|
from typing import List
|
|
4
2
|
from typing import Optional
|
|
5
|
-
from pydub import AudioSegment
|
|
6
3
|
from dataclasses import dataclass
|
|
7
4
|
|
|
8
5
|
from smallestai.waves.exceptions import ValidationError
|
|
@@ -10,7 +7,7 @@ from smallestai.waves.models import TTSModels, TTSLanguages_lightning, TTSLangua
|
|
|
10
7
|
|
|
11
8
|
|
|
12
9
|
API_BASE_URL = "https://waves-api.smallest.ai/api/v1"
|
|
13
|
-
|
|
10
|
+
WEBSOCKET_URL = "wss://waves-api.smallest.ai/api/v1/lightning-v2/get_speech/stream"
|
|
14
11
|
SAMPLE_WIDTH = 2
|
|
15
12
|
CHANNELS = 1
|
|
16
13
|
ALLOWED_AUDIO_EXTENSIONS = ['.mp3', '.wav']
|
|
@@ -22,11 +19,12 @@ class TTSOptions:
|
|
|
22
19
|
sample_rate: int
|
|
23
20
|
voice_id: str
|
|
24
21
|
api_key: str
|
|
25
|
-
add_wav_header: bool
|
|
26
22
|
speed: float
|
|
27
23
|
consistency: float
|
|
28
24
|
similarity: float
|
|
29
25
|
enhancement: int
|
|
26
|
+
language: str
|
|
27
|
+
output_format: str
|
|
30
28
|
|
|
31
29
|
|
|
32
30
|
def validate_input(text: str, model: str, sample_rate: int, speed: float, consistency: Optional[float] = None, similarity: Optional[float] = None, enhancement: Optional[int] = None):
|
|
@@ -46,50 +44,6 @@ def validate_input(text: str, model: str, sample_rate: int, speed: float, consis
|
|
|
46
44
|
raise ValidationError(f"Invalid enhancement: {enhancement}. Must be between 0 and 2.")
|
|
47
45
|
|
|
48
46
|
|
|
49
|
-
def add_wav_header(frame_input: bytes, sample_rate: int = 24000, sample_width: int = 2, channels: int = 1) -> bytes:
|
|
50
|
-
audio = AudioSegment(data=frame_input, sample_width=sample_width, frame_rate=sample_rate, channels=channels)
|
|
51
|
-
wav_buf = io.BytesIO()
|
|
52
|
-
audio.export(wav_buf, format="wav")
|
|
53
|
-
wav_buf.seek(0)
|
|
54
|
-
return wav_buf.read()
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
def preprocess_text(text: str) -> str:
|
|
58
|
-
text = text.replace("\n", " ").replace("\t", " ")
|
|
59
|
-
text = re.sub(r'\s+', ' ', text)
|
|
60
|
-
return text.strip()
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
def chunk_text(text: str, chunk_size: int = 250) -> List[str]:
|
|
64
|
-
chunks = []
|
|
65
|
-
while text:
|
|
66
|
-
if len(text) <= chunk_size:
|
|
67
|
-
chunks.append(text.strip())
|
|
68
|
-
break
|
|
69
|
-
|
|
70
|
-
chunk_text = text[:chunk_size]
|
|
71
|
-
last_break_index = -1
|
|
72
|
-
|
|
73
|
-
# Find last sentence boundary using regex
|
|
74
|
-
for i in range(len(chunk_text) - 1, -1, -1):
|
|
75
|
-
if SENTENCE_END_REGEX.match(chunk_text[:i + 1]):
|
|
76
|
-
last_break_index = i
|
|
77
|
-
break
|
|
78
|
-
|
|
79
|
-
if last_break_index == -1:
|
|
80
|
-
# Fallback to space if no sentence boundary found
|
|
81
|
-
last_space = chunk_text.rfind(' ')
|
|
82
|
-
if last_space != -1:
|
|
83
|
-
last_break_index = last_space
|
|
84
|
-
else:
|
|
85
|
-
last_break_index = chunk_size - 1
|
|
86
|
-
|
|
87
|
-
chunks.append(text[:last_break_index + 1].strip())
|
|
88
|
-
text = text[last_break_index + 1:].strip()
|
|
89
|
-
|
|
90
|
-
return chunks
|
|
91
|
-
|
|
92
|
-
|
|
93
47
|
def get_smallest_languages(model: str = 'lightning') -> List[str]:
|
|
94
48
|
if model == 'lightning':
|
|
95
49
|
return TTSLanguages_lightning
|
smallestai/waves/waves_client.py
CHANGED
|
@@ -1,13 +1,12 @@
|
|
|
1
1
|
import os
|
|
2
2
|
import json
|
|
3
|
-
import wave
|
|
4
3
|
import copy
|
|
5
4
|
import requests
|
|
6
|
-
from typing import Optional, Union, List
|
|
5
|
+
from typing import Optional, Union, List
|
|
7
6
|
|
|
8
7
|
from smallestai.waves.exceptions import TTSError, APIError
|
|
9
|
-
from smallestai.waves.utils import (TTSOptions, validate_input,
|
|
10
|
-
get_smallest_languages, get_smallest_models, ALLOWED_AUDIO_EXTENSIONS, API_BASE_URL)
|
|
8
|
+
from smallestai.waves.utils import (TTSOptions, validate_input,
|
|
9
|
+
get_smallest_languages, get_smallest_models, ALLOWED_AUDIO_EXTENSIONS, API_BASE_URL)
|
|
11
10
|
|
|
12
11
|
class WavesClient:
|
|
13
12
|
def __init__(
|
|
@@ -20,7 +19,8 @@ class WavesClient:
|
|
|
20
19
|
consistency: Optional[float] = 0.5,
|
|
21
20
|
similarity: Optional[float] = 0.0,
|
|
22
21
|
enhancement: Optional[int] = 1,
|
|
23
|
-
|
|
22
|
+
language: Optional[str] = "en",
|
|
23
|
+
output_format: Optional[str] = "wav"
|
|
24
24
|
) -> None:
|
|
25
25
|
"""
|
|
26
26
|
Smallest Instance for text-to-speech synthesis.
|
|
@@ -37,7 +37,8 @@ class WavesClient:
|
|
|
37
37
|
- consistency (float): This parameter controls word repetition and skipping. Decrease it to prevent skipped words, and increase it to prevent repetition. Only supported in `lightning-large` model. Range - [0, 1]
|
|
38
38
|
- similarity (float): This parameter controls the similarity between the synthesized audio and the reference audio. Increase it to make the speech more similar to the reference audio. Only supported in `lightning-large` model. Range - [0, 1]
|
|
39
39
|
- enhancement (int): Enhances speech quality at the cost of increased latency. Only supported in `lightning-large` model. Range - [0, 2].
|
|
40
|
-
-
|
|
40
|
+
- language (str): The language for synthesis. Default is "en".
|
|
41
|
+
- output_format (str): The output audio format. Options: "pcm", "mp3", "wav", "mulaw". Default is "pcm".
|
|
41
42
|
|
|
42
43
|
Methods:
|
|
43
44
|
- get_languages: Returns a list of available languages for synthesis.
|
|
@@ -58,11 +59,12 @@ class WavesClient:
|
|
|
58
59
|
sample_rate=sample_rate,
|
|
59
60
|
voice_id=voice_id,
|
|
60
61
|
api_key=self.api_key,
|
|
61
|
-
add_wav_header=add_wav_header,
|
|
62
62
|
speed=speed,
|
|
63
63
|
consistency=consistency,
|
|
64
64
|
similarity=similarity,
|
|
65
|
-
enhancement=enhancement
|
|
65
|
+
enhancement=enhancement,
|
|
66
|
+
language=language,
|
|
67
|
+
output_format=output_format
|
|
66
68
|
)
|
|
67
69
|
|
|
68
70
|
|
|
@@ -107,17 +109,13 @@ class WavesClient:
|
|
|
107
109
|
def synthesize(
|
|
108
110
|
self,
|
|
109
111
|
text: str,
|
|
110
|
-
stream: Optional[bool] = False,
|
|
111
|
-
save_as: Optional[str] = None,
|
|
112
112
|
**kwargs
|
|
113
|
-
) -> Union[bytes
|
|
113
|
+
) -> Union[bytes]:
|
|
114
114
|
"""
|
|
115
115
|
Synthesize speech from the provided text.
|
|
116
116
|
|
|
117
117
|
- text (str): The text to be converted to speech.
|
|
118
118
|
- stream (Optional[bool]): If True, returns an iterator yielding audio chunks instead of a full byte array.
|
|
119
|
-
- save_as (Optional[str]): If provided, the synthesized audio will be saved to this file path.
|
|
120
|
-
The file must have a .wav extension.
|
|
121
119
|
- kwargs: Additional optional parameters to override `__init__` options for this call.
|
|
122
120
|
|
|
123
121
|
Returns:
|
|
@@ -127,7 +125,7 @@ class WavesClient:
|
|
|
127
125
|
- Otherwise, returns the synthesized audio content as bytes.
|
|
128
126
|
|
|
129
127
|
Raises:
|
|
130
|
-
- TTSError: If the provided file name does not have a .wav extension when `save_as` is specified.
|
|
128
|
+
- TTSError: If the provided file name does not have a .wav or .mp3 extension when `save_as` is specified.
|
|
131
129
|
- APIError: If the API request fails or returns an error.
|
|
132
130
|
"""
|
|
133
131
|
opts = copy.deepcopy(self.opts)
|
|
@@ -140,64 +138,38 @@ class WavesClient:
|
|
|
140
138
|
for key, value in kwargs.items():
|
|
141
139
|
setattr(opts, key, value)
|
|
142
140
|
|
|
143
|
-
text = preprocess_text(text)
|
|
144
141
|
validate_input(text, opts.model, opts.sample_rate, opts.speed, opts.consistency, opts.similarity, opts.enhancement)
|
|
145
142
|
|
|
146
|
-
|
|
147
|
-
|
|
148
|
-
|
|
149
|
-
|
|
150
|
-
|
|
151
|
-
|
|
152
|
-
|
|
153
|
-
|
|
154
|
-
|
|
155
|
-
|
|
156
|
-
|
|
157
|
-
"voice_id": opts.voice_id,
|
|
158
|
-
"add_wav_header": False,
|
|
159
|
-
"speed": opts.speed,
|
|
160
|
-
}
|
|
161
|
-
|
|
162
|
-
if opts.model == "lightning-large" or opts.model == "lightning-v2":
|
|
163
|
-
if opts.consistency is not None:
|
|
164
|
-
payload["consistency"] = opts.consistency
|
|
165
|
-
if opts.similarity is not None:
|
|
166
|
-
payload["similarity"] = opts.similarity
|
|
167
|
-
if opts.enhancement is not None:
|
|
168
|
-
payload["enhancement"] = opts.enhancement
|
|
169
|
-
|
|
170
|
-
headers = {
|
|
171
|
-
"Authorization": f"Bearer {self.api_key}",
|
|
172
|
-
"Content-Type": "application/json",
|
|
173
|
-
}
|
|
174
|
-
|
|
175
|
-
res = requests.post(f"{API_BASE_URL}/{opts.model}/get_speech", json=payload, headers=headers)
|
|
176
|
-
if res.status_code != 200:
|
|
177
|
-
raise APIError(f"Failed to synthesize speech: {res.text}. Please check if you have set the correct API key. For more information, visit https://waves.smallest.ai/")
|
|
143
|
+
payload = {
|
|
144
|
+
"text": text,
|
|
145
|
+
"voice_id": opts.voice_id,
|
|
146
|
+
"sample_rate": opts.sample_rate,
|
|
147
|
+
"speed": opts.speed,
|
|
148
|
+
"consistency": opts.consistency,
|
|
149
|
+
"similarity": opts.similarity,
|
|
150
|
+
"enhancement": opts.enhancement,
|
|
151
|
+
"language": opts.language,
|
|
152
|
+
"output_format": opts.output_format
|
|
153
|
+
}
|
|
178
154
|
|
|
179
|
-
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
155
|
+
if opts.model == "lightning-large" or opts.model == "lightning-v2":
|
|
156
|
+
if opts.consistency is not None:
|
|
157
|
+
payload["consistency"] = opts.consistency
|
|
158
|
+
if opts.similarity is not None:
|
|
159
|
+
payload["similarity"] = opts.similarity
|
|
160
|
+
if opts.enhancement is not None:
|
|
161
|
+
payload["enhancement"] = opts.enhancement
|
|
162
|
+
|
|
163
|
+
headers = {
|
|
164
|
+
"Authorization": f"Bearer {self.api_key}",
|
|
165
|
+
"Content-Type": "application/json",
|
|
166
|
+
}
|
|
185
167
|
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
wf.setnchannels(1)
|
|
192
|
-
wf.setsampwidth(2)
|
|
193
|
-
wf.setframerate(opts.sample_rate)
|
|
194
|
-
wf.writeframes(audio_content)
|
|
195
|
-
return None
|
|
196
|
-
|
|
197
|
-
if opts.add_wav_header:
|
|
198
|
-
return add_wav_header(audio_content, opts.sample_rate)
|
|
199
|
-
|
|
200
|
-
return audio_content
|
|
168
|
+
res = requests.post(f"{API_BASE_URL}/{opts.model}/get_speech", json=payload, headers=headers)
|
|
169
|
+
if res.status_code != 200:
|
|
170
|
+
raise APIError(f"Failed to synthesize speech: {res.text}. Please check if you have set the correct API key. For more information, visit https://waves.smallest.ai/")
|
|
171
|
+
|
|
172
|
+
return res.content
|
|
201
173
|
|
|
202
174
|
|
|
203
175
|
def add_voice(self, display_name: str, file_path: str) -> str:
|
|
@@ -262,4 +234,4 @@ class WavesClient:
|
|
|
262
234
|
if response.status_code != 200:
|
|
263
235
|
raise APIError(f"Failed to delete voice: {response.text}. For more information, visit https://waves.smallest.ai/")
|
|
264
236
|
|
|
265
|
-
return json.dumps(response.json(), indent=4, ensure_ascii=False)
|
|
237
|
+
return json.dumps(response.json(), indent=4, ensure_ascii=False)
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: smallestai
|
|
3
|
-
Version:
|
|
3
|
+
Version: 4.0.0
|
|
4
4
|
Summary: Official Python client for the Smallest AI API
|
|
5
5
|
Author-email: Smallest <support@smallest.ai>
|
|
6
6
|
License: MIT
|
|
@@ -16,6 +16,7 @@ Requires-Dist: aiohttp
|
|
|
16
16
|
Requires-Dist: aiofiles
|
|
17
17
|
Requires-Dist: requests
|
|
18
18
|
Requires-Dist: pydub
|
|
19
|
+
Requires-Dist: websocket-client
|
|
19
20
|
Requires-Dist: urllib3<3.0.0,>=1.25.3
|
|
20
21
|
Requires-Dist: python-dateutil>=2.8.2
|
|
21
22
|
Requires-Dist: pydantic>=2
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
smallestai/__init__.py,sha256=
|
|
1
|
+
smallestai/__init__.py,sha256=lY4DcFTosH2W0KVVN7pSbJmZBu067wG_y1u3GwGNru8,2624
|
|
2
2
|
smallestai/atoms/__init__.py,sha256=cn5_9tVsUwFQ_zdAZv263P4ow4N7dxRWCYAz82GjwuI,9342
|
|
3
3
|
smallestai/atoms/api_client.py,sha256=EcyN6nFp9U4u8TPJx3a9ZvbM2T4a9xrHGopQGLZuJpw,27448
|
|
4
4
|
smallestai/atoms/api_response.py,sha256=eMxw1mpmJcoGZ3gs9z6jM4oYoZ10Gjk333s9sKxGv7s,652
|
|
@@ -73,15 +73,15 @@ smallestai/atoms/models/update_agent_request_synthesizer_voice_config.py,sha256=
|
|
|
73
73
|
smallestai/atoms/models/update_agent_request_synthesizer_voice_config_one_of.py,sha256=8nGPcJ_CRUlXXjy3vCjpmbHWVBwQo2ebFP1K0MZPAsk,3955
|
|
74
74
|
smallestai/atoms/models/update_agent_request_synthesizer_voice_config_one_of1.py,sha256=9AJxgngoNSMvDbceajIqnG23PY4rw84coTh7yUTNS3c,3487
|
|
75
75
|
smallestai/atoms/models/upload_text_to_knowledge_base_request.py,sha256=Sxg0vRv_naT15odE8fBUeyjwLpEYOmQwGcJuzRRr90A,2587
|
|
76
|
-
smallestai/waves/__init__.py,sha256=
|
|
77
|
-
smallestai/waves/async_waves_client.py,sha256=
|
|
76
|
+
smallestai/waves/__init__.py,sha256=hxyqisgFiKiroxupuZeNXpXFIbnivmdgPrid3CnLhh0,268
|
|
77
|
+
smallestai/waves/async_waves_client.py,sha256=BgiSqd2UjwECCPwuh2dyhLSBP0inIsbPUEbduWTJrmI,11704
|
|
78
78
|
smallestai/waves/exceptions.py,sha256=nY6I8fCXe2By54CytQ0-i3hFiYtt8TYAKj0g6OYsCjc,585
|
|
79
79
|
smallestai/waves/models.py,sha256=FaMVkOFyNCVpWvyMCmqkv3t1wmnfCs1HIULxLr1L8XE,283
|
|
80
|
-
smallestai/waves/stream_tts.py,sha256=
|
|
81
|
-
smallestai/waves/utils.py,sha256=
|
|
82
|
-
smallestai/waves/waves_client.py,sha256=
|
|
83
|
-
smallestai-
|
|
84
|
-
smallestai-
|
|
85
|
-
smallestai-
|
|
86
|
-
smallestai-
|
|
87
|
-
smallestai-
|
|
80
|
+
smallestai/waves/stream_tts.py,sha256=c9r8mZuuFjbyWsUrlZ1jb0WNX7-lR39EXDUqyF-5g14,6792
|
|
81
|
+
smallestai/waves/utils.py,sha256=sqDpfa5SC60C_kJZo4MKxlDfkX7RRzO6aJ2hKpNMemE,2273
|
|
82
|
+
smallestai/waves/waves_client.py,sha256=U6aqClYL49cTtYisvpUVhas2miGZiCfqwTU0eDUY548,9770
|
|
83
|
+
smallestai-4.0.0.dist-info/licenses/LICENSE,sha256=kK3HNKhN7luQhkjkNWIvy9_gizbEDUM4mSv_HWq9uuM,1068
|
|
84
|
+
smallestai-4.0.0.dist-info/METADATA,sha256=bk0xBChPACeJiL8j6zxbHGnty60N5cjDH1sLPx32hLM,20424
|
|
85
|
+
smallestai-4.0.0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
86
|
+
smallestai-4.0.0.dist-info/top_level.txt,sha256=pdJzm1VC2J6RxoobATz45L9U3cki4AFLigsfvETz7Io,11
|
|
87
|
+
smallestai-4.0.0.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|