@mastra/voice-openai-realtime 0.0.0-working-memory-per-user-20250620163010 → 0.0.0-zod-v4-compat-part-2-20250822105954
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +161 -7
- package/LICENSE.md +11 -42
- package/dist/index.cjs +7 -1
- package/dist/index.cjs.map +1 -0
- package/dist/index.d.ts +325 -1
- package/dist/index.d.ts.map +1 -0
- package/dist/index.js +7 -1
- package/dist/index.js.map +1 -0
- package/dist/utils.d.ts +18 -0
- package/dist/utils.d.ts.map +1 -0
- package/package.json +13 -13
- package/src/index.test.ts +22 -28
- package/src/index.ts +5 -1
- package/tsconfig.build.json +9 -0
- package/tsconfig.json +1 -1
- package/tsup.config.ts +22 -0
- package/dist/_tsup-dts-rollup.d.cts +0 -348
- package/dist/_tsup-dts-rollup.d.ts +0 -348
- package/dist/index.d.cts +0 -1
package/dist/index.d.ts
CHANGED
|
@@ -1 +1,325 @@
|
|
|
1
|
-
|
|
1
|
+
import type { ToolsInput } from '@mastra/core/agent';
|
|
2
|
+
import type { RuntimeContext } from '@mastra/core/runtime-context';
|
|
3
|
+
import { MastraVoice } from '@mastra/core/voice';
|
|
4
|
+
import type { Realtime } from 'openai-realtime-api';
|
|
5
|
+
/**
|
|
6
|
+
* Event callback function type
|
|
7
|
+
*/
|
|
8
|
+
type EventCallback = (...args: any[]) => void;
|
|
9
|
+
type TTools = ToolsInput;
|
|
10
|
+
/**
|
|
11
|
+
* OpenAIRealtimeVoice provides real-time voice interaction capabilities using OpenAI's
|
|
12
|
+
* WebSocket-based API. It supports:
|
|
13
|
+
* - Real-time text-to-speech
|
|
14
|
+
* - Speech-to-text (transcription)
|
|
15
|
+
* - Voice activity detection
|
|
16
|
+
* - Multiple voice options
|
|
17
|
+
* - Event-based audio streaming
|
|
18
|
+
*
|
|
19
|
+
* The class manages WebSocket connections, audio streaming, and event handling
|
|
20
|
+
* for seamless voice interactions.
|
|
21
|
+
*
|
|
22
|
+
* @extends MastraVoice
|
|
23
|
+
*
|
|
24
|
+
* @example
|
|
25
|
+
* ```typescript
|
|
26
|
+
* const voice = new OpenAIRealtimeVoice({
|
|
27
|
+
* apiKey: process.env.OPENAI_API_KEY,
|
|
28
|
+
* model: 'gpt-4o-mini-realtime'
|
|
29
|
+
* });
|
|
30
|
+
*
|
|
31
|
+
* await voice.open();
|
|
32
|
+
* voice.on('speaking', (audioData) => {
|
|
33
|
+
* // Handle audio data
|
|
34
|
+
* });
|
|
35
|
+
*
|
|
36
|
+
* await voice.speak('Hello, how can I help you today?');
|
|
37
|
+
* ```
|
|
38
|
+
*/
|
|
39
|
+
export declare class OpenAIRealtimeVoice extends MastraVoice {
|
|
40
|
+
private options;
|
|
41
|
+
private ws?;
|
|
42
|
+
private state;
|
|
43
|
+
private client;
|
|
44
|
+
private events;
|
|
45
|
+
private instructions?;
|
|
46
|
+
private tools?;
|
|
47
|
+
private debug;
|
|
48
|
+
private queue;
|
|
49
|
+
private transcriber;
|
|
50
|
+
private runtimeContext?;
|
|
51
|
+
/**
|
|
52
|
+
* Creates a new instance of OpenAIRealtimeVoice.
|
|
53
|
+
*
|
|
54
|
+
* @param options - Configuration options for the voice instance
|
|
55
|
+
* @param options.url - The base URL for the OpenAI Realtime API
|
|
56
|
+
* @param options.model - The model ID to use (defaults to GPT-4 Mini Realtime)
|
|
57
|
+
* @param options.apiKey - OpenAI API key. Falls back to process.env.OPENAI_API_KEY
|
|
58
|
+
* @param options.speaker - Voice ID to use (defaults to 'alloy')
|
|
59
|
+
* @param options.debug - Enable debug mode
|
|
60
|
+
*
|
|
61
|
+
* @example
|
|
62
|
+
* ```typescript
|
|
63
|
+
* const voice = new OpenAIRealtimeVoice({
|
|
64
|
+
* apiKey: 'your-api-key',
|
|
65
|
+
* model: 'gpt-4o-mini-realtime',
|
|
66
|
+
* speaker: 'alloy'
|
|
67
|
+
* });
|
|
68
|
+
* ```
|
|
69
|
+
*/
|
|
70
|
+
constructor(options?: {
|
|
71
|
+
model?: string;
|
|
72
|
+
url?: string;
|
|
73
|
+
apiKey?: string;
|
|
74
|
+
speaker?: Realtime.Voice;
|
|
75
|
+
transcriber?: Realtime.AudioTranscriptionModel;
|
|
76
|
+
debug?: boolean;
|
|
77
|
+
});
|
|
78
|
+
/**
|
|
79
|
+
* Returns a list of available voice speakers.
|
|
80
|
+
*
|
|
81
|
+
* @returns Promise resolving to an array of voice objects, each containing at least a voiceId
|
|
82
|
+
*
|
|
83
|
+
* @example
|
|
84
|
+
* ```typescript
|
|
85
|
+
* const speakers = await voice.getSpeakers();
|
|
86
|
+
* // speakers = [{ voiceId: 'alloy' }, { voiceId: 'echo' }, ...]
|
|
87
|
+
* ```
|
|
88
|
+
*/
|
|
89
|
+
getSpeakers(): Promise<Array<{
|
|
90
|
+
voiceId: string;
|
|
91
|
+
[key: string]: any;
|
|
92
|
+
}>>;
|
|
93
|
+
/**
|
|
94
|
+
* Disconnects from the OpenAI realtime session and cleans up resources.
|
|
95
|
+
* Should be called when you're done with the voice instance.
|
|
96
|
+
*
|
|
97
|
+
* @example
|
|
98
|
+
* ```typescript
|
|
99
|
+
* voice.close(); // Disconnects and cleans up
|
|
100
|
+
* ```
|
|
101
|
+
*/
|
|
102
|
+
close(): void;
|
|
103
|
+
/**
|
|
104
|
+
* Equips the voice instance with a set of instructions.
|
|
105
|
+
* Instructions allow the model to perform additional actions during conversations.
|
|
106
|
+
*
|
|
107
|
+
* @param instructions - Optional instructions to addInstructions
|
|
108
|
+
* @returns Transformed instructions ready for use with the model
|
|
109
|
+
*
|
|
110
|
+
* @example
|
|
111
|
+
* ```typescript
|
|
112
|
+
* voice.addInstructions('You are a helpful assistant.');
|
|
113
|
+
* ```
|
|
114
|
+
*/
|
|
115
|
+
addInstructions(instructions?: string): void;
|
|
116
|
+
/**
|
|
117
|
+
* Equips the voice instance with a set of tools.
|
|
118
|
+
* Tools allow the model to perform additional actions during conversations.
|
|
119
|
+
*
|
|
120
|
+
* @param tools - Optional tools configuration to addTools
|
|
121
|
+
* @returns Transformed tools configuration ready for use with the model
|
|
122
|
+
*
|
|
123
|
+
* @example
|
|
124
|
+
* ```typescript
|
|
125
|
+
* const tools = {
|
|
126
|
+
* search: async (query: string) => { ... },
|
|
127
|
+
* calculate: (expression: string) => { ... }
|
|
128
|
+
* };
|
|
129
|
+
* voice.addTools(tools);
|
|
130
|
+
* ```
|
|
131
|
+
*/
|
|
132
|
+
addTools(tools?: TTools): void;
|
|
133
|
+
/**
|
|
134
|
+
* Emits a speaking event using the configured voice model.
|
|
135
|
+
* Can accept either a string or a readable stream as input.
|
|
136
|
+
*
|
|
137
|
+
* @param input - The text to convert to speech, or a readable stream containing the text
|
|
138
|
+
* @param options - Optional configuration for this specific speech request
|
|
139
|
+
* @param options.speaker - Override the voice to use for this specific request
|
|
140
|
+
*
|
|
141
|
+
* @throws {Error} If the input text is empty
|
|
142
|
+
*
|
|
143
|
+
* @example
|
|
144
|
+
* ```typescript
|
|
145
|
+
* // Simple text to speech
|
|
146
|
+
* await voice.speak('Hello world');
|
|
147
|
+
*
|
|
148
|
+
* // With custom voice
|
|
149
|
+
* await voice.speak('Hello world', { speaker: 'echo' });
|
|
150
|
+
*
|
|
151
|
+
* // Using a stream
|
|
152
|
+
* const stream = fs.createReadStream('text.txt');
|
|
153
|
+
* await voice.speak(stream);
|
|
154
|
+
* ```
|
|
155
|
+
*/
|
|
156
|
+
speak(input: string | NodeJS.ReadableStream, options?: {
|
|
157
|
+
speaker?: Realtime.Voice;
|
|
158
|
+
}): Promise<void>;
|
|
159
|
+
/**
|
|
160
|
+
* Updates the session configuration for the voice instance.
|
|
161
|
+
* This can be used to modify voice settings, turn detection, and other parameters.
|
|
162
|
+
*
|
|
163
|
+
* @param sessionConfig - New session configuration to apply
|
|
164
|
+
*
|
|
165
|
+
* @example
|
|
166
|
+
* ```typescript
|
|
167
|
+
* voice.updateConfig({
|
|
168
|
+
* voice: 'echo',
|
|
169
|
+
* turn_detection: {
|
|
170
|
+
* type: 'server_vad',
|
|
171
|
+
* threshold: 0.5,
|
|
172
|
+
* silence_duration_ms: 1000
|
|
173
|
+
* }
|
|
174
|
+
* });
|
|
175
|
+
* ```
|
|
176
|
+
*/
|
|
177
|
+
updateConfig(sessionConfig: unknown): void;
|
|
178
|
+
/**
|
|
179
|
+
* Checks if listening capabilities are enabled.
|
|
180
|
+
*
|
|
181
|
+
* @returns {Promise<{ enabled: boolean }>}
|
|
182
|
+
*/
|
|
183
|
+
getListener(): Promise<{
|
|
184
|
+
enabled: boolean;
|
|
185
|
+
}>;
|
|
186
|
+
/**
|
|
187
|
+
* Processes audio input for speech recognition.
|
|
188
|
+
* Takes a readable stream of audio data and emits a writing event.
|
|
189
|
+
* The output of the writing event is int16 audio data.
|
|
190
|
+
*
|
|
191
|
+
* @param audioData - Readable stream containing the audio data to process
|
|
192
|
+
* @param options - Optional configuration for audio processing
|
|
193
|
+
*
|
|
194
|
+
* @throws {Error} If the audio data format is not supported
|
|
195
|
+
*
|
|
196
|
+
* @example
|
|
197
|
+
* ```typescript
|
|
198
|
+
* // Process audio from a file
|
|
199
|
+
* const audioStream = fs.createReadStream('audio.raw');
|
|
200
|
+
* await voice.listen(audioStream);
|
|
201
|
+
*
|
|
202
|
+
* // Process audio with options
|
|
203
|
+
* await voice.listen(microphoneStream, {
|
|
204
|
+
* format: 'int16',
|
|
205
|
+
* sampleRate: 24000
|
|
206
|
+
* });
|
|
207
|
+
* ```
|
|
208
|
+
*/
|
|
209
|
+
listen(audioData: NodeJS.ReadableStream): Promise<void>;
|
|
210
|
+
waitForOpen(): Promise<unknown>;
|
|
211
|
+
waitForSessionCreated(): Promise<unknown>;
|
|
212
|
+
/**
|
|
213
|
+
* Establishes a connection to the OpenAI realtime service.
|
|
214
|
+
* Must be called before using speak, listen, or relay functions.
|
|
215
|
+
*
|
|
216
|
+
* @throws {Error} If connection fails or session creation times out
|
|
217
|
+
*
|
|
218
|
+
* @example
|
|
219
|
+
* ```typescript
|
|
220
|
+
* await voice.open();
|
|
221
|
+
* // Now ready for voice interactions
|
|
222
|
+
* ```
|
|
223
|
+
*/
|
|
224
|
+
connect({ runtimeContext }?: {
|
|
225
|
+
runtimeContext?: RuntimeContext;
|
|
226
|
+
}): Promise<void>;
|
|
227
|
+
disconnect(): void;
|
|
228
|
+
/**
|
|
229
|
+
* Streams audio data in real-time to the OpenAI service.
|
|
230
|
+
* Useful for continuous audio streaming scenarios like live microphone input.
|
|
231
|
+
* Must be in 'open' state before calling this method.
|
|
232
|
+
*
|
|
233
|
+
* @param audioData - Readable stream of audio data to relay
|
|
234
|
+
* @throws {Error} If audio format is not supported
|
|
235
|
+
*
|
|
236
|
+
* @example
|
|
237
|
+
* ```typescript
|
|
238
|
+
* // First connect
|
|
239
|
+
* await voice.open();
|
|
240
|
+
*
|
|
241
|
+
* // Then relay audio
|
|
242
|
+
* const micStream = getMicrophoneStream();
|
|
243
|
+
* await voice.relay(micStream);
|
|
244
|
+
* ```
|
|
245
|
+
*/
|
|
246
|
+
send(audioData: NodeJS.ReadableStream | Int16Array, eventId?: string): Promise<void>;
|
|
247
|
+
/**
|
|
248
|
+
* Sends a response to the OpenAI Realtime API.
|
|
249
|
+
*
|
|
250
|
+
* Trigger a response to the real-time session.
|
|
251
|
+
*
|
|
252
|
+
* @param {Object} params - The parameters object
|
|
253
|
+
* @param {Realtime.ResponseConfig} params.options - Configuration options for the response
|
|
254
|
+
* @returns {Promise<void>} A promise that resolves when the response has been sent
|
|
255
|
+
*
|
|
256
|
+
* @example
|
|
257
|
+
* // Send a simple text response
|
|
258
|
+
* await realtimeVoice.answer({
|
|
259
|
+
* options: {
|
|
260
|
+
* content: "Hello, how can I help you today?",
|
|
261
|
+
* voice: "alloy"
|
|
262
|
+
* }
|
|
263
|
+
* });
|
|
264
|
+
*/
|
|
265
|
+
answer({ options }: {
|
|
266
|
+
options?: Realtime.ResponseConfig;
|
|
267
|
+
}): Promise<void>;
|
|
268
|
+
/**
|
|
269
|
+
* Registers an event listener for voice events.
|
|
270
|
+
* Available events: 'speaking', 'writing, 'error'
|
|
271
|
+
* Can listen to OpenAI Realtime events by prefixing with 'openAIRealtime:'
|
|
272
|
+
* Such as 'openAIRealtime:conversation.item.completed', 'openAIRealtime:conversation.updated', etc.
|
|
273
|
+
*
|
|
274
|
+
* @param event - Name of the event to listen for
|
|
275
|
+
* @param callback - Function to call when the event occurs
|
|
276
|
+
*
|
|
277
|
+
* @example
|
|
278
|
+
* ```typescript
|
|
279
|
+
* // Listen for speech events
|
|
280
|
+
* voice.on('speaking', (audioData: Int16Array) => {
|
|
281
|
+
* // Handle audio data
|
|
282
|
+
* });
|
|
283
|
+
*
|
|
284
|
+
* // Handle errors
|
|
285
|
+
* voice.on('error', (error: Error) => {
|
|
286
|
+
* console.error('Voice error:', error);
|
|
287
|
+
* });
|
|
288
|
+
* ```
|
|
289
|
+
*/
|
|
290
|
+
on(event: string, callback: EventCallback): void;
|
|
291
|
+
/**
|
|
292
|
+
* Removes a previously registered event listener.
|
|
293
|
+
*
|
|
294
|
+
* @param event - Name of the event to stop listening to
|
|
295
|
+
* @param callback - The specific callback function to remove
|
|
296
|
+
*
|
|
297
|
+
* @example
|
|
298
|
+
* ```typescript
|
|
299
|
+
* // Create event handler
|
|
300
|
+
* const handleSpeech = (audioData: Int16Array) => {
|
|
301
|
+
* // Handle audio data
|
|
302
|
+
* };
|
|
303
|
+
*
|
|
304
|
+
* // Add listener
|
|
305
|
+
* voice.on('speaking', handleSpeech);
|
|
306
|
+
*
|
|
307
|
+
* // Later, remove the listener
|
|
308
|
+
* voice.off('speaking', handleSpeech);
|
|
309
|
+
* ```
|
|
310
|
+
*/
|
|
311
|
+
off(event: string, callback: EventCallback): void;
|
|
312
|
+
/**
|
|
313
|
+
* Emit an event with arguments
|
|
314
|
+
* @param event Event name
|
|
315
|
+
* @param args Arguments to pass to the callbacks
|
|
316
|
+
*/
|
|
317
|
+
private emit;
|
|
318
|
+
private setupEventListeners;
|
|
319
|
+
private handleFunctionCalls;
|
|
320
|
+
private handleFunctionCall;
|
|
321
|
+
private int16ArrayToBase64;
|
|
322
|
+
private sendEvent;
|
|
323
|
+
}
|
|
324
|
+
export {};
|
|
325
|
+
//# sourceMappingURL=index.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAEA,OAAO,KAAK,EAAE,UAAU,EAAE,MAAM,oBAAoB,CAAC;AACrD,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,8BAA8B,CAAC;AACnE,OAAO,EAAE,WAAW,EAAE,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,EAAE,QAAQ,EAAwB,MAAM,qBAAqB,CAAC;AAI1E;;GAEG;AACH,KAAK,aAAa,GAAG,CAAC,GAAG,IAAI,EAAE,GAAG,EAAE,KAAK,IAAI,CAAC;AA8C9C,KAAK,MAAM,GAAG,UAAU,CAAC;AAmBzB;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA4BG;AACH,qBAAa,mBAAoB,SAAQ,WAAW;IA+BhD,OAAO,CAAC,OAAO;IA9BjB,OAAO,CAAC,EAAE,CAAC,CAAY;IACvB,OAAO,CAAC,KAAK,CAAmB;IAChC,OAAO,CAAC,MAAM,CAA6C;IAC3D,OAAO,CAAC,MAAM,CAAW;IACzB,OAAO,CAAC,YAAY,CAAC,CAAS;IAC9B,OAAO,CAAC,KAAK,CAAC,CAAS;IACvB,OAAO,CAAC,KAAK,CAAU;IACvB,OAAO,CAAC,KAAK,CAAiB;IAC9B,OAAO,CAAC,WAAW,CAAmC;IACtD,OAAO,CAAC,cAAc,CAAC,CAAiB;IACxC;;;;;;;;;;;;;;;;;;OAkBG;gBAEO,OAAO,GAAE;QACf,KAAK,CAAC,EAAE,MAAM,CAAC;QACf,GAAG,CAAC,EAAE,MAAM,CAAC;QACb,MAAM,CAAC,EAAE,MAAM,CAAC;QAChB,OAAO,CAAC,EAAE,QAAQ,CAAC,KAAK,CAAC;QACzB,WAAW,CAAC,EAAE,QAAQ,CAAC,uBAAuB,CAAC;QAC/C,KAAK,CAAC,EAAE,OAAO,CAAC;KACZ;IAYR;;;;;;;;;;OAUG;IACH,WAAW,IAAI,OAAO,CAAC,KAAK,CAAC;QAAE,OAAO,EAAE,MAAM,CAAC;QAAC,CAAC,GAAG,EAAE,MAAM,GAAG,GAAG,CAAA;KAAE,CAAC,CAAC;IAItE;;;;;;;;OAQG;IACH,KAAK;IAML;;;;;;;;;;;OAWG;IACH,eAAe,CAAC,YAAY,CAAC,EAAE,MAAM;IAIrC;;;;;;;;;;;;;;;OAeG;IACH,QAAQ,CAAC,KAAK,CAAC,EAAE,MAAM;IAIvB;;;;;;;;;;;;;;;;;;;;;;OAsBG;IACG,KAAK,CAAC,KAAK,EAAE,MAAM,GAAG,MAAM,CAAC,cAAc,EAAE,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,KAAK,CAAA;KAAE,GAAG,OAAO,CAAC,IAAI,CAAC;IAqBzG;;;;;;;;;;;;;;;;;OAiBG;IACH,YAAY,CAAC,aAAa,EAAE,OAAO,GAAG,IAAI;IAI1C;;;;OAIG;IACG,WAAW;;;IAIjB;;;;;;;;;;;;;;;;;;;;;;OAsBG;IACG,MAAM,CAAC,SAAS,EAAE,MAAM,CAAC,cAAc,GAAG,OAAO,CAAC,IAAI,CAAC;IA+B7D,WAAW;IAMX,qBAAqB;IAMrB;;;;;;;;;;;OAWG;IACG,OAAO,CAAC,EAAE,cAAc,EAAE,GAAE;QAAE,cAAc,CAAC,EAAE,cAAc,CAAA;KAAO;IA2B1E,UAAU;IAKV;;;;;;;;;;;;;;;;;OAiBG;IACG,IAAI,CAAC,SAAS,EAAE,MAAM,CAAC,cAAc,GAAG,UAAU,EAAE,OAAO,CAAC,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC;IA4B1F;;;;;;;;;;;;;;;;;OAiBG;IACG,MAAM,CAAC,EAAE,OAAO,EAAE,EAAE;QAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,cAAc,CAAA;KAAE;IAI/D;;;;;;;;;;;;;;;;;;;;;OAqBG;IACH,EAAE,CAAC,KAAK,EAAE,MAAM,EAAE,QAAQ,EAAE,aAAa,GAAG,IAAI;IAOhD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,GAAG,CAAC,KAAK,EAAE,MAAM,EAAE,QAAQ,EAAE,aAAa,GAAG,IAAI;IASjD;;;;OAIG;IACH,OAAO,CAAC,IAAI;IAQZ,OAAO,CAAC,mBAAmB;YA+Eb,mBAAmB;YAQnB,kBAAkB;IAwDhC,OAAO,CAAC,kBAAkB;IAc1B,OAAO,CAAC,SAAS;CAYlB"}
|
package/dist/index.js
CHANGED
|
@@ -379,7 +379,8 @@ var OpenAIRealtimeVoice = class extends MastraVoice {
|
|
|
379
379
|
});
|
|
380
380
|
} else if (audioData instanceof Int16Array) {
|
|
381
381
|
try {
|
|
382
|
-
this.
|
|
382
|
+
const base64Audio = this.int16ArrayToBase64(audioData);
|
|
383
|
+
this.sendEvent("input_audio_buffer.append", { audio: base64Audio, event_id: eventId });
|
|
383
384
|
} catch (err) {
|
|
384
385
|
this.emit("error", err);
|
|
385
386
|
}
|
|
@@ -538,6 +539,9 @@ var OpenAIRealtimeVoice = class extends MastraVoice {
|
|
|
538
539
|
this.emit("response.done", ev);
|
|
539
540
|
speakerStreams.delete(ev.response.id);
|
|
540
541
|
});
|
|
542
|
+
this.client.on("error", async (ev) => {
|
|
543
|
+
this.emit("error", ev);
|
|
544
|
+
});
|
|
541
545
|
}
|
|
542
546
|
async handleFunctionCalls(ev) {
|
|
543
547
|
for (const output of ev.response?.output ?? []) {
|
|
@@ -625,3 +629,5 @@ var OpenAIRealtimeVoice = class extends MastraVoice {
|
|
|
625
629
|
};
|
|
626
630
|
|
|
627
631
|
export { OpenAIRealtimeVoice };
|
|
632
|
+
//# sourceMappingURL=index.js.map
|
|
633
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/utils.ts","../src/index.ts"],"names":["buffer","ev"],"mappings":";;;;;;;AAeO,IAAM,cAAA,GAAiB,CAAC,KAAA,KAAmB;AAChD,EAAA,MAAM,cAAgF,EAAC;AACvF,EAAA,KAAA,MAAW,CAAC,MAAM,IAAI,CAAA,IAAK,OAAO,OAAA,CAAQ,KAAA,IAAS,EAAE,CAAA,EAAG;AACtD,IAAA,IAAI,UAAA;AAEJ,IAAA,IAAI,aAAA,IAAiB,IAAA,IAAQ,IAAA,CAAK,WAAA,EAAa;AAC7C,MAAA,IAAI,WAAA,CAAY,IAAA,CAAK,WAAW,CAAA,EAAG;AACjC,QAAA,UAAA,GAAa,eAAA,CAAgB,KAAK,WAAW,CAAA;AAC7C,QAAA,OAAO,UAAA,CAAW,OAAA;AAAA,MACpB,CAAA,MAAO;AACL,QAAA,UAAA,GAAa,IAAA,CAAK,WAAA;AAAA,MACpB;AAAA,IACF,CAAA,MAAA,IAAW,gBAAgB,IAAA,EAAM;AAC/B,MAAA,IAAI,WAAA,CAAY,IAAA,CAAK,UAAU,CAAA,EAAG;AAChC,QAAA,UAAA,GAAa,eAAA,CAAgB,KAAK,UAAU,CAAA;AAC5C,QAAA,OAAO,UAAA,CAAW,OAAA;AAAA,MACpB,CAAA,MAAO;AACL,QAAA,UAAA,GAAa,IAAA,CAAK,UAAA;AAAA,MACpB;AAAA,IACF,CAAA,MAAO;AACL,MAAA,OAAA,CAAQ,IAAA,CAAK,CAAA,KAAA,EAAQ,IAAI,CAAA,iDAAA,CAAmD,CAAA;AAC5E,MAAA;AAAA,IACF;AACA,IAAA,MAAM,UAAA,GAA6B;AAAA,MACjC,IAAA,EAAM,UAAA;AAAA,MACN,IAAA;AAAA,MACA,WAAA,EAAa,IAAA,CAAK,WAAA,IAAe,CAAA,MAAA,EAAS,IAAI,CAAA,CAAA;AAAA,MAC9C;AAAA,KACF;AAEA,IAAA,IAAI,KAAK,OAAA,EAAS;AAEhB,MAAA,MAAM,cAAA,GAAiB,OAAO,IAAA,KAAc;AAC1C,QAAA,IAAI;AACF,UAAA,IAAI,CAAC,KAAK,OAAA,EAAS;AACjB,YAAA,MAAM,IAAI,KAAA,CAAM,CAAA,KAAA,EAAQ,IAAI,CAAA,wBAAA,CAA0B,CAAA;AAAA,UACxD;AAGA,UAAA,IAAI,iBAAiB,IAAA,EAAM;AACzB,YAAA,OAAO,MAAM,IAAA,CAAK,OAAA,CAAQ,EAAE,OAAA,EAAS,MAAM,CAAA;AAAA,UAC7C,CAAA,MAEK;AAEH,YAAA,MAAM,OAAA,GAAU;AAAA,cACd,UAAA,EAAY,SAAA;AAAA,cACZ,UAAU;AAAC,aACb;AACA,YAAA,OAAO,MAAM,IAAA,CAAK,OAAA,CAAQ,IAAA,EAAM,OAAO,CAAA;AAAA,UACzC;AAAA,QACF,SAAS,KAAA,EAAO;AACd,UAAA,OAAA,CAAQ,KAAA,CAAM,CAAA,qBAAA,EAAwB,IAAI,CAAA,CAAA,CAAA,EAAK,KAAK,CAAA;AACpD,UAAA,MAAM,KAAA;AAAA,QACR;AAAA,MACF,CAAA;AACA,MAAA,WAAA,CAAY,IAAA,CAAK,EAAE,UAAA,EAAY,OAAA,EAAS,gBAAgB,CAAA;AAAA,IAC1D,CAAA,MAAO;AACL,MAAA,OAAA,CAAQ,IAAA,CAAK,CAAA,KAAA,EAAQ,IAAI,CAAA,kCAAA,CAAoC,CAAA;AAAA,IAC/D;AAAA,EACF;AACA,EAAA,OAAO,WAAA;AACT,CAAA;AAEO,IAAM,gBAAA,GAAmB,CAAC,GAAA,KAAiB;AAChD,EAAA,OACE,GAAA,IACA,GAAA,YAAe,QAAA,IACf,OAAO,GAAA,CAAI,IAAA,KAAS,UAAA,IACpB,OAAO,GAAA,CAAI,IAAA,KAAS,UAAA,IACpB,GAAA,CAAI,QAAA,KAAa,IAAA;AAErB,CAAA;AAEA,SAAS,YAAY,MAAA,EAAiB;AACpC,EAAA,OACE,CAAC,CAAC,MAAA,IACF,OAAO,MAAA,KAAW,QAAA,IAClB,UAAU,MAAA,IACV,MAAA,CAAO,QACP,OAAO,MAAA,CAAO,SAAS,QAAA,IACvB,UAAA,IAAc,OAAO,IAAA,IACrB,MAAA,CAAO,KAAK,QAAA,KAAa,WAAA;AAE7B;;;ACrEA,IAAM,aAAA,GAAgC,OAAA;AAEtC,IAAM,mBAAA,GAAwD,WAAA;AAE9D,IAAM,WAAA,GAAc,kCAAA;AAMpB,IAAM,aAAA,GAAgB,yCAAA;AA4BtB,IAAM,MAAA,GAAS,CAAC,OAAA,EAAS,KAAA,EAAO,UAAU,OAAA,EAAS,MAAA,EAAQ,MAAA,EAAQ,SAAA,EAAW,OAAO,CAAA;AAsC9E,IAAM,mBAAA,GAAN,cAAkC,WAAA,CAAY;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EA8BnD,WAAA,CACU,OAAA,GAOJ,EAAC,EACL;AACA,IAAA,KAAA,EAAM;AATE,IAAA,IAAA,CAAA,OAAA,GAAA,OAAA;AAWR,IAAA,IAAA,CAAK,MAAA,GAAS,IAAI,YAAA,EAAa;AAC/B,IAAA,IAAA,CAAK,KAAA,GAAQ,OAAA;AACb,IAAA,IAAA,CAAK,SAAS,EAAC;AACf,IAAA,IAAA,CAAK,OAAA,GAAU,QAAQ,OAAA,IAAW,aAAA;AAClC,IAAA,IAAA,CAAK,WAAA,GAAc,QAAQ,WAAA,IAAe,mBAAA;AAC1C,IAAA,IAAA,CAAK,KAAA,GAAQ,QAAQ,KAAA,IAAS,KAAA;AAAA,EAChC;AAAA,EA/CQ,EAAA;AAAA,EACA,KAAA;AAAA,EACA,MAAA;AAAA,EACA,MAAA;AAAA,EACA,YAAA;AAAA,EACA,KAAA;AAAA,EACA,KAAA;AAAA,EACA,QAAmB,EAAC;AAAA,EACpB,WAAA;AAAA,EACA,cAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAmDR,WAAA,GAAuE;AACrE,IAAA,OAAO,OAAA,CAAQ,QAAQ,MAAA,CAAO,GAAA,CAAI,QAAM,EAAE,OAAA,EAAS,CAAA,EAAE,CAAE,CAAC,CAAA;AAAA,EAC1D;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAWA,KAAA,GAAQ;AACN,IAAA,IAAI,CAAC,KAAK,EAAA,EAAI;AACd,IAAA,IAAA,CAAK,GAAG,KAAA,EAAM;AACd,IAAA,IAAA,CAAK,KAAA,GAAQ,OAAA;AAAA,EACf;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAcA,gBAAgB,YAAA,EAAuB;AACrC,IAAA,IAAA,CAAK,YAAA,GAAe,YAAA;AAAA,EACtB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAkBA,SAAS,KAAA,EAAgB;AACvB,IAAA,IAAA,CAAK,KAAA,GAAQ,SAAS,EAAC;AAAA,EACzB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAyBA,MAAM,KAAA,CAAM,KAAA,EAAuC,OAAA,EAAuD;AACxG,IAAA,IAAI,OAAO,UAAU,QAAA,EAAU;AAC7B,MAAA,MAAM,SAAmB,EAAC;AAC1B,MAAA,WAAA,MAAiB,SAAS,KAAA,EAAO;AAC/B,QAAA,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,QAAA,CAAS,KAAK,CAAA,GAAI,KAAA,GAAQ,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,KAAK,CAAC,CAAC,CAAA;AAAA,MACzE;AACA,MAAA,KAAA,GAAQ,MAAA,CAAO,MAAA,CAAO,MAAM,CAAA,CAAE,SAAS,OAAO,CAAA;AAAA,IAChD;AAEA,IAAA,IAAI,KAAA,CAAM,IAAA,EAAK,CAAE,MAAA,KAAW,CAAA,EAAG;AAC7B,MAAA,MAAM,IAAI,MAAM,qBAAqB,CAAA;AAAA,IACvC;AAEA,IAAA,IAAA,CAAK,UAAU,iBAAA,EAAmB;AAAA,MAChC,QAAA,EAAU;AAAA,QACR,YAAA,EAAc,8BAA8B,KAAK,CAAA,CAAA;AAAA,QACjD,KAAA,EAAO,OAAA,EAAS,OAAA,GAAU,OAAA,CAAQ,OAAA,GAAU;AAAA;AAC9C,KACD,CAAA;AAAA,EACH;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAoBA,aAAa,aAAA,EAA8B;AACzC,IAAA,IAAA,CAAK,SAAA,CAAU,gBAAA,EAAkB,EAAE,OAAA,EAAS,eAAe,CAAA;AAAA,EAC7D;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOA,MAAM,WAAA,GAAc;AAClB,IAAA,OAAO,EAAE,SAAS,IAAA,EAAK;AAAA,EACzB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAyBA,MAAM,OAAO,SAAA,EAAiD;AAC5D,IAAA,IAAI,gBAAA,CAAiB,SAAS,CAAA,EAAG;AAC/B,MAAA,MAAM,SAAmB,EAAC;AAC1B,MAAA,WAAA,MAAiB,SAAS,SAAA,EAAW;AACnC,QAAA,MAAMA,OAAAA,GAAS,OAAO,QAAA,CAAS,KAAK,IAAI,KAAA,GAAQ,MAAA,CAAO,KAAK,KAAK,CAAA;AACjE,QAAA,MAAA,CAAO,KAAKA,OAAM,CAAA;AAAA,MACpB;AAEA,MAAA,MAAM,MAAA,GAAS,MAAA,CAAO,MAAA,CAAO,MAAM,CAAA;AACnC,MAAA,MAAM,UAAA,GAAa,IAAI,UAAA,CAAW,MAAA,CAAO,MAAA,EAAQ,MAAA,CAAO,UAAA,IAAc,CAAA,EAAA,CAAI,MAAA,CAAO,UAAA,IAAc,CAAA,IAAK,CAAC,CAAA;AACrG,MAAA,MAAM,WAAA,GAAc,IAAA,CAAK,kBAAA,CAAmB,UAAU,CAAA;AAEtD,MAAA,IAAA,CAAK,UAAU,0BAAA,EAA4B;AAAA,QACzC,IAAA,EAAM;AAAA,UACJ,IAAA,EAAM,SAAA;AAAA,UACN,IAAA,EAAM,MAAA;AAAA,UACN,SAAS,CAAC,EAAE,MAAM,aAAA,EAAe,KAAA,EAAO,aAAa;AAAA;AACvD,OACD,CAAA;AAED,MAAA,IAAA,CAAK,UAAU,iBAAA,EAAmB;AAAA,QAChC,QAAA,EAAU;AAAA,UACR,UAAA,EAAY,CAAC,MAAM,CAAA;AAAA,UACnB,YAAA,EAAc,CAAA,kDAAA;AAAA;AAChB,OACD,CAAA;AAAA,IACH,CAAA,MAAO;AACL,MAAA,IAAA,CAAK,IAAA,CAAK,OAAA,EAAS,IAAI,KAAA,CAAM,+BAA+B,CAAC,CAAA;AAAA,IAC/D;AAAA,EACF;AAAA,EAEA,WAAA,GAAc;AACZ,IAAA,OAAO,IAAI,QAAQ,CAAA,OAAA,KAAW;AAC5B,MAAA,IAAA,CAAK,EAAA,EAAI,EAAA,CAAG,MAAA,EAAQ,OAAO,CAAA;AAAA,IAC7B,CAAC,CAAA;AAAA,EACH;AAAA,EAEA,qBAAA,GAAwB;AACtB,IAAA,OAAO,IAAI,QAAQ,CAAA,OAAA,KAAW;AAC5B,MAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,iBAAA,EAAmB,OAAO,CAAA;AAAA,IAC3C,CAAC,CAAA;AAAA,EACH;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAcA,MAAM,OAAA,CAAQ,EAAE,cAAA,EAAe,GAAyC,EAAC,EAAG;AAC1E,IAAA,MAAM,GAAA,GAAM,CAAA,EAAG,IAAA,CAAK,OAAA,CAAQ,GAAA,IAAO,WAAW,CAAA,OAAA,EAAU,IAAA,CAAK,OAAA,CAAQ,KAAA,IAAS,aAAa,CAAA,CAAA;AAC3F,IAAA,MAAM,MAAA,GAAS,IAAA,CAAK,OAAA,CAAQ,MAAA,IAAU,QAAQ,GAAA,CAAI,cAAA;AAClD,IAAA,IAAA,CAAK,cAAA,GAAiB,cAAA;AAEtB,IAAA,IAAA,CAAK,EAAA,GAAK,IAAI,SAAA,CAAU,GAAA,EAAK,MAAA,EAAW;AAAA,MACtC,OAAA,EAAS;AAAA,QACP,eAAe,SAAA,GAAY,MAAA;AAAA,QAC3B,aAAA,EAAe;AAAA;AACjB,KACD,CAAA;AAED,IAAA,IAAA,CAAK,mBAAA,EAAoB;AACzB,IAAA,MAAM,OAAA,CAAQ,IAAI,CAAC,IAAA,CAAK,aAAY,EAAG,IAAA,CAAK,qBAAA,EAAuB,CAAC,CAAA;AAEpE,IAAA,MAAM,WAAA,GAAc,cAAA,CAAe,IAAA,CAAK,KAAK,CAAA;AAC7C,IAAA,IAAA,CAAK,YAAA,CAAa;AAAA,MAChB,cAAc,IAAA,CAAK,YAAA;AAAA,MACnB,KAAA,EAAO,WAAA,CAAY,GAAA,CAAI,CAAA,CAAA,KAAK,EAAE,UAAU,CAAA;AAAA,MACxC,yBAAA,EAA2B;AAAA,QACzB,OAAO,IAAA,CAAK;AAAA,OACd;AAAA,MACA,OAAO,IAAA,CAAK;AAAA,KACb,CAAA;AACD,IAAA,IAAA,CAAK,KAAA,GAAQ,MAAA;AAAA,EACf;AAAA,EAEA,UAAA,GAAa;AACX,IAAA,IAAA,CAAK,KAAA,GAAQ,OAAA;AACb,IAAA,IAAA,CAAK,IAAI,KAAA,EAAM;AAAA,EACjB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAoBA,MAAM,IAAA,CAAK,SAAA,EAA+C,OAAA,EAAiC;AACzF,IAAA,IAAI,CAAC,IAAA,CAAK,KAAA,IAAS,IAAA,CAAK,UAAU,MAAA,EAAQ;AACxC,MAAA,OAAA,CAAQ,KAAK,sDAAsD,CAAA;AACnE,MAAA;AAAA,IACF;AAEA,IAAA,IAAI,gBAAA,CAAiB,SAAS,CAAA,EAAG;AAC/B,MAAA,MAAM,MAAA,GAAS,SAAA;AACf,MAAA,MAAA,CAAO,EAAA,CAAG,QAAQ,CAAA,KAAA,KAAS;AACzB,QAAA,IAAI;AACF,UAAA,MAAM,MAAA,GAAS,OAAO,QAAA,CAAS,KAAK,IAAI,KAAA,GAAQ,MAAA,CAAO,KAAK,KAAK,CAAA;AACjE,UAAA,IAAA,CAAK,SAAA,CAAU,2BAAA,EAA6B,EAAE,KAAA,EAAO,MAAA,CAAO,SAAS,QAAQ,CAAA,EAAG,QAAA,EAAU,OAAA,EAAS,CAAA;AAAA,QACrG,SAAS,GAAA,EAAK;AACZ,UAAA,IAAA,CAAK,IAAA,CAAK,SAAS,GAAG,CAAA;AAAA,QACxB;AAAA,MACF,CAAC,CAAA;AAAA,IACH,CAAA,MAAA,IAAW,qBAAqB,UAAA,EAAY;AAC1C,MAAA,IAAI;AACF,QAAA,MAAM,WAAA,GAAc,IAAA,CAAK,kBAAA,CAAmB,SAAS,CAAA;AACrD,QAAA,IAAA,CAAK,UAAU,2BAAA,EAA6B,EAAE,OAAO,WAAA,EAAa,QAAA,EAAU,SAAS,CAAA;AAAA,MACvF,SAAS,GAAA,EAAK;AACZ,QAAA,IAAA,CAAK,IAAA,CAAK,SAAS,GAAG,CAAA;AAAA,MACxB;AAAA,IACF,CAAA,MAAO;AACL,MAAA,IAAA,CAAK,IAAA,CAAK,OAAA,EAAS,IAAI,KAAA,CAAM,+BAA+B,CAAC,CAAA;AAAA,IAC/D;AAAA,EACF;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAoBA,MAAM,MAAA,CAAO,EAAE,OAAA,EAAQ,EAA0C;AAC/D,IAAA,IAAA,CAAK,UAAU,iBAAA,EAAmB,EAAE,UAAU,OAAA,IAAW,IAAI,CAAA;AAAA,EAC/D;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAwBA,EAAA,CAAG,OAAe,QAAA,EAA+B;AAC/C,IAAA,IAAI,CAAC,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,EAAG;AACvB,MAAA,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,GAAI,EAAC;AAAA,IACxB;AACA,IAAA,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,CAAE,IAAA,CAAK,QAAQ,CAAA;AAAA,EAClC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAsBA,GAAA,CAAI,OAAe,QAAA,EAA+B;AAChD,IAAA,IAAI,CAAC,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,EAAG;AAEzB,IAAA,MAAM,QAAQ,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,CAAE,QAAQ,QAAQ,CAAA;AACjD,IAAA,IAAI,UAAU,EAAA,EAAI;AAChB,MAAA,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,CAAE,MAAA,CAAO,OAAO,CAAC,CAAA;AAAA,IACpC;AAAA,EACF;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOQ,IAAA,CAAK,UAAkB,IAAA,EAAmB;AAChD,IAAA,IAAI,CAAC,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,EAAG;AAEzB,IAAA,KAAA,MAAW,QAAA,IAAY,IAAA,CAAK,MAAA,CAAO,KAAK,CAAA,EAAG;AACzC,MAAA,QAAA,CAAS,GAAG,IAAI,CAAA;AAAA,IAClB;AAAA,EACF;AAAA,EAEQ,mBAAA,GAA4B;AAClC,IAAA,MAAM,cAAA,uBAAqB,GAAA,EAA0B;AAErD,IAAA,IAAI,CAAC,KAAK,EAAA,EAAI;AACZ,MAAA,MAAM,IAAI,MAAM,2BAA2B,CAAA;AAAA,IAC7C;AAEA,IAAA,IAAA,CAAK,EAAA,CAAG,EAAA,CAAG,SAAA,EAAW,CAAA,OAAA,KAAW;AAC/B,MAAA,MAAM,IAAA,GAAO,IAAA,CAAK,KAAA,CAAM,OAAA,CAAQ,UAAU,CAAA;AAC1C,MAAA,IAAA,CAAK,MAAA,CAAO,IAAA,CAAK,IAAA,CAAK,IAAA,EAAM,IAAI,CAAA;AAEhC,MAAA,IAAI,KAAK,KAAA,EAAO;AACd,QAAA,MAAM,EAAE,KAAA,EAAO,GAAG,MAAA,EAAO,GAAI,IAAA;AAC7B,QAAA,OAAA,CAAQ,GAAA,CAAI,KAAK,IAAA,EAAM,MAAA,EAAQ,OAAO,MAAA,GAAS,GAAA,GAAM,QAAQ,EAAE,CAAA;AAAA,MACjE;AAAA,IACF,CAAC,CAAA;AAED,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,iBAAA,EAAmB,CAAA,EAAA,KAAM;AACtC,MAAA,IAAA,CAAK,IAAA,CAAK,mBAAmB,EAAE,CAAA;AAE/B,MAAA,MAAM,QAAQ,IAAA,CAAK,KAAA,CAAM,OAAO,CAAA,EAAG,IAAA,CAAK,MAAM,MAAM,CAAA;AACpD,MAAA,KAAA,MAAWC,OAAM,KAAA,EAAO;AACtB,QAAA,IAAA,CAAK,EAAA,EAAI,IAAA,CAAK,IAAA,CAAK,SAAA,CAAUA,GAAE,CAAC,CAAA;AAAA,MAClC;AAAA,IACF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,iBAAA,EAAmB,CAAA,EAAA,KAAM;AACtC,MAAA,IAAA,CAAK,IAAA,CAAK,mBAAmB,EAAE,CAAA;AAAA,IACjC,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,kBAAA,EAAoB,CAAA,EAAA,KAAM;AACvC,MAAA,IAAA,CAAK,IAAA,CAAK,oBAAoB,EAAE,CAAA;AAEhC,MAAA,MAAM,aAAA,GAAgB,IAAI,WAAA,EAAY;AAEtC,MAAA,aAAA,CAAc,EAAA,GAAK,GAAG,QAAA,CAAS,EAAA;AAE/B,MAAA,cAAA,CAAe,GAAA,CAAI,EAAA,CAAG,QAAA,CAAS,EAAA,EAAI,aAAa,CAAA;AAChD,MAAA,IAAA,CAAK,IAAA,CAAK,WAAW,aAAa,CAAA;AAAA,IACpC,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,mDAAA,EAAqD,CAAA,EAAA,KAAM;AACxE,MAAA,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,EAAE,IAAA,EAAM,EAAA,CAAG,KAAA,EAAO,WAAA,EAAa,EAAA,CAAG,WAAA,EAAa,IAAA,EAAM,MAAA,EAAQ,CAAA;AAAA,IACpF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,kDAAA,EAAoD,CAAA,EAAA,KAAM;AACvE,MAAA,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,EAAE,IAAA,EAAM,IAAA,EAAM,aAAa,EAAA,CAAG,WAAA,EAAa,IAAA,EAAM,MAAA,EAAQ,CAAA;AAAA,IAChF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,sBAAA,EAAwB,CAAA,EAAA,KAAM;AAC3C,MAAA,MAAM,KAAA,GAAQ,MAAA,CAAO,IAAA,CAAK,EAAA,CAAG,OAAO,QAAQ,CAAA;AAC5C,MAAA,IAAA,CAAK,KAAK,UAAA,EAAY,EAAE,OAAO,WAAA,EAAa,EAAA,CAAG,aAAa,CAAA;AAE5D,MAAA,MAAM,MAAA,GAAS,cAAA,CAAe,GAAA,CAAI,EAAA,CAAG,WAAW,CAAA;AAChD,MAAA,MAAA,EAAQ,MAAM,KAAK,CAAA;AAAA,IACrB,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,qBAAA,EAAuB,CAAA,EAAA,KAAM;AAC1C,MAAA,IAAA,CAAK,KAAK,eAAA,EAAiB,EAAE,WAAA,EAAa,EAAA,CAAG,aAAa,CAAA;AAE1D,MAAA,MAAM,MAAA,GAAS,cAAA,CAAe,GAAA,CAAI,EAAA,CAAG,WAAW,CAAA;AAChD,MAAA,MAAA,EAAQ,GAAA,EAAI;AAAA,IACd,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,iCAAA,EAAmC,CAAA,EAAA,KAAM;AACtD,MAAA,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,EAAE,IAAA,EAAM,EAAA,CAAG,KAAA,EAAO,WAAA,EAAa,EAAA,CAAG,WAAA,EAAa,IAAA,EAAM,WAAA,EAAa,CAAA;AAAA,IACzF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,gCAAA,EAAkC,CAAA,EAAA,KAAM;AACrD,MAAA,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,EAAE,IAAA,EAAM,IAAA,EAAM,aAAa,EAAA,CAAG,WAAA,EAAa,IAAA,EAAM,WAAA,EAAa,CAAA;AAAA,IACrF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,qBAAA,EAAuB,CAAA,EAAA,KAAM;AAC1C,MAAA,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,EAAE,IAAA,EAAM,EAAA,CAAG,KAAA,EAAO,WAAA,EAAa,EAAA,CAAG,WAAA,EAAa,IAAA,EAAM,WAAA,EAAa,CAAA;AAAA,IACzF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,oBAAA,EAAsB,CAAA,EAAA,KAAM;AACzC,MAAA,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,EAAE,IAAA,EAAM,IAAA,EAAM,aAAa,EAAA,CAAG,WAAA,EAAa,IAAA,EAAM,WAAA,EAAa,CAAA;AAAA,IACrF,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,eAAA,EAAiB,OAAM,EAAA,KAAM;AAC1C,MAAA,MAAM,IAAA,CAAK,oBAAoB,EAAE,CAAA;AACjC,MAAA,IAAA,CAAK,IAAA,CAAK,iBAAiB,EAAE,CAAA;AAC7B,MAAA,cAAA,CAAe,MAAA,CAAO,EAAA,CAAG,QAAA,CAAS,EAAE,CAAA;AAAA,IACtC,CAAC,CAAA;AACD,IAAA,IAAA,CAAK,MAAA,CAAO,EAAA,CAAG,OAAA,EAAS,OAAM,EAAA,KAAM;AAClC,MAAA,IAAA,CAAK,IAAA,CAAK,SAAS,EAAE,CAAA;AAAA,IACvB,CAAC,CAAA;AAAA,EACH;AAAA,EAEA,MAAc,oBAAoB,EAAA,EAAS;AACzC,IAAA,KAAA,MAAW,MAAA,IAAU,EAAA,CAAG,QAAA,EAAU,MAAA,IAAU,EAAC,EAAG;AAC9C,MAAA,IAAI,MAAA,CAAO,SAAS,eAAA,EAAiB;AACnC,QAAA,MAAM,IAAA,CAAK,mBAAmB,MAAM,CAAA;AAAA,MACtC;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAc,mBAAmB,MAAA,EAAa;AAC5C,IAAA,IAAI;AACF,MAAA,MAAM,OAAA,GAAU,IAAA,CAAK,KAAA,CAAM,MAAA,CAAO,SAAS,CAAA;AAC3C,MAAA,MAAM,IAAA,GAAO,IAAA,CAAK,KAAA,GAAQ,MAAA,CAAO,IAAI,CAAA;AACrC,MAAA,IAAI,CAAC,IAAA,EAAM;AACT,QAAA,OAAA,CAAQ,IAAA,CAAK,CAAA,MAAA,EAAS,MAAA,CAAO,IAAI,CAAA,WAAA,CAAa,CAAA;AAC9C,QAAA;AAAA,MACF;AAEA,MAAA,IAAI,MAAM,OAAA,EAAS;AACjB,QAAA,IAAA,CAAK,KAAK,iBAAA,EAAmB;AAAA,UAC3B,YAAY,MAAA,CAAO,OAAA;AAAA,UACnB,UAAU,MAAA,CAAO,IAAA;AAAA,UACjB,iBAAiB,IAAA,CAAK,WAAA;AAAA,UACtB,IAAA,EAAM;AAAA,SACP,CAAA;AAAA,MACH;AAEA,MAAA,MAAM,MAAA,GAAS,MAAM,IAAA,EAAM,OAAA;AAAA,QACzB,EAAE,OAAA,EAAS,cAAA,EAAgB,IAAA,CAAK,cAAA,EAAe;AAAA,QAC/C;AAAA,UACE,YAAY,MAAA,CAAO,OAAA;AAAA,UACnB,UAAU;AAAC;AACb,OACF;AAEA,MAAA,IAAA,CAAK,KAAK,kBAAA,EAAoB;AAAA,QAC5B,YAAY,MAAA,CAAO,OAAA;AAAA,QACnB,UAAU,MAAA,CAAO,IAAA;AAAA,QACjB,iBAAiB,IAAA,CAAK,WAAA;AAAA,QACtB,IAAA,EAAM,OAAA;AAAA,QACN;AAAA,OACD,CAAA;AAED,MAAA,IAAA,CAAK,UAAU,0BAAA,EAA4B;AAAA,QACzC,IAAA,EAAM;AAAA,UACJ,IAAA,EAAM,sBAAA;AAAA,UACN,SAAS,MAAA,CAAO,OAAA;AAAA,UAChB,MAAA,EAAQ,IAAA,CAAK,SAAA,CAAU,MAAM;AAAA;AAC/B,OACD,CAAA;AAAA,IACH,SAAS,CAAA,EAAG;AACV,MAAA,MAAM,GAAA,GAAM,CAAA;AACZ,MAAA,OAAA,CAAQ,KAAK,CAAA,oBAAA,EAAuB,MAAA,CAAO,IAAI,CAAA,EAAA,CAAA,EAAM,IAAI,OAAO,CAAA;AAChE,MAAA,IAAA,CAAK,UAAU,0BAAA,EAA4B;AAAA,QACzC,IAAA,EAAM;AAAA,UACJ,IAAA,EAAM,sBAAA;AAAA,UACN,SAAS,MAAA,CAAO,OAAA;AAAA,UAChB,QAAQ,IAAA,CAAK,SAAA,CAAU,EAAE,KAAA,EAAO,GAAA,CAAI,SAAS;AAAA;AAC/C,OACD,CAAA;AAAA,IACH,CAAA,SAAE;AACA,MAAA,IAAA,CAAK,SAAA,CAAU,iBAAA,EAAmB,EAAE,CAAA;AAAA,IACtC;AAAA,EACF;AAAA,EAEQ,mBAAmB,UAAA,EAAgC;AACzD,IAAA,MAAM,MAAA,GAAS,IAAI,WAAA,CAAY,UAAA,CAAW,SAAS,CAAC,CAAA;AACpD,IAAA,MAAM,IAAA,GAAO,IAAI,QAAA,CAAS,MAAM,CAAA;AAChC,IAAA,KAAA,IAAS,CAAA,GAAI,CAAA,EAAG,CAAA,GAAI,UAAA,CAAW,QAAQ,CAAA,EAAA,EAAK;AAC1C,MAAA,IAAA,CAAK,SAAS,CAAA,GAAI,CAAA,EAAG,UAAA,CAAW,CAAC,GAAI,IAAI,CAAA;AAAA,IAC3C;AACA,IAAA,MAAM,UAAA,GAAa,IAAI,UAAA,CAAW,MAAM,CAAA;AACxC,IAAA,IAAI,MAAA,GAAS,EAAA;AACb,IAAA,KAAA,IAAS,CAAA,GAAI,CAAA,EAAG,CAAA,GAAI,UAAA,CAAW,QAAQ,CAAA,EAAA,EAAK;AAC1C,MAAA,MAAA,IAAU,MAAA,CAAO,YAAA,CAAa,UAAA,CAAW,CAAC,CAAE,CAAA;AAAA,IAC9C;AACA,IAAA,OAAO,KAAK,MAAM,CAAA;AAAA,EACpB;AAAA,EAEQ,SAAA,CAAU,MAAc,IAAA,EAAW;AACzC,IAAA,IAAI,CAAC,KAAK,EAAA,IAAM,IAAA,CAAK,GAAG,UAAA,KAAe,IAAA,CAAK,GAAG,IAAA,EAAM;AACnD,MAAA,IAAA,CAAK,MAAM,IAAA,CAAK,EAAE,IAAA,EAAY,GAAG,MAAM,CAAA;AAAA,IACzC,CAAA,MAAO;AACL,MAAA,IAAA,CAAK,EAAA,EAAI,IAAA;AAAA,QACP,KAAK,SAAA,CAAU;AAAA,UACb,IAAA;AAAA,UACA,GAAG;AAAA,SACJ;AAAA,OACH;AAAA,IACF;AAAA,EACF;AACF","file":"index.js","sourcesContent":["import { Readable } from 'stream';\nimport type { ToolsInput } from '@mastra/core/agent';\nimport { zodToJsonSchema } from 'zod-to-json-schema';\n\nexport type OpenAIExecuteFunction = (args: any) => Promise<any>;\ntype ToolDefinition = {\n type: 'function';\n name: string;\n description: string;\n parameters: {\n [key: string]: any;\n };\n};\n\ntype TTools = ToolsInput;\nexport const transformTools = (tools?: TTools) => {\n const openaiTools: { openaiTool: ToolDefinition; execute: OpenAIExecuteFunction }[] = [];\n for (const [name, tool] of Object.entries(tools || {})) {\n let parameters: { [key: string]: any };\n\n if ('inputSchema' in tool && tool.inputSchema) {\n if (isZodObject(tool.inputSchema)) {\n parameters = zodToJsonSchema(tool.inputSchema);\n delete parameters.$schema;\n } else {\n parameters = tool.inputSchema;\n }\n } else if ('parameters' in tool) {\n if (isZodObject(tool.parameters)) {\n parameters = zodToJsonSchema(tool.parameters);\n delete parameters.$schema;\n } else {\n parameters = tool.parameters;\n }\n } else {\n console.warn(`Tool ${name} has neither inputSchema nor parameters, skipping`);\n continue;\n }\n const openaiTool: ToolDefinition = {\n type: 'function',\n name,\n description: tool.description || `Tool: ${name}`,\n parameters,\n };\n\n if (tool.execute) {\n // Create an adapter function that works with both ToolAction and VercelTool execute functions\n const executeAdapter = async (args: any) => {\n try {\n if (!tool.execute) {\n throw new Error(`Tool ${name} has no execute function`);\n }\n\n // For ToolAction, the first argument is a context object with the args in a 'context' property\n if ('inputSchema' in tool) {\n return await tool.execute({ context: args });\n }\n // For VercelTool, pass args directly\n else {\n // Create a minimal ToolExecutionOptions object with required properties\n const options = {\n toolCallId: 'unknown',\n messages: [],\n };\n return await tool.execute(args, options);\n }\n } catch (error) {\n console.error(`Error executing tool ${name}:`, error);\n throw error;\n }\n };\n openaiTools.push({ openaiTool, execute: executeAdapter });\n } else {\n console.warn(`Tool ${name} has no execute function, skipping`);\n }\n }\n return openaiTools;\n};\n\nexport const isReadableStream = (obj: unknown) => {\n return (\n obj &&\n obj instanceof Readable &&\n typeof obj.read === 'function' &&\n typeof obj.pipe === 'function' &&\n obj.readable === true\n );\n};\n\nfunction isZodObject(schema: unknown) {\n return (\n !!schema &&\n typeof schema === 'object' &&\n '_def' in schema &&\n schema._def &&\n typeof schema._def === 'object' &&\n 'typeName' in schema._def &&\n schema._def.typeName === 'ZodObject'\n );\n}\n","import { EventEmitter } from 'events';\nimport { PassThrough } from 'stream';\nimport type { ToolsInput } from '@mastra/core/agent';\nimport type { RuntimeContext } from '@mastra/core/runtime-context';\nimport { MastraVoice } from '@mastra/core/voice';\nimport type { Realtime, RealtimeServerEvents } from 'openai-realtime-api';\nimport { WebSocket } from 'ws';\nimport { isReadableStream, transformTools } from './utils';\n\n/**\n * Event callback function type\n */\ntype EventCallback = (...args: any[]) => void;\n\ntype StreamWithId = PassThrough & { id: string };\n\n/**\n * Map of event types to their callback arrays\n */\ntype EventMap = {\n transcribing: [{ text: string }];\n writing: [{ text: string }];\n speaking: [{ audio: string }];\n speaker: [StreamWithId];\n error: [Error];\n} & {\n [key: string]: EventCallback[];\n};\n\n/** Default voice for text-to-speech. 'alloy' provides a neutral, balanced voice suitable for most use cases */\nconst DEFAULT_VOICE: Realtime.Voice = 'alloy';\n\nconst DEFAULT_TRANSCRIBER: Realtime.AudioTranscriptionModel = 'whisper-1';\n\nconst DEFAULT_URL = 'wss://api.openai.com/v1/realtime';\n\n/**\n * Default model for real-time voice interactions.\n * This model is optimized for low-latency responses while maintaining high quality output.\n */\nconst DEFAULT_MODEL = 'gpt-4o-mini-realtime-preview-2024-12-17';\n\n// /**\n// * Default Voice Activity Detection (VAD) configuration.\n// * These settings control how the system detects speech segments.\n// *\n// * @property {string} type - Uses server-side VAD for better accuracy\n// * @property {number} threshold - Speech detection sensitivity (0.5 = balanced)\n// * @property {number} prefix_padding_ms - Includes 1 second of audio before speech\n// * @property {number} silence_duration_ms - Waits 1 second of silence before ending turn\n// */\n// const DEFAULT_VAD_CONFIG = {\n// type: 'server_vad',\n// threshold: 0.5,\n// prefix_padding_ms: 1000,\n// silence_duration_ms: 1000,\n// } as Realtime.TurnDetection;\n\ntype TTools = ToolsInput;\n\n/**\n * Available voice options for text-to-speech.\n * Each voice has unique characteristics suitable for different use cases:\n * - alloy: Neutral and balanced\n * - echo: Warm and natural\n * - shimmer: Clear and expressive\n * - And more...\n */\nconst VOICES = ['alloy', 'ash', 'ballad', 'coral', 'echo', 'sage', 'shimmer', 'verse'];\n\ntype RealtimeClientServerEventMap = {\n [K in RealtimeServerEvents.EventType]: [RealtimeServerEvents.EventMap[K]];\n} & {\n ['conversation.item.input_audio_transcription.delta']: [{ delta: string; response_id: string }];\n ['conversation.item.input_audio_transcription.done']: [{ response_id: string }];\n};\n\n/**\n * OpenAIRealtimeVoice provides real-time voice interaction capabilities using OpenAI's\n * WebSocket-based API. It supports:\n * - Real-time text-to-speech\n * - Speech-to-text (transcription)\n * - Voice activity detection\n * - Multiple voice options\n * - Event-based audio streaming\n *\n * The class manages WebSocket connections, audio streaming, and event handling\n * for seamless voice interactions.\n *\n * @extends MastraVoice\n *\n * @example\n * ```typescript\n * const voice = new OpenAIRealtimeVoice({\n * apiKey: process.env.OPENAI_API_KEY,\n * model: 'gpt-4o-mini-realtime'\n * });\n *\n * await voice.open();\n * voice.on('speaking', (audioData) => {\n * // Handle audio data\n * });\n *\n * await voice.speak('Hello, how can I help you today?');\n * ```\n */\nexport class OpenAIRealtimeVoice extends MastraVoice {\n private ws?: WebSocket;\n private state: 'close' | 'open';\n private client: EventEmitter<RealtimeClientServerEventMap>;\n private events: EventMap;\n private instructions?: string;\n private tools?: TTools;\n private debug: boolean;\n private queue: unknown[] = [];\n private transcriber: Realtime.AudioTranscriptionModel;\n private runtimeContext?: RuntimeContext;\n /**\n * Creates a new instance of OpenAIRealtimeVoice.\n *\n * @param options - Configuration options for the voice instance\n * @param options.url - The base URL for the OpenAI Realtime API\n * @param options.model - The model ID to use (defaults to GPT-4 Mini Realtime)\n * @param options.apiKey - OpenAI API key. Falls back to process.env.OPENAI_API_KEY\n * @param options.speaker - Voice ID to use (defaults to 'alloy')\n * @param options.debug - Enable debug mode\n *\n * @example\n * ```typescript\n * const voice = new OpenAIRealtimeVoice({\n * apiKey: 'your-api-key',\n * model: 'gpt-4o-mini-realtime',\n * speaker: 'alloy'\n * });\n * ```\n */\n constructor(\n private options: {\n model?: string;\n url?: string;\n apiKey?: string;\n speaker?: Realtime.Voice;\n transcriber?: Realtime.AudioTranscriptionModel;\n debug?: boolean;\n } = {},\n ) {\n super();\n\n this.client = new EventEmitter();\n this.state = 'close';\n this.events = {} as EventMap;\n this.speaker = options.speaker || DEFAULT_VOICE;\n this.transcriber = options.transcriber || DEFAULT_TRANSCRIBER;\n this.debug = options.debug || false;\n }\n\n /**\n * Returns a list of available voice speakers.\n *\n * @returns Promise resolving to an array of voice objects, each containing at least a voiceId\n *\n * @example\n * ```typescript\n * const speakers = await voice.getSpeakers();\n * // speakers = [{ voiceId: 'alloy' }, { voiceId: 'echo' }, ...]\n * ```\n */\n getSpeakers(): Promise<Array<{ voiceId: string; [key: string]: any }>> {\n return Promise.resolve(VOICES.map(v => ({ voiceId: v })));\n }\n\n /**\n * Disconnects from the OpenAI realtime session and cleans up resources.\n * Should be called when you're done with the voice instance.\n *\n * @example\n * ```typescript\n * voice.close(); // Disconnects and cleans up\n * ```\n */\n close() {\n if (!this.ws) return;\n this.ws.close();\n this.state = 'close';\n }\n\n /**\n * Equips the voice instance with a set of instructions.\n * Instructions allow the model to perform additional actions during conversations.\n *\n * @param instructions - Optional instructions to addInstructions\n * @returns Transformed instructions ready for use with the model\n *\n * @example\n * ```typescript\n * voice.addInstructions('You are a helpful assistant.');\n * ```\n */\n addInstructions(instructions?: string) {\n this.instructions = instructions;\n }\n\n /**\n * Equips the voice instance with a set of tools.\n * Tools allow the model to perform additional actions during conversations.\n *\n * @param tools - Optional tools configuration to addTools\n * @returns Transformed tools configuration ready for use with the model\n *\n * @example\n * ```typescript\n * const tools = {\n * search: async (query: string) => { ... },\n * calculate: (expression: string) => { ... }\n * };\n * voice.addTools(tools);\n * ```\n */\n addTools(tools?: TTools) {\n this.tools = tools || {};\n }\n\n /**\n * Emits a speaking event using the configured voice model.\n * Can accept either a string or a readable stream as input.\n *\n * @param input - The text to convert to speech, or a readable stream containing the text\n * @param options - Optional configuration for this specific speech request\n * @param options.speaker - Override the voice to use for this specific request\n *\n * @throws {Error} If the input text is empty\n *\n * @example\n * ```typescript\n * // Simple text to speech\n * await voice.speak('Hello world');\n *\n * // With custom voice\n * await voice.speak('Hello world', { speaker: 'echo' });\n *\n * // Using a stream\n * const stream = fs.createReadStream('text.txt');\n * await voice.speak(stream);\n * ```\n */\n async speak(input: string | NodeJS.ReadableStream, options?: { speaker?: Realtime.Voice }): Promise<void> {\n if (typeof input !== 'string') {\n const chunks: Buffer[] = [];\n for await (const chunk of input) {\n chunks.push(Buffer.isBuffer(chunk) ? chunk : Buffer.from(String(chunk)));\n }\n input = Buffer.concat(chunks).toString('utf-8');\n }\n\n if (input.trim().length === 0) {\n throw new Error('Input text is empty');\n }\n\n this.sendEvent('response.create', {\n response: {\n instructions: `Repeat the following text: ${input}`,\n voice: options?.speaker ? options.speaker : undefined,\n },\n });\n }\n\n /**\n * Updates the session configuration for the voice instance.\n * This can be used to modify voice settings, turn detection, and other parameters.\n *\n * @param sessionConfig - New session configuration to apply\n *\n * @example\n * ```typescript\n * voice.updateConfig({\n * voice: 'echo',\n * turn_detection: {\n * type: 'server_vad',\n * threshold: 0.5,\n * silence_duration_ms: 1000\n * }\n * });\n * ```\n */\n updateConfig(sessionConfig: unknown): void {\n this.sendEvent('session.update', { session: sessionConfig });\n }\n\n /**\n * Checks if listening capabilities are enabled.\n *\n * @returns {Promise<{ enabled: boolean }>}\n */\n async getListener() {\n return { enabled: true };\n }\n\n /**\n * Processes audio input for speech recognition.\n * Takes a readable stream of audio data and emits a writing event.\n * The output of the writing event is int16 audio data.\n *\n * @param audioData - Readable stream containing the audio data to process\n * @param options - Optional configuration for audio processing\n *\n * @throws {Error} If the audio data format is not supported\n *\n * @example\n * ```typescript\n * // Process audio from a file\n * const audioStream = fs.createReadStream('audio.raw');\n * await voice.listen(audioStream);\n *\n * // Process audio with options\n * await voice.listen(microphoneStream, {\n * format: 'int16',\n * sampleRate: 24000\n * });\n * ```\n */\n async listen(audioData: NodeJS.ReadableStream): Promise<void> {\n if (isReadableStream(audioData)) {\n const chunks: Buffer[] = [];\n for await (const chunk of audioData) {\n const buffer = Buffer.isBuffer(chunk) ? chunk : Buffer.from(chunk);\n chunks.push(buffer);\n }\n\n const buffer = Buffer.concat(chunks);\n const int16Array = new Int16Array(buffer.buffer, buffer.byteOffset ?? 0, (buffer.byteLength ?? 0) / 2);\n const base64Audio = this.int16ArrayToBase64(int16Array);\n\n this.sendEvent('conversation.item.create', {\n item: {\n type: 'message',\n role: 'user',\n content: [{ type: 'input_audio', audio: base64Audio }],\n },\n });\n\n this.sendEvent('response.create', {\n response: {\n modalities: ['text'],\n instructions: `ONLY repeat the input and DO NOT say anything else`,\n },\n });\n } else {\n this.emit('error', new Error('Unsupported audio data format'));\n }\n }\n\n waitForOpen() {\n return new Promise(resolve => {\n this.ws?.on('open', resolve);\n });\n }\n\n waitForSessionCreated() {\n return new Promise(resolve => {\n this.client.on('session.created', resolve);\n });\n }\n\n /**\n * Establishes a connection to the OpenAI realtime service.\n * Must be called before using speak, listen, or relay functions.\n *\n * @throws {Error} If connection fails or session creation times out\n *\n * @example\n * ```typescript\n * await voice.open();\n * // Now ready for voice interactions\n * ```\n */\n async connect({ runtimeContext }: { runtimeContext?: RuntimeContext } = {}) {\n const url = `${this.options.url || DEFAULT_URL}?model=${this.options.model || DEFAULT_MODEL}`;\n const apiKey = this.options.apiKey || process.env.OPENAI_API_KEY;\n this.runtimeContext = runtimeContext;\n\n this.ws = new WebSocket(url, undefined, {\n headers: {\n Authorization: 'Bearer ' + apiKey,\n 'OpenAI-Beta': 'realtime=v1',\n },\n });\n\n this.setupEventListeners();\n await Promise.all([this.waitForOpen(), this.waitForSessionCreated()]);\n\n const openaiTools = transformTools(this.tools);\n this.updateConfig({\n instructions: this.instructions,\n tools: openaiTools.map(t => t.openaiTool),\n input_audio_transcription: {\n model: this.transcriber,\n },\n voice: this.speaker,\n });\n this.state = 'open';\n }\n\n disconnect() {\n this.state = 'close';\n this.ws?.close();\n }\n\n /**\n * Streams audio data in real-time to the OpenAI service.\n * Useful for continuous audio streaming scenarios like live microphone input.\n * Must be in 'open' state before calling this method.\n *\n * @param audioData - Readable stream of audio data to relay\n * @throws {Error} If audio format is not supported\n *\n * @example\n * ```typescript\n * // First connect\n * await voice.open();\n *\n * // Then relay audio\n * const micStream = getMicrophoneStream();\n * await voice.relay(micStream);\n * ```\n */\n async send(audioData: NodeJS.ReadableStream | Int16Array, eventId?: string): Promise<void> {\n if (!this.state || this.state !== 'open') {\n console.warn('Cannot relay audio when not open. Call open() first.');\n return;\n }\n\n if (isReadableStream(audioData)) {\n const stream = audioData as NodeJS.ReadableStream;\n stream.on('data', chunk => {\n try {\n const buffer = Buffer.isBuffer(chunk) ? chunk : Buffer.from(chunk);\n this.sendEvent('input_audio_buffer.append', { audio: buffer.toString('base64'), event_id: eventId });\n } catch (err) {\n this.emit('error', err);\n }\n });\n } else if (audioData instanceof Int16Array) {\n try {\n const base64Audio = this.int16ArrayToBase64(audioData);\n this.sendEvent('input_audio_buffer.append', { audio: base64Audio, event_id: eventId });\n } catch (err) {\n this.emit('error', err);\n }\n } else {\n this.emit('error', new Error('Unsupported audio data format'));\n }\n }\n\n /**\n * Sends a response to the OpenAI Realtime API.\n *\n * Trigger a response to the real-time session.\n *\n * @param {Object} params - The parameters object\n * @param {Realtime.ResponseConfig} params.options - Configuration options for the response\n * @returns {Promise<void>} A promise that resolves when the response has been sent\n *\n * @example\n * // Send a simple text response\n * await realtimeVoice.answer({\n * options: {\n * content: \"Hello, how can I help you today?\",\n * voice: \"alloy\"\n * }\n * });\n */\n async answer({ options }: { options?: Realtime.ResponseConfig }) {\n this.sendEvent('response.create', { response: options ?? {} });\n }\n\n /**\n * Registers an event listener for voice events.\n * Available events: 'speaking', 'writing, 'error'\n * Can listen to OpenAI Realtime events by prefixing with 'openAIRealtime:'\n * Such as 'openAIRealtime:conversation.item.completed', 'openAIRealtime:conversation.updated', etc.\n *\n * @param event - Name of the event to listen for\n * @param callback - Function to call when the event occurs\n *\n * @example\n * ```typescript\n * // Listen for speech events\n * voice.on('speaking', (audioData: Int16Array) => {\n * // Handle audio data\n * });\n *\n * // Handle errors\n * voice.on('error', (error: Error) => {\n * console.error('Voice error:', error);\n * });\n * ```\n */\n on(event: string, callback: EventCallback): void {\n if (!this.events[event]) {\n this.events[event] = [];\n }\n this.events[event].push(callback);\n }\n\n /**\n * Removes a previously registered event listener.\n *\n * @param event - Name of the event to stop listening to\n * @param callback - The specific callback function to remove\n *\n * @example\n * ```typescript\n * // Create event handler\n * const handleSpeech = (audioData: Int16Array) => {\n * // Handle audio data\n * };\n *\n * // Add listener\n * voice.on('speaking', handleSpeech);\n *\n * // Later, remove the listener\n * voice.off('speaking', handleSpeech);\n * ```\n */\n off(event: string, callback: EventCallback): void {\n if (!this.events[event]) return;\n\n const index = this.events[event].indexOf(callback);\n if (index !== -1) {\n this.events[event].splice(index, 1);\n }\n }\n\n /**\n * Emit an event with arguments\n * @param event Event name\n * @param args Arguments to pass to the callbacks\n */\n private emit(event: string, ...args: any[]): void {\n if (!this.events[event]) return;\n\n for (const callback of this.events[event]) {\n callback(...args);\n }\n }\n\n private setupEventListeners(): void {\n const speakerStreams = new Map<string, StreamWithId>();\n\n if (!this.ws) {\n throw new Error('WebSocket not initialized');\n }\n\n this.ws.on('message', message => {\n const data = JSON.parse(message.toString());\n this.client.emit(data.type, data);\n\n if (this.debug) {\n const { delta, ...fields } = data;\n console.log(data.type, fields, delta?.length < 100 ? delta : '');\n }\n });\n\n this.client.on('session.created', ev => {\n this.emit('session.created', ev);\n\n const queue = this.queue.splice(0, this.queue.length);\n for (const ev of queue) {\n this.ws?.send(JSON.stringify(ev));\n }\n });\n this.client.on('session.updated', ev => {\n this.emit('session.updated', ev);\n });\n this.client.on('response.created', ev => {\n this.emit('response.created', ev);\n\n const speakerStream = new PassThrough() as StreamWithId;\n\n speakerStream.id = ev.response.id;\n\n speakerStreams.set(ev.response.id, speakerStream);\n this.emit('speaker', speakerStream);\n });\n this.client.on('conversation.item.input_audio_transcription.delta', ev => {\n this.emit('writing', { text: ev.delta, response_id: ev.response_id, role: 'user' });\n });\n this.client.on('conversation.item.input_audio_transcription.done', ev => {\n this.emit('writing', { text: '\\n', response_id: ev.response_id, role: 'user' });\n });\n this.client.on('response.audio.delta', ev => {\n const audio = Buffer.from(ev.delta, 'base64');\n this.emit('speaking', { audio, response_id: ev.response_id });\n\n const stream = speakerStreams.get(ev.response_id);\n stream?.write(audio);\n });\n this.client.on('response.audio.done', ev => {\n this.emit('speaking.done', { response_id: ev.response_id });\n\n const stream = speakerStreams.get(ev.response_id);\n stream?.end();\n });\n this.client.on('response.audio_transcript.delta', ev => {\n this.emit('writing', { text: ev.delta, response_id: ev.response_id, role: 'assistant' });\n });\n this.client.on('response.audio_transcript.done', ev => {\n this.emit('writing', { text: '\\n', response_id: ev.response_id, role: 'assistant' });\n });\n this.client.on('response.text.delta', ev => {\n this.emit('writing', { text: ev.delta, response_id: ev.response_id, role: 'assistant' });\n });\n this.client.on('response.text.done', ev => {\n this.emit('writing', { text: '\\n', response_id: ev.response_id, role: 'assistant' });\n });\n this.client.on('response.done', async ev => {\n await this.handleFunctionCalls(ev);\n this.emit('response.done', ev);\n speakerStreams.delete(ev.response.id);\n });\n this.client.on('error', async ev => {\n this.emit('error', ev);\n });\n }\n\n private async handleFunctionCalls(ev: any) {\n for (const output of ev.response?.output ?? []) {\n if (output.type === 'function_call') {\n await this.handleFunctionCall(output);\n }\n }\n }\n\n private async handleFunctionCall(output: any) {\n try {\n const context = JSON.parse(output.arguments);\n const tool = this.tools?.[output.name];\n if (!tool) {\n console.warn(`Tool \"${output.name}\" not found`);\n return;\n }\n\n if (tool?.execute) {\n this.emit('tool-call-start', {\n toolCallId: output.call_id,\n toolName: output.name,\n toolDescription: tool.description,\n args: context,\n });\n }\n\n const result = await tool?.execute?.(\n { context, runtimeContext: this.runtimeContext },\n {\n toolCallId: output.call_id,\n messages: [],\n },\n );\n\n this.emit('tool-call-result', {\n toolCallId: output.call_id,\n toolName: output.name,\n toolDescription: tool.description,\n args: context,\n result,\n });\n\n this.sendEvent('conversation.item.create', {\n item: {\n type: 'function_call_output',\n call_id: output.call_id,\n output: JSON.stringify(result),\n },\n });\n } catch (e) {\n const err = e as Error;\n console.warn(`Error calling tool \"${output.name}\":`, err.message);\n this.sendEvent('conversation.item.create', {\n item: {\n type: 'function_call_output',\n call_id: output.call_id,\n output: JSON.stringify({ error: err.message }),\n },\n });\n } finally {\n this.sendEvent('response.create', {});\n }\n }\n\n private int16ArrayToBase64(int16Array: Int16Array): string {\n const buffer = new ArrayBuffer(int16Array.length * 2);\n const view = new DataView(buffer);\n for (let i = 0; i < int16Array.length; i++) {\n view.setInt16(i * 2, int16Array[i]!, true);\n }\n const uint8Array = new Uint8Array(buffer);\n let binary = '';\n for (let i = 0; i < uint8Array.length; i++) {\n binary += String.fromCharCode(uint8Array[i]!);\n }\n return btoa(binary);\n }\n\n private sendEvent(type: string, data: any) {\n if (!this.ws || this.ws.readyState !== this.ws.OPEN) {\n this.queue.push({ type: type, ...data });\n } else {\n this.ws?.send(\n JSON.stringify({\n type: type,\n ...data,\n }),\n );\n }\n }\n}\n"]}
|
package/dist/utils.d.ts
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
1
|
+
import type { ToolsInput } from '@mastra/core/agent';
|
|
2
|
+
export type OpenAIExecuteFunction = (args: any) => Promise<any>;
|
|
3
|
+
type ToolDefinition = {
|
|
4
|
+
type: 'function';
|
|
5
|
+
name: string;
|
|
6
|
+
description: string;
|
|
7
|
+
parameters: {
|
|
8
|
+
[key: string]: any;
|
|
9
|
+
};
|
|
10
|
+
};
|
|
11
|
+
type TTools = ToolsInput;
|
|
12
|
+
export declare const transformTools: (tools?: TTools) => {
|
|
13
|
+
openaiTool: ToolDefinition;
|
|
14
|
+
execute: OpenAIExecuteFunction;
|
|
15
|
+
}[];
|
|
16
|
+
export declare const isReadableStream: (obj: unknown) => unknown;
|
|
17
|
+
export {};
|
|
18
|
+
//# sourceMappingURL=utils.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"utils.d.ts","sourceRoot":"","sources":["../src/utils.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,UAAU,EAAE,MAAM,oBAAoB,CAAC;AAGrD,MAAM,MAAM,qBAAqB,GAAG,CAAC,IAAI,EAAE,GAAG,KAAK,OAAO,CAAC,GAAG,CAAC,CAAC;AAChE,KAAK,cAAc,GAAG;IACpB,IAAI,EAAE,UAAU,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,WAAW,EAAE,MAAM,CAAC;IACpB,UAAU,EAAE;QACV,CAAC,GAAG,EAAE,MAAM,GAAG,GAAG,CAAC;KACpB,CAAC;CACH,CAAC;AAEF,KAAK,MAAM,GAAG,UAAU,CAAC;AACzB,eAAO,MAAM,cAAc,GAAI,QAAQ,MAAM;gBACV,cAAc;aAAW,qBAAqB;GA6DhF,CAAC;AAEF,eAAO,MAAM,gBAAgB,GAAI,KAAK,OAAO,YAQ5C,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/voice-openai-realtime",
|
|
3
|
-
"version": "0.0.0-
|
|
3
|
+
"version": "0.0.0-zod-v4-compat-part-2-20250822105954",
|
|
4
4
|
"description": "Mastra OpenAI Realtime API integration",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -18,31 +18,31 @@
|
|
|
18
18
|
},
|
|
19
19
|
"./package.json": "./package.json"
|
|
20
20
|
},
|
|
21
|
-
"license": "
|
|
21
|
+
"license": "Apache-2.0",
|
|
22
22
|
"dependencies": {
|
|
23
|
-
"openai-realtime-api": "^1.0.
|
|
24
|
-
"ws": "^8.18.
|
|
23
|
+
"openai-realtime-api": "^1.0.8",
|
|
24
|
+
"ws": "^8.18.3",
|
|
25
25
|
"zod-to-json-schema": "^3.24.5"
|
|
26
26
|
},
|
|
27
27
|
"devDependencies": {
|
|
28
28
|
"@microsoft/api-extractor": "^7.52.8",
|
|
29
29
|
"@types/node": "^20.19.0",
|
|
30
30
|
"@types/ws": "^8.18.1",
|
|
31
|
-
"eslint": "^9.
|
|
31
|
+
"eslint": "^9.30.1",
|
|
32
32
|
"tsup": "^8.5.0",
|
|
33
33
|
"typescript": "^5.8.3",
|
|
34
|
-
"vitest": "^2.
|
|
35
|
-
"zod": "^3.25.
|
|
36
|
-
"@internal/lint": "0.0.0-
|
|
37
|
-
"@mastra/core": "0.0.0-
|
|
34
|
+
"vitest": "^3.2.4",
|
|
35
|
+
"zod": "^3.25.67",
|
|
36
|
+
"@internal/lint": "0.0.0-zod-v4-compat-part-2-20250822105954",
|
|
37
|
+
"@mastra/core": "0.0.0-zod-v4-compat-part-2-20250822105954"
|
|
38
38
|
},
|
|
39
39
|
"peerDependencies": {
|
|
40
|
-
"zod": "^3.0.0",
|
|
41
|
-
"@mastra/core": "0.0.0-
|
|
40
|
+
"zod": "^3.0.0 || ^4.0.0",
|
|
41
|
+
"@mastra/core": "0.0.0-zod-v4-compat-part-2-20250822105954"
|
|
42
42
|
},
|
|
43
43
|
"scripts": {
|
|
44
|
-
"build": "tsup
|
|
45
|
-
"build:watch": "
|
|
44
|
+
"build": "tsup --silent --config tsup.config.ts",
|
|
45
|
+
"build:watch": "tsup --watch --silent --config tsup.config.ts",
|
|
46
46
|
"test": "vitest run",
|
|
47
47
|
"lint": "eslint ."
|
|
48
48
|
}
|
package/src/index.test.ts
CHANGED
|
@@ -9,9 +9,6 @@ vi.mock('openai-realtime-api', () => {
|
|
|
9
9
|
disconnect: vi.fn(),
|
|
10
10
|
waitForSessionCreated: vi.fn().mockResolvedValue(undefined),
|
|
11
11
|
updateSession: vi.fn(),
|
|
12
|
-
realtime: {
|
|
13
|
-
send: vi.fn(),
|
|
14
|
-
},
|
|
15
12
|
appendInputAudio: vi.fn(),
|
|
16
13
|
on: vi.fn(),
|
|
17
14
|
emit: vi.fn(),
|
|
@@ -19,26 +16,35 @@ vi.mock('openai-realtime-api', () => {
|
|
|
19
16
|
};
|
|
20
17
|
});
|
|
21
18
|
|
|
19
|
+
vi.mock('ws', () => {
|
|
20
|
+
return {
|
|
21
|
+
WebSocket: vi.fn().mockImplementation(() => ({
|
|
22
|
+
send: vi.fn(),
|
|
23
|
+
close: vi.fn(),
|
|
24
|
+
on: vi.fn(),
|
|
25
|
+
})),
|
|
26
|
+
};
|
|
27
|
+
});
|
|
28
|
+
|
|
22
29
|
describe('OpenAIRealtimeVoice', () => {
|
|
23
30
|
let voice: OpenAIRealtimeVoice;
|
|
24
|
-
let mockClient: any; // TODO: Replace with proper type once we have better type definitions
|
|
25
31
|
|
|
26
32
|
beforeEach(() => {
|
|
27
33
|
vi.clearAllMocks();
|
|
28
34
|
voice = new OpenAIRealtimeVoice({
|
|
29
35
|
apiKey: 'test-api-key',
|
|
30
36
|
});
|
|
31
|
-
|
|
37
|
+
voice.waitForOpen = () => Promise.resolve();
|
|
38
|
+
voice.waitForSessionCreated = () => Promise.resolve();
|
|
32
39
|
});
|
|
33
40
|
|
|
34
41
|
afterEach(() => {
|
|
35
|
-
voice?.
|
|
42
|
+
voice?.disconnect();
|
|
36
43
|
});
|
|
37
44
|
|
|
38
45
|
describe('initialization', () => {
|
|
39
46
|
it('should initialize with default values', () => {
|
|
40
47
|
expect(voice).toBeInstanceOf(OpenAIRealtimeVoice);
|
|
41
|
-
expect(mockClient).toBeDefined();
|
|
42
48
|
});
|
|
43
49
|
|
|
44
50
|
it('should initialize with custom speaker', () => {
|
|
@@ -58,31 +64,10 @@ describe('OpenAIRealtimeVoice', () => {
|
|
|
58
64
|
});
|
|
59
65
|
});
|
|
60
66
|
|
|
61
|
-
describe('huddle and leave', () => {
|
|
62
|
-
it('should connect and update state on huddle', async () => {
|
|
63
|
-
await voice.huddle();
|
|
64
|
-
expect(mockClient.connect).toHaveBeenCalled();
|
|
65
|
-
expect(mockClient.waitForSessionCreated).toHaveBeenCalled();
|
|
66
|
-
expect((voice as any).state).toBe('huddle');
|
|
67
|
-
});
|
|
68
|
-
|
|
69
|
-
it('should disconnect and update state on leave', () => {
|
|
70
|
-
voice.leave();
|
|
71
|
-
expect(mockClient.disconnect).toHaveBeenCalled();
|
|
72
|
-
expect((voice as any).state).toBe('leave');
|
|
73
|
-
});
|
|
74
|
-
});
|
|
75
|
-
|
|
76
67
|
describe('speak', () => {
|
|
77
68
|
it('should handle string input', async () => {
|
|
78
69
|
const testText = 'Hello, world!';
|
|
79
70
|
await voice.speak(testText);
|
|
80
|
-
expect(mockClient.realtime.send).toHaveBeenCalledWith('response.create', {
|
|
81
|
-
response: {
|
|
82
|
-
instructions: `Repeat the following text: ${testText}`,
|
|
83
|
-
voice: undefined,
|
|
84
|
-
},
|
|
85
|
-
});
|
|
86
71
|
});
|
|
87
72
|
|
|
88
73
|
it('should throw error on empty input', async () => {
|
|
@@ -90,6 +75,15 @@ describe('OpenAIRealtimeVoice', () => {
|
|
|
90
75
|
});
|
|
91
76
|
});
|
|
92
77
|
|
|
78
|
+
describe('send', () => {
|
|
79
|
+
it('should handle Int16Array input', async () => {
|
|
80
|
+
const testArray = new Int16Array([1, 2, 3]);
|
|
81
|
+
|
|
82
|
+
await voice.connect();
|
|
83
|
+
voice.send(testArray);
|
|
84
|
+
});
|
|
85
|
+
});
|
|
86
|
+
|
|
93
87
|
describe('event handling', () => {
|
|
94
88
|
it('should register and trigger event listeners', () => {
|
|
95
89
|
const mockCallback = vi.fn();
|
package/src/index.ts
CHANGED
|
@@ -441,7 +441,8 @@ export class OpenAIRealtimeVoice extends MastraVoice {
|
|
|
441
441
|
});
|
|
442
442
|
} else if (audioData instanceof Int16Array) {
|
|
443
443
|
try {
|
|
444
|
-
this.
|
|
444
|
+
const base64Audio = this.int16ArrayToBase64(audioData);
|
|
445
|
+
this.sendEvent('input_audio_buffer.append', { audio: base64Audio, event_id: eventId });
|
|
445
446
|
} catch (err) {
|
|
446
447
|
this.emit('error', err);
|
|
447
448
|
}
|
|
@@ -617,6 +618,9 @@ export class OpenAIRealtimeVoice extends MastraVoice {
|
|
|
617
618
|
this.emit('response.done', ev);
|
|
618
619
|
speakerStreams.delete(ev.response.id);
|
|
619
620
|
});
|
|
621
|
+
this.client.on('error', async ev => {
|
|
622
|
+
this.emit('error', ev);
|
|
623
|
+
});
|
|
620
624
|
}
|
|
621
625
|
|
|
622
626
|
private async handleFunctionCalls(ev: any) {
|