@mastra/mcp-docs-server 1.1.7 → 1.1.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (276) hide show
  1. package/.docs/docs/agents/agent-approval.md +61 -31
  2. package/.docs/docs/agents/supervisor-agents.md +1 -1
  3. package/.docs/docs/getting-started/manual-install.md +1 -1
  4. package/.docs/docs/index.md +68 -24
  5. package/.docs/docs/mastra-cloud/setup.md +1 -1
  6. package/.docs/docs/memory/observational-memory.md +9 -0
  7. package/.docs/docs/memory/semantic-recall.md +17 -1
  8. package/.docs/docs/workspace/skills.md +7 -5
  9. package/.docs/guides/deployment/aws-lambda.md +76 -165
  10. package/.docs/guides/deployment/azure-app-services.md +38 -61
  11. package/.docs/guides/deployment/cloudflare.md +1 -1
  12. package/.docs/guides/deployment/netlify.md +1 -1
  13. package/.docs/guides/deployment/vercel.md +1 -1
  14. package/.docs/models/gateways/netlify.md +5 -1
  15. package/.docs/models/index.md +1 -1
  16. package/.docs/models/providers/anthropic.md +9 -9
  17. package/.docs/models/providers/deepseek.md +1 -1
  18. package/.docs/models/providers/google.md +11 -11
  19. package/.docs/models/providers/openai.md +19 -19
  20. package/.docs/models/providers/opencode.md +33 -33
  21. package/.docs/models/providers/xai.md +3 -3
  22. package/.docs/reference/agents/agent.md +20 -20
  23. package/.docs/reference/agents/generate.md +200 -66
  24. package/.docs/reference/agents/generateLegacy.md +77 -35
  25. package/.docs/reference/agents/getDefaultGenerateOptions.md +4 -6
  26. package/.docs/reference/agents/getDefaultOptions.md +4 -6
  27. package/.docs/reference/agents/getDefaultStreamOptions.md +4 -6
  28. package/.docs/reference/agents/getDescription.md +1 -1
  29. package/.docs/reference/agents/getInstructions.md +4 -6
  30. package/.docs/reference/agents/getLLM.md +6 -8
  31. package/.docs/reference/agents/getMemory.md +4 -6
  32. package/.docs/reference/agents/getModel.md +4 -6
  33. package/.docs/reference/agents/getTools.md +5 -7
  34. package/.docs/reference/agents/getVoice.md +4 -6
  35. package/.docs/reference/agents/listAgents.md +4 -6
  36. package/.docs/reference/agents/listScorers.md +4 -6
  37. package/.docs/reference/agents/listTools.md +4 -6
  38. package/.docs/reference/agents/listWorkflows.md +4 -6
  39. package/.docs/reference/agents/network.md +69 -23
  40. package/.docs/reference/ai-sdk/chat-route.md +7 -7
  41. package/.docs/reference/ai-sdk/handle-chat-stream.md +12 -12
  42. package/.docs/reference/ai-sdk/handle-network-stream.md +4 -4
  43. package/.docs/reference/ai-sdk/handle-workflow-stream.md +11 -11
  44. package/.docs/reference/ai-sdk/network-route.md +3 -3
  45. package/.docs/reference/ai-sdk/to-ai-sdk-stream.md +9 -9
  46. package/.docs/reference/ai-sdk/to-ai-sdk-v4-messages.md +9 -9
  47. package/.docs/reference/ai-sdk/to-ai-sdk-v5-messages.md +5 -5
  48. package/.docs/reference/ai-sdk/with-mastra.md +12 -12
  49. package/.docs/reference/ai-sdk/workflow-route.md +3 -3
  50. package/.docs/reference/auth/auth0.md +6 -6
  51. package/.docs/reference/auth/better-auth.md +5 -5
  52. package/.docs/reference/auth/clerk.md +5 -5
  53. package/.docs/reference/auth/firebase.md +7 -7
  54. package/.docs/reference/auth/jwt.md +1 -1
  55. package/.docs/reference/auth/supabase.md +4 -4
  56. package/.docs/reference/auth/workos.md +6 -6
  57. package/.docs/reference/cli/create-mastra.md +14 -14
  58. package/.docs/reference/client-js/mastra-client.md +23 -23
  59. package/.docs/reference/client-js/workflows.md +3 -3
  60. package/.docs/reference/core/addGateway.md +2 -2
  61. package/.docs/reference/core/getAgent.md +2 -2
  62. package/.docs/reference/core/getAgentById.md +2 -2
  63. package/.docs/reference/core/getDeployer.md +1 -1
  64. package/.docs/reference/core/getGateway.md +2 -2
  65. package/.docs/reference/core/getGatewayById.md +2 -2
  66. package/.docs/reference/core/getLogger.md +1 -1
  67. package/.docs/reference/core/getMCPServer.md +2 -2
  68. package/.docs/reference/core/getMCPServerById.md +3 -3
  69. package/.docs/reference/core/getMemory.md +2 -2
  70. package/.docs/reference/core/getScorer.md +2 -2
  71. package/.docs/reference/core/getScorerById.md +2 -2
  72. package/.docs/reference/core/getServer.md +1 -1
  73. package/.docs/reference/core/getStorage.md +1 -1
  74. package/.docs/reference/core/getStoredAgentById.md +18 -20
  75. package/.docs/reference/core/getTelemetry.md +1 -1
  76. package/.docs/reference/core/getVector.md +2 -2
  77. package/.docs/reference/core/getWorkflow.md +3 -3
  78. package/.docs/reference/core/listAgents.md +1 -1
  79. package/.docs/reference/core/listGateways.md +1 -1
  80. package/.docs/reference/core/listLogs.md +9 -11
  81. package/.docs/reference/core/listLogsByRunId.md +9 -9
  82. package/.docs/reference/core/listMCPServers.md +1 -1
  83. package/.docs/reference/core/listMemory.md +1 -1
  84. package/.docs/reference/core/listScorers.md +1 -1
  85. package/.docs/reference/core/listStoredAgents.md +9 -11
  86. package/.docs/reference/core/listVectors.md +1 -1
  87. package/.docs/reference/core/listWorkflows.md +2 -2
  88. package/.docs/reference/core/mastra-class.md +17 -17
  89. package/.docs/reference/core/mastra-model-gateway.md +15 -15
  90. package/.docs/reference/core/setLogger.md +2 -4
  91. package/.docs/reference/core/setStorage.md +1 -1
  92. package/.docs/reference/datasets/addItem.md +20 -4
  93. package/.docs/reference/datasets/addItems.md +8 -2
  94. package/.docs/reference/datasets/compareExperiments.md +15 -3
  95. package/.docs/reference/datasets/create.md +6 -6
  96. package/.docs/reference/datasets/dataset.md +1 -1
  97. package/.docs/reference/datasets/delete.md +2 -2
  98. package/.docs/reference/datasets/deleteExperiment.md +2 -2
  99. package/.docs/reference/datasets/deleteItem.md +2 -2
  100. package/.docs/reference/datasets/deleteItems.md +2 -2
  101. package/.docs/reference/datasets/get.md +2 -2
  102. package/.docs/reference/datasets/getDetails.md +9 -9
  103. package/.docs/reference/datasets/getExperiment.md +2 -2
  104. package/.docs/reference/datasets/getItem.md +3 -3
  105. package/.docs/reference/datasets/getItemHistory.md +22 -2
  106. package/.docs/reference/datasets/list.md +7 -3
  107. package/.docs/reference/datasets/listExperimentResults.md +34 -4
  108. package/.docs/reference/datasets/listExperiments.md +41 -3
  109. package/.docs/reference/datasets/listItems.md +18 -6
  110. package/.docs/reference/datasets/listVersions.md +23 -3
  111. package/.docs/reference/datasets/startExperiment.md +62 -12
  112. package/.docs/reference/datasets/startExperimentAsync.md +5 -1
  113. package/.docs/reference/datasets/update.md +6 -6
  114. package/.docs/reference/datasets/updateItem.md +5 -5
  115. package/.docs/reference/deployer.md +8 -8
  116. package/.docs/reference/evals/answer-relevancy.md +11 -11
  117. package/.docs/reference/evals/answer-similarity.md +17 -19
  118. package/.docs/reference/evals/bias.md +10 -10
  119. package/.docs/reference/evals/completeness.md +3 -3
  120. package/.docs/reference/evals/content-similarity.md +6 -6
  121. package/.docs/reference/evals/context-precision.md +4 -4
  122. package/.docs/reference/evals/context-relevance.md +4 -4
  123. package/.docs/reference/evals/create-scorer.md +47 -49
  124. package/.docs/reference/evals/faithfulness.md +11 -11
  125. package/.docs/reference/evals/hallucination.md +17 -21
  126. package/.docs/reference/evals/keyword-coverage.md +4 -4
  127. package/.docs/reference/evals/mastra-scorer.md +14 -14
  128. package/.docs/reference/evals/noise-sensitivity.md +4 -4
  129. package/.docs/reference/evals/prompt-alignment.md +4 -4
  130. package/.docs/reference/evals/run-evals.md +16 -16
  131. package/.docs/reference/evals/scorer-utils.md +3 -3
  132. package/.docs/reference/evals/textual-difference.md +3 -3
  133. package/.docs/reference/evals/tone-consistency.md +3 -3
  134. package/.docs/reference/evals/tool-call-accuracy.md +5 -5
  135. package/.docs/reference/evals/toxicity.md +8 -8
  136. package/.docs/reference/harness/harness-class.md +34 -42
  137. package/.docs/reference/logging/pino-logger.md +5 -5
  138. package/.docs/reference/memory/clone-utilities.md +5 -5
  139. package/.docs/reference/memory/cloneThread.md +17 -21
  140. package/.docs/reference/memory/createThread.md +10 -10
  141. package/.docs/reference/memory/deleteMessages.md +2 -2
  142. package/.docs/reference/memory/getThreadById.md +2 -2
  143. package/.docs/reference/memory/listThreads.md +5 -5
  144. package/.docs/reference/memory/memory-class.md +12 -14
  145. package/.docs/reference/memory/observational-memory.md +102 -94
  146. package/.docs/reference/memory/recall.md +14 -16
  147. package/.docs/reference/observability/tracing/configuration.md +27 -10
  148. package/.docs/reference/observability/tracing/exporters/console-exporter.md +4 -7
  149. package/.docs/reference/processors/batch-parts-processor.md +8 -10
  150. package/.docs/reference/processors/language-detector.md +14 -16
  151. package/.docs/reference/processors/message-history-processor.md +7 -9
  152. package/.docs/reference/processors/moderation-processor.md +13 -15
  153. package/.docs/reference/processors/pii-detector.md +14 -16
  154. package/.docs/reference/processors/processor-interface.md +62 -62
  155. package/.docs/reference/processors/prompt-injection-detector.md +11 -13
  156. package/.docs/reference/processors/semantic-recall-processor.md +14 -16
  157. package/.docs/reference/processors/system-prompt-scrubber.md +12 -14
  158. package/.docs/reference/processors/token-limiter-processor.md +11 -13
  159. package/.docs/reference/processors/tool-call-filter.md +5 -7
  160. package/.docs/reference/processors/tool-search-processor.md +9 -11
  161. package/.docs/reference/processors/unicode-normalizer.md +8 -10
  162. package/.docs/reference/processors/working-memory-processor.md +14 -18
  163. package/.docs/reference/rag/chunk.md +38 -38
  164. package/.docs/reference/rag/database-config.md +11 -7
  165. package/.docs/reference/rag/document.md +2 -2
  166. package/.docs/reference/rag/embeddings.md +12 -12
  167. package/.docs/reference/rag/extract-params.md +23 -23
  168. package/.docs/reference/rag/graph-rag.md +12 -12
  169. package/.docs/reference/rag/rerank.md +25 -17
  170. package/.docs/reference/rag/rerankWithScorer.md +25 -17
  171. package/.docs/reference/server/create-route.md +14 -14
  172. package/.docs/reference/server/express-adapter.md +10 -10
  173. package/.docs/reference/server/fastify-adapter.md +10 -10
  174. package/.docs/reference/server/hono-adapter.md +10 -10
  175. package/.docs/reference/server/koa-adapter.md +10 -10
  176. package/.docs/reference/server/mastra-server.md +10 -10
  177. package/.docs/reference/server/register-api-route.md +13 -13
  178. package/.docs/reference/storage/cloudflare-d1.md +5 -5
  179. package/.docs/reference/storage/cloudflare.md +6 -6
  180. package/.docs/reference/storage/composite.md +9 -9
  181. package/.docs/reference/storage/convex.md +3 -3
  182. package/.docs/reference/storage/dynamodb.md +9 -9
  183. package/.docs/reference/storage/lance.md +3 -3
  184. package/.docs/reference/storage/libsql.md +2 -2
  185. package/.docs/reference/storage/mongodb.md +5 -5
  186. package/.docs/reference/storage/mssql.md +2 -2
  187. package/.docs/reference/storage/postgresql.md +25 -25
  188. package/.docs/reference/storage/upstash.md +3 -3
  189. package/.docs/reference/streaming/ChunkType.md +251 -59
  190. package/.docs/reference/streaming/agents/MastraModelOutput.md +86 -16
  191. package/.docs/reference/streaming/agents/stream.md +155 -43
  192. package/.docs/reference/streaming/agents/streamLegacy.md +79 -39
  193. package/.docs/reference/streaming/workflows/resumeStream.md +18 -8
  194. package/.docs/reference/streaming/workflows/stream.md +21 -9
  195. package/.docs/reference/streaming/workflows/timeTravelStream.md +4 -4
  196. package/.docs/reference/tools/create-tool.md +25 -21
  197. package/.docs/reference/tools/document-chunker-tool.md +7 -7
  198. package/.docs/reference/tools/graph-rag-tool.md +16 -18
  199. package/.docs/reference/tools/mcp-client.md +38 -27
  200. package/.docs/reference/tools/mcp-server.md +45 -45
  201. package/.docs/reference/tools/vector-query-tool.md +34 -22
  202. package/.docs/reference/vectors/astra.md +22 -22
  203. package/.docs/reference/vectors/chroma.md +44 -44
  204. package/.docs/reference/vectors/convex.md +26 -26
  205. package/.docs/reference/vectors/couchbase.md +30 -30
  206. package/.docs/reference/vectors/duckdb.md +29 -29
  207. package/.docs/reference/vectors/elasticsearch.md +27 -27
  208. package/.docs/reference/vectors/lance.md +46 -38
  209. package/.docs/reference/vectors/libsql.md +31 -31
  210. package/.docs/reference/vectors/mongodb.md +32 -32
  211. package/.docs/reference/vectors/opensearch.md +27 -27
  212. package/.docs/reference/vectors/pg.md +60 -44
  213. package/.docs/reference/vectors/pinecone.md +38 -38
  214. package/.docs/reference/vectors/qdrant.md +36 -36
  215. package/.docs/reference/vectors/s3vectors.md +24 -24
  216. package/.docs/reference/vectors/turbopuffer.md +29 -29
  217. package/.docs/reference/vectors/upstash.md +25 -25
  218. package/.docs/reference/vectors/vectorize.md +27 -27
  219. package/.docs/reference/voice/azure.md +19 -15
  220. package/.docs/reference/voice/cloudflare.md +10 -12
  221. package/.docs/reference/voice/composite-voice.md +10 -10
  222. package/.docs/reference/voice/deepgram.md +20 -14
  223. package/.docs/reference/voice/elevenlabs.md +20 -22
  224. package/.docs/reference/voice/google-gemini-live.md +42 -44
  225. package/.docs/reference/voice/google.md +25 -21
  226. package/.docs/reference/voice/mastra-voice.md +20 -20
  227. package/.docs/reference/voice/murf.md +24 -28
  228. package/.docs/reference/voice/openai-realtime.md +26 -26
  229. package/.docs/reference/voice/openai.md +14 -12
  230. package/.docs/reference/voice/playai.md +13 -15
  231. package/.docs/reference/voice/sarvam.md +21 -25
  232. package/.docs/reference/voice/speechify.md +11 -13
  233. package/.docs/reference/voice/voice.addInstructions.md +1 -1
  234. package/.docs/reference/voice/voice.addTools.md +1 -1
  235. package/.docs/reference/voice/voice.answer.md +1 -1
  236. package/.docs/reference/voice/voice.connect.md +3 -3
  237. package/.docs/reference/voice/voice.events.md +11 -11
  238. package/.docs/reference/voice/voice.getSpeakers.md +30 -30
  239. package/.docs/reference/voice/voice.listen.md +9 -9
  240. package/.docs/reference/voice/voice.off.md +2 -2
  241. package/.docs/reference/voice/voice.on.md +2 -2
  242. package/.docs/reference/voice/voice.send.md +1 -1
  243. package/.docs/reference/voice/voice.speak.md +11 -11
  244. package/.docs/reference/voice/voice.updateConfig.md +3 -3
  245. package/.docs/reference/workflows/run-methods/cancel.md +2 -2
  246. package/.docs/reference/workflows/run-methods/restart.md +17 -5
  247. package/.docs/reference/workflows/run-methods/resume.md +23 -9
  248. package/.docs/reference/workflows/run-methods/start.md +22 -8
  249. package/.docs/reference/workflows/run-methods/startAsync.md +12 -6
  250. package/.docs/reference/workflows/run-methods/timeTravel.md +29 -13
  251. package/.docs/reference/workflows/run.md +12 -12
  252. package/.docs/reference/workflows/step.md +24 -26
  253. package/.docs/reference/workflows/workflow-methods/branch.md +2 -2
  254. package/.docs/reference/workflows/workflow-methods/commit.md +1 -1
  255. package/.docs/reference/workflows/workflow-methods/create-run.md +4 -4
  256. package/.docs/reference/workflows/workflow-methods/dountil.md +3 -3
  257. package/.docs/reference/workflows/workflow-methods/dowhile.md +3 -3
  258. package/.docs/reference/workflows/workflow-methods/foreach.md +9 -9
  259. package/.docs/reference/workflows/workflow-methods/map.md +2 -2
  260. package/.docs/reference/workflows/workflow-methods/parallel.md +2 -2
  261. package/.docs/reference/workflows/workflow-methods/sleep.md +2 -2
  262. package/.docs/reference/workflows/workflow-methods/sleepUntil.md +2 -2
  263. package/.docs/reference/workflows/workflow-methods/then.md +2 -2
  264. package/.docs/reference/workflows/workflow.md +40 -50
  265. package/.docs/reference/workspace/daytona-sandbox.md +33 -33
  266. package/.docs/reference/workspace/e2b-sandbox.md +20 -20
  267. package/.docs/reference/workspace/filesystem.md +22 -22
  268. package/.docs/reference/workspace/gcs-filesystem.md +15 -15
  269. package/.docs/reference/workspace/local-filesystem.md +35 -35
  270. package/.docs/reference/workspace/local-sandbox.md +26 -26
  271. package/.docs/reference/workspace/s3-filesystem.md +18 -18
  272. package/.docs/reference/workspace/sandbox.md +8 -8
  273. package/.docs/reference/workspace/workspace-class.md +30 -34
  274. package/CHANGELOG.md +16 -0
  275. package/package.json +6 -6
  276. package/.docs/docs/getting-started/start.md +0 -28
@@ -57,21 +57,21 @@ voice.connect()
57
57
 
58
58
  ### Constructor Options
59
59
 
60
- **model?:** (`string`): The model ID to use for real-time voice interactions. (Default: `'gpt-5.1-realtime-preview-2024-12-17'`)
60
+ **model** (`string`): The model ID to use for real-time voice interactions. (Default: `'gpt-5.1-realtime-preview-2024-12-17'`)
61
61
 
62
- **apiKey?:** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
62
+ **apiKey** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
63
63
 
64
- **speaker?:** (`string`): Default voice ID for speech synthesis. (Default: `'alloy'`)
64
+ **speaker** (`string`): Default voice ID for speech synthesis. (Default: `'alloy'`)
65
65
 
66
66
  ### Voice Activity Detection (VAD) Configuration
67
67
 
68
- **type?:** (`string`): Type of VAD to use. Server-side VAD provides better accuracy. (Default: `'server_vad'`)
68
+ **type** (`string`): Type of VAD to use. Server-side VAD provides better accuracy. (Default: `'server_vad'`)
69
69
 
70
- **threshold?:** (`number`): Speech detection sensitivity (0.0-1.0). (Default: `0.5`)
70
+ **threshold** (`number`): Speech detection sensitivity (0.0-1.0). (Default: `0.5`)
71
71
 
72
- **prefix\_padding\_ms?:** (`number`): Milliseconds of audio to include before speech is detected. (Default: `1000`)
72
+ **prefix\_padding\_ms** (`number`): Milliseconds of audio to include before speech is detected. (Default: `1000`)
73
73
 
74
- **silence\_duration\_ms?:** (`number`): Milliseconds of silence before ending a turn. (Default: `1000`)
74
+ **silence\_duration\_ms** (`number`): Milliseconds of silence before ending a turn. (Default: `1000`)
75
75
 
76
76
  ## Methods
77
77
 
@@ -79,15 +79,15 @@ voice.connect()
79
79
 
80
80
  Establishes a connection to the OpenAI realtime service. Must be called before using speak, listen, or send functions.
81
81
 
82
- **returns:** (`Promise<void>`): Promise that resolves when the connection is established.
82
+ **returns** (`Promise<void>`): Promise that resolves when the connection is established.
83
83
 
84
84
  ### speak()
85
85
 
86
86
  Emits a speaking event using the configured voice model. Can accept either a string or a readable stream as input.
87
87
 
88
- **input:** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
88
+ **input** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
89
89
 
90
- **options.speaker?:** (`string`): Voice ID to use for this specific speech request. (Default: `Constructor's speaker value`)
90
+ **options.speaker** (`string`): Voice ID to use for this specific speech request. (Default: `Constructor's speaker value`)
91
91
 
92
92
  Returns: `Promise<void>`
93
93
 
@@ -95,7 +95,7 @@ Returns: `Promise<void>`
95
95
 
96
96
  Processes audio input for speech recognition. Takes a readable stream of audio data and emits a 'listening' event with the transcribed text.
97
97
 
98
- **audioData:** (`NodeJS.ReadableStream`): Audio stream to transcribe.
98
+ **audioData** (`NodeJS.ReadableStream`): Audio stream to transcribe.
99
99
 
100
100
  Returns: `Promise<void>`
101
101
 
@@ -103,7 +103,7 @@ Returns: `Promise<void>`
103
103
 
104
104
  Streams audio data in real-time to the OpenAI service for continuous audio streaming scenarios like live microphone input.
105
105
 
106
- **audioData:** (`NodeJS.ReadableStream`): Audio stream to send to the service.
106
+ **audioData** (`NodeJS.ReadableStream`): Audio stream to send to the service.
107
107
 
108
108
  Returns: `Promise<void>`
109
109
 
@@ -111,7 +111,7 @@ Returns: `Promise<void>`
111
111
 
112
112
  Updates the session configuration for the voice instance. This can be used to modify voice settings, turn detection, and other parameters.
113
113
 
114
- **sessionConfig:** (`Realtime.SessionConfig`): New session configuration to apply.
114
+ **sessionConfig** (`Realtime.SessionConfig`): New session configuration to apply.
115
115
 
116
116
  Returns: `void`
117
117
 
@@ -119,7 +119,7 @@ Returns: `void`
119
119
 
120
120
  Adds a set of tools to the voice instance. Tools allow the model to perform additional actions during conversations. When OpenAIRealtimeVoice is added to an Agent, any tools configured for the Agent will automatically be available to the voice interface.
121
121
 
122
- **tools?:** (`ToolsInput`): Tools configuration to equip.
122
+ **tools** (`ToolsInput`): Tools configuration to equip.
123
123
 
124
124
  Returns: `void`
125
125
 
@@ -139,9 +139,9 @@ Returns: `Promise<Array<{ voiceId: string; [key: string]: any }>>`
139
139
 
140
140
  Registers an event listener for voice events.
141
141
 
142
- **event:** (`string`): Name of the event to listen for.
142
+ **event** (`string`): Name of the event to listen for.
143
143
 
144
- **callback:** (`Function`): Function to call when the event occurs.
144
+ **callback** (`Function`): Function to call when the event occurs.
145
145
 
146
146
  Returns: `void`
147
147
 
@@ -149,9 +149,9 @@ Returns: `void`
149
149
 
150
150
  Removes a previously registered event listener.
151
151
 
152
- **event:** (`string`): Name of the event to stop listening to.
152
+ **event** (`string`): Name of the event to stop listening to.
153
153
 
154
- **callback:** (`Function`): The specific callback function to remove.
154
+ **callback** (`Function`): The specific callback function to remove.
155
155
 
156
156
  Returns: `void`
157
157
 
@@ -159,25 +159,25 @@ Returns: `void`
159
159
 
160
160
  The OpenAIRealtimeVoice class emits the following events:
161
161
 
162
- **speaking:** (`event`): Emitted when audio data is received from the model. Callback receives { audio: Int16Array }.
162
+ **speaking** (`event`): Emitted when audio data is received from the model. Callback receives { audio: Int16Array }.
163
163
 
164
- **writing:** (`event`): Emitted when transcribed text is available. Callback receives { text: string, role: string }.
164
+ **writing** (`event`): Emitted when transcribed text is available. Callback receives { text: string, role: string }.
165
165
 
166
- **error:** (`event`): Emitted when an error occurs. Callback receives the error object.
166
+ **error** (`event`): Emitted when an error occurs. Callback receives the error object.
167
167
 
168
168
  ### OpenAI Realtime Events
169
169
 
170
170
  You can also listen to [OpenAI Realtime utility events](https://github.com/openai/openai-realtime-api-beta#reference-client-utility-events) by prefixing with 'openAIRealtime:':
171
171
 
172
- **openAIRealtime:conversation.created:** (`event`): Emitted when a new conversation is created.
172
+ **openAIRealtime:conversation.created** (`event`): Emitted when a new conversation is created.
173
173
 
174
- **openAIRealtime:conversation.interrupted:** (`event`): Emitted when a conversation is interrupted.
174
+ **openAIRealtime:conversation.interrupted** (`event`): Emitted when a conversation is interrupted.
175
175
 
176
- **openAIRealtime:conversation.updated:** (`event`): Emitted when a conversation is updated.
176
+ **openAIRealtime:conversation.updated** (`event`): Emitted when a conversation is updated.
177
177
 
178
- **openAIRealtime:conversation.item.appended:** (`event`): Emitted when an item is appended to the conversation.
178
+ **openAIRealtime:conversation.item.appended** (`event`): Emitted when an item is appended to the conversation.
179
179
 
180
- **openAIRealtime:conversation.item.completed:** (`event`): Emitted when an item in the conversation is completed.
180
+ **openAIRealtime:conversation.item.completed** (`event`): Emitted when an item in the conversation is completed.
181
181
 
182
182
  ## Available Voices
183
183
 
@@ -39,17 +39,19 @@ const text = await voice.listen(audioStream, {
39
39
 
40
40
  ### Constructor Options
41
41
 
42
- **speechModel?:** (`OpenAIConfig`): Configuration for text-to-speech synthesis. (Default: `{ name: 'tts-1' }`)
42
+ **speechModel** (`OpenAIConfig`): Configuration for text-to-speech synthesis. (Default: `{ name: 'tts-1' }`)
43
43
 
44
- **listeningModel?:** (`OpenAIConfig`): Configuration for speech-to-text recognition. (Default: `{ name: 'whisper-1' }`)
44
+ **speechModel.name** (`'tts-1' | 'tts-1-hd' | 'whisper-1'`): Model name. Use 'tts-1-hd' for higher quality audio.
45
45
 
46
- **speaker?:** (`OpenAIVoiceId`): Default voice ID for speech synthesis. (Default: `'alloy'`)
46
+ **speechModel.apiKey** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
47
47
 
48
- ### OpenAIConfig
48
+ **listeningModel** (`OpenAIConfig`): Configuration for speech-to-text recognition. (Default: `{ name: 'whisper-1' }`)
49
49
 
50
- **name?:** (`'tts-1' | 'tts-1-hd' | 'whisper-1'`): Model name. Use 'tts-1-hd' for higher quality audio.
50
+ **listeningModel.name** (`'tts-1' | 'tts-1-hd' | 'whisper-1'`): Model name. Use 'tts-1-hd' for higher quality audio.
51
51
 
52
- **apiKey?:** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
52
+ **listeningModel.apiKey** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
53
+
54
+ **speaker** (`OpenAIVoiceId`): Default voice ID for speech synthesis. (Default: `'alloy'`)
53
55
 
54
56
  ## Methods
55
57
 
@@ -57,11 +59,11 @@ const text = await voice.listen(audioStream, {
57
59
 
58
60
  Converts text to speech using OpenAI's text-to-speech models.
59
61
 
60
- **input:** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
62
+ **input** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
61
63
 
62
- **options.speaker?:** (`OpenAIVoiceId`): Voice ID to use for speech synthesis. (Default: `Constructor's speaker value`)
64
+ **options.speaker** (`OpenAIVoiceId`): Voice ID to use for speech synthesis. (Default: `Constructor's speaker value`)
63
65
 
64
- **options.speed?:** (`number`): Speech speed multiplier. (Default: `1.0`)
66
+ **options.speed** (`number`): Speech speed multiplier. (Default: `1.0`)
65
67
 
66
68
  Returns: `Promise<NodeJS.ReadableStream>`
67
69
 
@@ -69,9 +71,9 @@ Returns: `Promise<NodeJS.ReadableStream>`
69
71
 
70
72
  Transcribes audio using OpenAI's Whisper model.
71
73
 
72
- **audioStream:** (`NodeJS.ReadableStream`): Audio stream to transcribe.
74
+ **audioStream** (`NodeJS.ReadableStream`): Audio stream to transcribe.
73
75
 
74
- **options.filetype?:** (`string`): Audio format of the input stream. (Default: `'mp3'`)
76
+ **options.filetype** (`string`): Audio format of the input stream. (Default: `'mp3'`)
75
77
 
76
78
  Returns: `Promise<string>`
77
79
 
@@ -79,7 +81,7 @@ Returns: `Promise<string>`
79
81
 
80
82
  Returns an array of available voice options, where each node contains:
81
83
 
82
- **voiceId:** (`string`): Unique identifier for the voice
84
+ **voiceId** (`string`): Unique identifier for the voice
83
85
 
84
86
  ## Notes
85
87
 
@@ -29,17 +29,15 @@ const audioStream = await voice.speak('Hello, world!', {
29
29
 
30
30
  ## Constructor Parameters
31
31
 
32
- **speechModel?:** (`PlayAIConfig`): Configuration for text-to-speech functionality (Default: `{ name: 'PlayDialog' }`)
32
+ **speechModel** (`PlayAIConfig`): Configuration for text-to-speech functionality (Default: `{ name: 'PlayDialog' }`)
33
33
 
34
- **speaker?:** (`string`): Default voice ID to use for speech synthesis (Default: `First available voice ID`)
34
+ **speechModel.name** (`'PlayDialog' | 'Play3.0-mini'`): The PlayAI model to use
35
35
 
36
- ### PlayAIConfig
36
+ **speechModel.apiKey** (`string`): PlayAI API key. Falls back to PLAYAI\_API\_KEY environment variable
37
37
 
38
- **name?:** (`'PlayDialog' | 'Play3.0-mini'`): The PlayAI model to use (Default: `'PlayDialog'`)
38
+ **speechModel.userId** (`string`): PlayAI user ID. Falls back to PLAYAI\_USER\_ID environment variable
39
39
 
40
- **apiKey?:** (`string`): PlayAI API key. Falls back to PLAYAI\_API\_KEY environment variable
41
-
42
- **userId?:** (`string`): PlayAI user ID. Falls back to PLAYAI\_USER\_ID environment variable
40
+ **speaker** (`string`): Default voice ID to use for speech synthesis (Default: `First available voice ID`)
43
41
 
44
42
  ## Methods
45
43
 
@@ -47,9 +45,9 @@ const audioStream = await voice.speak('Hello, world!', {
47
45
 
48
46
  Converts text to speech using the configured speech model and voice.
49
47
 
50
- **input:** (`string | NodeJS.ReadableStream`): Text to convert to speech. If a stream is provided, it will be converted to text first.
48
+ **input** (`string | NodeJS.ReadableStream`): Text to convert to speech. If a stream is provided, it will be converted to text first.
51
49
 
52
- **options.speaker?:** (`string`): Override the default speaker for this request (Default: `Constructor's speaker value`)
50
+ **options.speaker** (`string`): Override the default speaker for this request (Default: `Constructor's speaker value`)
53
51
 
54
52
  Returns: `Promise<NodeJS.ReadableStream>`.
55
53
 
@@ -57,17 +55,17 @@ Returns: `Promise<NodeJS.ReadableStream>`.
57
55
 
58
56
  Returns an array of available voice options, where each node contains:
59
57
 
60
- **name:** (`string`): Name of the voice
58
+ **name** (`string`): Name of the voice
61
59
 
62
- **accent:** (`string`): Accent of the voice (e.g., 'US', 'British', 'Australian')
60
+ **accent** (`string`): Accent of the voice (e.g., 'US', 'British', 'Australian')
63
61
 
64
- **gender:** (`'M' | 'F'`): Gender of the voice
62
+ **gender** (`'M' | 'F'`): Gender of the voice
65
63
 
66
- **age:** (`'Young' | 'Middle' | 'Old'`): Age category of the voice
64
+ **age** (`'Young' | 'Middle' | 'Old'`): Age category of the voice
67
65
 
68
- **style:** (`'Conversational' | 'Narrative'`): Speaking style of the voice
66
+ **style** (`'Conversational' | 'Narrative'`): Speaking style of the voice
69
67
 
70
- **voiceId:** (`string`): Unique identifier for the voice
68
+ **voiceId** (`string`): Unique identifier for the voice
71
69
 
72
70
  ### listen()
73
71
 
@@ -51,43 +51,39 @@ const text = await voice.listen(audioStream, {
51
51
 
52
52
  ### Constructor Options
53
53
 
54
- **speechModel?:** (`SarvamVoiceConfig`): Configuration for text-to-speech synthesis. (Default: `{ model: 'bulbul:v1', language: 'en-IN' }`)
54
+ **speechModel** (`SarvamVoiceConfig`): Configuration for text-to-speech synthesis. (Default: `{ model: 'bulbul:v1', language: 'en-IN' }`)
55
55
 
56
- **speaker?:** (`SarvamVoiceId`): The speaker to be used for the output audio. If not provided, Meera will be used as default. AvailableOptions - meera, pavithra, maitreyi, arvind, amol, amartya, diya, neel, misha, vian, arjun, maya (Default: `'meera'`)
56
+ **speechModel.apiKey** (`string`): Sarvam API key. Falls back to SARVAM\_API\_KEY environment variable.
57
57
 
58
- **listeningModel?:** (`SarvamListenOptions`): Configuration for speech-to-text recognition. (Default: `{ model: 'saarika:v2', language_code: 'unknown' }`)
58
+ **speechModel.model** (`SarvamTTSModel`): Specifies the model to use for text-to-speech conversion.
59
59
 
60
- ### SarvamVoiceConfig
60
+ **speechModel.language** (`SarvamTTSLanguage`): Target language for speech synthesis. Available options: hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN
61
61
 
62
- **apiKey?:** (`string`): Sarvam API key. Falls back to SARVAM\_API\_KEY environment variable.
62
+ **speechModel.properties** (`object`): Additional voice properties for customization.
63
63
 
64
- **model?:** (`SarvamTTSModel`): Specifies the model to use for text-to-speech conversion. (Default: `'bulbul:v1'`)
64
+ **speechModel.properties.pitch** (`number`): Controls the pitch of the audio. Lower values result in a deeper voice, while higher values make it sharper. The suitable range is between -0.75 and 0.75.
65
65
 
66
- **language:** (`SarvamTTSLanguage`): Target language for speech synthesis. Available options: hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN (Default: `'en-IN'`)
66
+ **speechModel.properties.pace** (`number`): Controls the speed of the audio. Lower values result in slower speech, while higher values make it faster. The suitable range is between 0.5 and 2.0. Default is 1.0. Required range: 0.3 <= x <= 3
67
67
 
68
- **properties?:** (`object`): Additional voice properties for customization.
68
+ **speechModel.properties.loudness** (`number`): Controls the loudness of the audio. Lower values result in quieter audio, while higher values make it louder. The suitable range is between 0.3 and 3.0. Required range: 0 <= x <= 3
69
69
 
70
- **properties.pitch?:** (`number`): Controls the pitch of the audio. Lower values result in a deeper voice, while higher values make it sharper. The suitable range is between -0.75 and 0.75.
70
+ **speechModel.properties.speech\_sample\_rate** (`8000 | 16000 | 22050`): Audio sample rate in Hz.
71
71
 
72
- **properties.pace?:** (`number`): Controls the speed of the audio. Lower values result in slower speech, while higher values make it faster. The suitable range is between 0.5 and 2.0. Default is 1.0. Required range: 0.3 <= x <= 3
72
+ **speechModel.properties.enable\_preprocessing** (`boolean`): Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
73
73
 
74
- **properties.loudness?:** (`number`): Controls the loudness of the audio. Lower values result in quieter audio, while higher values make it louder. The suitable range is between 0.3 and 3.0. Required range: 0 <= x <= 3
74
+ **speechModel.properties.eng\_interpolation\_wt** (`number`): Weight for interpolating with English speaker at encoder.
75
75
 
76
- **properties.speech\_sample\_rate?:** (`8000 | 16000 | 22050`): Audio sample rate in Hz.
76
+ **speaker** (`SarvamVoiceId`): The speaker to be used for the output audio. If not provided, Meera will be used as default. AvailableOptions - meera, pavithra, maitreyi, arvind, amol, amartya, diya, neel, misha, vian, arjun, maya (Default: `'meera'`)
77
77
 
78
- **properties.enable\_preprocessing?:** (`boolean`): Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
78
+ **listeningModel** (`SarvamListenOptions`): Configuration for speech-to-text recognition. (Default: `{ model: 'saarika:v2', language_code: 'unknown' }`)
79
79
 
80
- **properties.eng\_interpolation\_wt?:** (`number`): Weight for interpolating with English speaker at encoder.
80
+ **listeningModel.apiKey** (`string`): Sarvam API key. Falls back to SARVAM\_API\_KEY environment variable.
81
81
 
82
- ### SarvamListenOptions
82
+ **listeningModel.model** (`SarvamSTTModel`): Specifies the model to use for speech-to-text conversion. Note:- Default model is saarika:v2 . Available options: saarika:v1, saarika:v2, saarika:flash
83
83
 
84
- **apiKey?:** (`string`): Sarvam API key. Falls back to SARVAM\_API\_KEY environment variable.
84
+ **listeningModel.languageCode** (`SarvamSTTLanguage`): Specifies the language of the input audio. This parameter is required to ensure accurate transcription. For the saarika:v1 model, this parameter is mandatory. For the saarika:v2 model, it is optional. unknown: Use this when the language is not known; the API will detect it automatically. Note:- that the saarika:v1 model does not support unknown language code. Available options: unknown, hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN
85
85
 
86
- **model?:** (`SarvamSTTModel`): Specifies the model to use for speech-to-text conversion. Note:- Default model is saarika:v2 . Available options: saarika:v1, saarika:v2, saarika:flash (Default: `'saarika:v2'`)
87
-
88
- **languageCode?:** (`SarvamSTTLanguage`): Specifies the language of the input audio. This parameter is required to ensure accurate transcription. For the saarika:v1 model, this parameter is mandatory. For the saarika:v2 model, it is optional. unknown: Use this when the language is not known; the API will detect it automatically. Note:- that the saarika:v1 model does not support unknown language code. Available options: unknown, hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN (Default: `'unknown'`)
89
-
90
- **filetype?:** (`'mp3' | 'wav'`): Audio format of the input stream.
86
+ **listeningModel.filetype** (`'mp3' | 'wav'`): Audio format of the input stream.
91
87
 
92
88
  ## Methods
93
89
 
@@ -95,9 +91,9 @@ const text = await voice.listen(audioStream, {
95
91
 
96
92
  Converts text to speech using Sarvam's text-to-speech models.
97
93
 
98
- **input:** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
94
+ **input** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
99
95
 
100
- **options.speaker?:** (`SarvamVoiceId`): Voice ID to use for speech synthesis. (Default: `Constructor's speaker value`)
96
+ **options.speaker** (`SarvamVoiceId`): Voice ID to use for speech synthesis. (Default: `Constructor's speaker value`)
101
97
 
102
98
  Returns: `Promise<NodeJS.ReadableStream>`
103
99
 
@@ -105,9 +101,9 @@ Returns: `Promise<NodeJS.ReadableStream>`
105
101
 
106
102
  Transcribes audio using Sarvam's speech recognition models.
107
103
 
108
- **input:** (`NodeJS.ReadableStream`): Audio stream to transcribe.
104
+ **input** (`NodeJS.ReadableStream`): Audio stream to transcribe.
109
105
 
110
- **options?:** (`SarvamListenOptions`): Configuration options for speech recognition.
106
+ **options** (`SarvamListenOptions`): Configuration options for speech recognition.
111
107
 
112
108
  Returns: `Promise<string>`
113
109
 
@@ -27,15 +27,13 @@ const audioStream = await voice.speak('Hello, world!', {
27
27
 
28
28
  ## Constructor Parameters
29
29
 
30
- **speechModel?:** (`SpeechifyConfig`): Configuration for text-to-speech functionality (Default: `{ name: 'simba-english' }`)
30
+ **speechModel** (`SpeechifyConfig`): Configuration for text-to-speech functionality (Default: `{ name: 'simba-english' }`)
31
31
 
32
- **speaker?:** (`SpeechifyVoiceId`): Default voice ID to use for speech synthesis (Default: `'george'`)
32
+ **speechModel.name** (`VoiceModelName`): The Speechify model to use
33
33
 
34
- ### SpeechifyConfig
34
+ **speechModel.apiKey** (`string`): Speechify API key. Falls back to SPEECHIFY\_API\_KEY environment variable
35
35
 
36
- **name?:** (`VoiceModelName`): The Speechify model to use (Default: `'simba-english'`)
37
-
38
- **apiKey?:** (`string`): Speechify API key. Falls back to SPEECHIFY\_API\_KEY environment variable
36
+ **speaker** (`SpeechifyVoiceId`): Default voice ID to use for speech synthesis (Default: `'george'`)
39
37
 
40
38
  ## Methods
41
39
 
@@ -43,11 +41,11 @@ const audioStream = await voice.speak('Hello, world!', {
43
41
 
44
42
  Converts text to speech using the configured speech model and voice.
45
43
 
46
- **input:** (`string | NodeJS.ReadableStream`): Text to convert to speech. If a stream is provided, it will be converted to text first.
44
+ **input** (`string | NodeJS.ReadableStream`): Text to convert to speech. If a stream is provided, it will be converted to text first.
47
45
 
48
- **options.speaker?:** (`string`): Override the default speaker for this request (Default: `Constructor's speaker value`)
46
+ **options.speaker** (`string`): Override the default speaker for this request (Default: `Constructor's speaker value`)
49
47
 
50
- **options.model?:** (`VoiceModelName`): Override the default model for this request (Default: `Constructor's model value`)
48
+ **options.model** (`VoiceModelName`): Override the default model for this request (Default: `Constructor's model value`)
51
49
 
52
50
  Returns: `Promise<NodeJS.ReadableStream>`
53
51
 
@@ -55,13 +53,13 @@ Returns: `Promise<NodeJS.ReadableStream>`
55
53
 
56
54
  Returns an array of available voice options, where each node contains:
57
55
 
58
- **voiceId:** (`string`): Unique identifier for the voice
56
+ **voiceId** (`string`): Unique identifier for the voice
59
57
 
60
- **name:** (`string`): Display name of the voice
58
+ **name** (`string`): Display name of the voice
61
59
 
62
- **language:** (`string`): Language code for the voice
60
+ **language** (`string`): Language code for the voice
63
61
 
64
- **gender:** (`string`): Gender of the voice
62
+ **gender** (`string`): Gender of the voice
65
63
 
66
64
  ### listen()
67
65
 
@@ -39,7 +39,7 @@ await voice.connect()
39
39
 
40
40
  ## Parameters
41
41
 
42
- **instructions:** (`string`): Instructions to guide the voice model's behavior
42
+ **instructions** (`string`): Instructions to guide the voice model's behavior
43
43
 
44
44
  ## Return Value
45
45
 
@@ -50,7 +50,7 @@ await voice.connect()
50
50
 
51
51
  ## Parameters
52
52
 
53
- **tools:** (`ToolsInput`): Object containing tool definitions that can be called by the voice model
53
+ **tools** (`ToolsInput`): Object containing tool definitions that can be called by the voice model
54
54
 
55
55
  ## Return Value
56
56
 
@@ -39,7 +39,7 @@ await voice.answer()
39
39
 
40
40
  ## Parameters
41
41
 
42
- **options?:** (`Record<string, unknown>`): Provider-specific options for the response
42
+ **options** (`Record<string, unknown>`): Provider-specific options for the response
43
43
 
44
44
  ## Return Value
45
45
 
@@ -46,7 +46,7 @@ await voice.connect({
46
46
 
47
47
  ## Parameters
48
48
 
49
- **options?:** (`Record<string, unknown>`): Provider-specific connection options
49
+ **options** (`Record<string, unknown>`): Provider-specific connection options
50
50
 
51
51
  ## Return Value
52
52
 
@@ -58,9 +58,9 @@ Each real-time voice provider may support different options for the `connect()`
58
58
 
59
59
  ### OpenAI Realtime
60
60
 
61
- **options.timeout?:** (`number`): Connection timeout in milliseconds (Default: `30000`)
61
+ **options.timeout** (`number`): Connection timeout in milliseconds (Default: `30000`)
62
62
 
63
- **options.reconnect?:** (`boolean`): Whether to automatically reconnect on connection loss (Default: `false`)
63
+ **options.reconnect** (`boolean`): Whether to automatically reconnect on connection loss (Default: `false`)
64
64
 
65
65
  ## Using with CompositeVoice
66
66
 
@@ -6,27 +6,27 @@ Voice providers emit various events during real-time voice interactions. These e
6
6
 
7
7
  These events are commonly implemented across real-time voice providers:
8
8
 
9
- **error:** (`Error`): Emitted when an error occurs during voice processing or when audio data format is unsupported
9
+ **error** (`Error`): Emitted when an error occurs during voice processing or when audio data format is unsupported
10
10
 
11
- **session.created:** (`object`): Emitted when a new session is created with the OpenAI service
11
+ **session.created** (`object`): Emitted when a new session is created with the OpenAI service
12
12
 
13
- **session.updated:** (`object`): Emitted when the session configuration is updated
13
+ **session.updated** (`object`): Emitted when the session configuration is updated
14
14
 
15
- **response.created:** (`object`): Emitted when a new response is created by the AI assistant
15
+ **response.created** (`object`): Emitted when a new response is created by the AI assistant
16
16
 
17
- **response.done:** (`object`): Emitted when the AI assistant has completed its response
17
+ **response.done** (`object`): Emitted when the AI assistant has completed its response
18
18
 
19
- **speaker:** (`StreamWithId`): Emitted with a new audio stream that can be piped to an audio output
19
+ **speaker** (`StreamWithId`): Emitted with a new audio stream that can be piped to an audio output
20
20
 
21
- **writing:** (`object`): Emitted when text is being transcribed (user) or generated (assistant)
21
+ **writing** (`object`): Emitted when text is being transcribed (user) or generated (assistant)
22
22
 
23
- **speaking:** (`object`): Emitted when audio data is available from the voice provider
23
+ **speaking** (`object`): Emitted when audio data is available from the voice provider
24
24
 
25
- **speaking.done:** (`object`): Emitted when the voice provider has finished speaking
25
+ **speaking.done** (`object`): Emitted when the voice provider has finished speaking
26
26
 
27
- **tool-call-start:** (`object`): Emitted when the AI assistant starts executing a tool
27
+ **tool-call-start** (`object`): Emitted when the AI assistant starts executing a tool
28
28
 
29
- **tool-call-result:** (`object`): Emitted when a tool execution is complete with its result
29
+ **tool-call-result** (`object`): Emitted when a tool execution is complete with its result
30
30
 
31
31
  ## Notes
32
32
 
@@ -36,7 +36,7 @@ This method does not accept any parameters.
36
36
 
37
37
  ## Return Value
38
38
 
39
- **Promise\<Array<{ voiceId: string } & TSpeakerMetadata>>:** (`Promise`): A promise that resolves to an array of voice options, where each option contains at least a voiceId property and may include additional provider-specific metadata.
39
+ **Promise\<Array<{ voiceId: string } & TSpeakerMetadata>>** (`Promise`): A promise that resolves to an array of voice options, where each option contains at least a voiceId property and may include additional provider-specific metadata.
40
40
 
41
41
  ## Provider-Specific Metadata
42
42
 
@@ -44,81 +44,81 @@ Different voice providers return different metadata for their voices:
44
44
 
45
45
  **OpenAI**:
46
46
 
47
- **voiceId:** (`string`): Unique identifier for the voice (e.g., 'alloy', 'echo', 'fable', 'onyx', 'nova', 'shimmer')
47
+ **voiceId** (`string`): Unique identifier for the voice (e.g., 'alloy', 'echo', 'fable', 'onyx', 'nova', 'shimmer')
48
48
 
49
49
  **OpenAI Realtime**:
50
50
 
51
- **voiceId:** (`string`): Unique identifier for the voice (e.g., 'alloy', 'echo', 'fable', 'onyx', 'nova', 'shimmer')
51
+ **voiceId** (`string`): Unique identifier for the voice (e.g., 'alloy', 'echo', 'fable', 'onyx', 'nova', 'shimmer')
52
52
 
53
53
  **Deepgram**:
54
54
 
55
- **voiceId:** (`string`): Unique identifier for the voice
55
+ **voiceId** (`string`): Unique identifier for the voice
56
56
 
57
- **language:** (`string`): Language code embedded in the voice ID (e.g., 'en')
57
+ **language** (`string`): Language code embedded in the voice ID (e.g., 'en')
58
58
 
59
59
  **ElevenLabs**:
60
60
 
61
- **voiceId:** (`string`): Unique identifier for the voice
61
+ **voiceId** (`string`): Unique identifier for the voice
62
62
 
63
- **name:** (`string`): Human-readable name of the voice
63
+ **name** (`string`): Human-readable name of the voice
64
64
 
65
- **category:** (`string`): Category of the voice (e.g., 'premade', 'cloned')
65
+ **category** (`string`): Category of the voice (e.g., 'premade', 'cloned')
66
66
 
67
67
  **Google**:
68
68
 
69
- **voiceId:** (`string`): Unique identifier for the voice
69
+ **voiceId** (`string`): Unique identifier for the voice
70
70
 
71
- **languageCodes:** (`string[]`): Array of language codes supported by the voice (e.g., \['en-US'])
71
+ **languageCodes** (`string[]`): Array of language codes supported by the voice (e.g., \['en-US'])
72
72
 
73
73
  **Azure**:
74
74
 
75
- **voiceId:** (`string`): Unique identifier for the voice
75
+ **voiceId** (`string`): Unique identifier for the voice
76
76
 
77
- **language:** (`string`): Language code extracted from the voice ID (e.g., 'en')
77
+ **language** (`string`): Language code extracted from the voice ID (e.g., 'en')
78
78
 
79
- **region:** (`string`): Region code extracted from the voice ID (e.g., 'US')
79
+ **region** (`string`): Region code extracted from the voice ID (e.g., 'US')
80
80
 
81
81
  **Murf**:
82
82
 
83
- **voiceId:** (`string`): Unique identifier for the voice
83
+ **voiceId** (`string`): Unique identifier for the voice
84
84
 
85
- **name:** (`string`): Name of the voice (same as voiceId)
85
+ **name** (`string`): Name of the voice (same as voiceId)
86
86
 
87
- **language:** (`string`): Language code extracted from the voice ID (e.g., 'en')
87
+ **language** (`string`): Language code extracted from the voice ID (e.g., 'en')
88
88
 
89
- **gender:** (`string`): Gender of the voice (always 'neutral' in current implementation)
89
+ **gender** (`string`): Gender of the voice (always 'neutral' in current implementation)
90
90
 
91
91
  **PlayAI**:
92
92
 
93
- **voiceId:** (`string`): Unique identifier for the voice (S3 URL to manifest.json)
93
+ **voiceId** (`string`): Unique identifier for the voice (S3 URL to manifest.json)
94
94
 
95
- **name:** (`string`): Human-readable name of the voice (e.g., 'Angelo', 'Arsenio')
95
+ **name** (`string`): Human-readable name of the voice (e.g., 'Angelo', 'Arsenio')
96
96
 
97
- **accent:** (`string`): Accent of the voice (e.g., 'US', 'Irish', 'US African American')
97
+ **accent** (`string`): Accent of the voice (e.g., 'US', 'Irish', 'US African American')
98
98
 
99
- **gender:** (`string`): Gender of the voice ('M' or 'F')
99
+ **gender** (`string`): Gender of the voice ('M' or 'F')
100
100
 
101
- **age:** (`string`): Age category of the voice (e.g., 'Young', 'Middle')
101
+ **age** (`string`): Age category of the voice (e.g., 'Young', 'Middle')
102
102
 
103
- **style:** (`string`): Speaking style of the voice (e.g., 'Conversational')
103
+ **style** (`string`): Speaking style of the voice (e.g., 'Conversational')
104
104
 
105
105
  **Speechify**:
106
106
 
107
- **voiceId:** (`string`): Unique identifier for the voice
107
+ **voiceId** (`string`): Unique identifier for the voice
108
108
 
109
- **name:** (`string`): Human-readable name of the voice
109
+ **name** (`string`): Human-readable name of the voice
110
110
 
111
- **language:** (`string`): Language code of the voice (e.g., 'en-US')
111
+ **language** (`string`): Language code of the voice (e.g., 'en-US')
112
112
 
113
113
  **Sarvam**:
114
114
 
115
- **voiceId:** (`string`): Unique identifier for the voice
115
+ **voiceId** (`string`): Unique identifier for the voice
116
116
 
117
- **name:** (`string`): Human-readable name of the voice
117
+ **name** (`string`): Human-readable name of the voice
118
118
 
119
- **language:** (`string`): Language of the voice (e.g., 'english', 'hindi')
119
+ **language** (`string`): Language of the voice (e.g., 'english', 'hindi')
120
120
 
121
- **gender:** (`string`): Gender of the voice ('male' or 'female')
121
+ **gender** (`string`): Gender of the voice ('male' or 'female')
122
122
 
123
123
  ## Notes
124
124