@mastra/mcp-docs-server 1.1.7 → 1.1.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/agents/agent-approval.md +61 -31
- package/.docs/docs/agents/supervisor-agents.md +1 -1
- package/.docs/docs/getting-started/manual-install.md +1 -1
- package/.docs/docs/index.md +68 -24
- package/.docs/docs/mastra-cloud/setup.md +1 -1
- package/.docs/docs/memory/observational-memory.md +9 -0
- package/.docs/docs/memory/semantic-recall.md +17 -1
- package/.docs/docs/workspace/skills.md +7 -5
- package/.docs/guides/deployment/aws-lambda.md +76 -165
- package/.docs/guides/deployment/azure-app-services.md +38 -61
- package/.docs/guides/deployment/cloudflare.md +1 -1
- package/.docs/guides/deployment/netlify.md +1 -1
- package/.docs/guides/deployment/vercel.md +1 -1
- package/.docs/models/gateways/netlify.md +5 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/anthropic.md +9 -9
- package/.docs/models/providers/deepseek.md +1 -1
- package/.docs/models/providers/google.md +11 -11
- package/.docs/models/providers/openai.md +19 -19
- package/.docs/models/providers/opencode.md +33 -33
- package/.docs/models/providers/xai.md +3 -3
- package/.docs/reference/agents/agent.md +20 -20
- package/.docs/reference/agents/generate.md +200 -66
- package/.docs/reference/agents/generateLegacy.md +77 -35
- package/.docs/reference/agents/getDefaultGenerateOptions.md +4 -6
- package/.docs/reference/agents/getDefaultOptions.md +4 -6
- package/.docs/reference/agents/getDefaultStreamOptions.md +4 -6
- package/.docs/reference/agents/getDescription.md +1 -1
- package/.docs/reference/agents/getInstructions.md +4 -6
- package/.docs/reference/agents/getLLM.md +6 -8
- package/.docs/reference/agents/getMemory.md +4 -6
- package/.docs/reference/agents/getModel.md +4 -6
- package/.docs/reference/agents/getTools.md +5 -7
- package/.docs/reference/agents/getVoice.md +4 -6
- package/.docs/reference/agents/listAgents.md +4 -6
- package/.docs/reference/agents/listScorers.md +4 -6
- package/.docs/reference/agents/listTools.md +4 -6
- package/.docs/reference/agents/listWorkflows.md +4 -6
- package/.docs/reference/agents/network.md +69 -23
- package/.docs/reference/ai-sdk/chat-route.md +7 -7
- package/.docs/reference/ai-sdk/handle-chat-stream.md +12 -12
- package/.docs/reference/ai-sdk/handle-network-stream.md +4 -4
- package/.docs/reference/ai-sdk/handle-workflow-stream.md +11 -11
- package/.docs/reference/ai-sdk/network-route.md +3 -3
- package/.docs/reference/ai-sdk/to-ai-sdk-stream.md +9 -9
- package/.docs/reference/ai-sdk/to-ai-sdk-v4-messages.md +9 -9
- package/.docs/reference/ai-sdk/to-ai-sdk-v5-messages.md +5 -5
- package/.docs/reference/ai-sdk/with-mastra.md +12 -12
- package/.docs/reference/ai-sdk/workflow-route.md +3 -3
- package/.docs/reference/auth/auth0.md +6 -6
- package/.docs/reference/auth/better-auth.md +5 -5
- package/.docs/reference/auth/clerk.md +5 -5
- package/.docs/reference/auth/firebase.md +7 -7
- package/.docs/reference/auth/jwt.md +1 -1
- package/.docs/reference/auth/supabase.md +4 -4
- package/.docs/reference/auth/workos.md +6 -6
- package/.docs/reference/cli/create-mastra.md +14 -14
- package/.docs/reference/client-js/mastra-client.md +23 -23
- package/.docs/reference/client-js/workflows.md +3 -3
- package/.docs/reference/core/addGateway.md +2 -2
- package/.docs/reference/core/getAgent.md +2 -2
- package/.docs/reference/core/getAgentById.md +2 -2
- package/.docs/reference/core/getDeployer.md +1 -1
- package/.docs/reference/core/getGateway.md +2 -2
- package/.docs/reference/core/getGatewayById.md +2 -2
- package/.docs/reference/core/getLogger.md +1 -1
- package/.docs/reference/core/getMCPServer.md +2 -2
- package/.docs/reference/core/getMCPServerById.md +3 -3
- package/.docs/reference/core/getMemory.md +2 -2
- package/.docs/reference/core/getScorer.md +2 -2
- package/.docs/reference/core/getScorerById.md +2 -2
- package/.docs/reference/core/getServer.md +1 -1
- package/.docs/reference/core/getStorage.md +1 -1
- package/.docs/reference/core/getStoredAgentById.md +18 -20
- package/.docs/reference/core/getTelemetry.md +1 -1
- package/.docs/reference/core/getVector.md +2 -2
- package/.docs/reference/core/getWorkflow.md +3 -3
- package/.docs/reference/core/listAgents.md +1 -1
- package/.docs/reference/core/listGateways.md +1 -1
- package/.docs/reference/core/listLogs.md +9 -11
- package/.docs/reference/core/listLogsByRunId.md +9 -9
- package/.docs/reference/core/listMCPServers.md +1 -1
- package/.docs/reference/core/listMemory.md +1 -1
- package/.docs/reference/core/listScorers.md +1 -1
- package/.docs/reference/core/listStoredAgents.md +9 -11
- package/.docs/reference/core/listVectors.md +1 -1
- package/.docs/reference/core/listWorkflows.md +2 -2
- package/.docs/reference/core/mastra-class.md +17 -17
- package/.docs/reference/core/mastra-model-gateway.md +15 -15
- package/.docs/reference/core/setLogger.md +2 -4
- package/.docs/reference/core/setStorage.md +1 -1
- package/.docs/reference/datasets/addItem.md +20 -4
- package/.docs/reference/datasets/addItems.md +8 -2
- package/.docs/reference/datasets/compareExperiments.md +15 -3
- package/.docs/reference/datasets/create.md +6 -6
- package/.docs/reference/datasets/dataset.md +1 -1
- package/.docs/reference/datasets/delete.md +2 -2
- package/.docs/reference/datasets/deleteExperiment.md +2 -2
- package/.docs/reference/datasets/deleteItem.md +2 -2
- package/.docs/reference/datasets/deleteItems.md +2 -2
- package/.docs/reference/datasets/get.md +2 -2
- package/.docs/reference/datasets/getDetails.md +9 -9
- package/.docs/reference/datasets/getExperiment.md +2 -2
- package/.docs/reference/datasets/getItem.md +3 -3
- package/.docs/reference/datasets/getItemHistory.md +22 -2
- package/.docs/reference/datasets/list.md +7 -3
- package/.docs/reference/datasets/listExperimentResults.md +34 -4
- package/.docs/reference/datasets/listExperiments.md +41 -3
- package/.docs/reference/datasets/listItems.md +18 -6
- package/.docs/reference/datasets/listVersions.md +23 -3
- package/.docs/reference/datasets/startExperiment.md +62 -12
- package/.docs/reference/datasets/startExperimentAsync.md +5 -1
- package/.docs/reference/datasets/update.md +6 -6
- package/.docs/reference/datasets/updateItem.md +5 -5
- package/.docs/reference/deployer.md +8 -8
- package/.docs/reference/evals/answer-relevancy.md +11 -11
- package/.docs/reference/evals/answer-similarity.md +17 -19
- package/.docs/reference/evals/bias.md +10 -10
- package/.docs/reference/evals/completeness.md +3 -3
- package/.docs/reference/evals/content-similarity.md +6 -6
- package/.docs/reference/evals/context-precision.md +4 -4
- package/.docs/reference/evals/context-relevance.md +4 -4
- package/.docs/reference/evals/create-scorer.md +47 -49
- package/.docs/reference/evals/faithfulness.md +11 -11
- package/.docs/reference/evals/hallucination.md +17 -21
- package/.docs/reference/evals/keyword-coverage.md +4 -4
- package/.docs/reference/evals/mastra-scorer.md +14 -14
- package/.docs/reference/evals/noise-sensitivity.md +4 -4
- package/.docs/reference/evals/prompt-alignment.md +4 -4
- package/.docs/reference/evals/run-evals.md +16 -16
- package/.docs/reference/evals/scorer-utils.md +3 -3
- package/.docs/reference/evals/textual-difference.md +3 -3
- package/.docs/reference/evals/tone-consistency.md +3 -3
- package/.docs/reference/evals/tool-call-accuracy.md +5 -5
- package/.docs/reference/evals/toxicity.md +8 -8
- package/.docs/reference/harness/harness-class.md +34 -42
- package/.docs/reference/logging/pino-logger.md +5 -5
- package/.docs/reference/memory/clone-utilities.md +5 -5
- package/.docs/reference/memory/cloneThread.md +17 -21
- package/.docs/reference/memory/createThread.md +10 -10
- package/.docs/reference/memory/deleteMessages.md +2 -2
- package/.docs/reference/memory/getThreadById.md +2 -2
- package/.docs/reference/memory/listThreads.md +5 -5
- package/.docs/reference/memory/memory-class.md +12 -14
- package/.docs/reference/memory/observational-memory.md +102 -94
- package/.docs/reference/memory/recall.md +14 -16
- package/.docs/reference/observability/tracing/configuration.md +27 -10
- package/.docs/reference/observability/tracing/exporters/console-exporter.md +4 -7
- package/.docs/reference/processors/batch-parts-processor.md +8 -10
- package/.docs/reference/processors/language-detector.md +14 -16
- package/.docs/reference/processors/message-history-processor.md +7 -9
- package/.docs/reference/processors/moderation-processor.md +13 -15
- package/.docs/reference/processors/pii-detector.md +14 -16
- package/.docs/reference/processors/processor-interface.md +62 -62
- package/.docs/reference/processors/prompt-injection-detector.md +11 -13
- package/.docs/reference/processors/semantic-recall-processor.md +14 -16
- package/.docs/reference/processors/system-prompt-scrubber.md +12 -14
- package/.docs/reference/processors/token-limiter-processor.md +11 -13
- package/.docs/reference/processors/tool-call-filter.md +5 -7
- package/.docs/reference/processors/tool-search-processor.md +9 -11
- package/.docs/reference/processors/unicode-normalizer.md +8 -10
- package/.docs/reference/processors/working-memory-processor.md +14 -18
- package/.docs/reference/rag/chunk.md +38 -38
- package/.docs/reference/rag/database-config.md +11 -7
- package/.docs/reference/rag/document.md +2 -2
- package/.docs/reference/rag/embeddings.md +12 -12
- package/.docs/reference/rag/extract-params.md +23 -23
- package/.docs/reference/rag/graph-rag.md +12 -12
- package/.docs/reference/rag/rerank.md +25 -17
- package/.docs/reference/rag/rerankWithScorer.md +25 -17
- package/.docs/reference/server/create-route.md +14 -14
- package/.docs/reference/server/express-adapter.md +10 -10
- package/.docs/reference/server/fastify-adapter.md +10 -10
- package/.docs/reference/server/hono-adapter.md +10 -10
- package/.docs/reference/server/koa-adapter.md +10 -10
- package/.docs/reference/server/mastra-server.md +10 -10
- package/.docs/reference/server/register-api-route.md +13 -13
- package/.docs/reference/storage/cloudflare-d1.md +5 -5
- package/.docs/reference/storage/cloudflare.md +6 -6
- package/.docs/reference/storage/composite.md +9 -9
- package/.docs/reference/storage/convex.md +3 -3
- package/.docs/reference/storage/dynamodb.md +9 -9
- package/.docs/reference/storage/lance.md +3 -3
- package/.docs/reference/storage/libsql.md +2 -2
- package/.docs/reference/storage/mongodb.md +5 -5
- package/.docs/reference/storage/mssql.md +2 -2
- package/.docs/reference/storage/postgresql.md +25 -25
- package/.docs/reference/storage/upstash.md +3 -3
- package/.docs/reference/streaming/ChunkType.md +251 -59
- package/.docs/reference/streaming/agents/MastraModelOutput.md +86 -16
- package/.docs/reference/streaming/agents/stream.md +155 -43
- package/.docs/reference/streaming/agents/streamLegacy.md +79 -39
- package/.docs/reference/streaming/workflows/resumeStream.md +18 -8
- package/.docs/reference/streaming/workflows/stream.md +21 -9
- package/.docs/reference/streaming/workflows/timeTravelStream.md +4 -4
- package/.docs/reference/tools/create-tool.md +25 -21
- package/.docs/reference/tools/document-chunker-tool.md +7 -7
- package/.docs/reference/tools/graph-rag-tool.md +16 -18
- package/.docs/reference/tools/mcp-client.md +38 -27
- package/.docs/reference/tools/mcp-server.md +45 -45
- package/.docs/reference/tools/vector-query-tool.md +34 -22
- package/.docs/reference/vectors/astra.md +22 -22
- package/.docs/reference/vectors/chroma.md +44 -44
- package/.docs/reference/vectors/convex.md +26 -26
- package/.docs/reference/vectors/couchbase.md +30 -30
- package/.docs/reference/vectors/duckdb.md +29 -29
- package/.docs/reference/vectors/elasticsearch.md +27 -27
- package/.docs/reference/vectors/lance.md +46 -38
- package/.docs/reference/vectors/libsql.md +31 -31
- package/.docs/reference/vectors/mongodb.md +32 -32
- package/.docs/reference/vectors/opensearch.md +27 -27
- package/.docs/reference/vectors/pg.md +60 -44
- package/.docs/reference/vectors/pinecone.md +38 -38
- package/.docs/reference/vectors/qdrant.md +36 -36
- package/.docs/reference/vectors/s3vectors.md +24 -24
- package/.docs/reference/vectors/turbopuffer.md +29 -29
- package/.docs/reference/vectors/upstash.md +25 -25
- package/.docs/reference/vectors/vectorize.md +27 -27
- package/.docs/reference/voice/azure.md +19 -15
- package/.docs/reference/voice/cloudflare.md +10 -12
- package/.docs/reference/voice/composite-voice.md +10 -10
- package/.docs/reference/voice/deepgram.md +20 -14
- package/.docs/reference/voice/elevenlabs.md +20 -22
- package/.docs/reference/voice/google-gemini-live.md +42 -44
- package/.docs/reference/voice/google.md +25 -21
- package/.docs/reference/voice/mastra-voice.md +20 -20
- package/.docs/reference/voice/murf.md +24 -28
- package/.docs/reference/voice/openai-realtime.md +26 -26
- package/.docs/reference/voice/openai.md +14 -12
- package/.docs/reference/voice/playai.md +13 -15
- package/.docs/reference/voice/sarvam.md +21 -25
- package/.docs/reference/voice/speechify.md +11 -13
- package/.docs/reference/voice/voice.addInstructions.md +1 -1
- package/.docs/reference/voice/voice.addTools.md +1 -1
- package/.docs/reference/voice/voice.answer.md +1 -1
- package/.docs/reference/voice/voice.connect.md +3 -3
- package/.docs/reference/voice/voice.events.md +11 -11
- package/.docs/reference/voice/voice.getSpeakers.md +30 -30
- package/.docs/reference/voice/voice.listen.md +9 -9
- package/.docs/reference/voice/voice.off.md +2 -2
- package/.docs/reference/voice/voice.on.md +2 -2
- package/.docs/reference/voice/voice.send.md +1 -1
- package/.docs/reference/voice/voice.speak.md +11 -11
- package/.docs/reference/voice/voice.updateConfig.md +3 -3
- package/.docs/reference/workflows/run-methods/cancel.md +2 -2
- package/.docs/reference/workflows/run-methods/restart.md +17 -5
- package/.docs/reference/workflows/run-methods/resume.md +23 -9
- package/.docs/reference/workflows/run-methods/start.md +22 -8
- package/.docs/reference/workflows/run-methods/startAsync.md +12 -6
- package/.docs/reference/workflows/run-methods/timeTravel.md +29 -13
- package/.docs/reference/workflows/run.md +12 -12
- package/.docs/reference/workflows/step.md +24 -26
- package/.docs/reference/workflows/workflow-methods/branch.md +2 -2
- package/.docs/reference/workflows/workflow-methods/commit.md +1 -1
- package/.docs/reference/workflows/workflow-methods/create-run.md +4 -4
- package/.docs/reference/workflows/workflow-methods/dountil.md +3 -3
- package/.docs/reference/workflows/workflow-methods/dowhile.md +3 -3
- package/.docs/reference/workflows/workflow-methods/foreach.md +9 -9
- package/.docs/reference/workflows/workflow-methods/map.md +2 -2
- package/.docs/reference/workflows/workflow-methods/parallel.md +2 -2
- package/.docs/reference/workflows/workflow-methods/sleep.md +2 -2
- package/.docs/reference/workflows/workflow-methods/sleepUntil.md +2 -2
- package/.docs/reference/workflows/workflow-methods/then.md +2 -2
- package/.docs/reference/workflows/workflow.md +40 -50
- package/.docs/reference/workspace/daytona-sandbox.md +33 -33
- package/.docs/reference/workspace/e2b-sandbox.md +20 -20
- package/.docs/reference/workspace/filesystem.md +22 -22
- package/.docs/reference/workspace/gcs-filesystem.md +15 -15
- package/.docs/reference/workspace/local-filesystem.md +35 -35
- package/.docs/reference/workspace/local-sandbox.md +26 -26
- package/.docs/reference/workspace/s3-filesystem.md +18 -18
- package/.docs/reference/workspace/sandbox.md +8 -8
- package/.docs/reference/workspace/workspace-class.md +30 -34
- package/CHANGELOG.md +16 -0
- package/package.json +6 -6
- package/.docs/docs/getting-started/start.md +0 -28
|
@@ -57,21 +57,21 @@ voice.connect()
|
|
|
57
57
|
|
|
58
58
|
### Constructor Options
|
|
59
59
|
|
|
60
|
-
**model
|
|
60
|
+
**model** (`string`): The model ID to use for real-time voice interactions. (Default: `'gpt-5.1-realtime-preview-2024-12-17'`)
|
|
61
61
|
|
|
62
|
-
**apiKey
|
|
62
|
+
**apiKey** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
|
|
63
63
|
|
|
64
|
-
**speaker
|
|
64
|
+
**speaker** (`string`): Default voice ID for speech synthesis. (Default: `'alloy'`)
|
|
65
65
|
|
|
66
66
|
### Voice Activity Detection (VAD) Configuration
|
|
67
67
|
|
|
68
|
-
**type
|
|
68
|
+
**type** (`string`): Type of VAD to use. Server-side VAD provides better accuracy. (Default: `'server_vad'`)
|
|
69
69
|
|
|
70
|
-
**threshold
|
|
70
|
+
**threshold** (`number`): Speech detection sensitivity (0.0-1.0). (Default: `0.5`)
|
|
71
71
|
|
|
72
|
-
**prefix\_padding\_ms
|
|
72
|
+
**prefix\_padding\_ms** (`number`): Milliseconds of audio to include before speech is detected. (Default: `1000`)
|
|
73
73
|
|
|
74
|
-
**silence\_duration\_ms
|
|
74
|
+
**silence\_duration\_ms** (`number`): Milliseconds of silence before ending a turn. (Default: `1000`)
|
|
75
75
|
|
|
76
76
|
## Methods
|
|
77
77
|
|
|
@@ -79,15 +79,15 @@ voice.connect()
|
|
|
79
79
|
|
|
80
80
|
Establishes a connection to the OpenAI realtime service. Must be called before using speak, listen, or send functions.
|
|
81
81
|
|
|
82
|
-
**returns
|
|
82
|
+
**returns** (`Promise<void>`): Promise that resolves when the connection is established.
|
|
83
83
|
|
|
84
84
|
### speak()
|
|
85
85
|
|
|
86
86
|
Emits a speaking event using the configured voice model. Can accept either a string or a readable stream as input.
|
|
87
87
|
|
|
88
|
-
**input
|
|
88
|
+
**input** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
|
|
89
89
|
|
|
90
|
-
**options.speaker
|
|
90
|
+
**options.speaker** (`string`): Voice ID to use for this specific speech request. (Default: `Constructor's speaker value`)
|
|
91
91
|
|
|
92
92
|
Returns: `Promise<void>`
|
|
93
93
|
|
|
@@ -95,7 +95,7 @@ Returns: `Promise<void>`
|
|
|
95
95
|
|
|
96
96
|
Processes audio input for speech recognition. Takes a readable stream of audio data and emits a 'listening' event with the transcribed text.
|
|
97
97
|
|
|
98
|
-
**audioData
|
|
98
|
+
**audioData** (`NodeJS.ReadableStream`): Audio stream to transcribe.
|
|
99
99
|
|
|
100
100
|
Returns: `Promise<void>`
|
|
101
101
|
|
|
@@ -103,7 +103,7 @@ Returns: `Promise<void>`
|
|
|
103
103
|
|
|
104
104
|
Streams audio data in real-time to the OpenAI service for continuous audio streaming scenarios like live microphone input.
|
|
105
105
|
|
|
106
|
-
**audioData
|
|
106
|
+
**audioData** (`NodeJS.ReadableStream`): Audio stream to send to the service.
|
|
107
107
|
|
|
108
108
|
Returns: `Promise<void>`
|
|
109
109
|
|
|
@@ -111,7 +111,7 @@ Returns: `Promise<void>`
|
|
|
111
111
|
|
|
112
112
|
Updates the session configuration for the voice instance. This can be used to modify voice settings, turn detection, and other parameters.
|
|
113
113
|
|
|
114
|
-
**sessionConfig
|
|
114
|
+
**sessionConfig** (`Realtime.SessionConfig`): New session configuration to apply.
|
|
115
115
|
|
|
116
116
|
Returns: `void`
|
|
117
117
|
|
|
@@ -119,7 +119,7 @@ Returns: `void`
|
|
|
119
119
|
|
|
120
120
|
Adds a set of tools to the voice instance. Tools allow the model to perform additional actions during conversations. When OpenAIRealtimeVoice is added to an Agent, any tools configured for the Agent will automatically be available to the voice interface.
|
|
121
121
|
|
|
122
|
-
**tools
|
|
122
|
+
**tools** (`ToolsInput`): Tools configuration to equip.
|
|
123
123
|
|
|
124
124
|
Returns: `void`
|
|
125
125
|
|
|
@@ -139,9 +139,9 @@ Returns: `Promise<Array<{ voiceId: string; [key: string]: any }>>`
|
|
|
139
139
|
|
|
140
140
|
Registers an event listener for voice events.
|
|
141
141
|
|
|
142
|
-
**event
|
|
142
|
+
**event** (`string`): Name of the event to listen for.
|
|
143
143
|
|
|
144
|
-
**callback
|
|
144
|
+
**callback** (`Function`): Function to call when the event occurs.
|
|
145
145
|
|
|
146
146
|
Returns: `void`
|
|
147
147
|
|
|
@@ -149,9 +149,9 @@ Returns: `void`
|
|
|
149
149
|
|
|
150
150
|
Removes a previously registered event listener.
|
|
151
151
|
|
|
152
|
-
**event
|
|
152
|
+
**event** (`string`): Name of the event to stop listening to.
|
|
153
153
|
|
|
154
|
-
**callback
|
|
154
|
+
**callback** (`Function`): The specific callback function to remove.
|
|
155
155
|
|
|
156
156
|
Returns: `void`
|
|
157
157
|
|
|
@@ -159,25 +159,25 @@ Returns: `void`
|
|
|
159
159
|
|
|
160
160
|
The OpenAIRealtimeVoice class emits the following events:
|
|
161
161
|
|
|
162
|
-
**speaking
|
|
162
|
+
**speaking** (`event`): Emitted when audio data is received from the model. Callback receives { audio: Int16Array }.
|
|
163
163
|
|
|
164
|
-
**writing
|
|
164
|
+
**writing** (`event`): Emitted when transcribed text is available. Callback receives { text: string, role: string }.
|
|
165
165
|
|
|
166
|
-
**error
|
|
166
|
+
**error** (`event`): Emitted when an error occurs. Callback receives the error object.
|
|
167
167
|
|
|
168
168
|
### OpenAI Realtime Events
|
|
169
169
|
|
|
170
170
|
You can also listen to [OpenAI Realtime utility events](https://github.com/openai/openai-realtime-api-beta#reference-client-utility-events) by prefixing with 'openAIRealtime:':
|
|
171
171
|
|
|
172
|
-
**openAIRealtime:conversation.created
|
|
172
|
+
**openAIRealtime:conversation.created** (`event`): Emitted when a new conversation is created.
|
|
173
173
|
|
|
174
|
-
**openAIRealtime:conversation.interrupted
|
|
174
|
+
**openAIRealtime:conversation.interrupted** (`event`): Emitted when a conversation is interrupted.
|
|
175
175
|
|
|
176
|
-
**openAIRealtime:conversation.updated
|
|
176
|
+
**openAIRealtime:conversation.updated** (`event`): Emitted when a conversation is updated.
|
|
177
177
|
|
|
178
|
-
**openAIRealtime:conversation.item.appended
|
|
178
|
+
**openAIRealtime:conversation.item.appended** (`event`): Emitted when an item is appended to the conversation.
|
|
179
179
|
|
|
180
|
-
**openAIRealtime:conversation.item.completed
|
|
180
|
+
**openAIRealtime:conversation.item.completed** (`event`): Emitted when an item in the conversation is completed.
|
|
181
181
|
|
|
182
182
|
## Available Voices
|
|
183
183
|
|
|
@@ -39,17 +39,19 @@ const text = await voice.listen(audioStream, {
|
|
|
39
39
|
|
|
40
40
|
### Constructor Options
|
|
41
41
|
|
|
42
|
-
**speechModel
|
|
42
|
+
**speechModel** (`OpenAIConfig`): Configuration for text-to-speech synthesis. (Default: `{ name: 'tts-1' }`)
|
|
43
43
|
|
|
44
|
-
**
|
|
44
|
+
**speechModel.name** (`'tts-1' | 'tts-1-hd' | 'whisper-1'`): Model name. Use 'tts-1-hd' for higher quality audio.
|
|
45
45
|
|
|
46
|
-
**
|
|
46
|
+
**speechModel.apiKey** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
|
|
47
47
|
|
|
48
|
-
|
|
48
|
+
**listeningModel** (`OpenAIConfig`): Configuration for speech-to-text recognition. (Default: `{ name: 'whisper-1' }`)
|
|
49
49
|
|
|
50
|
-
**name
|
|
50
|
+
**listeningModel.name** (`'tts-1' | 'tts-1-hd' | 'whisper-1'`): Model name. Use 'tts-1-hd' for higher quality audio.
|
|
51
51
|
|
|
52
|
-
**apiKey
|
|
52
|
+
**listeningModel.apiKey** (`string`): OpenAI API key. Falls back to OPENAI\_API\_KEY environment variable.
|
|
53
|
+
|
|
54
|
+
**speaker** (`OpenAIVoiceId`): Default voice ID for speech synthesis. (Default: `'alloy'`)
|
|
53
55
|
|
|
54
56
|
## Methods
|
|
55
57
|
|
|
@@ -57,11 +59,11 @@ const text = await voice.listen(audioStream, {
|
|
|
57
59
|
|
|
58
60
|
Converts text to speech using OpenAI's text-to-speech models.
|
|
59
61
|
|
|
60
|
-
**input
|
|
62
|
+
**input** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
|
|
61
63
|
|
|
62
|
-
**options.speaker
|
|
64
|
+
**options.speaker** (`OpenAIVoiceId`): Voice ID to use for speech synthesis. (Default: `Constructor's speaker value`)
|
|
63
65
|
|
|
64
|
-
**options.speed
|
|
66
|
+
**options.speed** (`number`): Speech speed multiplier. (Default: `1.0`)
|
|
65
67
|
|
|
66
68
|
Returns: `Promise<NodeJS.ReadableStream>`
|
|
67
69
|
|
|
@@ -69,9 +71,9 @@ Returns: `Promise<NodeJS.ReadableStream>`
|
|
|
69
71
|
|
|
70
72
|
Transcribes audio using OpenAI's Whisper model.
|
|
71
73
|
|
|
72
|
-
**audioStream
|
|
74
|
+
**audioStream** (`NodeJS.ReadableStream`): Audio stream to transcribe.
|
|
73
75
|
|
|
74
|
-
**options.filetype
|
|
76
|
+
**options.filetype** (`string`): Audio format of the input stream. (Default: `'mp3'`)
|
|
75
77
|
|
|
76
78
|
Returns: `Promise<string>`
|
|
77
79
|
|
|
@@ -79,7 +81,7 @@ Returns: `Promise<string>`
|
|
|
79
81
|
|
|
80
82
|
Returns an array of available voice options, where each node contains:
|
|
81
83
|
|
|
82
|
-
**voiceId
|
|
84
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
83
85
|
|
|
84
86
|
## Notes
|
|
85
87
|
|
|
@@ -29,17 +29,15 @@ const audioStream = await voice.speak('Hello, world!', {
|
|
|
29
29
|
|
|
30
30
|
## Constructor Parameters
|
|
31
31
|
|
|
32
|
-
**speechModel
|
|
32
|
+
**speechModel** (`PlayAIConfig`): Configuration for text-to-speech functionality (Default: `{ name: 'PlayDialog' }`)
|
|
33
33
|
|
|
34
|
-
**
|
|
34
|
+
**speechModel.name** (`'PlayDialog' | 'Play3.0-mini'`): The PlayAI model to use
|
|
35
35
|
|
|
36
|
-
|
|
36
|
+
**speechModel.apiKey** (`string`): PlayAI API key. Falls back to PLAYAI\_API\_KEY environment variable
|
|
37
37
|
|
|
38
|
-
**
|
|
38
|
+
**speechModel.userId** (`string`): PlayAI user ID. Falls back to PLAYAI\_USER\_ID environment variable
|
|
39
39
|
|
|
40
|
-
**
|
|
41
|
-
|
|
42
|
-
**userId?:** (`string`): PlayAI user ID. Falls back to PLAYAI\_USER\_ID environment variable
|
|
40
|
+
**speaker** (`string`): Default voice ID to use for speech synthesis (Default: `First available voice ID`)
|
|
43
41
|
|
|
44
42
|
## Methods
|
|
45
43
|
|
|
@@ -47,9 +45,9 @@ const audioStream = await voice.speak('Hello, world!', {
|
|
|
47
45
|
|
|
48
46
|
Converts text to speech using the configured speech model and voice.
|
|
49
47
|
|
|
50
|
-
**input
|
|
48
|
+
**input** (`string | NodeJS.ReadableStream`): Text to convert to speech. If a stream is provided, it will be converted to text first.
|
|
51
49
|
|
|
52
|
-
**options.speaker
|
|
50
|
+
**options.speaker** (`string`): Override the default speaker for this request (Default: `Constructor's speaker value`)
|
|
53
51
|
|
|
54
52
|
Returns: `Promise<NodeJS.ReadableStream>`.
|
|
55
53
|
|
|
@@ -57,17 +55,17 @@ Returns: `Promise<NodeJS.ReadableStream>`.
|
|
|
57
55
|
|
|
58
56
|
Returns an array of available voice options, where each node contains:
|
|
59
57
|
|
|
60
|
-
**name
|
|
58
|
+
**name** (`string`): Name of the voice
|
|
61
59
|
|
|
62
|
-
**accent
|
|
60
|
+
**accent** (`string`): Accent of the voice (e.g., 'US', 'British', 'Australian')
|
|
63
61
|
|
|
64
|
-
**gender
|
|
62
|
+
**gender** (`'M' | 'F'`): Gender of the voice
|
|
65
63
|
|
|
66
|
-
**age
|
|
64
|
+
**age** (`'Young' | 'Middle' | 'Old'`): Age category of the voice
|
|
67
65
|
|
|
68
|
-
**style
|
|
66
|
+
**style** (`'Conversational' | 'Narrative'`): Speaking style of the voice
|
|
69
67
|
|
|
70
|
-
**voiceId
|
|
68
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
71
69
|
|
|
72
70
|
### listen()
|
|
73
71
|
|
|
@@ -51,43 +51,39 @@ const text = await voice.listen(audioStream, {
|
|
|
51
51
|
|
|
52
52
|
### Constructor Options
|
|
53
53
|
|
|
54
|
-
**speechModel
|
|
54
|
+
**speechModel** (`SarvamVoiceConfig`): Configuration for text-to-speech synthesis. (Default: `{ model: 'bulbul:v1', language: 'en-IN' }`)
|
|
55
55
|
|
|
56
|
-
**
|
|
56
|
+
**speechModel.apiKey** (`string`): Sarvam API key. Falls back to SARVAM\_API\_KEY environment variable.
|
|
57
57
|
|
|
58
|
-
**
|
|
58
|
+
**speechModel.model** (`SarvamTTSModel`): Specifies the model to use for text-to-speech conversion.
|
|
59
59
|
|
|
60
|
-
|
|
60
|
+
**speechModel.language** (`SarvamTTSLanguage`): Target language for speech synthesis. Available options: hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN
|
|
61
61
|
|
|
62
|
-
**
|
|
62
|
+
**speechModel.properties** (`object`): Additional voice properties for customization.
|
|
63
63
|
|
|
64
|
-
**
|
|
64
|
+
**speechModel.properties.pitch** (`number`): Controls the pitch of the audio. Lower values result in a deeper voice, while higher values make it sharper. The suitable range is between -0.75 and 0.75.
|
|
65
65
|
|
|
66
|
-
**
|
|
66
|
+
**speechModel.properties.pace** (`number`): Controls the speed of the audio. Lower values result in slower speech, while higher values make it faster. The suitable range is between 0.5 and 2.0. Default is 1.0. Required range: 0.3 <= x <= 3
|
|
67
67
|
|
|
68
|
-
**properties
|
|
68
|
+
**speechModel.properties.loudness** (`number`): Controls the loudness of the audio. Lower values result in quieter audio, while higher values make it louder. The suitable range is between 0.3 and 3.0. Required range: 0 <= x <= 3
|
|
69
69
|
|
|
70
|
-
**properties.
|
|
70
|
+
**speechModel.properties.speech\_sample\_rate** (`8000 | 16000 | 22050`): Audio sample rate in Hz.
|
|
71
71
|
|
|
72
|
-
**properties.
|
|
72
|
+
**speechModel.properties.enable\_preprocessing** (`boolean`): Controls whether normalization of English words and numeric entities (e.g., numbers, dates) is performed. Set to true for better handling of mixed-language text. Default is false.
|
|
73
73
|
|
|
74
|
-
**properties.
|
|
74
|
+
**speechModel.properties.eng\_interpolation\_wt** (`number`): Weight for interpolating with English speaker at encoder.
|
|
75
75
|
|
|
76
|
-
**
|
|
76
|
+
**speaker** (`SarvamVoiceId`): The speaker to be used for the output audio. If not provided, Meera will be used as default. AvailableOptions - meera, pavithra, maitreyi, arvind, amol, amartya, diya, neel, misha, vian, arjun, maya (Default: `'meera'`)
|
|
77
77
|
|
|
78
|
-
**
|
|
78
|
+
**listeningModel** (`SarvamListenOptions`): Configuration for speech-to-text recognition. (Default: `{ model: 'saarika:v2', language_code: 'unknown' }`)
|
|
79
79
|
|
|
80
|
-
**
|
|
80
|
+
**listeningModel.apiKey** (`string`): Sarvam API key. Falls back to SARVAM\_API\_KEY environment variable.
|
|
81
81
|
|
|
82
|
-
|
|
82
|
+
**listeningModel.model** (`SarvamSTTModel`): Specifies the model to use for speech-to-text conversion. Note:- Default model is saarika:v2 . Available options: saarika:v1, saarika:v2, saarika:flash
|
|
83
83
|
|
|
84
|
-
**
|
|
84
|
+
**listeningModel.languageCode** (`SarvamSTTLanguage`): Specifies the language of the input audio. This parameter is required to ensure accurate transcription. For the saarika:v1 model, this parameter is mandatory. For the saarika:v2 model, it is optional. unknown: Use this when the language is not known; the API will detect it automatically. Note:- that the saarika:v1 model does not support unknown language code. Available options: unknown, hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN
|
|
85
85
|
|
|
86
|
-
**
|
|
87
|
-
|
|
88
|
-
**languageCode?:** (`SarvamSTTLanguage`): Specifies the language of the input audio. This parameter is required to ensure accurate transcription. For the saarika:v1 model, this parameter is mandatory. For the saarika:v2 model, it is optional. unknown: Use this when the language is not known; the API will detect it automatically. Note:- that the saarika:v1 model does not support unknown language code. Available options: unknown, hi-IN, bn-IN, kn-IN, ml-IN, mr-IN, od-IN, pa-IN, ta-IN, te-IN, en-IN, gu-IN (Default: `'unknown'`)
|
|
89
|
-
|
|
90
|
-
**filetype?:** (`'mp3' | 'wav'`): Audio format of the input stream.
|
|
86
|
+
**listeningModel.filetype** (`'mp3' | 'wav'`): Audio format of the input stream.
|
|
91
87
|
|
|
92
88
|
## Methods
|
|
93
89
|
|
|
@@ -95,9 +91,9 @@ const text = await voice.listen(audioStream, {
|
|
|
95
91
|
|
|
96
92
|
Converts text to speech using Sarvam's text-to-speech models.
|
|
97
93
|
|
|
98
|
-
**input
|
|
94
|
+
**input** (`string | NodeJS.ReadableStream`): Text or text stream to convert to speech.
|
|
99
95
|
|
|
100
|
-
**options.speaker
|
|
96
|
+
**options.speaker** (`SarvamVoiceId`): Voice ID to use for speech synthesis. (Default: `Constructor's speaker value`)
|
|
101
97
|
|
|
102
98
|
Returns: `Promise<NodeJS.ReadableStream>`
|
|
103
99
|
|
|
@@ -105,9 +101,9 @@ Returns: `Promise<NodeJS.ReadableStream>`
|
|
|
105
101
|
|
|
106
102
|
Transcribes audio using Sarvam's speech recognition models.
|
|
107
103
|
|
|
108
|
-
**input
|
|
104
|
+
**input** (`NodeJS.ReadableStream`): Audio stream to transcribe.
|
|
109
105
|
|
|
110
|
-
**options
|
|
106
|
+
**options** (`SarvamListenOptions`): Configuration options for speech recognition.
|
|
111
107
|
|
|
112
108
|
Returns: `Promise<string>`
|
|
113
109
|
|
|
@@ -27,15 +27,13 @@ const audioStream = await voice.speak('Hello, world!', {
|
|
|
27
27
|
|
|
28
28
|
## Constructor Parameters
|
|
29
29
|
|
|
30
|
-
**speechModel
|
|
30
|
+
**speechModel** (`SpeechifyConfig`): Configuration for text-to-speech functionality (Default: `{ name: 'simba-english' }`)
|
|
31
31
|
|
|
32
|
-
**
|
|
32
|
+
**speechModel.name** (`VoiceModelName`): The Speechify model to use
|
|
33
33
|
|
|
34
|
-
|
|
34
|
+
**speechModel.apiKey** (`string`): Speechify API key. Falls back to SPEECHIFY\_API\_KEY environment variable
|
|
35
35
|
|
|
36
|
-
**
|
|
37
|
-
|
|
38
|
-
**apiKey?:** (`string`): Speechify API key. Falls back to SPEECHIFY\_API\_KEY environment variable
|
|
36
|
+
**speaker** (`SpeechifyVoiceId`): Default voice ID to use for speech synthesis (Default: `'george'`)
|
|
39
37
|
|
|
40
38
|
## Methods
|
|
41
39
|
|
|
@@ -43,11 +41,11 @@ const audioStream = await voice.speak('Hello, world!', {
|
|
|
43
41
|
|
|
44
42
|
Converts text to speech using the configured speech model and voice.
|
|
45
43
|
|
|
46
|
-
**input
|
|
44
|
+
**input** (`string | NodeJS.ReadableStream`): Text to convert to speech. If a stream is provided, it will be converted to text first.
|
|
47
45
|
|
|
48
|
-
**options.speaker
|
|
46
|
+
**options.speaker** (`string`): Override the default speaker for this request (Default: `Constructor's speaker value`)
|
|
49
47
|
|
|
50
|
-
**options.model
|
|
48
|
+
**options.model** (`VoiceModelName`): Override the default model for this request (Default: `Constructor's model value`)
|
|
51
49
|
|
|
52
50
|
Returns: `Promise<NodeJS.ReadableStream>`
|
|
53
51
|
|
|
@@ -55,13 +53,13 @@ Returns: `Promise<NodeJS.ReadableStream>`
|
|
|
55
53
|
|
|
56
54
|
Returns an array of available voice options, where each node contains:
|
|
57
55
|
|
|
58
|
-
**voiceId
|
|
56
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
59
57
|
|
|
60
|
-
**name
|
|
58
|
+
**name** (`string`): Display name of the voice
|
|
61
59
|
|
|
62
|
-
**language
|
|
60
|
+
**language** (`string`): Language code for the voice
|
|
63
61
|
|
|
64
|
-
**gender
|
|
62
|
+
**gender** (`string`): Gender of the voice
|
|
65
63
|
|
|
66
64
|
### listen()
|
|
67
65
|
|
|
@@ -50,7 +50,7 @@ await voice.connect()
|
|
|
50
50
|
|
|
51
51
|
## Parameters
|
|
52
52
|
|
|
53
|
-
**tools
|
|
53
|
+
**tools** (`ToolsInput`): Object containing tool definitions that can be called by the voice model
|
|
54
54
|
|
|
55
55
|
## Return Value
|
|
56
56
|
|
|
@@ -46,7 +46,7 @@ await voice.connect({
|
|
|
46
46
|
|
|
47
47
|
## Parameters
|
|
48
48
|
|
|
49
|
-
**options
|
|
49
|
+
**options** (`Record<string, unknown>`): Provider-specific connection options
|
|
50
50
|
|
|
51
51
|
## Return Value
|
|
52
52
|
|
|
@@ -58,9 +58,9 @@ Each real-time voice provider may support different options for the `connect()`
|
|
|
58
58
|
|
|
59
59
|
### OpenAI Realtime
|
|
60
60
|
|
|
61
|
-
**options.timeout
|
|
61
|
+
**options.timeout** (`number`): Connection timeout in milliseconds (Default: `30000`)
|
|
62
62
|
|
|
63
|
-
**options.reconnect
|
|
63
|
+
**options.reconnect** (`boolean`): Whether to automatically reconnect on connection loss (Default: `false`)
|
|
64
64
|
|
|
65
65
|
## Using with CompositeVoice
|
|
66
66
|
|
|
@@ -6,27 +6,27 @@ Voice providers emit various events during real-time voice interactions. These e
|
|
|
6
6
|
|
|
7
7
|
These events are commonly implemented across real-time voice providers:
|
|
8
8
|
|
|
9
|
-
**error
|
|
9
|
+
**error** (`Error`): Emitted when an error occurs during voice processing or when audio data format is unsupported
|
|
10
10
|
|
|
11
|
-
**session.created
|
|
11
|
+
**session.created** (`object`): Emitted when a new session is created with the OpenAI service
|
|
12
12
|
|
|
13
|
-
**session.updated
|
|
13
|
+
**session.updated** (`object`): Emitted when the session configuration is updated
|
|
14
14
|
|
|
15
|
-
**response.created
|
|
15
|
+
**response.created** (`object`): Emitted when a new response is created by the AI assistant
|
|
16
16
|
|
|
17
|
-
**response.done
|
|
17
|
+
**response.done** (`object`): Emitted when the AI assistant has completed its response
|
|
18
18
|
|
|
19
|
-
**speaker
|
|
19
|
+
**speaker** (`StreamWithId`): Emitted with a new audio stream that can be piped to an audio output
|
|
20
20
|
|
|
21
|
-
**writing
|
|
21
|
+
**writing** (`object`): Emitted when text is being transcribed (user) or generated (assistant)
|
|
22
22
|
|
|
23
|
-
**speaking
|
|
23
|
+
**speaking** (`object`): Emitted when audio data is available from the voice provider
|
|
24
24
|
|
|
25
|
-
**speaking.done
|
|
25
|
+
**speaking.done** (`object`): Emitted when the voice provider has finished speaking
|
|
26
26
|
|
|
27
|
-
**tool-call-start
|
|
27
|
+
**tool-call-start** (`object`): Emitted when the AI assistant starts executing a tool
|
|
28
28
|
|
|
29
|
-
**tool-call-result
|
|
29
|
+
**tool-call-result** (`object`): Emitted when a tool execution is complete with its result
|
|
30
30
|
|
|
31
31
|
## Notes
|
|
32
32
|
|
|
@@ -36,7 +36,7 @@ This method does not accept any parameters.
|
|
|
36
36
|
|
|
37
37
|
## Return Value
|
|
38
38
|
|
|
39
|
-
**Promise\<Array<{ voiceId: string } & TSpeakerMetadata
|
|
39
|
+
**Promise\<Array<{ voiceId: string } & TSpeakerMetadata>>** (`Promise`): A promise that resolves to an array of voice options, where each option contains at least a voiceId property and may include additional provider-specific metadata.
|
|
40
40
|
|
|
41
41
|
## Provider-Specific Metadata
|
|
42
42
|
|
|
@@ -44,81 +44,81 @@ Different voice providers return different metadata for their voices:
|
|
|
44
44
|
|
|
45
45
|
**OpenAI**:
|
|
46
46
|
|
|
47
|
-
**voiceId
|
|
47
|
+
**voiceId** (`string`): Unique identifier for the voice (e.g., 'alloy', 'echo', 'fable', 'onyx', 'nova', 'shimmer')
|
|
48
48
|
|
|
49
49
|
**OpenAI Realtime**:
|
|
50
50
|
|
|
51
|
-
**voiceId
|
|
51
|
+
**voiceId** (`string`): Unique identifier for the voice (e.g., 'alloy', 'echo', 'fable', 'onyx', 'nova', 'shimmer')
|
|
52
52
|
|
|
53
53
|
**Deepgram**:
|
|
54
54
|
|
|
55
|
-
**voiceId
|
|
55
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
56
56
|
|
|
57
|
-
**language
|
|
57
|
+
**language** (`string`): Language code embedded in the voice ID (e.g., 'en')
|
|
58
58
|
|
|
59
59
|
**ElevenLabs**:
|
|
60
60
|
|
|
61
|
-
**voiceId
|
|
61
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
62
62
|
|
|
63
|
-
**name
|
|
63
|
+
**name** (`string`): Human-readable name of the voice
|
|
64
64
|
|
|
65
|
-
**category
|
|
65
|
+
**category** (`string`): Category of the voice (e.g., 'premade', 'cloned')
|
|
66
66
|
|
|
67
67
|
**Google**:
|
|
68
68
|
|
|
69
|
-
**voiceId
|
|
69
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
70
70
|
|
|
71
|
-
**languageCodes
|
|
71
|
+
**languageCodes** (`string[]`): Array of language codes supported by the voice (e.g., \['en-US'])
|
|
72
72
|
|
|
73
73
|
**Azure**:
|
|
74
74
|
|
|
75
|
-
**voiceId
|
|
75
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
76
76
|
|
|
77
|
-
**language
|
|
77
|
+
**language** (`string`): Language code extracted from the voice ID (e.g., 'en')
|
|
78
78
|
|
|
79
|
-
**region
|
|
79
|
+
**region** (`string`): Region code extracted from the voice ID (e.g., 'US')
|
|
80
80
|
|
|
81
81
|
**Murf**:
|
|
82
82
|
|
|
83
|
-
**voiceId
|
|
83
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
84
84
|
|
|
85
|
-
**name
|
|
85
|
+
**name** (`string`): Name of the voice (same as voiceId)
|
|
86
86
|
|
|
87
|
-
**language
|
|
87
|
+
**language** (`string`): Language code extracted from the voice ID (e.g., 'en')
|
|
88
88
|
|
|
89
|
-
**gender
|
|
89
|
+
**gender** (`string`): Gender of the voice (always 'neutral' in current implementation)
|
|
90
90
|
|
|
91
91
|
**PlayAI**:
|
|
92
92
|
|
|
93
|
-
**voiceId
|
|
93
|
+
**voiceId** (`string`): Unique identifier for the voice (S3 URL to manifest.json)
|
|
94
94
|
|
|
95
|
-
**name
|
|
95
|
+
**name** (`string`): Human-readable name of the voice (e.g., 'Angelo', 'Arsenio')
|
|
96
96
|
|
|
97
|
-
**accent
|
|
97
|
+
**accent** (`string`): Accent of the voice (e.g., 'US', 'Irish', 'US African American')
|
|
98
98
|
|
|
99
|
-
**gender
|
|
99
|
+
**gender** (`string`): Gender of the voice ('M' or 'F')
|
|
100
100
|
|
|
101
|
-
**age
|
|
101
|
+
**age** (`string`): Age category of the voice (e.g., 'Young', 'Middle')
|
|
102
102
|
|
|
103
|
-
**style
|
|
103
|
+
**style** (`string`): Speaking style of the voice (e.g., 'Conversational')
|
|
104
104
|
|
|
105
105
|
**Speechify**:
|
|
106
106
|
|
|
107
|
-
**voiceId
|
|
107
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
108
108
|
|
|
109
|
-
**name
|
|
109
|
+
**name** (`string`): Human-readable name of the voice
|
|
110
110
|
|
|
111
|
-
**language
|
|
111
|
+
**language** (`string`): Language code of the voice (e.g., 'en-US')
|
|
112
112
|
|
|
113
113
|
**Sarvam**:
|
|
114
114
|
|
|
115
|
-
**voiceId
|
|
115
|
+
**voiceId** (`string`): Unique identifier for the voice
|
|
116
116
|
|
|
117
|
-
**name
|
|
117
|
+
**name** (`string`): Human-readable name of the voice
|
|
118
118
|
|
|
119
|
-
**language
|
|
119
|
+
**language** (`string`): Language of the voice (e.g., 'english', 'hindi')
|
|
120
120
|
|
|
121
|
-
**gender
|
|
121
|
+
**gender** (`string`): Gender of the voice ('male' or 'female')
|
|
122
122
|
|
|
123
123
|
## Notes
|
|
124
124
|
|