hume 0.12.2 → 0.13.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.mock/definition/empathic-voice/__package__.yml +688 -735
- package/.mock/definition/empathic-voice/chat.yml +29 -23
- package/.mock/definition/empathic-voice/chatWebhooks.yml +8 -12
- package/.mock/definition/empathic-voice/prompts.yml +2 -2
- package/.mock/definition/empathic-voice/tools.yml +2 -2
- package/.mock/definition/tts/__package__.yml +70 -87
- package/.mock/fern.config.json +1 -1
- package/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/index.js +0 -15
- package/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/api/resources/empathicVoice/errors/index.js +0 -1
- package/api/resources/empathicVoice/resources/chat/client/Client.d.ts +2 -0
- package/api/resources/empathicVoice/resources/prompts/client/requests/PostedPrompt.d.ts +1 -1
- package/api/resources/empathicVoice/resources/prompts/client/requests/PostedPromptVersion.d.ts +1 -1
- package/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedTool.d.ts +1 -1
- package/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedToolVersion.d.ts +1 -1
- package/api/resources/empathicVoice/types/AssistantEnd.d.ts +3 -3
- package/api/resources/empathicVoice/types/AssistantInput.d.ts +3 -3
- package/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/api/resources/empathicVoice/types/AudioInput.d.ts +7 -7
- package/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +3 -3
- package/api/resources/empathicVoice/types/PostedLanguageModel.d.ts +1 -1
- package/api/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +2 -2
- package/api/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +2 -2
- package/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/api/resources/empathicVoice/types/ReturnLanguageModel.d.ts +1 -1
- package/api/resources/empathicVoice/types/ReturnPrompt.d.ts +4 -6
- package/api/resources/empathicVoice/types/ReturnUserDefinedTool.d.ts +1 -1
- package/api/resources/empathicVoice/types/SessionSettings.d.ts +30 -30
- package/api/resources/empathicVoice/types/Tool.d.ts +7 -7
- package/api/resources/empathicVoice/types/ToolCallMessage.d.ts +7 -7
- package/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/UserInput.d.ts +3 -3
- package/api/resources/empathicVoice/types/UserInterruption.d.ts +5 -5
- package/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/api/resources/empathicVoice/types/index.js +16 -17
- package/api/resources/tts/client/Client.d.ts +21 -23
- package/api/resources/tts/client/Client.js +50 -58
- package/api/resources/tts/client/index.d.ts +0 -1
- package/api/resources/tts/client/index.js +0 -15
- package/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/index.js +0 -15
- package/dist/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/errors/index.js +0 -1
- package/dist/api/resources/empathicVoice/resources/chat/client/Client.d.ts +2 -0
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PostedPrompt.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PostedPromptVersion.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedTool.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedToolVersion.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/AssistantEnd.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/AssistantInput.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/AudioInput.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/dist/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/dist/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/dist/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/dist/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/dist/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/PostedLanguageModel.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ReturnLanguageModel.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ReturnPrompt.d.ts +4 -6
- package/dist/api/resources/empathicVoice/types/ReturnUserDefinedTool.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/SessionSettings.d.ts +30 -30
- package/dist/api/resources/empathicVoice/types/Tool.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ToolCallMessage.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/dist/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/UserInput.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/UserInterruption.d.ts +5 -5
- package/dist/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/dist/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/dist/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/dist/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/dist/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/api/resources/empathicVoice/types/index.js +16 -17
- package/dist/api/resources/tts/client/Client.d.ts +21 -23
- package/dist/api/resources/tts/client/Client.js +50 -58
- package/dist/api/resources/tts/client/index.d.ts +0 -1
- package/dist/api/resources/tts/client/index.js +0 -15
- package/dist/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/dist/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/dist/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/dist/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/index.js +0 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/serialization/resources/empathicVoice/types/index.js +16 -17
- package/dist/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/dist/serialization/resources/tts/types/PostedTts.js +3 -3
- package/dist/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/dist/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/dist/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/dist/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/dist/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/dist/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/dist/serialization/resources/tts/types/Snippet.js +3 -3
- package/dist/version.d.ts +1 -1
- package/dist/version.js +1 -1
- package/package.json +1 -1
- package/reference.md +24 -78
- package/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/index.js +0 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.js +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/serialization/resources/empathicVoice/types/index.js +16 -17
- package/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/serialization/resources/tts/types/PostedTts.js +3 -3
- package/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/serialization/resources/tts/types/Snippet.js +3 -3
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/api/resources/tts/client/requests/index.d.ts +0 -1
- package/api/resources/tts/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/dist/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/dist/api/resources/tts/client/requests/index.d.ts +0 -1
- package/dist/api/resources/tts/client/requests/index.js +0 -2
- package/dist/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/index.js +0 -17
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/dist/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
- package/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/index.js +0 -17
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
|
@@ -1,73 +1,20 @@
|
|
|
1
1
|
errors:
|
|
2
|
-
UnprocessableEntityError:
|
|
3
|
-
status-code: 422
|
|
4
|
-
type: HTTPValidationError
|
|
5
|
-
docs: Validation Error
|
|
6
|
-
examples:
|
|
7
|
-
- value: {}
|
|
8
2
|
BadRequestError:
|
|
9
3
|
status-code: 400
|
|
10
4
|
type: ErrorResponse
|
|
11
5
|
docs: Bad Request
|
|
12
6
|
examples:
|
|
13
7
|
- value: {}
|
|
14
|
-
service:
|
|
15
|
-
auth: false
|
|
16
|
-
base-path: ''
|
|
17
|
-
endpoints:
|
|
18
|
-
custom_language_model_supports_tool_use_v0_evi_custom_language_model_supports_tool_use_post:
|
|
19
|
-
path: /v0/evi/custom_language_model_supports_tool_use
|
|
20
|
-
method: POST
|
|
21
|
-
auth: true
|
|
22
|
-
source:
|
|
23
|
-
openapi: evi-openapi.json
|
|
24
|
-
display-name: Custom Language Model Supports Tool Use
|
|
25
|
-
request:
|
|
26
|
-
name: >-
|
|
27
|
-
BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost
|
|
28
|
-
body:
|
|
29
|
-
properties:
|
|
30
|
-
model_resource: string
|
|
31
|
-
content-type: application/json
|
|
32
|
-
response:
|
|
33
|
-
docs: Successful Response
|
|
34
|
-
type: SupportsToolUse
|
|
35
|
-
status-code: 200
|
|
36
|
-
errors:
|
|
37
|
-
- UnprocessableEntityError
|
|
38
|
-
examples:
|
|
39
|
-
- request:
|
|
40
|
-
model_resource: model_resource
|
|
41
|
-
response:
|
|
42
|
-
body:
|
|
43
|
-
model_resource: model_resource
|
|
44
|
-
supports_tool_use: true
|
|
45
|
-
source:
|
|
46
|
-
openapi: evi-openapi.json
|
|
47
8
|
types:
|
|
48
|
-
AssistantEnd:
|
|
49
|
-
docs: When provided, the output is an assistant end message.
|
|
50
|
-
properties:
|
|
51
|
-
custom_session_id:
|
|
52
|
-
type: optional<string>
|
|
53
|
-
docs: >-
|
|
54
|
-
Used to manage conversational state, correlate frontend and backend
|
|
55
|
-
data, and persist conversations across EVI sessions.
|
|
56
|
-
type:
|
|
57
|
-
type: literal<"assistant_end">
|
|
58
|
-
docs: >-
|
|
59
|
-
The type of message sent through the socket; for an Assistant End
|
|
60
|
-
message, this must be `assistant_end`.
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
This message indicates the conclusion of the assistant’s response,
|
|
64
|
-
signaling that the assistant has finished speaking for the current
|
|
65
|
-
conversational turn.
|
|
66
|
-
source:
|
|
67
|
-
openapi: evi-asyncapi.json
|
|
68
9
|
AssistantInput:
|
|
69
10
|
docs: When provided, the input is spoken by EVI.
|
|
70
11
|
properties:
|
|
12
|
+
type:
|
|
13
|
+
type: literal<"assistant_input">
|
|
14
|
+
docs: >-
|
|
15
|
+
The type of message sent through the socket; must be `assistant_input`
|
|
16
|
+
for our server to correctly identify and process it as an Assistant
|
|
17
|
+
Input message.
|
|
71
18
|
custom_session_id:
|
|
72
19
|
type: optional<string>
|
|
73
20
|
docs: >-
|
|
@@ -82,88 +29,20 @@ types:
|
|
|
82
29
|
|
|
83
30
|
EVI uses this text to generate spoken audio using our proprietary
|
|
84
31
|
expressive text-to-speech model. Our model adds appropriate emotional
|
|
85
|
-
inflections and tones to the text based on the user
|
|
32
|
+
inflections and tones to the text based on the user's expressions and
|
|
86
33
|
the context of the conversation. The synthesized audio is streamed
|
|
87
34
|
back to the user as an [Assistant
|
|
88
35
|
Message](/reference/empathic-voice-interface-evi/chat/chat#receive.AssistantMessage.type).
|
|
89
|
-
type:
|
|
90
|
-
type: literal<"assistant_input">
|
|
91
|
-
docs: >-
|
|
92
|
-
The type of message sent through the socket; must be `assistant_input`
|
|
93
|
-
for our server to correctly identify and process it as an Assistant
|
|
94
|
-
Input message.
|
|
95
|
-
source:
|
|
96
|
-
openapi: evi-asyncapi.json
|
|
97
|
-
AssistantMessage:
|
|
98
|
-
docs: When provided, the output is an assistant message.
|
|
99
|
-
properties:
|
|
100
|
-
custom_session_id:
|
|
101
|
-
type: optional<string>
|
|
102
|
-
docs: >-
|
|
103
|
-
Used to manage conversational state, correlate frontend and backend
|
|
104
|
-
data, and persist conversations across EVI sessions.
|
|
105
|
-
from_text:
|
|
106
|
-
type: boolean
|
|
107
|
-
docs: >-
|
|
108
|
-
Indicates if this message was inserted into the conversation as text
|
|
109
|
-
from an [Assistant Input
|
|
110
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#send.AssistantInput.text).
|
|
111
|
-
id:
|
|
112
|
-
type: optional<string>
|
|
113
|
-
docs: >-
|
|
114
|
-
ID of the assistant message. Allows the Assistant Message to be
|
|
115
|
-
tracked and referenced.
|
|
116
|
-
message:
|
|
117
|
-
type: ChatMessage
|
|
118
|
-
docs: Transcript of the message.
|
|
119
|
-
models:
|
|
120
|
-
type: Inference
|
|
121
|
-
docs: Inference model results.
|
|
122
|
-
type:
|
|
123
|
-
type: literal<"assistant_message">
|
|
124
|
-
docs: >-
|
|
125
|
-
The type of message sent through the socket; for an Assistant Message,
|
|
126
|
-
this must be `assistant_message`.
|
|
127
|
-
|
|
128
|
-
|
|
129
|
-
This message contains both a transcript of the assistant’s response
|
|
130
|
-
and the expression measurement predictions of the assistant’s audio
|
|
131
|
-
output.
|
|
132
|
-
source:
|
|
133
|
-
openapi: evi-asyncapi.json
|
|
134
|
-
AssistantProsody:
|
|
135
|
-
docs: When provided, the output is an Assistant Prosody message.
|
|
136
|
-
properties:
|
|
137
|
-
custom_session_id:
|
|
138
|
-
type: optional<string>
|
|
139
|
-
docs: >-
|
|
140
|
-
Used to manage conversational state, correlate frontend and backend
|
|
141
|
-
data, and persist conversations across EVI sessions.
|
|
142
|
-
id:
|
|
143
|
-
type: optional<string>
|
|
144
|
-
docs: Unique identifier for the segment.
|
|
145
|
-
models:
|
|
146
|
-
type: Inference
|
|
147
|
-
docs: Inference model results.
|
|
148
|
-
type:
|
|
149
|
-
type: literal<"assistant_prosody">
|
|
150
|
-
docs: >-
|
|
151
|
-
The type of message sent through the socket; for an Assistant Prosody
|
|
152
|
-
message, this must be `assistant_PROSODY`.
|
|
153
|
-
|
|
154
|
-
|
|
155
|
-
This message the expression measurement predictions of the assistant's
|
|
156
|
-
audio output.
|
|
157
36
|
source:
|
|
158
37
|
openapi: evi-asyncapi.json
|
|
159
38
|
AudioConfiguration:
|
|
160
39
|
properties:
|
|
161
|
-
channels:
|
|
162
|
-
type: integer
|
|
163
|
-
docs: Number of audio channels.
|
|
164
40
|
encoding:
|
|
165
41
|
type: Encoding
|
|
166
42
|
docs: Encoding format of the audio input, such as `linear16`.
|
|
43
|
+
channels:
|
|
44
|
+
type: integer
|
|
45
|
+
docs: Number of audio channels.
|
|
167
46
|
sample_rate:
|
|
168
47
|
type: integer
|
|
169
48
|
docs: >-
|
|
@@ -174,6 +53,17 @@ types:
|
|
|
174
53
|
AudioInput:
|
|
175
54
|
docs: When provided, the input is audio.
|
|
176
55
|
properties:
|
|
56
|
+
type:
|
|
57
|
+
type: literal<"audio_input">
|
|
58
|
+
docs: >-
|
|
59
|
+
The type of message sent through the socket; must be `audio_input` for
|
|
60
|
+
our server to correctly identify and process it as an Audio Input
|
|
61
|
+
message.
|
|
62
|
+
|
|
63
|
+
|
|
64
|
+
This message is used for sending audio input data to EVI for
|
|
65
|
+
processing and expression measurement. Audio data should be sent as a
|
|
66
|
+
continuous stream, encoded in Base64.
|
|
177
67
|
custom_session_id:
|
|
178
68
|
type: optional<string>
|
|
179
69
|
docs: >-
|
|
@@ -185,7 +75,7 @@ types:
|
|
|
185
75
|
Base64 encoded audio input to insert into the conversation.
|
|
186
76
|
|
|
187
77
|
|
|
188
|
-
The content of an Audio Input message is treated as the user
|
|
78
|
+
The content of an Audio Input message is treated as the user's speech
|
|
189
79
|
to EVI and must be streamed continuously. Pre-recorded audio files are
|
|
190
80
|
not supported.
|
|
191
81
|
|
|
@@ -196,48 +86,6 @@ types:
|
|
|
196
86
|
|
|
197
87
|
Hume recommends streaming audio with a buffer window of 20
|
|
198
88
|
milliseconds (ms), or 100 milliseconds (ms) for web applications.
|
|
199
|
-
type:
|
|
200
|
-
type: literal<"audio_input">
|
|
201
|
-
docs: >-
|
|
202
|
-
The type of message sent through the socket; must be `audio_input` for
|
|
203
|
-
our server to correctly identify and process it as an Audio Input
|
|
204
|
-
message.
|
|
205
|
-
|
|
206
|
-
|
|
207
|
-
This message is used for sending audio input data to EVI for
|
|
208
|
-
processing and expression measurement. Audio data should be sent as a
|
|
209
|
-
continuous stream, encoded in Base64.
|
|
210
|
-
source:
|
|
211
|
-
openapi: evi-asyncapi.json
|
|
212
|
-
AudioOutput:
|
|
213
|
-
docs: >-
|
|
214
|
-
The type of message sent through the socket; for an Audio Output message,
|
|
215
|
-
this must be `audio_output`.
|
|
216
|
-
properties:
|
|
217
|
-
custom_session_id:
|
|
218
|
-
type: optional<string>
|
|
219
|
-
docs: >-
|
|
220
|
-
Used to manage conversational state, correlate frontend and backend
|
|
221
|
-
data, and persist conversations across EVI sessions.
|
|
222
|
-
data:
|
|
223
|
-
type: string
|
|
224
|
-
docs: >-
|
|
225
|
-
Base64 encoded audio output. This encoded audio is transmitted to the
|
|
226
|
-
client, where it can be decoded and played back as part of the user
|
|
227
|
-
interaction.
|
|
228
|
-
id:
|
|
229
|
-
type: string
|
|
230
|
-
docs: >-
|
|
231
|
-
ID of the audio output. Allows the Audio Output message to be tracked
|
|
232
|
-
and referenced.
|
|
233
|
-
index:
|
|
234
|
-
type: integer
|
|
235
|
-
docs: Index of the chunk of audio relative to the whole audio segment.
|
|
236
|
-
type:
|
|
237
|
-
type: literal<"audio_output">
|
|
238
|
-
docs: >-
|
|
239
|
-
The type of message sent through the socket; for an Audio Output
|
|
240
|
-
message, this must be `audio_output`.
|
|
241
89
|
source:
|
|
242
90
|
openapi: evi-asyncapi.json
|
|
243
91
|
BuiltInTool:
|
|
@@ -251,87 +99,31 @@ types:
|
|
|
251
99
|
openapi: evi-asyncapi.json
|
|
252
100
|
BuiltinToolConfig:
|
|
253
101
|
properties:
|
|
102
|
+
name:
|
|
103
|
+
type: BuiltInTool
|
|
254
104
|
fallback_content:
|
|
255
105
|
type: optional<string>
|
|
256
106
|
docs: >-
|
|
257
107
|
Optional text passed to the supplemental LLM if the tool call fails.
|
|
258
108
|
The LLM then uses this text to generate a response back to the user,
|
|
259
109
|
ensuring continuity in the conversation.
|
|
260
|
-
name:
|
|
261
|
-
type: BuiltInTool
|
|
262
|
-
source:
|
|
263
|
-
openapi: evi-asyncapi.json
|
|
264
|
-
ChatMessageToolResult:
|
|
265
|
-
discriminated: false
|
|
266
|
-
docs: Function call response from client.
|
|
267
|
-
union:
|
|
268
|
-
- type: ToolResponseMessage
|
|
269
|
-
- type: ToolErrorMessage
|
|
270
|
-
source:
|
|
271
|
-
openapi: evi-asyncapi.json
|
|
272
|
-
inline: true
|
|
273
|
-
ChatMessage:
|
|
274
|
-
properties:
|
|
275
|
-
content:
|
|
276
|
-
type: optional<string>
|
|
277
|
-
docs: Transcript of the message.
|
|
278
|
-
role:
|
|
279
|
-
type: Role
|
|
280
|
-
docs: Role of who is providing the message.
|
|
281
|
-
tool_call:
|
|
282
|
-
type: optional<ToolCallMessage>
|
|
283
|
-
docs: Function call name and arguments.
|
|
284
|
-
tool_result:
|
|
285
|
-
type: optional<ChatMessageToolResult>
|
|
286
|
-
docs: Function call response from client.
|
|
287
110
|
source:
|
|
288
111
|
openapi: evi-asyncapi.json
|
|
289
|
-
|
|
290
|
-
docs: When provided, the output is a chat metadata message.
|
|
112
|
+
Context:
|
|
291
113
|
properties:
|
|
292
|
-
|
|
293
|
-
type:
|
|
114
|
+
type:
|
|
115
|
+
type: optional<ContextType>
|
|
294
116
|
docs: >-
|
|
295
|
-
|
|
296
|
-
|
|
297
|
-
|
|
298
|
-
Used to resume a Chat when passed in the
|
|
299
|
-
[resumed_chat_group_id](/reference/empathic-voice-interface-evi/chat/chat#request.query.resumed_chat_group_id)
|
|
300
|
-
query parameter of a subsequent connection request. This allows EVI to
|
|
301
|
-
continue the conversation from where it left off within the Chat
|
|
302
|
-
Group.
|
|
117
|
+
The persistence level of the injected context. Specifies how long the
|
|
118
|
+
injected context will remain active in the session.
|
|
303
119
|
|
|
304
120
|
|
|
305
|
-
|
|
306
|
-
|
|
307
|
-
from the EVI FAQ.
|
|
308
|
-
chat_id:
|
|
309
|
-
type: string
|
|
310
|
-
docs: >-
|
|
311
|
-
ID of the Chat session. Allows the Chat session to be tracked and
|
|
312
|
-
referenced.
|
|
313
|
-
custom_session_id:
|
|
314
|
-
type: optional<string>
|
|
315
|
-
docs: >-
|
|
316
|
-
Used to manage conversational state, correlate frontend and backend
|
|
317
|
-
data, and persist conversations across EVI sessions.
|
|
318
|
-
request_id:
|
|
319
|
-
type: optional<string>
|
|
320
|
-
docs: ID of the initiating request.
|
|
321
|
-
type:
|
|
322
|
-
type: literal<"chat_metadata">
|
|
323
|
-
docs: >-
|
|
324
|
-
The type of message sent through the socket; for a Chat Metadata
|
|
325
|
-
message, this must be `chat_metadata`.
|
|
121
|
+
- **Temporary**: Context that is only applied to the following
|
|
122
|
+
assistant response.
|
|
326
123
|
|
|
327
124
|
|
|
328
|
-
|
|
329
|
-
|
|
330
|
-
for the current Chat session.
|
|
331
|
-
source:
|
|
332
|
-
openapi: evi-asyncapi.json
|
|
333
|
-
Context:
|
|
334
|
-
properties:
|
|
125
|
+
- **Persistent**: Context that is applied to all subsequent assistant
|
|
126
|
+
responses for the remainder of the Chat.
|
|
335
127
|
text:
|
|
336
128
|
type: string
|
|
337
129
|
docs: >-
|
|
@@ -340,163 +132,29 @@ types:
|
|
|
340
132
|
conversation.
|
|
341
133
|
|
|
342
134
|
|
|
343
|
-
This text will be appended to the end of
|
|
344
|
-
|
|
345
|
-
|
|
346
|
-
|
|
347
|
-
|
|
348
|
-
|
|
349
|
-
type: optional<ContextType>
|
|
350
|
-
docs: >-
|
|
351
|
-
The persistence level of the injected context. Specifies how long the
|
|
352
|
-
injected context will remain active in the session.
|
|
353
|
-
|
|
354
|
-
|
|
355
|
-
There are three possible context types:
|
|
356
|
-
|
|
357
|
-
|
|
358
|
-
- **Persistent**: The context is appended to all user messages for the
|
|
359
|
-
duration of the session.
|
|
360
|
-
|
|
361
|
-
|
|
362
|
-
- **Temporary**: The context is appended only to the next user
|
|
363
|
-
message.
|
|
364
|
-
|
|
365
|
-
- **Editable**: The original context is updated to reflect the new context.
|
|
366
|
-
|
|
367
|
-
If the type is not specified, it will default to `temporary`.
|
|
135
|
+
This text will be appended to the end of
|
|
136
|
+
[user_messages](/reference/speech-to-speech-evi/chat#receive.UserMessage.message.content)
|
|
137
|
+
based on the chosen persistence level. For example, if you want to
|
|
138
|
+
remind EVI of its role as a helpful weather assistant, the context you
|
|
139
|
+
insert will be appended to the end of user messages as `{Context: You
|
|
140
|
+
are a helpful weather assistant}`.
|
|
368
141
|
source:
|
|
369
142
|
openapi: evi-asyncapi.json
|
|
370
143
|
ContextType:
|
|
371
144
|
enum:
|
|
372
|
-
- editable
|
|
373
|
-
- persistent
|
|
374
145
|
- temporary
|
|
146
|
+
- persistent
|
|
375
147
|
source:
|
|
376
148
|
openapi: evi-asyncapi.json
|
|
377
|
-
|
|
149
|
+
Encoding:
|
|
150
|
+
type: literal<"linear16">
|
|
151
|
+
ErrorLevel:
|
|
152
|
+
type: literal<"warn">
|
|
153
|
+
PauseAssistantMessage:
|
|
154
|
+
docs: >-
|
|
155
|
+
Pause responses from EVI. Chat history is still saved and sent after
|
|
156
|
+
resuming.
|
|
378
157
|
properties:
|
|
379
|
-
Admiration: double
|
|
380
|
-
Adoration: double
|
|
381
|
-
Aesthetic Appreciation: double
|
|
382
|
-
Amusement: double
|
|
383
|
-
Anger: double
|
|
384
|
-
Anxiety: double
|
|
385
|
-
Awe: double
|
|
386
|
-
Awkwardness: double
|
|
387
|
-
Boredom: double
|
|
388
|
-
Calmness: double
|
|
389
|
-
Concentration: double
|
|
390
|
-
Confusion: double
|
|
391
|
-
Contemplation: double
|
|
392
|
-
Contempt: double
|
|
393
|
-
Contentment: double
|
|
394
|
-
Craving: double
|
|
395
|
-
Desire: double
|
|
396
|
-
Determination: double
|
|
397
|
-
Disappointment: double
|
|
398
|
-
Disgust: double
|
|
399
|
-
Distress: double
|
|
400
|
-
Doubt: double
|
|
401
|
-
Ecstasy: double
|
|
402
|
-
Embarrassment: double
|
|
403
|
-
Empathic Pain: double
|
|
404
|
-
Entrancement: double
|
|
405
|
-
Envy: double
|
|
406
|
-
Excitement: double
|
|
407
|
-
Fear: double
|
|
408
|
-
Guilt: double
|
|
409
|
-
Horror: double
|
|
410
|
-
Interest: double
|
|
411
|
-
Joy: double
|
|
412
|
-
Love: double
|
|
413
|
-
Nostalgia: double
|
|
414
|
-
Pain: double
|
|
415
|
-
Pride: double
|
|
416
|
-
Realization: double
|
|
417
|
-
Relief: double
|
|
418
|
-
Romance: double
|
|
419
|
-
Sadness: double
|
|
420
|
-
Satisfaction: double
|
|
421
|
-
Shame: double
|
|
422
|
-
Surprise (negative): double
|
|
423
|
-
Surprise (positive): double
|
|
424
|
-
Sympathy: double
|
|
425
|
-
Tiredness: double
|
|
426
|
-
Triumph: double
|
|
427
|
-
source:
|
|
428
|
-
openapi: evi-asyncapi.json
|
|
429
|
-
Encoding:
|
|
430
|
-
type: literal<"linear16">
|
|
431
|
-
WebSocketError:
|
|
432
|
-
docs: When provided, the output is an error message.
|
|
433
|
-
properties:
|
|
434
|
-
code:
|
|
435
|
-
type: string
|
|
436
|
-
docs: Error code. Identifies the type of error encountered.
|
|
437
|
-
custom_session_id:
|
|
438
|
-
type: optional<string>
|
|
439
|
-
docs: >-
|
|
440
|
-
Used to manage conversational state, correlate frontend and backend
|
|
441
|
-
data, and persist conversations across EVI sessions.
|
|
442
|
-
message:
|
|
443
|
-
type: string
|
|
444
|
-
docs: Detailed description of the error.
|
|
445
|
-
request_id:
|
|
446
|
-
type: optional<string>
|
|
447
|
-
docs: ID of the initiating request.
|
|
448
|
-
slug:
|
|
449
|
-
type: string
|
|
450
|
-
docs: >-
|
|
451
|
-
Short, human-readable identifier and description for the error. See a
|
|
452
|
-
complete list of error slugs on the [Errors
|
|
453
|
-
page](/docs/resources/errors).
|
|
454
|
-
type:
|
|
455
|
-
type: literal<"error">
|
|
456
|
-
docs: >-
|
|
457
|
-
The type of message sent through the socket; for a Web Socket Error
|
|
458
|
-
message, this must be `error`.
|
|
459
|
-
|
|
460
|
-
|
|
461
|
-
This message indicates a disruption in the WebSocket connection, such
|
|
462
|
-
as an unexpected disconnection, protocol error, or data transmission
|
|
463
|
-
issue.
|
|
464
|
-
source:
|
|
465
|
-
openapi: evi-asyncapi.json
|
|
466
|
-
ErrorLevel:
|
|
467
|
-
type: literal<"warn">
|
|
468
|
-
Inference:
|
|
469
|
-
properties:
|
|
470
|
-
prosody:
|
|
471
|
-
type: optional<ProsodyInference>
|
|
472
|
-
docs: >-
|
|
473
|
-
Prosody model inference results.
|
|
474
|
-
|
|
475
|
-
|
|
476
|
-
EVI uses the prosody model to measure 48 emotions related to speech
|
|
477
|
-
and vocal characteristics within a given expression.
|
|
478
|
-
source:
|
|
479
|
-
openapi: evi-asyncapi.json
|
|
480
|
-
MillisecondInterval:
|
|
481
|
-
properties:
|
|
482
|
-
begin:
|
|
483
|
-
type: integer
|
|
484
|
-
docs: Start time of the interval in milliseconds.
|
|
485
|
-
end:
|
|
486
|
-
type: integer
|
|
487
|
-
docs: End time of the interval in milliseconds.
|
|
488
|
-
source:
|
|
489
|
-
openapi: evi-asyncapi.json
|
|
490
|
-
PauseAssistantMessage:
|
|
491
|
-
docs: >-
|
|
492
|
-
Pause responses from EVI. Chat history is still saved and sent after
|
|
493
|
-
resuming.
|
|
494
|
-
properties:
|
|
495
|
-
custom_session_id:
|
|
496
|
-
type: optional<string>
|
|
497
|
-
docs: >-
|
|
498
|
-
Used to manage conversational state, correlate frontend and backend
|
|
499
|
-
data, and persist conversations across EVI sessions.
|
|
500
158
|
type:
|
|
501
159
|
type: literal<"pause_assistant_message">
|
|
502
160
|
docs: >-
|
|
@@ -508,26 +166,13 @@ types:
|
|
|
508
166
|
Once this message is sent, EVI will not respond until a [Resume
|
|
509
167
|
Assistant
|
|
510
168
|
message](/reference/empathic-voice-interface-evi/chat/chat#send.ResumeAssistantMessage.type)
|
|
511
|
-
is sent. When paused, EVI won
|
|
169
|
+
is sent. When paused, EVI won't respond, but transcriptions of your
|
|
512
170
|
audio inputs will still be recorded.
|
|
513
|
-
|
|
514
|
-
|
|
515
|
-
ProsodyInference:
|
|
516
|
-
properties:
|
|
517
|
-
scores:
|
|
518
|
-
type: EmotionScores
|
|
171
|
+
custom_session_id:
|
|
172
|
+
type: optional<string>
|
|
519
173
|
docs: >-
|
|
520
|
-
|
|
521
|
-
|
|
522
|
-
|
|
523
|
-
|
|
524
|
-
Scores typically range from 0 to 1, with higher values indicating a
|
|
525
|
-
stronger confidence level in the measured attribute.
|
|
526
|
-
|
|
527
|
-
|
|
528
|
-
See our guide on [interpreting expression measurement
|
|
529
|
-
results](/docs/expression-measurement/faq#how-do-i-interpret-my-results)
|
|
530
|
-
to learn more.
|
|
174
|
+
Used to manage conversational state, correlate frontend and backend
|
|
175
|
+
data, and persist conversations across EVI sessions.
|
|
531
176
|
source:
|
|
532
177
|
openapi: evi-asyncapi.json
|
|
533
178
|
ResumeAssistantMessage:
|
|
@@ -535,11 +180,6 @@ types:
|
|
|
535
180
|
Resume responses from EVI. Chat history sent while paused will now be
|
|
536
181
|
sent.
|
|
537
182
|
properties:
|
|
538
|
-
custom_session_id:
|
|
539
|
-
type: optional<string>
|
|
540
|
-
docs: >-
|
|
541
|
-
Used to manage conversational state, correlate frontend and backend
|
|
542
|
-
data, and persist conversations across EVI sessions.
|
|
543
183
|
type:
|
|
544
184
|
type: literal<"resume_assistant_message">
|
|
545
185
|
docs: >-
|
|
@@ -554,15 +194,11 @@ types:
|
|
|
554
194
|
then send a `resume_assistant_message`, EVI will respond to the second
|
|
555
195
|
question and have added the first question to its conversation
|
|
556
196
|
context.)
|
|
557
|
-
|
|
558
|
-
|
|
559
|
-
|
|
560
|
-
|
|
561
|
-
|
|
562
|
-
- system
|
|
563
|
-
- user
|
|
564
|
-
- all
|
|
565
|
-
- tool
|
|
197
|
+
custom_session_id:
|
|
198
|
+
type: optional<string>
|
|
199
|
+
docs: >-
|
|
200
|
+
Used to manage conversational state, correlate frontend and backend
|
|
201
|
+
data, and persist conversations across EVI sessions.
|
|
566
202
|
source:
|
|
567
203
|
openapi: evi-asyncapi.json
|
|
568
204
|
SessionSettingsVariablesValue:
|
|
@@ -577,48 +213,22 @@ types:
|
|
|
577
213
|
SessionSettings:
|
|
578
214
|
docs: Settings for this chat session.
|
|
579
215
|
properties:
|
|
580
|
-
|
|
581
|
-
type:
|
|
582
|
-
docs: >-
|
|
583
|
-
Configuration details for the audio input used during the session.
|
|
584
|
-
Ensures the audio is being correctly set up for processing.
|
|
585
|
-
|
|
586
|
-
|
|
587
|
-
This optional field is only required when the audio input is encoded
|
|
588
|
-
in PCM Linear 16 (16-bit, little-endian, signed PCM WAV data). For
|
|
589
|
-
detailed instructions on how to configure session settings for PCM
|
|
590
|
-
Linear 16 audio, please refer to the [Session Settings
|
|
591
|
-
guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
|
|
592
|
-
builtin_tools:
|
|
593
|
-
type: optional<list<BuiltinToolConfig>>
|
|
594
|
-
docs: >-
|
|
595
|
-
List of built-in tools to enable for the session.
|
|
596
|
-
|
|
597
|
-
|
|
598
|
-
Tools are resources used by EVI to perform various tasks, such as
|
|
599
|
-
searching the web or calling external APIs. Built-in tools, like web
|
|
600
|
-
search, are natively integrated, while user-defined tools are created
|
|
601
|
-
and invoked by the user. To learn more, see our [Tool Use
|
|
602
|
-
Guide](/docs/empathic-voice-interface-evi/features/tool-use).
|
|
603
|
-
|
|
604
|
-
|
|
605
|
-
Currently, the only built-in tool Hume provides is **Web Search**.
|
|
606
|
-
When enabled, Web Search equips EVI with the ability to search the web
|
|
607
|
-
for up-to-date information.
|
|
608
|
-
context:
|
|
609
|
-
type: optional<Context>
|
|
216
|
+
type:
|
|
217
|
+
type: literal<"session_settings">
|
|
610
218
|
docs: >-
|
|
611
|
-
|
|
612
|
-
|
|
219
|
+
The type of message sent through the socket; must be
|
|
220
|
+
`session_settings` for our server to correctly identify and process it
|
|
221
|
+
as a Session Settings message.
|
|
613
222
|
|
|
614
223
|
|
|
615
|
-
|
|
616
|
-
|
|
617
|
-
|
|
618
|
-
|
|
224
|
+
Session settings are temporary and apply only to the current Chat
|
|
225
|
+
session. These settings can be adjusted dynamically based on the
|
|
226
|
+
requirements of each session to ensure optimal performance and user
|
|
227
|
+
experience.
|
|
619
228
|
|
|
620
229
|
|
|
621
|
-
|
|
230
|
+
For more information, please refer to the [Session Settings
|
|
231
|
+
guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
|
|
622
232
|
custom_session_id:
|
|
623
233
|
type: optional<string>
|
|
624
234
|
docs: >-
|
|
@@ -637,21 +247,10 @@ types:
|
|
|
637
247
|
language
|
|
638
248
|
model](/docs/empathic-voice-interface-evi/guides/custom-language-model)
|
|
639
249
|
with EVI to learn more.
|
|
640
|
-
language_model_api_key:
|
|
641
|
-
type: optional<string>
|
|
642
|
-
docs: >-
|
|
643
|
-
Third party API key for the supplemental language model.
|
|
644
|
-
|
|
645
|
-
|
|
646
|
-
When provided, EVI will use this key instead of Hume’s API key for the
|
|
647
|
-
supplemental LLM. This allows you to bypass rate limits and utilize
|
|
648
|
-
your own API key as needed.
|
|
649
|
-
metadata:
|
|
650
|
-
type: optional<map<string, unknown>>
|
|
651
250
|
system_prompt:
|
|
652
251
|
type: optional<string>
|
|
653
252
|
docs: >-
|
|
654
|
-
Instructions used to shape EVI
|
|
253
|
+
Instructions used to shape EVI's behavior, responses, and style for
|
|
655
254
|
the session.
|
|
656
255
|
|
|
657
256
|
|
|
@@ -670,6 +269,41 @@ types:
|
|
|
670
269
|
|
|
671
270
|
For help writing a system prompt, see our [Prompting
|
|
672
271
|
Guide](/docs/empathic-voice-interface-evi/guides/prompting).
|
|
272
|
+
context:
|
|
273
|
+
type: optional<Context>
|
|
274
|
+
docs: >-
|
|
275
|
+
Field for injecting additional context into the conversation, which is
|
|
276
|
+
appended to the end of user messages for the session.
|
|
277
|
+
|
|
278
|
+
|
|
279
|
+
When included in a Session Settings message, the provided context can
|
|
280
|
+
be used to remind the LLM of its role in every user message, prevent
|
|
281
|
+
it from forgetting important details, or add new relevant information
|
|
282
|
+
to the conversation.
|
|
283
|
+
|
|
284
|
+
|
|
285
|
+
Set to `null` to clear injected context.
|
|
286
|
+
audio:
|
|
287
|
+
type: optional<AudioConfiguration>
|
|
288
|
+
docs: >-
|
|
289
|
+
Configuration details for the audio input used during the session.
|
|
290
|
+
Ensures the audio is being correctly set up for processing.
|
|
291
|
+
|
|
292
|
+
|
|
293
|
+
This optional field is only required when the audio input is encoded
|
|
294
|
+
in PCM Linear 16 (16-bit, little-endian, signed PCM WAV data). For
|
|
295
|
+
detailed instructions on how to configure session settings for PCM
|
|
296
|
+
Linear 16 audio, please refer to the [Session Settings
|
|
297
|
+
guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
|
|
298
|
+
language_model_api_key:
|
|
299
|
+
type: optional<string>
|
|
300
|
+
docs: >-
|
|
301
|
+
Third party API key for the supplemental language model.
|
|
302
|
+
|
|
303
|
+
|
|
304
|
+
When provided, EVI will use this key instead of Hume's API key for the
|
|
305
|
+
supplemental LLM. This allows you to bypass rate limits and utilize
|
|
306
|
+
your own API key as needed.
|
|
673
307
|
tools:
|
|
674
308
|
type: optional<list<Tool>>
|
|
675
309
|
docs: >-
|
|
@@ -681,22 +315,24 @@ types:
|
|
|
681
315
|
search, are natively integrated, while user-defined tools are created
|
|
682
316
|
and invoked by the user. To learn more, see our [Tool Use
|
|
683
317
|
Guide](/docs/empathic-voice-interface-evi/features/tool-use).
|
|
684
|
-
|
|
685
|
-
type:
|
|
318
|
+
builtin_tools:
|
|
319
|
+
type: optional<list<BuiltinToolConfig>>
|
|
686
320
|
docs: >-
|
|
687
|
-
|
|
688
|
-
`session_settings` for our server to correctly identify and process it
|
|
689
|
-
as a Session Settings message.
|
|
690
|
-
|
|
691
|
-
|
|
692
|
-
Session settings are temporary and apply only to the current Chat
|
|
693
|
-
session. These settings can be adjusted dynamically based on the
|
|
694
|
-
requirements of each session to ensure optimal performance and user
|
|
695
|
-
experience.
|
|
321
|
+
List of built-in tools to enable for the session.
|
|
696
322
|
|
|
697
323
|
|
|
698
|
-
|
|
699
|
-
|
|
324
|
+
Tools are resources used by EVI to perform various tasks, such as
|
|
325
|
+
searching the web or calling external APIs. Built-in tools, like web
|
|
326
|
+
search, are natively integrated, while user-defined tools are created
|
|
327
|
+
and invoked by the user. To learn more, see our [Tool Use
|
|
328
|
+
Guide](/docs/empathic-voice-interface-evi/features/tool-use).
|
|
329
|
+
|
|
330
|
+
|
|
331
|
+
Currently, the only built-in tool Hume provides is **Web Search**.
|
|
332
|
+
When enabled, Web Search equips EVI with the ability to search the web
|
|
333
|
+
for up-to-date information.
|
|
334
|
+
metadata:
|
|
335
|
+
type: optional<map<string, unknown>>
|
|
700
336
|
variables:
|
|
701
337
|
type: optional<map<string, SessionSettingsVariablesValue>>
|
|
702
338
|
docs: >-
|
|
@@ -719,17 +355,9 @@ types:
|
|
|
719
355
|
openapi: evi-asyncapi.json
|
|
720
356
|
Tool:
|
|
721
357
|
properties:
|
|
722
|
-
|
|
723
|
-
type:
|
|
724
|
-
docs:
|
|
725
|
-
An optional description of what the tool does, used by the
|
|
726
|
-
supplemental LLM to choose when and how to call the function.
|
|
727
|
-
fallback_content:
|
|
728
|
-
type: optional<string>
|
|
729
|
-
docs: >-
|
|
730
|
-
Optional text passed to the supplemental LLM if the tool call fails.
|
|
731
|
-
The LLM then uses this text to generate a response back to the user,
|
|
732
|
-
ensuring continuity in the conversation.
|
|
358
|
+
type:
|
|
359
|
+
type: ToolType
|
|
360
|
+
docs: Type of tool. Set to `function` for user-defined tools.
|
|
733
361
|
name:
|
|
734
362
|
type: string
|
|
735
363
|
docs: Name of the user-defined tool to be enabled.
|
|
@@ -739,131 +367,91 @@ types:
|
|
|
739
367
|
Parameters of the tool. Is a stringified JSON schema.
|
|
740
368
|
|
|
741
369
|
|
|
742
|
-
These parameters define the inputs needed for the tool
|
|
370
|
+
These parameters define the inputs needed for the tool's execution,
|
|
743
371
|
including the expected data type and description for each input field.
|
|
744
372
|
Structured as a JSON schema, this format ensures the tool receives
|
|
745
373
|
data in the expected format.
|
|
746
|
-
|
|
747
|
-
type:
|
|
748
|
-
docs:
|
|
374
|
+
description:
|
|
375
|
+
type: optional<string>
|
|
376
|
+
docs: >-
|
|
377
|
+
An optional description of what the tool does, used by the
|
|
378
|
+
supplemental LLM to choose when and how to call the function.
|
|
379
|
+
fallback_content:
|
|
380
|
+
type: optional<string>
|
|
381
|
+
docs: >-
|
|
382
|
+
Optional text passed to the supplemental LLM if the tool call fails.
|
|
383
|
+
The LLM then uses this text to generate a response back to the user,
|
|
384
|
+
ensuring continuity in the conversation.
|
|
749
385
|
source:
|
|
750
386
|
openapi: evi-asyncapi.json
|
|
751
|
-
|
|
752
|
-
docs: When provided, the output is a
|
|
387
|
+
ToolErrorMessage:
|
|
388
|
+
docs: When provided, the output is a function call error.
|
|
753
389
|
properties:
|
|
390
|
+
type:
|
|
391
|
+
type: literal<"tool_error">
|
|
392
|
+
docs: >-
|
|
393
|
+
The type of message sent through the socket; for a Tool Error message,
|
|
394
|
+
this must be `tool_error`.
|
|
395
|
+
|
|
396
|
+
|
|
397
|
+
Upon receiving a [Tool Call
|
|
398
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
|
|
399
|
+
and failing to invoke the function, this message is sent to notify EVI
|
|
400
|
+
of the tool's failure.
|
|
754
401
|
custom_session_id:
|
|
755
402
|
type: optional<string>
|
|
756
403
|
docs: >-
|
|
757
404
|
Used to manage conversational state, correlate frontend and backend
|
|
758
405
|
data, and persist conversations across EVI sessions.
|
|
759
|
-
name:
|
|
760
|
-
type: string
|
|
761
|
-
docs: Name of the tool called.
|
|
762
|
-
parameters:
|
|
763
|
-
type: string
|
|
764
|
-
docs: >-
|
|
765
|
-
Parameters of the tool.
|
|
766
|
-
|
|
767
|
-
|
|
768
|
-
These parameters define the inputs needed for the tool’s execution,
|
|
769
|
-
including the expected data type and description for each input field.
|
|
770
|
-
Structured as a stringified JSON schema, this format ensures the tool
|
|
771
|
-
receives data in the expected format.
|
|
772
|
-
response_required:
|
|
773
|
-
type: boolean
|
|
774
|
-
docs: >-
|
|
775
|
-
Indicates whether a response to the tool call is required from the
|
|
776
|
-
developer, either in the form of a [Tool Response
|
|
777
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolResponseMessage.type)
|
|
778
|
-
or a [Tool Error
|
|
779
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolErrorMessage.type).
|
|
780
|
-
tool_call_id:
|
|
781
|
-
type: string
|
|
782
|
-
docs: >-
|
|
783
|
-
The unique identifier for a specific tool call instance.
|
|
784
|
-
|
|
785
|
-
|
|
786
|
-
This ID is used to track the request and response of a particular tool
|
|
787
|
-
invocation, ensuring that the correct response is linked to the
|
|
788
|
-
appropriate request.
|
|
789
406
|
tool_type:
|
|
790
407
|
type: optional<ToolType>
|
|
791
408
|
docs: >-
|
|
792
409
|
Type of tool called. Either `builtin` for natively implemented tools,
|
|
793
410
|
like web search, or `function` for user-defined tools.
|
|
794
|
-
|
|
795
|
-
type:
|
|
411
|
+
tool_call_id:
|
|
412
|
+
type: string
|
|
796
413
|
docs: >-
|
|
797
|
-
The
|
|
798
|
-
this must be `tool_call`.
|
|
414
|
+
The unique identifier for a specific tool call instance.
|
|
799
415
|
|
|
800
416
|
|
|
801
|
-
This
|
|
802
|
-
|
|
803
|
-
|
|
804
|
-
|
|
805
|
-
|
|
806
|
-
docs: When provided, the output is a function call error.
|
|
807
|
-
properties:
|
|
808
|
-
code:
|
|
809
|
-
type: optional<string>
|
|
810
|
-
docs: Error code. Identifies the type of error encountered.
|
|
417
|
+
This ID is used to track the request and response of a particular tool
|
|
418
|
+
invocation, ensuring that the Tool Error message is linked to the
|
|
419
|
+
appropriate tool call request. The specified `tool_call_id` must match
|
|
420
|
+
the one received in the [Tool Call
|
|
421
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type).
|
|
811
422
|
content:
|
|
812
423
|
type: optional<string>
|
|
813
424
|
docs: >-
|
|
814
425
|
Optional text passed to the supplemental LLM in place of the tool call
|
|
815
426
|
result. The LLM then uses this text to generate a response back to the
|
|
816
427
|
user, ensuring continuity in the conversation if the tool errors.
|
|
817
|
-
custom_session_id:
|
|
818
|
-
type: optional<string>
|
|
819
|
-
docs: >-
|
|
820
|
-
Used to manage conversational state, correlate frontend and backend
|
|
821
|
-
data, and persist conversations across EVI sessions.
|
|
822
428
|
error:
|
|
823
429
|
type: string
|
|
824
430
|
docs: Error message from the tool call, not exposed to the LLM or user.
|
|
431
|
+
code:
|
|
432
|
+
type: optional<string>
|
|
433
|
+
docs: Error code. Identifies the type of error encountered.
|
|
825
434
|
level:
|
|
826
435
|
type: optional<ErrorLevel>
|
|
827
436
|
docs: >-
|
|
828
437
|
Indicates the severity of an error; for a Tool Error message, this
|
|
829
438
|
must be `warn` to signal an unexpected event.
|
|
830
|
-
tool_call_id:
|
|
831
|
-
type: string
|
|
832
|
-
docs: >-
|
|
833
|
-
The unique identifier for a specific tool call instance.
|
|
834
|
-
|
|
835
|
-
|
|
836
|
-
This ID is used to track the request and response of a particular tool
|
|
837
|
-
invocation, ensuring that the Tool Error message is linked to the
|
|
838
|
-
appropriate tool call request. The specified `tool_call_id` must match
|
|
839
|
-
the one received in the [Tool Call
|
|
840
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type).
|
|
841
|
-
tool_type:
|
|
842
|
-
type: optional<ToolType>
|
|
843
|
-
docs: >-
|
|
844
|
-
Type of tool called. Either `builtin` for natively implemented tools,
|
|
845
|
-
like web search, or `function` for user-defined tools.
|
|
846
|
-
type:
|
|
847
|
-
type: literal<"tool_error">
|
|
848
|
-
docs: >-
|
|
849
|
-
The type of message sent through the socket; for a Tool Error message,
|
|
850
|
-
this must be `tool_error`.
|
|
851
|
-
|
|
852
|
-
|
|
853
|
-
Upon receiving a [Tool Call
|
|
854
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
|
|
855
|
-
and failing to invoke the function, this message is sent to notify EVI
|
|
856
|
-
of the tool's failure.
|
|
857
439
|
source:
|
|
858
440
|
openapi: evi-asyncapi.json
|
|
859
441
|
ToolResponseMessage:
|
|
860
442
|
docs: When provided, the output is a function call response.
|
|
861
443
|
properties:
|
|
862
|
-
|
|
863
|
-
type:
|
|
444
|
+
type:
|
|
445
|
+
type: literal<"tool_response">
|
|
864
446
|
docs: >-
|
|
865
|
-
|
|
866
|
-
|
|
447
|
+
The type of message sent through the socket; for a Tool Response
|
|
448
|
+
message, this must be `tool_response`.
|
|
449
|
+
|
|
450
|
+
|
|
451
|
+
Upon receiving a [Tool Call
|
|
452
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
|
|
453
|
+
and successfully invoking the function, this message is sent to convey
|
|
454
|
+
the result of the function call back to EVI.
|
|
867
455
|
custom_session_id:
|
|
868
456
|
type: optional<string>
|
|
869
457
|
docs: >-
|
|
@@ -880,6 +468,11 @@ types:
|
|
|
880
468
|
appropriate request. The specified `tool_call_id` must match the one
|
|
881
469
|
received in the [Tool Call
|
|
882
470
|
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.tool_call_id).
|
|
471
|
+
content:
|
|
472
|
+
type: string
|
|
473
|
+
docs: >-
|
|
474
|
+
Return value of the tool call. Contains the output generated by the
|
|
475
|
+
tool to pass back to EVI.
|
|
883
476
|
tool_name:
|
|
884
477
|
type: optional<string>
|
|
885
478
|
docs: >-
|
|
@@ -895,70 +488,423 @@ types:
|
|
|
895
488
|
docs: >-
|
|
896
489
|
Type of tool called. Either `builtin` for natively implemented tools,
|
|
897
490
|
like web search, or `function` for user-defined tools.
|
|
491
|
+
source:
|
|
492
|
+
openapi: evi-asyncapi.json
|
|
493
|
+
ToolType:
|
|
494
|
+
enum:
|
|
495
|
+
- builtin
|
|
496
|
+
- function
|
|
497
|
+
source:
|
|
498
|
+
openapi: evi-asyncapi.json
|
|
499
|
+
UserInput:
|
|
500
|
+
docs: >-
|
|
501
|
+
User text to insert into the conversation. Text sent through a User Input
|
|
502
|
+
message is treated as the user's speech to EVI. EVI processes this input
|
|
503
|
+
and provides a corresponding response.
|
|
504
|
+
|
|
505
|
+
|
|
506
|
+
Expression measurement results are not available for User Input messages,
|
|
507
|
+
as the prosody model relies on audio input and cannot process text alone.
|
|
508
|
+
properties:
|
|
898
509
|
type:
|
|
899
|
-
type: literal<"
|
|
510
|
+
type: literal<"user_input">
|
|
511
|
+
docs: >-
|
|
512
|
+
The type of message sent through the socket; must be `user_input` for
|
|
513
|
+
our server to correctly identify and process it as a User Input
|
|
514
|
+
message.
|
|
515
|
+
custom_session_id:
|
|
516
|
+
type: optional<string>
|
|
517
|
+
docs: >-
|
|
518
|
+
Used to manage conversational state, correlate frontend and backend
|
|
519
|
+
data, and persist conversations across EVI sessions.
|
|
520
|
+
text:
|
|
521
|
+
type: string
|
|
522
|
+
docs: >-
|
|
523
|
+
User text to insert into the conversation. Text sent through a User
|
|
524
|
+
Input message is treated as the user's speech to EVI. EVI processes
|
|
525
|
+
this input and provides a corresponding response.
|
|
526
|
+
|
|
527
|
+
|
|
528
|
+
Expression measurement results are not available for User Input
|
|
529
|
+
messages, as the prosody model relies on audio input and cannot
|
|
530
|
+
process text alone.
|
|
531
|
+
source:
|
|
532
|
+
openapi: evi-asyncapi.json
|
|
533
|
+
AssistantEnd:
|
|
534
|
+
docs: When provided, the output is an assistant end message.
|
|
535
|
+
properties:
|
|
536
|
+
type:
|
|
537
|
+
type: literal<"assistant_end">
|
|
538
|
+
docs: >-
|
|
539
|
+
The type of message sent through the socket; for an Assistant End
|
|
540
|
+
message, this must be `assistant_end`.
|
|
541
|
+
|
|
542
|
+
|
|
543
|
+
This message indicates the conclusion of the assistant's response,
|
|
544
|
+
signaling that the assistant has finished speaking for the current
|
|
545
|
+
conversational turn.
|
|
546
|
+
custom_session_id:
|
|
547
|
+
type: optional<string>
|
|
548
|
+
docs: >-
|
|
549
|
+
Used to manage conversational state, correlate frontend and backend
|
|
550
|
+
data, and persist conversations across EVI sessions.
|
|
551
|
+
source:
|
|
552
|
+
openapi: evi-asyncapi.json
|
|
553
|
+
AssistantMessage:
|
|
554
|
+
docs: When provided, the output is an assistant message.
|
|
555
|
+
properties:
|
|
556
|
+
type:
|
|
557
|
+
type: literal<"assistant_message">
|
|
558
|
+
docs: >-
|
|
559
|
+
The type of message sent through the socket; for an Assistant Message,
|
|
560
|
+
this must be `assistant_message`.
|
|
561
|
+
|
|
562
|
+
|
|
563
|
+
This message contains both a transcript of the assistant's response
|
|
564
|
+
and the expression measurement predictions of the assistant's audio
|
|
565
|
+
output.
|
|
566
|
+
custom_session_id:
|
|
567
|
+
type: optional<string>
|
|
568
|
+
docs: >-
|
|
569
|
+
Used to manage conversational state, correlate frontend and backend
|
|
570
|
+
data, and persist conversations across EVI sessions.
|
|
571
|
+
id:
|
|
572
|
+
type: optional<string>
|
|
573
|
+
docs: >-
|
|
574
|
+
ID of the assistant message. Allows the Assistant Message to be
|
|
575
|
+
tracked and referenced.
|
|
576
|
+
message:
|
|
577
|
+
type: ChatMessage
|
|
578
|
+
docs: Transcript of the message.
|
|
579
|
+
models:
|
|
580
|
+
type: Inference
|
|
581
|
+
docs: Inference model results.
|
|
582
|
+
from_text:
|
|
583
|
+
type: boolean
|
|
584
|
+
docs: >-
|
|
585
|
+
Indicates if this message was inserted into the conversation as text
|
|
586
|
+
from an [Assistant Input
|
|
587
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#send.AssistantInput.text).
|
|
588
|
+
source:
|
|
589
|
+
openapi: evi-asyncapi.json
|
|
590
|
+
AssistantProsody:
|
|
591
|
+
docs: When provided, the output is an Assistant Prosody message.
|
|
592
|
+
properties:
|
|
593
|
+
type:
|
|
594
|
+
type: literal<"assistant_prosody">
|
|
595
|
+
docs: >-
|
|
596
|
+
The type of message sent through the socket; for an Assistant Prosody
|
|
597
|
+
message, this must be `assistant_PROSODY`.
|
|
598
|
+
|
|
599
|
+
|
|
600
|
+
This message the expression measurement predictions of the assistant's
|
|
601
|
+
audio output.
|
|
602
|
+
custom_session_id:
|
|
603
|
+
type: optional<string>
|
|
604
|
+
docs: >-
|
|
605
|
+
Used to manage conversational state, correlate frontend and backend
|
|
606
|
+
data, and persist conversations across EVI sessions.
|
|
607
|
+
models:
|
|
608
|
+
type: Inference
|
|
609
|
+
docs: Inference model results.
|
|
610
|
+
id:
|
|
611
|
+
type: optional<string>
|
|
612
|
+
docs: Unique identifier for the segment.
|
|
613
|
+
source:
|
|
614
|
+
openapi: evi-asyncapi.json
|
|
615
|
+
AudioOutput:
|
|
616
|
+
docs: >-
|
|
617
|
+
The type of message sent through the socket; for an Audio Output message,
|
|
618
|
+
this must be `audio_output`.
|
|
619
|
+
properties:
|
|
620
|
+
type:
|
|
621
|
+
type: literal<"audio_output">
|
|
622
|
+
docs: >-
|
|
623
|
+
The type of message sent through the socket; for an Audio Output
|
|
624
|
+
message, this must be `audio_output`.
|
|
625
|
+
custom_session_id:
|
|
626
|
+
type: optional<string>
|
|
627
|
+
docs: >-
|
|
628
|
+
Used to manage conversational state, correlate frontend and backend
|
|
629
|
+
data, and persist conversations across EVI sessions.
|
|
630
|
+
id:
|
|
631
|
+
type: string
|
|
632
|
+
docs: >-
|
|
633
|
+
ID of the audio output. Allows the Audio Output message to be tracked
|
|
634
|
+
and referenced.
|
|
635
|
+
index:
|
|
636
|
+
type: integer
|
|
637
|
+
docs: Index of the chunk of audio relative to the whole audio segment.
|
|
638
|
+
data:
|
|
639
|
+
type: string
|
|
640
|
+
docs: >-
|
|
641
|
+
Base64 encoded audio output. This encoded audio is transmitted to the
|
|
642
|
+
client, where it can be decoded and played back as part of the user
|
|
643
|
+
interaction.
|
|
644
|
+
source:
|
|
645
|
+
openapi: evi-asyncapi.json
|
|
646
|
+
ChatMessageToolResult:
|
|
647
|
+
discriminated: false
|
|
648
|
+
docs: Function call response from client.
|
|
649
|
+
union:
|
|
650
|
+
- type: ToolResponseMessage
|
|
651
|
+
- type: ToolErrorMessage
|
|
652
|
+
source:
|
|
653
|
+
openapi: evi-asyncapi.json
|
|
654
|
+
inline: true
|
|
655
|
+
ChatMessage:
|
|
656
|
+
properties:
|
|
657
|
+
role:
|
|
658
|
+
type: Role
|
|
659
|
+
docs: Role of who is providing the message.
|
|
660
|
+
content:
|
|
661
|
+
type: optional<string>
|
|
662
|
+
docs: Transcript of the message.
|
|
663
|
+
tool_call:
|
|
664
|
+
type: optional<ToolCallMessage>
|
|
665
|
+
docs: Function call name and arguments.
|
|
666
|
+
tool_result:
|
|
667
|
+
type: optional<ChatMessageToolResult>
|
|
668
|
+
docs: Function call response from client.
|
|
669
|
+
source:
|
|
670
|
+
openapi: evi-asyncapi.json
|
|
671
|
+
ChatMetadata:
|
|
672
|
+
docs: When provided, the output is a chat metadata message.
|
|
673
|
+
properties:
|
|
674
|
+
type:
|
|
675
|
+
type: literal<"chat_metadata">
|
|
676
|
+
docs: >-
|
|
677
|
+
The type of message sent through the socket; for a Chat Metadata
|
|
678
|
+
message, this must be `chat_metadata`.
|
|
679
|
+
|
|
680
|
+
|
|
681
|
+
The Chat Metadata message is the first message you receive after
|
|
682
|
+
establishing a connection with EVI and contains important identifiers
|
|
683
|
+
for the current Chat session.
|
|
684
|
+
custom_session_id:
|
|
685
|
+
type: optional<string>
|
|
686
|
+
docs: >-
|
|
687
|
+
Used to manage conversational state, correlate frontend and backend
|
|
688
|
+
data, and persist conversations across EVI sessions.
|
|
689
|
+
chat_group_id:
|
|
690
|
+
type: string
|
|
691
|
+
docs: >-
|
|
692
|
+
ID of the Chat Group.
|
|
693
|
+
|
|
694
|
+
|
|
695
|
+
Used to resume a Chat when passed in the
|
|
696
|
+
[resumed_chat_group_id](/reference/empathic-voice-interface-evi/chat/chat#request.query.resumed_chat_group_id)
|
|
697
|
+
query parameter of a subsequent connection request. This allows EVI to
|
|
698
|
+
continue the conversation from where it left off within the Chat
|
|
699
|
+
Group.
|
|
700
|
+
|
|
701
|
+
|
|
702
|
+
Learn more about [supporting chat
|
|
703
|
+
resumability](/docs/empathic-voice-interface-evi/faq#does-evi-support-chat-resumability)
|
|
704
|
+
from the EVI FAQ.
|
|
705
|
+
chat_id:
|
|
706
|
+
type: string
|
|
707
|
+
docs: >-
|
|
708
|
+
ID of the Chat session. Allows the Chat session to be tracked and
|
|
709
|
+
referenced.
|
|
710
|
+
request_id:
|
|
711
|
+
type: optional<string>
|
|
712
|
+
docs: ID of the initiating request.
|
|
713
|
+
source:
|
|
714
|
+
openapi: evi-asyncapi.json
|
|
715
|
+
EmotionScores:
|
|
716
|
+
properties:
|
|
717
|
+
Admiration: double
|
|
718
|
+
Adoration: double
|
|
719
|
+
Aesthetic Appreciation: double
|
|
720
|
+
Amusement: double
|
|
721
|
+
Anger: double
|
|
722
|
+
Anxiety: double
|
|
723
|
+
Awe: double
|
|
724
|
+
Awkwardness: double
|
|
725
|
+
Boredom: double
|
|
726
|
+
Calmness: double
|
|
727
|
+
Concentration: double
|
|
728
|
+
Confusion: double
|
|
729
|
+
Contemplation: double
|
|
730
|
+
Contempt: double
|
|
731
|
+
Contentment: double
|
|
732
|
+
Craving: double
|
|
733
|
+
Desire: double
|
|
734
|
+
Determination: double
|
|
735
|
+
Disappointment: double
|
|
736
|
+
Disgust: double
|
|
737
|
+
Distress: double
|
|
738
|
+
Doubt: double
|
|
739
|
+
Ecstasy: double
|
|
740
|
+
Embarrassment: double
|
|
741
|
+
Empathic Pain: double
|
|
742
|
+
Entrancement: double
|
|
743
|
+
Envy: double
|
|
744
|
+
Excitement: double
|
|
745
|
+
Fear: double
|
|
746
|
+
Guilt: double
|
|
747
|
+
Horror: double
|
|
748
|
+
Interest: double
|
|
749
|
+
Joy: double
|
|
750
|
+
Love: double
|
|
751
|
+
Nostalgia: double
|
|
752
|
+
Pain: double
|
|
753
|
+
Pride: double
|
|
754
|
+
Realization: double
|
|
755
|
+
Relief: double
|
|
756
|
+
Romance: double
|
|
757
|
+
Sadness: double
|
|
758
|
+
Satisfaction: double
|
|
759
|
+
Shame: double
|
|
760
|
+
Surprise (negative): double
|
|
761
|
+
Surprise (positive): double
|
|
762
|
+
Sympathy: double
|
|
763
|
+
Tiredness: double
|
|
764
|
+
Triumph: double
|
|
765
|
+
source:
|
|
766
|
+
openapi: evi-asyncapi.json
|
|
767
|
+
WebSocketError:
|
|
768
|
+
docs: When provided, the output is an error message.
|
|
769
|
+
properties:
|
|
770
|
+
type:
|
|
771
|
+
type: literal<"error">
|
|
772
|
+
docs: >-
|
|
773
|
+
The type of message sent through the socket; for a Web Socket Error
|
|
774
|
+
message, this must be `error`.
|
|
775
|
+
|
|
776
|
+
|
|
777
|
+
This message indicates a disruption in the WebSocket connection, such
|
|
778
|
+
as an unexpected disconnection, protocol error, or data transmission
|
|
779
|
+
issue.
|
|
780
|
+
custom_session_id:
|
|
781
|
+
type: optional<string>
|
|
782
|
+
docs: >-
|
|
783
|
+
Used to manage conversational state, correlate frontend and backend
|
|
784
|
+
data, and persist conversations across EVI sessions.
|
|
785
|
+
code:
|
|
786
|
+
type: string
|
|
787
|
+
docs: Error code. Identifies the type of error encountered.
|
|
788
|
+
slug:
|
|
789
|
+
type: string
|
|
790
|
+
docs: >-
|
|
791
|
+
Short, human-readable identifier and description for the error. See a
|
|
792
|
+
complete list of error slugs on the [Errors
|
|
793
|
+
page](/docs/resources/errors).
|
|
794
|
+
message:
|
|
795
|
+
type: string
|
|
796
|
+
docs: Detailed description of the error.
|
|
797
|
+
request_id:
|
|
798
|
+
type: optional<string>
|
|
799
|
+
docs: ID of the initiating request.
|
|
800
|
+
source:
|
|
801
|
+
openapi: evi-asyncapi.json
|
|
802
|
+
Inference:
|
|
803
|
+
properties:
|
|
804
|
+
prosody:
|
|
805
|
+
type: optional<ProsodyInference>
|
|
806
|
+
docs: >-
|
|
807
|
+
Prosody model inference results.
|
|
808
|
+
|
|
809
|
+
|
|
810
|
+
EVI uses the prosody model to measure 48 emotions related to speech
|
|
811
|
+
and vocal characteristics within a given expression.
|
|
812
|
+
source:
|
|
813
|
+
openapi: evi-asyncapi.json
|
|
814
|
+
MillisecondInterval:
|
|
815
|
+
properties:
|
|
816
|
+
begin:
|
|
817
|
+
type: integer
|
|
818
|
+
docs: Start time of the interval in milliseconds.
|
|
819
|
+
end:
|
|
820
|
+
type: integer
|
|
821
|
+
docs: End time of the interval in milliseconds.
|
|
822
|
+
source:
|
|
823
|
+
openapi: evi-asyncapi.json
|
|
824
|
+
ProsodyInference:
|
|
825
|
+
properties:
|
|
826
|
+
scores:
|
|
827
|
+
type: EmotionScores
|
|
828
|
+
docs: >-
|
|
829
|
+
The confidence scores for 48 emotions within the detected expression
|
|
830
|
+
of an audio sample.
|
|
831
|
+
|
|
832
|
+
|
|
833
|
+
Scores typically range from 0 to 1, with higher values indicating a
|
|
834
|
+
stronger confidence level in the measured attribute.
|
|
835
|
+
|
|
836
|
+
|
|
837
|
+
See our guide on [interpreting expression measurement
|
|
838
|
+
results](/docs/expression-measurement/faq#how-do-i-interpret-my-results)
|
|
839
|
+
to learn more.
|
|
840
|
+
source:
|
|
841
|
+
openapi: evi-asyncapi.json
|
|
842
|
+
Role:
|
|
843
|
+
enum:
|
|
844
|
+
- assistant
|
|
845
|
+
- system
|
|
846
|
+
- user
|
|
847
|
+
- all
|
|
848
|
+
- tool
|
|
849
|
+
source:
|
|
850
|
+
openapi: evi-asyncapi.json
|
|
851
|
+
ToolCallMessage:
|
|
852
|
+
docs: When provided, the output is a tool call.
|
|
853
|
+
properties:
|
|
854
|
+
name:
|
|
855
|
+
type: string
|
|
856
|
+
docs: Name of the tool called.
|
|
857
|
+
parameters:
|
|
858
|
+
type: string
|
|
859
|
+
docs: >-
|
|
860
|
+
Parameters of the tool.
|
|
861
|
+
|
|
862
|
+
|
|
863
|
+
These parameters define the inputs needed for the tool's execution,
|
|
864
|
+
including the expected data type and description for each input field.
|
|
865
|
+
Structured as a stringified JSON schema, this format ensures the tool
|
|
866
|
+
receives data in the expected format.
|
|
867
|
+
tool_call_id:
|
|
868
|
+
type: string
|
|
869
|
+
docs: >-
|
|
870
|
+
The unique identifier for a specific tool call instance.
|
|
871
|
+
|
|
872
|
+
|
|
873
|
+
This ID is used to track the request and response of a particular tool
|
|
874
|
+
invocation, ensuring that the correct response is linked to the
|
|
875
|
+
appropriate request.
|
|
876
|
+
type:
|
|
877
|
+
type: literal<"tool_call">
|
|
900
878
|
docs: >-
|
|
901
|
-
The type of message sent through the socket; for a Tool
|
|
902
|
-
|
|
903
|
-
|
|
904
|
-
|
|
905
|
-
Upon receiving a [Tool Call
|
|
906
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
|
|
907
|
-
and successfully invoking the function, this message is sent to convey
|
|
908
|
-
the result of the function call back to EVI.
|
|
909
|
-
source:
|
|
910
|
-
openapi: evi-asyncapi.json
|
|
911
|
-
ToolType:
|
|
912
|
-
enum:
|
|
913
|
-
- builtin
|
|
914
|
-
- function
|
|
915
|
-
source:
|
|
916
|
-
openapi: evi-asyncapi.json
|
|
917
|
-
UserInput:
|
|
918
|
-
docs: >-
|
|
919
|
-
User text to insert into the conversation. Text sent through a User Input
|
|
920
|
-
message is treated as the user's speech to EVI. EVI processes this input
|
|
921
|
-
and provides a corresponding response.
|
|
879
|
+
The type of message sent through the socket; for a Tool Call message,
|
|
880
|
+
this must be `tool_call`.
|
|
922
881
|
|
|
923
882
|
|
|
924
|
-
|
|
925
|
-
|
|
926
|
-
properties:
|
|
883
|
+
This message indicates that the supplemental LLM has detected a need
|
|
884
|
+
to invoke the specified tool.
|
|
927
885
|
custom_session_id:
|
|
928
886
|
type: optional<string>
|
|
929
887
|
docs: >-
|
|
930
888
|
Used to manage conversational state, correlate frontend and backend
|
|
931
889
|
data, and persist conversations across EVI sessions.
|
|
932
|
-
|
|
933
|
-
type:
|
|
890
|
+
tool_type:
|
|
891
|
+
type: optional<ToolType>
|
|
934
892
|
docs: >-
|
|
935
|
-
|
|
936
|
-
|
|
937
|
-
|
|
938
|
-
|
|
939
|
-
|
|
940
|
-
Expression measurement results are not available for User Input
|
|
941
|
-
messages, as the prosody model relies on audio input and cannot
|
|
942
|
-
process text alone.
|
|
943
|
-
type:
|
|
944
|
-
type: literal<"user_input">
|
|
893
|
+
Type of tool called. Either `builtin` for natively implemented tools,
|
|
894
|
+
like web search, or `function` for user-defined tools.
|
|
895
|
+
response_required:
|
|
896
|
+
type: boolean
|
|
945
897
|
docs: >-
|
|
946
|
-
|
|
947
|
-
|
|
948
|
-
message.
|
|
898
|
+
Indicates whether a response to the tool call is required from the
|
|
899
|
+
developer, either in the form of a [Tool Response
|
|
900
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolResponseMessage.type)
|
|
901
|
+
or a [Tool Error
|
|
902
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolErrorMessage.type).
|
|
949
903
|
source:
|
|
950
904
|
openapi: evi-asyncapi.json
|
|
951
905
|
UserInterruption:
|
|
952
906
|
docs: When provided, the output is an interruption.
|
|
953
907
|
properties:
|
|
954
|
-
custom_session_id:
|
|
955
|
-
type: optional<string>
|
|
956
|
-
docs: >-
|
|
957
|
-
Used to manage conversational state, correlate frontend and backend
|
|
958
|
-
data, and persist conversations across EVI sessions.
|
|
959
|
-
time:
|
|
960
|
-
type: integer
|
|
961
|
-
docs: Unix timestamp of the detected user interruption.
|
|
962
908
|
type:
|
|
963
909
|
type: literal<"user_interruption">
|
|
964
910
|
docs: >-
|
|
@@ -966,21 +912,53 @@ types:
|
|
|
966
912
|
message, this must be `user_interruption`.
|
|
967
913
|
|
|
968
914
|
|
|
969
|
-
This message indicates the user has interrupted the assistant
|
|
915
|
+
This message indicates the user has interrupted the assistant's
|
|
970
916
|
response. EVI detects the interruption in real-time and sends this
|
|
971
917
|
message to signal the interruption event. This message allows the
|
|
972
918
|
system to stop the current audio playback, clear the audio queue, and
|
|
973
919
|
prepare to handle new user input.
|
|
920
|
+
custom_session_id:
|
|
921
|
+
type: optional<string>
|
|
922
|
+
docs: >-
|
|
923
|
+
Used to manage conversational state, correlate frontend and backend
|
|
924
|
+
data, and persist conversations across EVI sessions.
|
|
925
|
+
time:
|
|
926
|
+
type: integer
|
|
927
|
+
docs: Unix timestamp of the detected user interruption.
|
|
974
928
|
source:
|
|
975
929
|
openapi: evi-asyncapi.json
|
|
976
930
|
UserMessage:
|
|
977
931
|
docs: When provided, the output is a user message.
|
|
978
932
|
properties:
|
|
933
|
+
type:
|
|
934
|
+
type: literal<"user_message">
|
|
935
|
+
docs: >-
|
|
936
|
+
The type of message sent through the socket; for a User Message, this
|
|
937
|
+
must be `user_message`.
|
|
938
|
+
|
|
939
|
+
|
|
940
|
+
This message contains both a transcript of the user's input and the
|
|
941
|
+
expression measurement predictions if the input was sent as an [Audio
|
|
942
|
+
Input
|
|
943
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#send.AudioInput.type).
|
|
944
|
+
Expression measurement predictions are not provided for a [User Input
|
|
945
|
+
message](/reference/empathic-voice-interface-evi/chat/chat#send.UserInput.type),
|
|
946
|
+
as the prosody model relies on audio input and cannot process text
|
|
947
|
+
alone.
|
|
979
948
|
custom_session_id:
|
|
980
949
|
type: optional<string>
|
|
981
950
|
docs: >-
|
|
982
951
|
Used to manage conversational state, correlate frontend and backend
|
|
983
952
|
data, and persist conversations across EVI sessions.
|
|
953
|
+
message:
|
|
954
|
+
type: ChatMessage
|
|
955
|
+
docs: Transcript of the message.
|
|
956
|
+
models:
|
|
957
|
+
type: Inference
|
|
958
|
+
docs: Inference model results.
|
|
959
|
+
time:
|
|
960
|
+
type: MillisecondInterval
|
|
961
|
+
docs: Start and End time of user message.
|
|
984
962
|
from_text:
|
|
985
963
|
type: boolean
|
|
986
964
|
docs: >-
|
|
@@ -1005,30 +983,6 @@ types:
|
|
|
1005
983
|
Interim transcripts are only sent when the
|
|
1006
984
|
[`verbose_transcription`](/reference/empathic-voice-interface-evi/chat/chat#request.query.verbose_transcription)
|
|
1007
985
|
query parameter is set to `true` in the initial handshake.
|
|
1008
|
-
message:
|
|
1009
|
-
type: ChatMessage
|
|
1010
|
-
docs: Transcript of the message.
|
|
1011
|
-
models:
|
|
1012
|
-
type: Inference
|
|
1013
|
-
docs: Inference model results.
|
|
1014
|
-
time:
|
|
1015
|
-
type: MillisecondInterval
|
|
1016
|
-
docs: Start and End time of user message.
|
|
1017
|
-
type:
|
|
1018
|
-
type: literal<"user_message">
|
|
1019
|
-
docs: >-
|
|
1020
|
-
The type of message sent through the socket; for a User Message, this
|
|
1021
|
-
must be `user_message`.
|
|
1022
|
-
|
|
1023
|
-
|
|
1024
|
-
This message contains both a transcript of the user’s input and the
|
|
1025
|
-
expression measurement predictions if the input was sent as an [Audio
|
|
1026
|
-
Input
|
|
1027
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#send.AudioInput.type).
|
|
1028
|
-
Expression measurement predictions are not provided for a [User Input
|
|
1029
|
-
message](/reference/empathic-voice-interface-evi/chat/chat#send.UserInput.type),
|
|
1030
|
-
as the prosody model relies on audio input and cannot process text
|
|
1031
|
-
alone.
|
|
1032
986
|
source:
|
|
1033
987
|
openapi: evi-asyncapi.json
|
|
1034
988
|
JsonMessage:
|
|
@@ -1112,6 +1066,12 @@ types:
|
|
|
1112
1066
|
name: Gpt4OMini
|
|
1113
1067
|
- value: gpt-4.1
|
|
1114
1068
|
name: Gpt41
|
|
1069
|
+
- value: gpt-5
|
|
1070
|
+
name: Gpt5
|
|
1071
|
+
- value: gpt-5-mini
|
|
1072
|
+
name: Gpt5Mini
|
|
1073
|
+
- value: gpt-5-nano
|
|
1074
|
+
name: Gpt5Nano
|
|
1115
1075
|
- value: gemma-7b-it
|
|
1116
1076
|
name: Gemma7BIt
|
|
1117
1077
|
- value: llama3-8b-8192
|
|
@@ -1166,12 +1126,6 @@ types:
|
|
|
1166
1126
|
- CEREBRAS
|
|
1167
1127
|
source:
|
|
1168
1128
|
openapi: evi-openapi.json
|
|
1169
|
-
SupportsToolUse:
|
|
1170
|
-
properties:
|
|
1171
|
-
model_resource: string
|
|
1172
|
-
supports_tool_use: boolean
|
|
1173
|
-
source:
|
|
1174
|
-
openapi: evi-openapi.json
|
|
1175
1129
|
ValidationErrorLocItem:
|
|
1176
1130
|
discriminated: false
|
|
1177
1131
|
union:
|
|
@@ -1188,29 +1142,20 @@ types:
|
|
|
1188
1142
|
type: string
|
|
1189
1143
|
source:
|
|
1190
1144
|
openapi: evi-openapi.json
|
|
1191
|
-
WebhookEventBase:
|
|
1192
|
-
docs: Represents the fields common to all webhook events.
|
|
1193
|
-
properties:
|
|
1194
|
-
chat_group_id:
|
|
1195
|
-
type: string
|
|
1196
|
-
docs: Unique ID of the **Chat Group** associated with the **Chat** session.
|
|
1197
|
-
chat_id:
|
|
1198
|
-
type: string
|
|
1199
|
-
docs: Unique ID of the **Chat** session.
|
|
1200
|
-
config_id:
|
|
1201
|
-
type: optional<string>
|
|
1202
|
-
docs: Unique ID of the EVI **Config** used for the session.
|
|
1203
|
-
source:
|
|
1204
|
-
openapi: evi-openapi.json
|
|
1205
|
-
WebhookEvent:
|
|
1206
|
-
discriminated: false
|
|
1207
|
-
union:
|
|
1208
|
-
- WebhookEventChatStarted
|
|
1209
|
-
- WebhookEventChatEnded
|
|
1210
|
-
source:
|
|
1211
|
-
openapi: evi-openapi.json
|
|
1212
1145
|
WebhookEventChatEnded:
|
|
1213
1146
|
properties:
|
|
1147
|
+
event_name:
|
|
1148
|
+
type: optional<literal<"chat_ended">>
|
|
1149
|
+
docs: Always `chat_ended`.
|
|
1150
|
+
end_time:
|
|
1151
|
+
type: integer
|
|
1152
|
+
docs: Unix timestamp (in milliseconds) indicating when the session ended.
|
|
1153
|
+
duration_seconds:
|
|
1154
|
+
type: integer
|
|
1155
|
+
docs: Total duration of the session in seconds.
|
|
1156
|
+
end_reason:
|
|
1157
|
+
type: WebhookEventChatStatus
|
|
1158
|
+
docs: Reason for the session's termination.
|
|
1214
1159
|
caller_number:
|
|
1215
1160
|
type: optional<string>
|
|
1216
1161
|
docs: >-
|
|
@@ -1225,18 +1170,6 @@ types:
|
|
|
1225
1170
|
language
|
|
1226
1171
|
model](/docs/empathic-voice-interface-evi/custom-language-model) in
|
|
1227
1172
|
the EVI Config.
|
|
1228
|
-
duration_seconds:
|
|
1229
|
-
type: integer
|
|
1230
|
-
docs: Total duration of the session in seconds.
|
|
1231
|
-
end_reason:
|
|
1232
|
-
type: WebhookEventChatStatus
|
|
1233
|
-
docs: Reason for the session's termination.
|
|
1234
|
-
end_time:
|
|
1235
|
-
type: integer
|
|
1236
|
-
docs: Unix timestamp (in milliseconds) indicating when the session ended.
|
|
1237
|
-
event_name:
|
|
1238
|
-
type: optional<literal<"chat_ended">>
|
|
1239
|
-
docs: Always `chat_ended`.
|
|
1240
1173
|
extends:
|
|
1241
1174
|
- WebhookEventBase
|
|
1242
1175
|
source:
|
|
@@ -1249,6 +1182,18 @@ types:
|
|
|
1249
1182
|
openapi: evi-openapi.json
|
|
1250
1183
|
WebhookEventChatStarted:
|
|
1251
1184
|
properties:
|
|
1185
|
+
event_name:
|
|
1186
|
+
type: optional<literal<"chat_started">>
|
|
1187
|
+
docs: Always `chat_started`.
|
|
1188
|
+
start_time:
|
|
1189
|
+
type: integer
|
|
1190
|
+
docs: Unix timestamp (in milliseconds) indicating when the session started.
|
|
1191
|
+
chat_start_type:
|
|
1192
|
+
type: WebhookEventChatStartType
|
|
1193
|
+
docs: >-
|
|
1194
|
+
Indicates whether the chat is the first in a new Chat Group
|
|
1195
|
+
(`new_chat_group`) or the continuation of an existing chat group
|
|
1196
|
+
(`resumed_chat_group`).
|
|
1252
1197
|
caller_number:
|
|
1253
1198
|
type: optional<string>
|
|
1254
1199
|
docs: >-
|
|
@@ -1256,12 +1201,6 @@ types:
|
|
|
1256
1201
|
This field is included only if the Chat was created via the [Twilio
|
|
1257
1202
|
phone calling](/docs/empathic-voice-interface-evi/phone-calling)
|
|
1258
1203
|
integration.
|
|
1259
|
-
chat_start_type:
|
|
1260
|
-
type: WebhookEventChatStartType
|
|
1261
|
-
docs: >-
|
|
1262
|
-
Indicates whether the chat is the first in a new Chat Group
|
|
1263
|
-
(`new_chat_group`) or the continuation of an existing chat group
|
|
1264
|
-
(`resumed_chat_group`).
|
|
1265
1204
|
custom_session_id:
|
|
1266
1205
|
type: optional<string>
|
|
1267
1206
|
docs: >-
|
|
@@ -1269,12 +1208,6 @@ types:
|
|
|
1269
1208
|
language
|
|
1270
1209
|
model](/docs/empathic-voice-interface-evi/custom-language-model) in
|
|
1271
1210
|
the EVI Config.
|
|
1272
|
-
event_name:
|
|
1273
|
-
type: optional<literal<"chat_started">>
|
|
1274
|
-
docs: Always `chat_started`.
|
|
1275
|
-
start_time:
|
|
1276
|
-
type: integer
|
|
1277
|
-
docs: Unix timestamp (in milliseconds) indicating when the session started.
|
|
1278
1211
|
extends:
|
|
1279
1212
|
- WebhookEventBase
|
|
1280
1213
|
source:
|
|
@@ -1290,6 +1223,27 @@ types:
|
|
|
1290
1223
|
- ERROR
|
|
1291
1224
|
source:
|
|
1292
1225
|
openapi: evi-openapi.json
|
|
1226
|
+
WebhookEvent:
|
|
1227
|
+
discriminated: false
|
|
1228
|
+
union:
|
|
1229
|
+
- WebhookEventChatStarted
|
|
1230
|
+
- WebhookEventChatEnded
|
|
1231
|
+
source:
|
|
1232
|
+
openapi: evi-openapi.json
|
|
1233
|
+
WebhookEventBase:
|
|
1234
|
+
docs: Represents the fields common to all webhook events.
|
|
1235
|
+
properties:
|
|
1236
|
+
chat_group_id:
|
|
1237
|
+
type: string
|
|
1238
|
+
docs: Unique ID of the **Chat Group** associated with the **Chat** session.
|
|
1239
|
+
chat_id:
|
|
1240
|
+
type: string
|
|
1241
|
+
docs: Unique ID of the **Chat** session.
|
|
1242
|
+
config_id:
|
|
1243
|
+
type: optional<string>
|
|
1244
|
+
docs: Unique ID of the EVI **Config** used for the session.
|
|
1245
|
+
source:
|
|
1246
|
+
openapi: evi-openapi.json
|
|
1293
1247
|
ErrorResponse:
|
|
1294
1248
|
properties:
|
|
1295
1249
|
error: optional<string>
|
|
@@ -1409,7 +1363,7 @@ types:
|
|
|
1409
1363
|
Tool.
|
|
1410
1364
|
|
|
1411
1365
|
|
|
1412
|
-
These parameters define the inputs needed for the Tool
|
|
1366
|
+
These parameters define the inputs needed for the Tool's execution,
|
|
1413
1367
|
including the expected data type and description for each input field.
|
|
1414
1368
|
Structured as a stringified JSON schema, this format ensures the tool
|
|
1415
1369
|
receives data in the expected format.
|
|
@@ -1456,17 +1410,15 @@ types:
|
|
|
1456
1410
|
text:
|
|
1457
1411
|
type: string
|
|
1458
1412
|
docs: >-
|
|
1459
|
-
Instructions used to shape EVI
|
|
1413
|
+
Instructions used to shape EVI's behavior, responses, and style.
|
|
1460
1414
|
|
|
1461
1415
|
|
|
1462
1416
|
You can use the Prompt to define a specific goal or role for EVI,
|
|
1463
1417
|
specifying how it should act or what it should focus on during the
|
|
1464
1418
|
conversation. For example, EVI can be instructed to act as a customer
|
|
1465
1419
|
support representative, a fitness coach, or a travel advisor, each
|
|
1466
|
-
with its own set of behaviors and response styles.
|
|
1467
|
-
|
|
1468
|
-
|
|
1469
|
-
For help writing a system prompt, see our [Prompting
|
|
1420
|
+
with its own set of behaviors and response styles. For help writing a
|
|
1421
|
+
system prompt, see our [Prompting
|
|
1470
1422
|
Guide](/docs/speech-to-speech-evi/guides/prompting).
|
|
1471
1423
|
version:
|
|
1472
1424
|
type: integer
|
|
@@ -1483,14 +1435,14 @@ types:
|
|
|
1483
1435
|
Version numbers are integer values representing different iterations
|
|
1484
1436
|
of the Prompt. Each update to the Prompt increments its version
|
|
1485
1437
|
number.
|
|
1438
|
+
version_description:
|
|
1439
|
+
type: optional<string>
|
|
1440
|
+
docs: An optional description of the Prompt version.
|
|
1486
1441
|
version_type:
|
|
1487
1442
|
type: ReturnPromptVersionType
|
|
1488
1443
|
docs: >-
|
|
1489
1444
|
Versioning method for a Prompt. Either `FIXED` for using a fixed
|
|
1490
1445
|
version number or `LATEST` for auto-updating to the latest version.
|
|
1491
|
-
version_description:
|
|
1492
|
-
type: optional<string>
|
|
1493
|
-
docs: An optional description of the Prompt version.
|
|
1494
1446
|
created_on:
|
|
1495
1447
|
type: long
|
|
1496
1448
|
docs: >-
|
|
@@ -1556,6 +1508,12 @@ types:
|
|
|
1556
1508
|
Version numbers are integer values representing different iterations
|
|
1557
1509
|
of the Config. Each update to the Config increments its version
|
|
1558
1510
|
number.
|
|
1511
|
+
tools:
|
|
1512
|
+
type: optional<list<optional<ReturnUserDefinedTool>>>
|
|
1513
|
+
docs: List of user-defined tools associated with this Config.
|
|
1514
|
+
version_description:
|
|
1515
|
+
type: optional<string>
|
|
1516
|
+
docs: An optional description of the Config version.
|
|
1559
1517
|
language_model:
|
|
1560
1518
|
type: optional<ReturnLanguageModel>
|
|
1561
1519
|
docs: >-
|
|
@@ -1581,6 +1539,7 @@ types:
|
|
|
1581
1539
|
3](/docs/speech-to-speech-evi/configuration/evi-version#migrating-to-evi-3)
|
|
1582
1540
|
before then.
|
|
1583
1541
|
timeouts: optional<ReturnTimeoutSpecs>
|
|
1542
|
+
nudges: optional<ReturnNudgeSpec>
|
|
1584
1543
|
event_messages: optional<ReturnEventMessageSpecs>
|
|
1585
1544
|
ellm_model:
|
|
1586
1545
|
type: optional<ReturnEllmModel>
|
|
@@ -1592,12 +1551,11 @@ types:
|
|
|
1592
1551
|
model that takes into account both expression measures and language.
|
|
1593
1552
|
The eLLM generates short, empathic language responses and guides
|
|
1594
1553
|
text-to-speech (TTS) prosody.
|
|
1554
|
+
voice: optional<unknown>
|
|
1555
|
+
prompt: optional<ReturnPrompt>
|
|
1595
1556
|
webhooks:
|
|
1596
1557
|
type: optional<list<optional<ReturnWebhookSpec>>>
|
|
1597
1558
|
docs: Map of webhooks associated with this config.
|
|
1598
|
-
version_description:
|
|
1599
|
-
type: optional<string>
|
|
1600
|
-
docs: An optional description of the Config version.
|
|
1601
1559
|
created_on:
|
|
1602
1560
|
type: optional<long>
|
|
1603
1561
|
docs: >-
|
|
@@ -1608,12 +1566,6 @@ types:
|
|
|
1608
1566
|
docs: >-
|
|
1609
1567
|
Time at which the Config was last modified. Measured in seconds since
|
|
1610
1568
|
the Unix epoch.
|
|
1611
|
-
nudges: optional<ReturnNudgeSpec>
|
|
1612
|
-
voice: optional<unknown>
|
|
1613
|
-
prompt: optional<ReturnPrompt>
|
|
1614
|
-
tools:
|
|
1615
|
-
type: optional<list<optional<ReturnUserDefinedTool>>>
|
|
1616
|
-
docs: List of user-defined tools associated with this Config.
|
|
1617
1569
|
source:
|
|
1618
1570
|
openapi: evi-openapi.json
|
|
1619
1571
|
ReturnPagedChatsPaginationDirection:
|
|
@@ -2188,7 +2140,7 @@ types:
|
|
|
2188
2140
|
The model temperature, with values between 0 to 1 (inclusive).
|
|
2189
2141
|
|
|
2190
2142
|
|
|
2191
|
-
Controls the randomness of the LLM
|
|
2143
|
+
Controls the randomness of the LLM's output, with values closer to 0
|
|
2192
2144
|
yielding focused, deterministic responses and values closer to 1
|
|
2193
2145
|
producing more creative, diverse responses.
|
|
2194
2146
|
source:
|
|
@@ -2339,6 +2291,11 @@ types:
|
|
|
2339
2291
|
Accepts a minimum value of 30 seconds and a maximum value of 1,800
|
|
2340
2292
|
seconds.
|
|
2341
2293
|
properties:
|
|
2294
|
+
duration_secs:
|
|
2295
|
+
type: optional<integer>
|
|
2296
|
+
docs: >-
|
|
2297
|
+
Duration in seconds for the timeout (e.g. 600 seconds represents 10
|
|
2298
|
+
minutes).
|
|
2342
2299
|
enabled:
|
|
2343
2300
|
type: boolean
|
|
2344
2301
|
docs: >-
|
|
@@ -2349,11 +2306,6 @@ types:
|
|
|
2349
2306
|
user inactivity being reached. However, the conversation will
|
|
2350
2307
|
eventually disconnect after 1,800 seconds (30 minutes), which is the
|
|
2351
2308
|
maximum WebSocket duration limit for EVI.
|
|
2352
|
-
duration_secs:
|
|
2353
|
-
type: optional<integer>
|
|
2354
|
-
docs: >-
|
|
2355
|
-
Duration in seconds for the timeout (e.g. 600 seconds represents 10
|
|
2356
|
-
minutes).
|
|
2357
2309
|
source:
|
|
2358
2310
|
openapi: evi-openapi.json
|
|
2359
2311
|
inline: true
|
|
@@ -2367,6 +2319,11 @@ types:
|
|
|
2367
2319
|
Accepts a minimum value of 30 seconds and a maximum value of 1,800
|
|
2368
2320
|
seconds.
|
|
2369
2321
|
properties:
|
|
2322
|
+
duration_secs:
|
|
2323
|
+
type: optional<integer>
|
|
2324
|
+
docs: >-
|
|
2325
|
+
Duration in seconds for the timeout (e.g. 600 seconds represents 10
|
|
2326
|
+
minutes).
|
|
2370
2327
|
enabled:
|
|
2371
2328
|
type: boolean
|
|
2372
2329
|
docs: >-
|
|
@@ -2377,11 +2334,6 @@ types:
|
|
|
2377
2334
|
duration being reached. However, the conversation will eventually
|
|
2378
2335
|
disconnect after 1,800 seconds (30 minutes), which is the maximum
|
|
2379
2336
|
WebSocket duration limit for EVI.
|
|
2380
|
-
duration_secs:
|
|
2381
|
-
type: optional<integer>
|
|
2382
|
-
docs: >-
|
|
2383
|
-
Duration in seconds for the timeout (e.g. 600 seconds represents 10
|
|
2384
|
-
minutes).
|
|
2385
2337
|
source:
|
|
2386
2338
|
openapi: evi-openapi.json
|
|
2387
2339
|
inline: true
|
|
@@ -2460,7 +2412,7 @@ types:
|
|
|
2460
2412
|
The model temperature, with values between 0 to 1 (inclusive).
|
|
2461
2413
|
|
|
2462
2414
|
|
|
2463
|
-
Controls the randomness of the LLM
|
|
2415
|
+
Controls the randomness of the LLM's output, with values closer to 0
|
|
2464
2416
|
yielding focused, deterministic responses and values closer to 1
|
|
2465
2417
|
producing more creative, diverse responses.
|
|
2466
2418
|
source:
|
|
@@ -2980,6 +2932,7 @@ types:
|
|
|
2980
2932
|
enum:
|
|
2981
2933
|
- HUME_AI
|
|
2982
2934
|
- CUSTOM_VOICE
|
|
2935
|
+
- OCTAVE_COMBINED
|
|
2983
2936
|
source:
|
|
2984
2937
|
openapi: evi-openapi.json
|
|
2985
2938
|
VoiceId:
|