hume 0.12.2 → 0.13.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.mock/definition/empathic-voice/__package__.yml +688 -735
- package/.mock/definition/empathic-voice/chat.yml +29 -23
- package/.mock/definition/empathic-voice/chatWebhooks.yml +8 -12
- package/.mock/definition/empathic-voice/prompts.yml +2 -2
- package/.mock/definition/empathic-voice/tools.yml +2 -2
- package/.mock/definition/tts/__package__.yml +70 -87
- package/.mock/fern.config.json +1 -1
- package/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/index.js +0 -15
- package/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/api/resources/empathicVoice/errors/index.js +0 -1
- package/api/resources/empathicVoice/resources/chat/client/Client.d.ts +2 -0
- package/api/resources/empathicVoice/resources/prompts/client/requests/PostedPrompt.d.ts +1 -1
- package/api/resources/empathicVoice/resources/prompts/client/requests/PostedPromptVersion.d.ts +1 -1
- package/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedTool.d.ts +1 -1
- package/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedToolVersion.d.ts +1 -1
- package/api/resources/empathicVoice/types/AssistantEnd.d.ts +3 -3
- package/api/resources/empathicVoice/types/AssistantInput.d.ts +3 -3
- package/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/api/resources/empathicVoice/types/AudioInput.d.ts +7 -7
- package/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +3 -3
- package/api/resources/empathicVoice/types/PostedLanguageModel.d.ts +1 -1
- package/api/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +2 -2
- package/api/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +2 -2
- package/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/api/resources/empathicVoice/types/ReturnLanguageModel.d.ts +1 -1
- package/api/resources/empathicVoice/types/ReturnPrompt.d.ts +4 -6
- package/api/resources/empathicVoice/types/ReturnUserDefinedTool.d.ts +1 -1
- package/api/resources/empathicVoice/types/SessionSettings.d.ts +30 -30
- package/api/resources/empathicVoice/types/Tool.d.ts +7 -7
- package/api/resources/empathicVoice/types/ToolCallMessage.d.ts +7 -7
- package/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/UserInput.d.ts +3 -3
- package/api/resources/empathicVoice/types/UserInterruption.d.ts +5 -5
- package/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/api/resources/empathicVoice/types/index.js +16 -17
- package/api/resources/tts/client/Client.d.ts +21 -23
- package/api/resources/tts/client/Client.js +50 -58
- package/api/resources/tts/client/index.d.ts +0 -1
- package/api/resources/tts/client/index.js +0 -15
- package/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/index.js +0 -15
- package/dist/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/errors/index.js +0 -1
- package/dist/api/resources/empathicVoice/resources/chat/client/Client.d.ts +2 -0
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PostedPrompt.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PostedPromptVersion.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedTool.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedToolVersion.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/AssistantEnd.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/AssistantInput.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/AudioInput.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/dist/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/dist/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/dist/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/dist/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/dist/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/PostedLanguageModel.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ReturnLanguageModel.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ReturnPrompt.d.ts +4 -6
- package/dist/api/resources/empathicVoice/types/ReturnUserDefinedTool.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/SessionSettings.d.ts +30 -30
- package/dist/api/resources/empathicVoice/types/Tool.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ToolCallMessage.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/dist/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/UserInput.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/UserInterruption.d.ts +5 -5
- package/dist/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/dist/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/dist/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/dist/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/dist/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/api/resources/empathicVoice/types/index.js +16 -17
- package/dist/api/resources/tts/client/Client.d.ts +21 -23
- package/dist/api/resources/tts/client/Client.js +50 -58
- package/dist/api/resources/tts/client/index.d.ts +0 -1
- package/dist/api/resources/tts/client/index.js +0 -15
- package/dist/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/dist/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/dist/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/dist/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/index.js +0 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/serialization/resources/empathicVoice/types/index.js +16 -17
- package/dist/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/dist/serialization/resources/tts/types/PostedTts.js +3 -3
- package/dist/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/dist/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/dist/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/dist/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/dist/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/dist/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/dist/serialization/resources/tts/types/Snippet.js +3 -3
- package/dist/version.d.ts +1 -1
- package/dist/version.js +1 -1
- package/package.json +1 -1
- package/reference.md +24 -78
- package/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/index.js +0 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.js +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/serialization/resources/empathicVoice/types/index.js +16 -17
- package/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/serialization/resources/tts/types/PostedTts.js +3 -3
- package/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/serialization/resources/tts/types/Snippet.js +3 -3
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/api/resources/tts/client/requests/index.d.ts +0 -1
- package/api/resources/tts/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/dist/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/dist/api/resources/tts/client/requests/index.d.ts +0 -1
- package/dist/api/resources/tts/client/requests/index.js +0 -2
- package/dist/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/index.js +0 -17
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/dist/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
- package/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/index.js +0 -17
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
|
@@ -5,22 +5,22 @@
|
|
|
5
5
|
* When provided, the input is audio.
|
|
6
6
|
*/
|
|
7
7
|
export interface AudioInput {
|
|
8
|
+
/**
|
|
9
|
+
* The type of message sent through the socket; must be `audio_input` for our server to correctly identify and process it as an Audio Input message.
|
|
10
|
+
*
|
|
11
|
+
* This message is used for sending audio input data to EVI for processing and expression measurement. Audio data should be sent as a continuous stream, encoded in Base64.
|
|
12
|
+
*/
|
|
13
|
+
type: "audio_input";
|
|
8
14
|
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
9
15
|
customSessionId?: string;
|
|
10
16
|
/**
|
|
11
17
|
* Base64 encoded audio input to insert into the conversation.
|
|
12
18
|
*
|
|
13
|
-
* The content of an Audio Input message is treated as the user
|
|
19
|
+
* The content of an Audio Input message is treated as the user's speech to EVI and must be streamed continuously. Pre-recorded audio files are not supported.
|
|
14
20
|
*
|
|
15
21
|
* For optimal transcription quality, the audio data should be transmitted in small chunks.
|
|
16
22
|
*
|
|
17
23
|
* Hume recommends streaming audio with a buffer window of 20 milliseconds (ms), or 100 milliseconds (ms) for web applications.
|
|
18
24
|
*/
|
|
19
25
|
data: string;
|
|
20
|
-
/**
|
|
21
|
-
* The type of message sent through the socket; must be `audio_input` for our server to correctly identify and process it as an Audio Input message.
|
|
22
|
-
*
|
|
23
|
-
* This message is used for sending audio input data to EVI for processing and expression measurement. Audio data should be sent as a continuous stream, encoded in Base64.
|
|
24
|
-
*/
|
|
25
|
-
type: "audio_input";
|
|
26
26
|
}
|
|
@@ -5,14 +5,14 @@
|
|
|
5
5
|
* The type of message sent through the socket; for an Audio Output message, this must be `audio_output`.
|
|
6
6
|
*/
|
|
7
7
|
export interface AudioOutput {
|
|
8
|
+
/** The type of message sent through the socket; for an Audio Output message, this must be `audio_output`. */
|
|
9
|
+
type: "audio_output";
|
|
8
10
|
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
9
11
|
customSessionId?: string;
|
|
10
|
-
/** Base64 encoded audio output. This encoded audio is transmitted to the client, where it can be decoded and played back as part of the user interaction. */
|
|
11
|
-
data: string;
|
|
12
12
|
/** ID of the audio output. Allows the Audio Output message to be tracked and referenced. */
|
|
13
13
|
id: string;
|
|
14
14
|
/** Index of the chunk of audio relative to the whole audio segment. */
|
|
15
15
|
index: number;
|
|
16
|
-
/**
|
|
17
|
-
|
|
16
|
+
/** Base64 encoded audio output. This encoded audio is transmitted to the client, where it can be decoded and played back as part of the user interaction. */
|
|
17
|
+
data: string;
|
|
18
18
|
}
|
|
@@ -3,7 +3,7 @@
|
|
|
3
3
|
*/
|
|
4
4
|
import * as Hume from "../../../index";
|
|
5
5
|
export interface BuiltinToolConfig {
|
|
6
|
+
name: Hume.empathicVoice.BuiltInTool;
|
|
6
7
|
/** Optional text passed to the supplemental LLM if the tool call fails. The LLM then uses this text to generate a response back to the user, ensuring continuity in the conversation. */
|
|
7
8
|
fallbackContent?: string;
|
|
8
|
-
name: Hume.empathicVoice.BuiltInTool;
|
|
9
9
|
}
|
|
@@ -3,10 +3,10 @@
|
|
|
3
3
|
*/
|
|
4
4
|
import * as Hume from "../../../index";
|
|
5
5
|
export interface ChatMessage {
|
|
6
|
-
/** Transcript of the message. */
|
|
7
|
-
content?: string;
|
|
8
6
|
/** Role of who is providing the message. */
|
|
9
7
|
role: Hume.empathicVoice.Role;
|
|
8
|
+
/** Transcript of the message. */
|
|
9
|
+
content?: string;
|
|
10
10
|
/** Function call name and arguments. */
|
|
11
11
|
toolCall?: Hume.empathicVoice.ToolCallMessage;
|
|
12
12
|
/** Function call response from client. */
|
|
@@ -5,6 +5,14 @@
|
|
|
5
5
|
* When provided, the output is a chat metadata message.
|
|
6
6
|
*/
|
|
7
7
|
export interface ChatMetadata {
|
|
8
|
+
/**
|
|
9
|
+
* The type of message sent through the socket; for a Chat Metadata message, this must be `chat_metadata`.
|
|
10
|
+
*
|
|
11
|
+
* The Chat Metadata message is the first message you receive after establishing a connection with EVI and contains important identifiers for the current Chat session.
|
|
12
|
+
*/
|
|
13
|
+
type: "chat_metadata";
|
|
14
|
+
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
15
|
+
customSessionId?: string;
|
|
8
16
|
/**
|
|
9
17
|
* ID of the Chat Group.
|
|
10
18
|
*
|
|
@@ -15,14 +23,6 @@ export interface ChatMetadata {
|
|
|
15
23
|
chatGroupId: string;
|
|
16
24
|
/** ID of the Chat session. Allows the Chat session to be tracked and referenced. */
|
|
17
25
|
chatId: string;
|
|
18
|
-
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
19
|
-
customSessionId?: string;
|
|
20
26
|
/** ID of the initiating request. */
|
|
21
27
|
requestId?: string;
|
|
22
|
-
/**
|
|
23
|
-
* The type of message sent through the socket; for a Chat Metadata message, this must be `chat_metadata`.
|
|
24
|
-
*
|
|
25
|
-
* The Chat Metadata message is the first message you receive after establishing a connection with EVI and contains important identifiers for the current Chat session.
|
|
26
|
-
*/
|
|
27
|
-
type: "chat_metadata";
|
|
28
28
|
}
|
|
@@ -3,24 +3,18 @@
|
|
|
3
3
|
*/
|
|
4
4
|
import * as Hume from "../../../index";
|
|
5
5
|
export interface Context {
|
|
6
|
-
/**
|
|
7
|
-
* The context to be injected into the conversation. Helps inform the LLM's response by providing relevant information about the ongoing conversation.
|
|
8
|
-
*
|
|
9
|
-
* This text will be appended to the end of user messages based on the chosen persistence level. For example, if you want to remind EVI of its role as a helpful weather assistant, the context you insert will be appended to the end of user messages as `{Context: You are a helpful weather assistant}`.
|
|
10
|
-
*/
|
|
11
|
-
text: string;
|
|
12
6
|
/**
|
|
13
7
|
* The persistence level of the injected context. Specifies how long the injected context will remain active in the session.
|
|
14
8
|
*
|
|
15
|
-
*
|
|
16
|
-
*
|
|
17
|
-
* - **Persistent**: The context is appended to all user messages for the duration of the session.
|
|
18
|
-
*
|
|
19
|
-
* - **Temporary**: The context is appended only to the next user message.
|
|
9
|
+
* - **Temporary**: Context that is only applied to the following assistant response.
|
|
20
10
|
*
|
|
21
|
-
*
|
|
22
|
-
*
|
|
23
|
-
* If the type is not specified, it will default to `temporary`.
|
|
11
|
+
* - **Persistent**: Context that is applied to all subsequent assistant responses for the remainder of the Chat.
|
|
24
12
|
*/
|
|
25
13
|
type?: Hume.empathicVoice.ContextType;
|
|
14
|
+
/**
|
|
15
|
+
* The context to be injected into the conversation. Helps inform the LLM's response by providing relevant information about the ongoing conversation.
|
|
16
|
+
*
|
|
17
|
+
* This text will be appended to the end of [user_messages](/reference/speech-to-speech-evi/chat#receive.UserMessage.message.content) based on the chosen persistence level. For example, if you want to remind EVI of its role as a helpful weather assistant, the context you insert will be appended to the end of user messages as `{Context: You are a helpful weather assistant}`.
|
|
18
|
+
*/
|
|
19
|
+
text: string;
|
|
26
20
|
}
|
|
@@ -1,9 +1,8 @@
|
|
|
1
1
|
/**
|
|
2
2
|
* This file was auto-generated by Fern from our API Definition.
|
|
3
3
|
*/
|
|
4
|
-
export type ContextType = "
|
|
4
|
+
export type ContextType = "temporary" | "persistent";
|
|
5
5
|
export declare const ContextType: {
|
|
6
|
-
readonly Editable: "editable";
|
|
7
|
-
readonly Persistent: "persistent";
|
|
8
6
|
readonly Temporary: "temporary";
|
|
7
|
+
readonly Persistent: "persistent";
|
|
9
8
|
};
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
/**
|
|
2
2
|
* This file was auto-generated by Fern from our API Definition.
|
|
3
3
|
*/
|
|
4
|
-
export type LanguageModelType = "claude-3-7-sonnet-latest" | "claude-3-5-sonnet-latest" | "claude-3-5-haiku-latest" | "claude-3-5-sonnet-20240620" | "claude-3-opus-20240229" | "claude-3-sonnet-20240229" | "claude-3-haiku-20240307" | "claude-sonnet-4-20250514" | "us.anthropic.claude-3-5-haiku-20241022-v1:0" | "us.anthropic.claude-3-5-sonnet-20240620-v1:0" | "us.anthropic.claude-3-haiku-20240307-v1:0" | "gpt-oss-120b" | "qwen-3-235b-a22b" | "qwen-3-235b-a22b-instruct-2507" | "qwen-3-235b-a22b-thinking-2507" | "gemini-1.5-pro" | "gemini-1.5-flash" | "gemini-1.5-pro-002" | "gemini-1.5-flash-002" | "gemini-2.0-flash" | "gemini-2.5-flash" | "gemini-2.5-flash-preview-04-17" | "gpt-4-turbo" | "gpt-4-turbo-preview" | "gpt-3.5-turbo-0125" | "gpt-3.5-turbo" | "gpt-4o" | "gpt-4o-mini" | "gpt-4.1" | "gemma-7b-it" | "llama3-8b-8192" | "llama3-70b-8192" | "llama-3.1-70b-versatile" | "llama-3.3-70b-versatile" | "llama-3.1-8b-instant" | "moonshotai/kimi-k2-instruct" | "accounts/fireworks/models/mixtral-8x7b-instruct" | "accounts/fireworks/models/llama-v3p1-405b-instruct" | "accounts/fireworks/models/llama-v3p1-70b-instruct" | "accounts/fireworks/models/llama-v3p1-8b-instruct" | "sonar" | "sonar-pro" | "sambanova" | "DeepSeek-R1-Distill-Llama-70B" | "Llama-4-Maverick-17B-128E-Instruct" | "Qwen3-32B" | "ellm" | "custom-language-model" | "hume-evi-3-web-search";
|
|
4
|
+
export type LanguageModelType = "claude-3-7-sonnet-latest" | "claude-3-5-sonnet-latest" | "claude-3-5-haiku-latest" | "claude-3-5-sonnet-20240620" | "claude-3-opus-20240229" | "claude-3-sonnet-20240229" | "claude-3-haiku-20240307" | "claude-sonnet-4-20250514" | "us.anthropic.claude-3-5-haiku-20241022-v1:0" | "us.anthropic.claude-3-5-sonnet-20240620-v1:0" | "us.anthropic.claude-3-haiku-20240307-v1:0" | "gpt-oss-120b" | "qwen-3-235b-a22b" | "qwen-3-235b-a22b-instruct-2507" | "qwen-3-235b-a22b-thinking-2507" | "gemini-1.5-pro" | "gemini-1.5-flash" | "gemini-1.5-pro-002" | "gemini-1.5-flash-002" | "gemini-2.0-flash" | "gemini-2.5-flash" | "gemini-2.5-flash-preview-04-17" | "gpt-4-turbo" | "gpt-4-turbo-preview" | "gpt-3.5-turbo-0125" | "gpt-3.5-turbo" | "gpt-4o" | "gpt-4o-mini" | "gpt-4.1" | "gpt-5" | "gpt-5-mini" | "gpt-5-nano" | "gemma-7b-it" | "llama3-8b-8192" | "llama3-70b-8192" | "llama-3.1-70b-versatile" | "llama-3.3-70b-versatile" | "llama-3.1-8b-instant" | "moonshotai/kimi-k2-instruct" | "accounts/fireworks/models/mixtral-8x7b-instruct" | "accounts/fireworks/models/llama-v3p1-405b-instruct" | "accounts/fireworks/models/llama-v3p1-70b-instruct" | "accounts/fireworks/models/llama-v3p1-8b-instruct" | "sonar" | "sonar-pro" | "sambanova" | "DeepSeek-R1-Distill-Llama-70B" | "Llama-4-Maverick-17B-128E-Instruct" | "Qwen3-32B" | "ellm" | "custom-language-model" | "hume-evi-3-web-search";
|
|
5
5
|
export declare const LanguageModelType: {
|
|
6
6
|
readonly Claude37SonnetLatest: "claude-3-7-sonnet-latest";
|
|
7
7
|
readonly Claude35SonnetLatest: "claude-3-5-sonnet-latest";
|
|
@@ -32,6 +32,9 @@ export declare const LanguageModelType: {
|
|
|
32
32
|
readonly Gpt4O: "gpt-4o";
|
|
33
33
|
readonly Gpt4OMini: "gpt-4o-mini";
|
|
34
34
|
readonly Gpt41: "gpt-4.1";
|
|
35
|
+
readonly Gpt5: "gpt-5";
|
|
36
|
+
readonly Gpt5Mini: "gpt-5-mini";
|
|
37
|
+
readonly Gpt5Nano: "gpt-5-nano";
|
|
35
38
|
readonly Gemma7BIt: "gemma-7b-it";
|
|
36
39
|
readonly Llama38B8192: "llama3-8b-8192";
|
|
37
40
|
readonly Llama370B8192: "llama3-70b-8192";
|
|
@@ -34,6 +34,9 @@ exports.LanguageModelType = {
|
|
|
34
34
|
Gpt4O: "gpt-4o",
|
|
35
35
|
Gpt4OMini: "gpt-4o-mini",
|
|
36
36
|
Gpt41: "gpt-4.1",
|
|
37
|
+
Gpt5: "gpt-5",
|
|
38
|
+
Gpt5Mini: "gpt-5-mini",
|
|
39
|
+
Gpt5Nano: "gpt-5-nano",
|
|
37
40
|
Gemma7BIt: "gemma-7b-it",
|
|
38
41
|
Llama38B8192: "llama3-8b-8192",
|
|
39
42
|
Llama370B8192: "llama3-70b-8192",
|
|
@@ -5,12 +5,12 @@
|
|
|
5
5
|
* Pause responses from EVI. Chat history is still saved and sent after resuming.
|
|
6
6
|
*/
|
|
7
7
|
export interface PauseAssistantMessage {
|
|
8
|
-
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
9
|
-
customSessionId?: string;
|
|
10
8
|
/**
|
|
11
9
|
* The type of message sent through the socket; must be `pause_assistant_message` for our server to correctly identify and process it as a Pause Assistant message.
|
|
12
10
|
*
|
|
13
|
-
* Once this message is sent, EVI will not respond until a [Resume Assistant message](/reference/empathic-voice-interface-evi/chat/chat#send.ResumeAssistantMessage.type) is sent. When paused, EVI won
|
|
11
|
+
* Once this message is sent, EVI will not respond until a [Resume Assistant message](/reference/empathic-voice-interface-evi/chat/chat#send.ResumeAssistantMessage.type) is sent. When paused, EVI won't respond, but transcriptions of your audio inputs will still be recorded.
|
|
14
12
|
*/
|
|
15
13
|
type: "pause_assistant_message";
|
|
14
|
+
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
15
|
+
customSessionId?: string;
|
|
16
16
|
}
|
|
@@ -13,7 +13,7 @@ export interface PostedLanguageModel {
|
|
|
13
13
|
/**
|
|
14
14
|
* The model temperature, with values between 0 to 1 (inclusive).
|
|
15
15
|
*
|
|
16
|
-
* Controls the randomness of the LLM
|
|
16
|
+
* Controls the randomness of the LLM's output, with values closer to 0 yielding focused, deterministic responses and values closer to 1 producing more creative, diverse responses.
|
|
17
17
|
*/
|
|
18
18
|
temperature?: number;
|
|
19
19
|
}
|
|
@@ -7,12 +7,12 @@
|
|
|
7
7
|
* Accepts a minimum value of 30 seconds and a maximum value of 1,800 seconds.
|
|
8
8
|
*/
|
|
9
9
|
export interface PostedTimeoutSpecsInactivity {
|
|
10
|
+
/** Duration in seconds for the timeout (e.g. 600 seconds represents 10 minutes). */
|
|
11
|
+
durationSecs?: number;
|
|
10
12
|
/**
|
|
11
13
|
* Boolean indicating if this timeout is enabled.
|
|
12
14
|
*
|
|
13
15
|
* If set to false, EVI will not timeout due to a specified duration of user inactivity being reached. However, the conversation will eventually disconnect after 1,800 seconds (30 minutes), which is the maximum WebSocket duration limit for EVI.
|
|
14
16
|
*/
|
|
15
17
|
enabled: boolean;
|
|
16
|
-
/** Duration in seconds for the timeout (e.g. 600 seconds represents 10 minutes). */
|
|
17
|
-
durationSecs?: number;
|
|
18
18
|
}
|
|
@@ -7,12 +7,12 @@
|
|
|
7
7
|
* Accepts a minimum value of 30 seconds and a maximum value of 1,800 seconds.
|
|
8
8
|
*/
|
|
9
9
|
export interface PostedTimeoutSpecsMaxDuration {
|
|
10
|
+
/** Duration in seconds for the timeout (e.g. 600 seconds represents 10 minutes). */
|
|
11
|
+
durationSecs?: number;
|
|
10
12
|
/**
|
|
11
13
|
* Boolean indicating if this timeout is enabled.
|
|
12
14
|
*
|
|
13
15
|
* If set to false, EVI will not timeout due to a specified maximum duration being reached. However, the conversation will eventually disconnect after 1,800 seconds (30 minutes), which is the maximum WebSocket duration limit for EVI.
|
|
14
16
|
*/
|
|
15
17
|
enabled: boolean;
|
|
16
|
-
/** Duration in seconds for the timeout (e.g. 600 seconds represents 10 minutes). */
|
|
17
|
-
durationSecs?: number;
|
|
18
18
|
}
|
|
@@ -5,12 +5,12 @@
|
|
|
5
5
|
* Resume responses from EVI. Chat history sent while paused will now be sent.
|
|
6
6
|
*/
|
|
7
7
|
export interface ResumeAssistantMessage {
|
|
8
|
-
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
9
|
-
customSessionId?: string;
|
|
10
8
|
/**
|
|
11
9
|
* The type of message sent through the socket; must be `resume_assistant_message` for our server to correctly identify and process it as a Resume Assistant message.
|
|
12
10
|
*
|
|
13
11
|
* Upon resuming, if any audio input was sent during the pause, EVI will retain context from all messages sent but only respond to the last user message. (e.g., If you ask EVI two questions while paused and then send a `resume_assistant_message`, EVI will respond to the second question and have added the first question to its conversation context.)
|
|
14
12
|
*/
|
|
15
13
|
type: "resume_assistant_message";
|
|
14
|
+
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
15
|
+
customSessionId?: string;
|
|
16
16
|
}
|
|
@@ -18,6 +18,10 @@ export interface ReturnConfig {
|
|
|
18
18
|
* Version numbers are integer values representing different iterations of the Config. Each update to the Config increments its version number.
|
|
19
19
|
*/
|
|
20
20
|
version?: number;
|
|
21
|
+
/** List of user-defined tools associated with this Config. */
|
|
22
|
+
tools?: (Hume.empathicVoice.ReturnUserDefinedTool | undefined)[];
|
|
23
|
+
/** An optional description of the Config version. */
|
|
24
|
+
versionDescription?: string;
|
|
21
25
|
/**
|
|
22
26
|
* The supplemental language model associated with this Config.
|
|
23
27
|
*
|
|
@@ -33,6 +37,7 @@ export interface ReturnConfig {
|
|
|
33
37
|
*/
|
|
34
38
|
eviVersion?: string;
|
|
35
39
|
timeouts?: Hume.empathicVoice.ReturnTimeoutSpecs;
|
|
40
|
+
nudges?: Hume.empathicVoice.ReturnNudgeSpec;
|
|
36
41
|
eventMessages?: Hume.empathicVoice.ReturnEventMessageSpecs;
|
|
37
42
|
/**
|
|
38
43
|
* The eLLM setup associated with this Config.
|
|
@@ -40,17 +45,12 @@ export interface ReturnConfig {
|
|
|
40
45
|
* Hume's eLLM (empathic Large Language Model) is a multimodal language model that takes into account both expression measures and language. The eLLM generates short, empathic language responses and guides text-to-speech (TTS) prosody.
|
|
41
46
|
*/
|
|
42
47
|
ellmModel?: Hume.empathicVoice.ReturnEllmModel;
|
|
48
|
+
voice?: unknown;
|
|
49
|
+
prompt?: Hume.empathicVoice.ReturnPrompt;
|
|
43
50
|
/** Map of webhooks associated with this config. */
|
|
44
51
|
webhooks?: (Hume.empathicVoice.ReturnWebhookSpec | undefined)[];
|
|
45
|
-
/** An optional description of the Config version. */
|
|
46
|
-
versionDescription?: string;
|
|
47
52
|
/** Time at which the Config was created. Measured in seconds since the Unix epoch. */
|
|
48
53
|
createdOn?: number;
|
|
49
54
|
/** Time at which the Config was last modified. Measured in seconds since the Unix epoch. */
|
|
50
55
|
modifiedOn?: number;
|
|
51
|
-
nudges?: Hume.empathicVoice.ReturnNudgeSpec;
|
|
52
|
-
voice?: unknown;
|
|
53
|
-
prompt?: Hume.empathicVoice.ReturnPrompt;
|
|
54
|
-
/** List of user-defined tools associated with this Config. */
|
|
55
|
-
tools?: (Hume.empathicVoice.ReturnUserDefinedTool | undefined)[];
|
|
56
56
|
}
|
|
@@ -13,7 +13,7 @@ export interface ReturnLanguageModel {
|
|
|
13
13
|
/**
|
|
14
14
|
* The model temperature, with values between 0 to 1 (inclusive).
|
|
15
15
|
*
|
|
16
|
-
* Controls the randomness of the LLM
|
|
16
|
+
* Controls the randomness of the LLM's output, with values closer to 0 yielding focused, deterministic responses and values closer to 1 producing more creative, diverse responses.
|
|
17
17
|
*/
|
|
18
18
|
temperature?: number;
|
|
19
19
|
}
|
|
@@ -11,11 +11,9 @@ export interface ReturnPrompt {
|
|
|
11
11
|
/** Identifier for a Prompt. Formatted as a UUID. */
|
|
12
12
|
id: string;
|
|
13
13
|
/**
|
|
14
|
-
* Instructions used to shape EVI
|
|
14
|
+
* Instructions used to shape EVI's behavior, responses, and style.
|
|
15
15
|
*
|
|
16
|
-
* You can use the Prompt to define a specific goal or role for EVI, specifying how it should act or what it should focus on during the conversation. For example, EVI can be instructed to act as a customer support representative, a fitness coach, or a travel advisor, each with its own set of behaviors and response styles.
|
|
17
|
-
*
|
|
18
|
-
* For help writing a system prompt, see our [Prompting Guide](/docs/speech-to-speech-evi/guides/prompting).
|
|
16
|
+
* You can use the Prompt to define a specific goal or role for EVI, specifying how it should act or what it should focus on during the conversation. For example, EVI can be instructed to act as a customer support representative, a fitness coach, or a travel advisor, each with its own set of behaviors and response styles. For help writing a system prompt, see our [Prompting Guide](/docs/speech-to-speech-evi/guides/prompting).
|
|
19
17
|
*/
|
|
20
18
|
text: string;
|
|
21
19
|
/**
|
|
@@ -26,10 +24,10 @@ export interface ReturnPrompt {
|
|
|
26
24
|
* Version numbers are integer values representing different iterations of the Prompt. Each update to the Prompt increments its version number.
|
|
27
25
|
*/
|
|
28
26
|
version: number;
|
|
29
|
-
/** Versioning method for a Prompt. Either `FIXED` for using a fixed version number or `LATEST` for auto-updating to the latest version. */
|
|
30
|
-
versionType: Hume.empathicVoice.ReturnPromptVersionType;
|
|
31
27
|
/** An optional description of the Prompt version. */
|
|
32
28
|
versionDescription?: string;
|
|
29
|
+
/** Versioning method for a Prompt. Either `FIXED` for using a fixed version number or `LATEST` for auto-updating to the latest version. */
|
|
30
|
+
versionType: Hume.empathicVoice.ReturnPromptVersionType;
|
|
33
31
|
/** Time at which the Prompt was created. Measured in seconds since the Unix epoch. */
|
|
34
32
|
createdOn: number;
|
|
35
33
|
/** Time at which the Prompt was last modified. Measured in seconds since the Unix epoch. */
|
|
@@ -35,7 +35,7 @@ export interface ReturnUserDefinedTool {
|
|
|
35
35
|
/**
|
|
36
36
|
* Stringified JSON defining the parameters used by this version of the Tool.
|
|
37
37
|
*
|
|
38
|
-
* These parameters define the inputs needed for the Tool
|
|
38
|
+
* These parameters define the inputs needed for the Tool's execution, including the expected data type and description for each input field. Structured as a stringified JSON schema, this format ensures the tool receives data in the expected format.
|
|
39
39
|
*/
|
|
40
40
|
parameters: string;
|
|
41
41
|
}
|
|
@@ -7,27 +7,13 @@ import * as Hume from "../../../index";
|
|
|
7
7
|
*/
|
|
8
8
|
export interface SessionSettings {
|
|
9
9
|
/**
|
|
10
|
-
*
|
|
11
|
-
*
|
|
12
|
-
* This optional field is only required when the audio input is encoded in PCM Linear 16 (16-bit, little-endian, signed PCM WAV data). For detailed instructions on how to configure session settings for PCM Linear 16 audio, please refer to the [Session Settings guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
|
|
13
|
-
*/
|
|
14
|
-
audio?: Hume.empathicVoice.AudioConfiguration;
|
|
15
|
-
/**
|
|
16
|
-
* List of built-in tools to enable for the session.
|
|
17
|
-
*
|
|
18
|
-
* Tools are resources used by EVI to perform various tasks, such as searching the web or calling external APIs. Built-in tools, like web search, are natively integrated, while user-defined tools are created and invoked by the user. To learn more, see our [Tool Use Guide](/docs/empathic-voice-interface-evi/features/tool-use).
|
|
19
|
-
*
|
|
20
|
-
* Currently, the only built-in tool Hume provides is **Web Search**. When enabled, Web Search equips EVI with the ability to search the web for up-to-date information.
|
|
21
|
-
*/
|
|
22
|
-
builtinTools?: Hume.empathicVoice.BuiltinToolConfig[];
|
|
23
|
-
/**
|
|
24
|
-
* Allows developers to inject additional context into the conversation, which is appended to the end of user messages for the session.
|
|
10
|
+
* The type of message sent through the socket; must be `session_settings` for our server to correctly identify and process it as a Session Settings message.
|
|
25
11
|
*
|
|
26
|
-
*
|
|
12
|
+
* Session settings are temporary and apply only to the current Chat session. These settings can be adjusted dynamically based on the requirements of each session to ensure optimal performance and user experience.
|
|
27
13
|
*
|
|
28
|
-
*
|
|
14
|
+
* For more information, please refer to the [Session Settings guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
|
|
29
15
|
*/
|
|
30
|
-
|
|
16
|
+
type: "session_settings";
|
|
31
17
|
/**
|
|
32
18
|
* Unique identifier for the session. Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions.
|
|
33
19
|
*
|
|
@@ -37,14 +23,7 @@ export interface SessionSettings {
|
|
|
37
23
|
*/
|
|
38
24
|
customSessionId?: string;
|
|
39
25
|
/**
|
|
40
|
-
*
|
|
41
|
-
*
|
|
42
|
-
* When provided, EVI will use this key instead of Hume’s API key for the supplemental LLM. This allows you to bypass rate limits and utilize your own API key as needed.
|
|
43
|
-
*/
|
|
44
|
-
languageModelApiKey?: string;
|
|
45
|
-
metadata?: Record<string, unknown>;
|
|
46
|
-
/**
|
|
47
|
-
* Instructions used to shape EVI’s behavior, responses, and style for the session.
|
|
26
|
+
* Instructions used to shape EVI's behavior, responses, and style for the session.
|
|
48
27
|
*
|
|
49
28
|
* When included in a Session Settings message, the provided Prompt overrides the existing one specified in the EVI configuration. If no Prompt was defined in the configuration, this Prompt will be the one used for the session.
|
|
50
29
|
*
|
|
@@ -53,6 +32,26 @@ export interface SessionSettings {
|
|
|
53
32
|
* For help writing a system prompt, see our [Prompting Guide](/docs/empathic-voice-interface-evi/guides/prompting).
|
|
54
33
|
*/
|
|
55
34
|
systemPrompt?: string;
|
|
35
|
+
/**
|
|
36
|
+
* Field for injecting additional context into the conversation, which is appended to the end of user messages for the session.
|
|
37
|
+
*
|
|
38
|
+
* When included in a Session Settings message, the provided context can be used to remind the LLM of its role in every user message, prevent it from forgetting important details, or add new relevant information to the conversation.
|
|
39
|
+
*
|
|
40
|
+
* Set to `null` to clear injected context.
|
|
41
|
+
*/
|
|
42
|
+
context?: Hume.empathicVoice.Context;
|
|
43
|
+
/**
|
|
44
|
+
* Configuration details for the audio input used during the session. Ensures the audio is being correctly set up for processing.
|
|
45
|
+
*
|
|
46
|
+
* This optional field is only required when the audio input is encoded in PCM Linear 16 (16-bit, little-endian, signed PCM WAV data). For detailed instructions on how to configure session settings for PCM Linear 16 audio, please refer to the [Session Settings guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
|
|
47
|
+
*/
|
|
48
|
+
audio?: Hume.empathicVoice.AudioConfiguration;
|
|
49
|
+
/**
|
|
50
|
+
* Third party API key for the supplemental language model.
|
|
51
|
+
*
|
|
52
|
+
* When provided, EVI will use this key instead of Hume's API key for the supplemental LLM. This allows you to bypass rate limits and utilize your own API key as needed.
|
|
53
|
+
*/
|
|
54
|
+
languageModelApiKey?: string;
|
|
56
55
|
/**
|
|
57
56
|
* List of user-defined tools to enable for the session.
|
|
58
57
|
*
|
|
@@ -60,13 +59,14 @@ export interface SessionSettings {
|
|
|
60
59
|
*/
|
|
61
60
|
tools?: Hume.empathicVoice.Tool[];
|
|
62
61
|
/**
|
|
63
|
-
*
|
|
62
|
+
* List of built-in tools to enable for the session.
|
|
64
63
|
*
|
|
65
|
-
*
|
|
64
|
+
* Tools are resources used by EVI to perform various tasks, such as searching the web or calling external APIs. Built-in tools, like web search, are natively integrated, while user-defined tools are created and invoked by the user. To learn more, see our [Tool Use Guide](/docs/empathic-voice-interface-evi/features/tool-use).
|
|
66
65
|
*
|
|
67
|
-
*
|
|
66
|
+
* Currently, the only built-in tool Hume provides is **Web Search**. When enabled, Web Search equips EVI with the ability to search the web for up-to-date information.
|
|
68
67
|
*/
|
|
69
|
-
|
|
68
|
+
builtinTools?: Hume.empathicVoice.BuiltinToolConfig[];
|
|
69
|
+
metadata?: Record<string, unknown>;
|
|
70
70
|
/**
|
|
71
71
|
* This field allows you to assign values to dynamic variables referenced in your system prompt.
|
|
72
72
|
*
|
|
@@ -3,18 +3,18 @@
|
|
|
3
3
|
*/
|
|
4
4
|
import * as Hume from "../../../index";
|
|
5
5
|
export interface Tool {
|
|
6
|
-
/**
|
|
7
|
-
|
|
8
|
-
/** Optional text passed to the supplemental LLM if the tool call fails. The LLM then uses this text to generate a response back to the user, ensuring continuity in the conversation. */
|
|
9
|
-
fallbackContent?: string;
|
|
6
|
+
/** Type of tool. Set to `function` for user-defined tools. */
|
|
7
|
+
type: Hume.empathicVoice.ToolType;
|
|
10
8
|
/** Name of the user-defined tool to be enabled. */
|
|
11
9
|
name: string;
|
|
12
10
|
/**
|
|
13
11
|
* Parameters of the tool. Is a stringified JSON schema.
|
|
14
12
|
*
|
|
15
|
-
* These parameters define the inputs needed for the tool
|
|
13
|
+
* These parameters define the inputs needed for the tool's execution, including the expected data type and description for each input field. Structured as a JSON schema, this format ensures the tool receives data in the expected format.
|
|
16
14
|
*/
|
|
17
15
|
parameters: string;
|
|
18
|
-
/**
|
|
19
|
-
|
|
16
|
+
/** An optional description of what the tool does, used by the supplemental LLM to choose when and how to call the function. */
|
|
17
|
+
description?: string;
|
|
18
|
+
/** Optional text passed to the supplemental LLM if the tool call fails. The LLM then uses this text to generate a response back to the user, ensuring continuity in the conversation. */
|
|
19
|
+
fallbackContent?: string;
|
|
20
20
|
}
|
|
@@ -6,30 +6,30 @@ import * as Hume from "../../../index";
|
|
|
6
6
|
* When provided, the output is a tool call.
|
|
7
7
|
*/
|
|
8
8
|
export interface ToolCallMessage {
|
|
9
|
-
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
10
|
-
customSessionId?: string;
|
|
11
9
|
/** Name of the tool called. */
|
|
12
10
|
name: string;
|
|
13
11
|
/**
|
|
14
12
|
* Parameters of the tool.
|
|
15
13
|
*
|
|
16
|
-
* These parameters define the inputs needed for the tool
|
|
14
|
+
* These parameters define the inputs needed for the tool's execution, including the expected data type and description for each input field. Structured as a stringified JSON schema, this format ensures the tool receives data in the expected format.
|
|
17
15
|
*/
|
|
18
16
|
parameters: string;
|
|
19
|
-
/** Indicates whether a response to the tool call is required from the developer, either in the form of a [Tool Response message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolResponseMessage.type) or a [Tool Error message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolErrorMessage.type). */
|
|
20
|
-
responseRequired: boolean;
|
|
21
17
|
/**
|
|
22
18
|
* The unique identifier for a specific tool call instance.
|
|
23
19
|
*
|
|
24
20
|
* This ID is used to track the request and response of a particular tool invocation, ensuring that the correct response is linked to the appropriate request.
|
|
25
21
|
*/
|
|
26
22
|
toolCallId: string;
|
|
27
|
-
/** Type of tool called. Either `builtin` for natively implemented tools, like web search, or `function` for user-defined tools. */
|
|
28
|
-
toolType?: Hume.empathicVoice.ToolType;
|
|
29
23
|
/**
|
|
30
24
|
* The type of message sent through the socket; for a Tool Call message, this must be `tool_call`.
|
|
31
25
|
*
|
|
32
26
|
* This message indicates that the supplemental LLM has detected a need to invoke the specified tool.
|
|
33
27
|
*/
|
|
34
28
|
type: "tool_call";
|
|
29
|
+
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
30
|
+
customSessionId?: string;
|
|
31
|
+
/** Type of tool called. Either `builtin` for natively implemented tools, like web search, or `function` for user-defined tools. */
|
|
32
|
+
toolType?: Hume.empathicVoice.ToolType;
|
|
33
|
+
/** Indicates whether a response to the tool call is required from the developer, either in the form of a [Tool Response message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolResponseMessage.type) or a [Tool Error message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolErrorMessage.type). */
|
|
34
|
+
responseRequired: boolean;
|
|
35
35
|
}
|
|
@@ -6,28 +6,28 @@ import * as Hume from "../../../index";
|
|
|
6
6
|
* When provided, the output is a function call error.
|
|
7
7
|
*/
|
|
8
8
|
export interface ToolErrorMessage {
|
|
9
|
-
/** Error code. Identifies the type of error encountered. */
|
|
10
|
-
code?: string;
|
|
11
|
-
/** Optional text passed to the supplemental LLM in place of the tool call result. The LLM then uses this text to generate a response back to the user, ensuring continuity in the conversation if the tool errors. */
|
|
12
|
-
content?: string;
|
|
13
|
-
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
14
|
-
customSessionId?: string;
|
|
15
|
-
/** Error message from the tool call, not exposed to the LLM or user. */
|
|
16
|
-
error: string;
|
|
17
|
-
/** Indicates the severity of an error; for a Tool Error message, this must be `warn` to signal an unexpected event. */
|
|
18
|
-
level?: Hume.empathicVoice.ErrorLevel;
|
|
19
9
|
/**
|
|
20
|
-
* The
|
|
10
|
+
* The type of message sent through the socket; for a Tool Error message, this must be `tool_error`.
|
|
21
11
|
*
|
|
22
|
-
*
|
|
12
|
+
* Upon receiving a [Tool Call message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type) and failing to invoke the function, this message is sent to notify EVI of the tool's failure.
|
|
23
13
|
*/
|
|
24
|
-
|
|
14
|
+
type: "tool_error";
|
|
15
|
+
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
16
|
+
customSessionId?: string;
|
|
25
17
|
/** Type of tool called. Either `builtin` for natively implemented tools, like web search, or `function` for user-defined tools. */
|
|
26
18
|
toolType?: Hume.empathicVoice.ToolType;
|
|
27
19
|
/**
|
|
28
|
-
* The
|
|
20
|
+
* The unique identifier for a specific tool call instance.
|
|
29
21
|
*
|
|
30
|
-
*
|
|
22
|
+
* This ID is used to track the request and response of a particular tool invocation, ensuring that the Tool Error message is linked to the appropriate tool call request. The specified `tool_call_id` must match the one received in the [Tool Call message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type).
|
|
31
23
|
*/
|
|
32
|
-
|
|
24
|
+
toolCallId: string;
|
|
25
|
+
/** Optional text passed to the supplemental LLM in place of the tool call result. The LLM then uses this text to generate a response back to the user, ensuring continuity in the conversation if the tool errors. */
|
|
26
|
+
content?: string;
|
|
27
|
+
/** Error message from the tool call, not exposed to the LLM or user. */
|
|
28
|
+
error: string;
|
|
29
|
+
/** Error code. Identifies the type of error encountered. */
|
|
30
|
+
code?: string;
|
|
31
|
+
/** Indicates the severity of an error; for a Tool Error message, this must be `warn` to signal an unexpected event. */
|
|
32
|
+
level?: Hume.empathicVoice.ErrorLevel;
|
|
33
33
|
}
|
|
@@ -6,8 +6,12 @@ import * as Hume from "../../../index";
|
|
|
6
6
|
* When provided, the output is a function call response.
|
|
7
7
|
*/
|
|
8
8
|
export interface ToolResponseMessage {
|
|
9
|
-
/**
|
|
10
|
-
|
|
9
|
+
/**
|
|
10
|
+
* The type of message sent through the socket; for a Tool Response message, this must be `tool_response`.
|
|
11
|
+
*
|
|
12
|
+
* Upon receiving a [Tool Call message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type) and successfully invoking the function, this message is sent to convey the result of the function call back to EVI.
|
|
13
|
+
*/
|
|
14
|
+
type: "tool_response";
|
|
11
15
|
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
|
|
12
16
|
customSessionId?: string;
|
|
13
17
|
/**
|
|
@@ -16,6 +20,8 @@ export interface ToolResponseMessage {
|
|
|
16
20
|
* This ID is used to track the request and response of a particular tool invocation, ensuring that the correct response is linked to the appropriate request. The specified `tool_call_id` must match the one received in the [Tool Call message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.tool_call_id).
|
|
17
21
|
*/
|
|
18
22
|
toolCallId: string;
|
|
23
|
+
/** Return value of the tool call. Contains the output generated by the tool to pass back to EVI. */
|
|
24
|
+
content: string;
|
|
19
25
|
/**
|
|
20
26
|
* Name of the tool.
|
|
21
27
|
*
|
|
@@ -24,10 +30,4 @@ export interface ToolResponseMessage {
|
|
|
24
30
|
toolName?: string;
|
|
25
31
|
/** Type of tool called. Either `builtin` for natively implemented tools, like web search, or `function` for user-defined tools. */
|
|
26
32
|
toolType?: Hume.empathicVoice.ToolType;
|
|
27
|
-
/**
|
|
28
|
-
* The type of message sent through the socket; for a Tool Response message, this must be `tool_response`.
|
|
29
|
-
*
|
|
30
|
-
* Upon receiving a [Tool Call message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type) and successfully invoking the function, this message is sent to convey the result of the function call back to EVI.
|
|
31
|
-
*/
|
|
32
|
-
type: "tool_response";
|
|
33
33
|
}
|