hume 0.12.2 → 0.13.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.mock/definition/empathic-voice/__package__.yml +688 -735
- package/.mock/definition/empathic-voice/chat.yml +29 -23
- package/.mock/definition/empathic-voice/chatWebhooks.yml +8 -12
- package/.mock/definition/empathic-voice/prompts.yml +2 -2
- package/.mock/definition/empathic-voice/tools.yml +2 -2
- package/.mock/definition/tts/__package__.yml +70 -87
- package/.mock/fern.config.json +1 -1
- package/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/index.js +0 -15
- package/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/api/resources/empathicVoice/errors/index.js +0 -1
- package/api/resources/empathicVoice/resources/chat/client/Client.d.ts +2 -0
- package/api/resources/empathicVoice/resources/prompts/client/requests/PostedPrompt.d.ts +1 -1
- package/api/resources/empathicVoice/resources/prompts/client/requests/PostedPromptVersion.d.ts +1 -1
- package/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedTool.d.ts +1 -1
- package/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedToolVersion.d.ts +1 -1
- package/api/resources/empathicVoice/types/AssistantEnd.d.ts +3 -3
- package/api/resources/empathicVoice/types/AssistantInput.d.ts +3 -3
- package/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/api/resources/empathicVoice/types/AudioInput.d.ts +7 -7
- package/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +3 -3
- package/api/resources/empathicVoice/types/PostedLanguageModel.d.ts +1 -1
- package/api/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +2 -2
- package/api/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +2 -2
- package/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/api/resources/empathicVoice/types/ReturnLanguageModel.d.ts +1 -1
- package/api/resources/empathicVoice/types/ReturnPrompt.d.ts +4 -6
- package/api/resources/empathicVoice/types/ReturnUserDefinedTool.d.ts +1 -1
- package/api/resources/empathicVoice/types/SessionSettings.d.ts +30 -30
- package/api/resources/empathicVoice/types/Tool.d.ts +7 -7
- package/api/resources/empathicVoice/types/ToolCallMessage.d.ts +7 -7
- package/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/UserInput.d.ts +3 -3
- package/api/resources/empathicVoice/types/UserInterruption.d.ts +5 -5
- package/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/api/resources/empathicVoice/types/index.js +16 -17
- package/api/resources/tts/client/Client.d.ts +21 -23
- package/api/resources/tts/client/Client.js +50 -58
- package/api/resources/tts/client/index.d.ts +0 -1
- package/api/resources/tts/client/index.js +0 -15
- package/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/index.js +0 -15
- package/dist/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/errors/index.js +0 -1
- package/dist/api/resources/empathicVoice/resources/chat/client/Client.d.ts +2 -0
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PostedPrompt.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PostedPromptVersion.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedTool.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/PostedUserDefinedToolVersion.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/AssistantEnd.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/AssistantInput.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/AudioInput.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/dist/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/dist/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/dist/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/dist/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/dist/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/PostedLanguageModel.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ReturnLanguageModel.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ReturnPrompt.d.ts +4 -6
- package/dist/api/resources/empathicVoice/types/ReturnUserDefinedTool.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/SessionSettings.d.ts +30 -30
- package/dist/api/resources/empathicVoice/types/Tool.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ToolCallMessage.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/dist/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/UserInput.d.ts +3 -3
- package/dist/api/resources/empathicVoice/types/UserInterruption.d.ts +5 -5
- package/dist/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/dist/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/dist/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/dist/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/dist/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/api/resources/empathicVoice/types/index.js +16 -17
- package/dist/api/resources/tts/client/Client.d.ts +21 -23
- package/dist/api/resources/tts/client/Client.js +50 -58
- package/dist/api/resources/tts/client/index.d.ts +0 -1
- package/dist/api/resources/tts/client/index.js +0 -15
- package/dist/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/dist/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/dist/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/dist/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/index.js +0 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/serialization/resources/empathicVoice/types/index.js +16 -17
- package/dist/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/dist/serialization/resources/tts/types/PostedTts.js +3 -3
- package/dist/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/dist/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/dist/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/dist/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/dist/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/dist/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/dist/serialization/resources/tts/types/Snippet.js +3 -3
- package/dist/version.d.ts +1 -1
- package/dist/version.js +1 -1
- package/package.json +1 -1
- package/reference.md +24 -78
- package/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/index.js +0 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsInactivity.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedTimeoutSpecsMaxDuration.js +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/serialization/resources/empathicVoice/types/index.js +16 -17
- package/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/serialization/resources/tts/types/PostedTts.js +3 -3
- package/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/serialization/resources/tts/types/Snippet.js +3 -3
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/api/resources/tts/client/requests/index.d.ts +0 -1
- package/api/resources/tts/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/dist/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/dist/api/resources/tts/client/requests/index.d.ts +0 -1
- package/dist/api/resources/tts/client/requests/index.js +0 -2
- package/dist/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/index.js +0 -17
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/dist/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
- package/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/index.js +0 -17
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
|
@@ -42,11 +42,11 @@ const ChatMessage_1 = require("./ChatMessage");
|
|
|
42
42
|
const Inference_1 = require("./Inference");
|
|
43
43
|
const MillisecondInterval_1 = require("./MillisecondInterval");
|
|
44
44
|
exports.UserMessage = core.serialization.object({
|
|
45
|
+
type: core.serialization.stringLiteral("user_message"),
|
|
45
46
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
46
|
-
fromText: core.serialization.property("from_text", core.serialization.boolean()),
|
|
47
|
-
interim: core.serialization.boolean(),
|
|
48
47
|
message: ChatMessage_1.ChatMessage,
|
|
49
48
|
models: Inference_1.Inference,
|
|
50
49
|
time: MillisecondInterval_1.MillisecondInterval,
|
|
51
|
-
|
|
50
|
+
fromText: core.serialization.property("from_text", core.serialization.boolean()),
|
|
51
|
+
interim: core.serialization.boolean(),
|
|
52
52
|
});
|
|
@@ -6,5 +6,5 @@ import * as Hume from "../../../../api/index";
|
|
|
6
6
|
import * as core from "../../../../core";
|
|
7
7
|
export declare const VoiceProvider: core.serialization.Schema<serializers.empathicVoice.VoiceProvider.Raw, Hume.empathicVoice.VoiceProvider>;
|
|
8
8
|
export declare namespace VoiceProvider {
|
|
9
|
-
type Raw = "HUME_AI" | "CUSTOM_VOICE";
|
|
9
|
+
type Raw = "HUME_AI" | "CUSTOM_VOICE" | "OCTAVE_COMBINED";
|
|
10
10
|
}
|
|
@@ -38,4 +38,4 @@ var __importStar = (this && this.__importStar) || (function () {
|
|
|
38
38
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
39
39
|
exports.VoiceProvider = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
|
-
exports.VoiceProvider = core.serialization.enum_(["HUME_AI", "CUSTOM_VOICE"]);
|
|
41
|
+
exports.VoiceProvider = core.serialization.enum_(["HUME_AI", "CUSTOM_VOICE", "OCTAVE_COMBINED"]);
|
|
@@ -7,11 +7,11 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const WebSocketError: core.serialization.ObjectSchema<serializers.empathicVoice.WebSocketError.Raw, Hume.empathicVoice.WebSocketError>;
|
|
8
8
|
export declare namespace WebSocketError {
|
|
9
9
|
interface Raw {
|
|
10
|
-
|
|
10
|
+
type: "error";
|
|
11
11
|
custom_session_id?: string | null;
|
|
12
|
+
code: string;
|
|
13
|
+
slug: string;
|
|
12
14
|
message: string;
|
|
13
15
|
request_id?: string | null;
|
|
14
|
-
slug: string;
|
|
15
|
-
type: "error";
|
|
16
16
|
}
|
|
17
17
|
}
|
|
@@ -39,10 +39,10 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.WebSocketError = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.WebSocketError = core.serialization.object({
|
|
42
|
-
|
|
42
|
+
type: core.serialization.stringLiteral("error"),
|
|
43
43
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
44
|
+
code: core.serialization.string(),
|
|
45
|
+
slug: core.serialization.string(),
|
|
44
46
|
message: core.serialization.string(),
|
|
45
47
|
requestId: core.serialization.property("request_id", core.serialization.string().optional()),
|
|
46
|
-
slug: core.serialization.string(),
|
|
47
|
-
type: core.serialization.stringLiteral("error"),
|
|
48
48
|
});
|
|
@@ -9,11 +9,11 @@ import { WebhookEventBase } from "./WebhookEventBase";
|
|
|
9
9
|
export declare const WebhookEventChatEnded: core.serialization.ObjectSchema<serializers.empathicVoice.WebhookEventChatEnded.Raw, Hume.empathicVoice.WebhookEventChatEnded>;
|
|
10
10
|
export declare namespace WebhookEventChatEnded {
|
|
11
11
|
interface Raw extends WebhookEventBase.Raw {
|
|
12
|
-
|
|
13
|
-
|
|
12
|
+
event_name?: "chat_ended" | null;
|
|
13
|
+
end_time: number;
|
|
14
14
|
duration_seconds: number;
|
|
15
15
|
end_reason: WebhookEventChatStatus.Raw;
|
|
16
|
-
|
|
17
|
-
|
|
16
|
+
caller_number?: string | null;
|
|
17
|
+
custom_session_id?: string | null;
|
|
18
18
|
}
|
|
19
19
|
}
|
|
@@ -42,11 +42,11 @@ const WebhookEventChatStatus_1 = require("./WebhookEventChatStatus");
|
|
|
42
42
|
const WebhookEventBase_1 = require("./WebhookEventBase");
|
|
43
43
|
exports.WebhookEventChatEnded = core.serialization
|
|
44
44
|
.object({
|
|
45
|
-
|
|
46
|
-
|
|
45
|
+
eventName: core.serialization.property("event_name", core.serialization.stringLiteral("chat_ended").optional()),
|
|
46
|
+
endTime: core.serialization.property("end_time", core.serialization.number()),
|
|
47
47
|
durationSeconds: core.serialization.property("duration_seconds", core.serialization.number()),
|
|
48
48
|
endReason: core.serialization.property("end_reason", WebhookEventChatStatus_1.WebhookEventChatStatus),
|
|
49
|
-
|
|
50
|
-
|
|
49
|
+
callerNumber: core.serialization.property("caller_number", core.serialization.string().optional()),
|
|
50
|
+
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
51
51
|
})
|
|
52
52
|
.extend(WebhookEventBase_1.WebhookEventBase);
|
|
@@ -9,10 +9,10 @@ import { WebhookEventBase } from "./WebhookEventBase";
|
|
|
9
9
|
export declare const WebhookEventChatStarted: core.serialization.ObjectSchema<serializers.empathicVoice.WebhookEventChatStarted.Raw, Hume.empathicVoice.WebhookEventChatStarted>;
|
|
10
10
|
export declare namespace WebhookEventChatStarted {
|
|
11
11
|
interface Raw extends WebhookEventBase.Raw {
|
|
12
|
-
caller_number?: string | null;
|
|
13
|
-
chat_start_type: WebhookEventChatStartType.Raw;
|
|
14
|
-
custom_session_id?: string | null;
|
|
15
12
|
event_name?: "chat_started" | null;
|
|
16
13
|
start_time: number;
|
|
14
|
+
chat_start_type: WebhookEventChatStartType.Raw;
|
|
15
|
+
caller_number?: string | null;
|
|
16
|
+
custom_session_id?: string | null;
|
|
17
17
|
}
|
|
18
18
|
}
|
|
@@ -42,10 +42,10 @@ const WebhookEventChatStartType_1 = require("./WebhookEventChatStartType");
|
|
|
42
42
|
const WebhookEventBase_1 = require("./WebhookEventBase");
|
|
43
43
|
exports.WebhookEventChatStarted = core.serialization
|
|
44
44
|
.object({
|
|
45
|
-
callerNumber: core.serialization.property("caller_number", core.serialization.string().optional()),
|
|
46
|
-
chatStartType: core.serialization.property("chat_start_type", WebhookEventChatStartType_1.WebhookEventChatStartType),
|
|
47
|
-
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
48
45
|
eventName: core.serialization.property("event_name", core.serialization.stringLiteral("chat_started").optional()),
|
|
49
46
|
startTime: core.serialization.property("start_time", core.serialization.number()),
|
|
47
|
+
chatStartType: core.serialization.property("chat_start_type", WebhookEventChatStartType_1.WebhookEventChatStartType),
|
|
48
|
+
callerNumber: core.serialization.property("caller_number", core.serialization.string().optional()),
|
|
49
|
+
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
50
50
|
})
|
|
51
51
|
.extend(WebhookEventBase_1.WebhookEventBase);
|
|
@@ -1,50 +1,49 @@
|
|
|
1
|
-
export * from "./AssistantEnd";
|
|
2
1
|
export * from "./AssistantInput";
|
|
3
|
-
export * from "./AssistantMessage";
|
|
4
|
-
export * from "./AssistantProsody";
|
|
5
2
|
export * from "./AudioConfiguration";
|
|
6
3
|
export * from "./AudioInput";
|
|
7
|
-
export * from "./AudioOutput";
|
|
8
4
|
export * from "./BuiltInTool";
|
|
9
5
|
export * from "./BuiltinToolConfig";
|
|
10
|
-
export * from "./ChatMessageToolResult";
|
|
11
|
-
export * from "./ChatMessage";
|
|
12
|
-
export * from "./ChatMetadata";
|
|
13
6
|
export * from "./Context";
|
|
14
7
|
export * from "./ContextType";
|
|
15
|
-
export * from "./EmotionScores";
|
|
16
8
|
export * from "./Encoding";
|
|
17
|
-
export * from "./WebSocketError";
|
|
18
9
|
export * from "./ErrorLevel";
|
|
19
|
-
export * from "./Inference";
|
|
20
|
-
export * from "./MillisecondInterval";
|
|
21
10
|
export * from "./PauseAssistantMessage";
|
|
22
|
-
export * from "./ProsodyInference";
|
|
23
11
|
export * from "./ResumeAssistantMessage";
|
|
24
|
-
export * from "./Role";
|
|
25
12
|
export * from "./SessionSettingsVariablesValue";
|
|
26
13
|
export * from "./SessionSettings";
|
|
27
14
|
export * from "./Tool";
|
|
28
|
-
export * from "./ToolCallMessage";
|
|
29
15
|
export * from "./ToolErrorMessage";
|
|
30
16
|
export * from "./ToolResponseMessage";
|
|
31
17
|
export * from "./ToolType";
|
|
32
18
|
export * from "./UserInput";
|
|
19
|
+
export * from "./AssistantEnd";
|
|
20
|
+
export * from "./AssistantMessage";
|
|
21
|
+
export * from "./AssistantProsody";
|
|
22
|
+
export * from "./AudioOutput";
|
|
23
|
+
export * from "./ChatMessageToolResult";
|
|
24
|
+
export * from "./ChatMessage";
|
|
25
|
+
export * from "./ChatMetadata";
|
|
26
|
+
export * from "./EmotionScores";
|
|
27
|
+
export * from "./WebSocketError";
|
|
28
|
+
export * from "./Inference";
|
|
29
|
+
export * from "./MillisecondInterval";
|
|
30
|
+
export * from "./ProsodyInference";
|
|
31
|
+
export * from "./Role";
|
|
32
|
+
export * from "./ToolCallMessage";
|
|
33
33
|
export * from "./UserInterruption";
|
|
34
34
|
export * from "./UserMessage";
|
|
35
35
|
export * from "./JsonMessage";
|
|
36
36
|
export * from "./HttpValidationError";
|
|
37
37
|
export * from "./LanguageModelType";
|
|
38
38
|
export * from "./ModelProviderEnum";
|
|
39
|
-
export * from "./SupportsToolUse";
|
|
40
39
|
export * from "./ValidationErrorLocItem";
|
|
41
40
|
export * from "./ValidationError";
|
|
42
|
-
export * from "./WebhookEventBase";
|
|
43
|
-
export * from "./WebhookEvent";
|
|
44
41
|
export * from "./WebhookEventChatEnded";
|
|
45
42
|
export * from "./WebhookEventChatStartType";
|
|
46
43
|
export * from "./WebhookEventChatStarted";
|
|
47
44
|
export * from "./WebhookEventChatStatus";
|
|
45
|
+
export * from "./WebhookEvent";
|
|
46
|
+
export * from "./WebhookEventBase";
|
|
48
47
|
export * from "./ErrorResponse";
|
|
49
48
|
export * from "./ReturnPagedUserDefinedTools";
|
|
50
49
|
export * from "./ReturnUserDefinedToolToolType";
|
|
@@ -14,53 +14,52 @@ var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
|
14
14
|
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
15
15
|
};
|
|
16
16
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
|
-
__exportStar(require("./AssistantEnd"), exports);
|
|
18
17
|
__exportStar(require("./AssistantInput"), exports);
|
|
19
|
-
__exportStar(require("./AssistantMessage"), exports);
|
|
20
|
-
__exportStar(require("./AssistantProsody"), exports);
|
|
21
18
|
__exportStar(require("./AudioConfiguration"), exports);
|
|
22
19
|
__exportStar(require("./AudioInput"), exports);
|
|
23
|
-
__exportStar(require("./AudioOutput"), exports);
|
|
24
20
|
__exportStar(require("./BuiltInTool"), exports);
|
|
25
21
|
__exportStar(require("./BuiltinToolConfig"), exports);
|
|
26
|
-
__exportStar(require("./ChatMessageToolResult"), exports);
|
|
27
|
-
__exportStar(require("./ChatMessage"), exports);
|
|
28
|
-
__exportStar(require("./ChatMetadata"), exports);
|
|
29
22
|
__exportStar(require("./Context"), exports);
|
|
30
23
|
__exportStar(require("./ContextType"), exports);
|
|
31
|
-
__exportStar(require("./EmotionScores"), exports);
|
|
32
24
|
__exportStar(require("./Encoding"), exports);
|
|
33
|
-
__exportStar(require("./WebSocketError"), exports);
|
|
34
25
|
__exportStar(require("./ErrorLevel"), exports);
|
|
35
|
-
__exportStar(require("./Inference"), exports);
|
|
36
|
-
__exportStar(require("./MillisecondInterval"), exports);
|
|
37
26
|
__exportStar(require("./PauseAssistantMessage"), exports);
|
|
38
|
-
__exportStar(require("./ProsodyInference"), exports);
|
|
39
27
|
__exportStar(require("./ResumeAssistantMessage"), exports);
|
|
40
|
-
__exportStar(require("./Role"), exports);
|
|
41
28
|
__exportStar(require("./SessionSettingsVariablesValue"), exports);
|
|
42
29
|
__exportStar(require("./SessionSettings"), exports);
|
|
43
30
|
__exportStar(require("./Tool"), exports);
|
|
44
|
-
__exportStar(require("./ToolCallMessage"), exports);
|
|
45
31
|
__exportStar(require("./ToolErrorMessage"), exports);
|
|
46
32
|
__exportStar(require("./ToolResponseMessage"), exports);
|
|
47
33
|
__exportStar(require("./ToolType"), exports);
|
|
48
34
|
__exportStar(require("./UserInput"), exports);
|
|
35
|
+
__exportStar(require("./AssistantEnd"), exports);
|
|
36
|
+
__exportStar(require("./AssistantMessage"), exports);
|
|
37
|
+
__exportStar(require("./AssistantProsody"), exports);
|
|
38
|
+
__exportStar(require("./AudioOutput"), exports);
|
|
39
|
+
__exportStar(require("./ChatMessageToolResult"), exports);
|
|
40
|
+
__exportStar(require("./ChatMessage"), exports);
|
|
41
|
+
__exportStar(require("./ChatMetadata"), exports);
|
|
42
|
+
__exportStar(require("./EmotionScores"), exports);
|
|
43
|
+
__exportStar(require("./WebSocketError"), exports);
|
|
44
|
+
__exportStar(require("./Inference"), exports);
|
|
45
|
+
__exportStar(require("./MillisecondInterval"), exports);
|
|
46
|
+
__exportStar(require("./ProsodyInference"), exports);
|
|
47
|
+
__exportStar(require("./Role"), exports);
|
|
48
|
+
__exportStar(require("./ToolCallMessage"), exports);
|
|
49
49
|
__exportStar(require("./UserInterruption"), exports);
|
|
50
50
|
__exportStar(require("./UserMessage"), exports);
|
|
51
51
|
__exportStar(require("./JsonMessage"), exports);
|
|
52
52
|
__exportStar(require("./HttpValidationError"), exports);
|
|
53
53
|
__exportStar(require("./LanguageModelType"), exports);
|
|
54
54
|
__exportStar(require("./ModelProviderEnum"), exports);
|
|
55
|
-
__exportStar(require("./SupportsToolUse"), exports);
|
|
56
55
|
__exportStar(require("./ValidationErrorLocItem"), exports);
|
|
57
56
|
__exportStar(require("./ValidationError"), exports);
|
|
58
|
-
__exportStar(require("./WebhookEventBase"), exports);
|
|
59
|
-
__exportStar(require("./WebhookEvent"), exports);
|
|
60
57
|
__exportStar(require("./WebhookEventChatEnded"), exports);
|
|
61
58
|
__exportStar(require("./WebhookEventChatStartType"), exports);
|
|
62
59
|
__exportStar(require("./WebhookEventChatStarted"), exports);
|
|
63
60
|
__exportStar(require("./WebhookEventChatStatus"), exports);
|
|
61
|
+
__exportStar(require("./WebhookEvent"), exports);
|
|
62
|
+
__exportStar(require("./WebhookEventBase"), exports);
|
|
64
63
|
__exportStar(require("./ErrorResponse"), exports);
|
|
65
64
|
__exportStar(require("./ReturnPagedUserDefinedTools"), exports);
|
|
66
65
|
__exportStar(require("./ReturnUserDefinedToolToolType"), exports);
|
|
@@ -5,17 +5,17 @@ import * as serializers from "../../../index";
|
|
|
5
5
|
import * as Hume from "../../../../api/index";
|
|
6
6
|
import * as core from "../../../../core";
|
|
7
7
|
import { PostedContext } from "./PostedContext";
|
|
8
|
-
import { Format } from "./Format";
|
|
9
8
|
import { PostedUtterance } from "./PostedUtterance";
|
|
9
|
+
import { Format } from "./Format";
|
|
10
10
|
export declare const PostedTts: core.serialization.ObjectSchema<serializers.tts.PostedTts.Raw, Hume.tts.PostedTts>;
|
|
11
11
|
export declare namespace PostedTts {
|
|
12
12
|
interface Raw {
|
|
13
13
|
context?: PostedContext.Raw | null;
|
|
14
|
-
|
|
14
|
+
utterances: PostedUtterance.Raw[];
|
|
15
15
|
num_generations?: number | null;
|
|
16
|
+
format?: Format.Raw | null;
|
|
16
17
|
split_utterances?: boolean | null;
|
|
17
18
|
strip_headers?: boolean | null;
|
|
18
|
-
utterances: PostedUtterance.Raw[];
|
|
19
19
|
instant_mode?: boolean | null;
|
|
20
20
|
}
|
|
21
21
|
}
|
|
@@ -39,14 +39,14 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.PostedTts = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const PostedContext_1 = require("./PostedContext");
|
|
42
|
-
const Format_1 = require("./Format");
|
|
43
42
|
const PostedUtterance_1 = require("./PostedUtterance");
|
|
43
|
+
const Format_1 = require("./Format");
|
|
44
44
|
exports.PostedTts = core.serialization.object({
|
|
45
45
|
context: PostedContext_1.PostedContext.optional(),
|
|
46
|
-
|
|
46
|
+
utterances: core.serialization.list(PostedUtterance_1.PostedUtterance),
|
|
47
47
|
numGenerations: core.serialization.property("num_generations", core.serialization.number().optional()),
|
|
48
|
+
format: Format_1.Format.optional(),
|
|
48
49
|
splitUtterances: core.serialization.property("split_utterances", core.serialization.boolean().optional()),
|
|
49
50
|
stripHeaders: core.serialization.property("strip_headers", core.serialization.boolean().optional()),
|
|
50
|
-
utterances: core.serialization.list(PostedUtterance_1.PostedUtterance),
|
|
51
51
|
instantMode: core.serialization.property("instant_mode", core.serialization.boolean().optional()),
|
|
52
52
|
});
|
|
@@ -8,10 +8,10 @@ import { PostedUtteranceVoice } from "./PostedUtteranceVoice";
|
|
|
8
8
|
export declare const PostedUtterance: core.serialization.ObjectSchema<serializers.tts.PostedUtterance.Raw, Hume.tts.PostedUtterance>;
|
|
9
9
|
export declare namespace PostedUtterance {
|
|
10
10
|
interface Raw {
|
|
11
|
+
text: string;
|
|
11
12
|
description?: string | null;
|
|
13
|
+
voice?: PostedUtteranceVoice.Raw | null;
|
|
12
14
|
speed?: number | null;
|
|
13
|
-
text: string;
|
|
14
15
|
trailing_silence?: number | null;
|
|
15
|
-
voice?: PostedUtteranceVoice.Raw | null;
|
|
16
16
|
}
|
|
17
17
|
}
|
|
@@ -40,9 +40,9 @@ exports.PostedUtterance = void 0;
|
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const PostedUtteranceVoice_1 = require("./PostedUtteranceVoice");
|
|
42
42
|
exports.PostedUtterance = core.serialization.object({
|
|
43
|
+
text: core.serialization.string(),
|
|
43
44
|
description: core.serialization.string().optional(),
|
|
45
|
+
voice: PostedUtteranceVoice_1.PostedUtteranceVoice.optional(),
|
|
44
46
|
speed: core.serialization.number().optional(),
|
|
45
|
-
text: core.serialization.string(),
|
|
46
47
|
trailingSilence: core.serialization.property("trailing_silence", core.serialization.number().optional()),
|
|
47
|
-
voice: PostedUtteranceVoice_1.PostedUtteranceVoice.optional(),
|
|
48
48
|
});
|
|
@@ -9,11 +9,11 @@ import { Snippet } from "./Snippet";
|
|
|
9
9
|
export declare const ReturnGeneration: core.serialization.ObjectSchema<serializers.tts.ReturnGeneration.Raw, Hume.tts.ReturnGeneration>;
|
|
10
10
|
export declare namespace ReturnGeneration {
|
|
11
11
|
interface Raw {
|
|
12
|
-
|
|
12
|
+
generation_id: string;
|
|
13
13
|
duration: number;
|
|
14
|
-
encoding: AudioEncoding.Raw;
|
|
15
14
|
file_size: number;
|
|
16
|
-
|
|
15
|
+
encoding: AudioEncoding.Raw;
|
|
16
|
+
audio: string;
|
|
17
17
|
snippets: Snippet.Raw[][];
|
|
18
18
|
}
|
|
19
19
|
}
|
|
@@ -41,10 +41,10 @@ const core = __importStar(require("../../../../core"));
|
|
|
41
41
|
const AudioEncoding_1 = require("./AudioEncoding");
|
|
42
42
|
const Snippet_1 = require("./Snippet");
|
|
43
43
|
exports.ReturnGeneration = core.serialization.object({
|
|
44
|
-
|
|
44
|
+
generationId: core.serialization.property("generation_id", core.serialization.string()),
|
|
45
45
|
duration: core.serialization.number(),
|
|
46
|
-
encoding: AudioEncoding_1.AudioEncoding,
|
|
47
46
|
fileSize: core.serialization.property("file_size", core.serialization.number()),
|
|
48
|
-
|
|
47
|
+
encoding: AudioEncoding_1.AudioEncoding,
|
|
48
|
+
audio: core.serialization.string(),
|
|
49
49
|
snippets: core.serialization.list(core.serialization.list(Snippet_1.Snippet)),
|
|
50
50
|
});
|
|
@@ -8,7 +8,7 @@ import { ReturnGeneration } from "./ReturnGeneration";
|
|
|
8
8
|
export declare const ReturnTts: core.serialization.ObjectSchema<serializers.tts.ReturnTts.Raw, Hume.tts.ReturnTts>;
|
|
9
9
|
export declare namespace ReturnTts {
|
|
10
10
|
interface Raw {
|
|
11
|
-
generations: ReturnGeneration.Raw[];
|
|
12
11
|
request_id?: string | null;
|
|
12
|
+
generations: ReturnGeneration.Raw[];
|
|
13
13
|
}
|
|
14
14
|
}
|
|
@@ -40,6 +40,6 @@ exports.ReturnTts = void 0;
|
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const ReturnGeneration_1 = require("./ReturnGeneration");
|
|
42
42
|
exports.ReturnTts = core.serialization.object({
|
|
43
|
-
generations: core.serialization.list(ReturnGeneration_1.ReturnGeneration),
|
|
44
43
|
requestId: core.serialization.property("request_id", core.serialization.string().optional()),
|
|
44
|
+
generations: core.serialization.list(ReturnGeneration_1.ReturnGeneration),
|
|
45
45
|
});
|
|
@@ -7,11 +7,11 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const Snippet: core.serialization.ObjectSchema<serializers.tts.Snippet.Raw, Hume.tts.Snippet>;
|
|
8
8
|
export declare namespace Snippet {
|
|
9
9
|
interface Raw {
|
|
10
|
-
audio: string;
|
|
11
|
-
generation_id: string;
|
|
12
10
|
id: string;
|
|
13
11
|
text: string;
|
|
14
|
-
|
|
12
|
+
generation_id: string;
|
|
15
13
|
utterance_index?: number | null;
|
|
14
|
+
transcribed_text?: string | null;
|
|
15
|
+
audio: string;
|
|
16
16
|
}
|
|
17
17
|
}
|
|
@@ -39,10 +39,10 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.Snippet = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.Snippet = core.serialization.object({
|
|
42
|
-
audio: core.serialization.string(),
|
|
43
|
-
generationId: core.serialization.property("generation_id", core.serialization.string()),
|
|
44
42
|
id: core.serialization.string(),
|
|
45
43
|
text: core.serialization.string(),
|
|
46
|
-
|
|
44
|
+
generationId: core.serialization.property("generation_id", core.serialization.string()),
|
|
47
45
|
utteranceIndex: core.serialization.property("utterance_index", core.serialization.number().optional()),
|
|
46
|
+
transcribedText: core.serialization.property("transcribed_text", core.serialization.string().optional()),
|
|
47
|
+
audio: core.serialization.string(),
|
|
48
48
|
});
|
package/dist/version.d.ts
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
export declare const SDK_VERSION = "0.
|
|
1
|
+
export declare const SDK_VERSION = "0.13.1";
|
package/dist/version.js
CHANGED
package/package.json
CHANGED
package/reference.md
CHANGED
|
@@ -33,28 +33,26 @@ The response includes the base64-encoded audio and metadata in JSON format.
|
|
|
33
33
|
|
|
34
34
|
```typescript
|
|
35
35
|
await client.tts.synthesizeJson({
|
|
36
|
-
|
|
36
|
+
utterances: [
|
|
37
|
+
{
|
|
38
|
+
text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
|
|
39
|
+
description:
|
|
40
|
+
"Middle-aged masculine voice with a clear, rhythmic Scots lilt, rounded vowels, and a warm, steady tone with an articulate, academic quality.",
|
|
41
|
+
},
|
|
42
|
+
],
|
|
43
|
+
context: {
|
|
37
44
|
utterances: [
|
|
38
45
|
{
|
|
39
|
-
text: "
|
|
46
|
+
text: "How can people see beauty so differently?",
|
|
40
47
|
description:
|
|
41
|
-
"
|
|
48
|
+
"A curious student with a clear and respectful tone, seeking clarification on Hume's ideas with a straightforward question.",
|
|
42
49
|
},
|
|
43
50
|
],
|
|
44
|
-
context: {
|
|
45
|
-
utterances: [
|
|
46
|
-
{
|
|
47
|
-
text: "How can people see beauty so differently?",
|
|
48
|
-
description:
|
|
49
|
-
"A curious student with a clear and respectful tone, seeking clarification on Hume's ideas with a straightforward question.",
|
|
50
|
-
},
|
|
51
|
-
],
|
|
52
|
-
},
|
|
53
|
-
format: {
|
|
54
|
-
type: "mp3",
|
|
55
|
-
},
|
|
56
|
-
numGenerations: 1,
|
|
57
51
|
},
|
|
52
|
+
format: {
|
|
53
|
+
type: "mp3",
|
|
54
|
+
},
|
|
55
|
+
numGenerations: 1,
|
|
58
56
|
});
|
|
59
57
|
```
|
|
60
58
|
|
|
@@ -71,7 +69,7 @@ await client.tts.synthesizeJson({
|
|
|
71
69
|
<dl>
|
|
72
70
|
<dd>
|
|
73
71
|
|
|
74
|
-
**request:** `Hume.
|
|
72
|
+
**request:** `Hume.PostedTts`
|
|
75
73
|
|
|
76
74
|
</dd>
|
|
77
75
|
</dl>
|
|
@@ -170,7 +168,7 @@ await client.tts.synthesizeFile({
|
|
|
170
168
|
</dl>
|
|
171
169
|
</details>
|
|
172
170
|
|
|
173
|
-
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">
|
|
171
|
+
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">synthesizeJsonStreaming</a>({ ...params }) -> core.Stream<Hume.SnippetAudioChunk></code></summary>
|
|
174
172
|
<dl>
|
|
175
173
|
<dd>
|
|
176
174
|
|
|
@@ -184,6 +182,8 @@ await client.tts.synthesizeFile({
|
|
|
184
182
|
|
|
185
183
|
Streams synthesized speech using the specified voice. If no voice is provided, a novel voice will be generated dynamically. Optionally, additional context can be included to influence the speech's style and prosody.
|
|
186
184
|
|
|
185
|
+
The response is a stream of JSON objects including audio encoded in base64.
|
|
186
|
+
|
|
187
187
|
</dd>
|
|
188
188
|
</dl>
|
|
189
189
|
</dd>
|
|
@@ -198,7 +198,7 @@ Streams synthesized speech using the specified voice. If no voice is provided, a
|
|
|
198
198
|
<dd>
|
|
199
199
|
|
|
200
200
|
```typescript
|
|
201
|
-
await client.tts.
|
|
201
|
+
const response = await client.tts.synthesizeJsonStreaming({
|
|
202
202
|
utterances: [
|
|
203
203
|
{
|
|
204
204
|
text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
|
|
@@ -209,6 +209,9 @@ await client.tts.synthesizeFileStreaming({
|
|
|
209
209
|
},
|
|
210
210
|
],
|
|
211
211
|
});
|
|
212
|
+
for await (const item of response) {
|
|
213
|
+
console.log(item);
|
|
214
|
+
}
|
|
212
215
|
```
|
|
213
216
|
|
|
214
217
|
</dd>
|
|
@@ -243,7 +246,7 @@ await client.tts.synthesizeFileStreaming({
|
|
|
243
246
|
</dl>
|
|
244
247
|
</details>
|
|
245
248
|
|
|
246
|
-
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">
|
|
249
|
+
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">synthesizeFileStreaming</a>({ ...params }) -> stream.Readable</code></summary>
|
|
247
250
|
<dl>
|
|
248
251
|
<dd>
|
|
249
252
|
|
|
@@ -257,8 +260,6 @@ await client.tts.synthesizeFileStreaming({
|
|
|
257
260
|
|
|
258
261
|
Streams synthesized speech using the specified voice. If no voice is provided, a novel voice will be generated dynamically. Optionally, additional context can be included to influence the speech's style and prosody.
|
|
259
262
|
|
|
260
|
-
The response is a stream of JSON objects including audio encoded in base64.
|
|
261
|
-
|
|
262
263
|
</dd>
|
|
263
264
|
</dl>
|
|
264
265
|
</dd>
|
|
@@ -273,7 +274,7 @@ The response is a stream of JSON objects including audio encoded in base64.
|
|
|
273
274
|
<dd>
|
|
274
275
|
|
|
275
276
|
```typescript
|
|
276
|
-
|
|
277
|
+
await client.tts.synthesizeFileStreaming({
|
|
277
278
|
utterances: [
|
|
278
279
|
{
|
|
279
280
|
text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
|
|
@@ -284,9 +285,6 @@ const response = await client.tts.synthesizeJsonStreaming({
|
|
|
284
285
|
},
|
|
285
286
|
],
|
|
286
287
|
});
|
|
287
|
-
for await (const item of response) {
|
|
288
|
-
console.log(item);
|
|
289
|
-
}
|
|
290
288
|
```
|
|
291
289
|
|
|
292
290
|
</dd>
|
|
@@ -532,58 +530,6 @@ await client.tts.voices.delete({
|
|
|
532
530
|
</dl>
|
|
533
531
|
</details>
|
|
534
532
|
|
|
535
|
-
## EmpathicVoice
|
|
536
|
-
|
|
537
|
-
<details><summary><code>client.empathicVoice.<a href="/src/api/resources/empathicVoice/client/Client.ts">customLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost</a>({ ...params }) -> Hume.SupportsToolUse</code></summary>
|
|
538
|
-
<dl>
|
|
539
|
-
<dd>
|
|
540
|
-
|
|
541
|
-
#### 🔌 Usage
|
|
542
|
-
|
|
543
|
-
<dl>
|
|
544
|
-
<dd>
|
|
545
|
-
|
|
546
|
-
<dl>
|
|
547
|
-
<dd>
|
|
548
|
-
|
|
549
|
-
```typescript
|
|
550
|
-
await client.empathicVoice.customLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost({
|
|
551
|
-
modelResource: "model_resource",
|
|
552
|
-
});
|
|
553
|
-
```
|
|
554
|
-
|
|
555
|
-
</dd>
|
|
556
|
-
</dl>
|
|
557
|
-
</dd>
|
|
558
|
-
</dl>
|
|
559
|
-
|
|
560
|
-
#### ⚙️ Parameters
|
|
561
|
-
|
|
562
|
-
<dl>
|
|
563
|
-
<dd>
|
|
564
|
-
|
|
565
|
-
<dl>
|
|
566
|
-
<dd>
|
|
567
|
-
|
|
568
|
-
**request:** `Hume.empathicVoice.BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost`
|
|
569
|
-
|
|
570
|
-
</dd>
|
|
571
|
-
</dl>
|
|
572
|
-
|
|
573
|
-
<dl>
|
|
574
|
-
<dd>
|
|
575
|
-
|
|
576
|
-
**requestOptions:** `EmpathicVoice.RequestOptions`
|
|
577
|
-
|
|
578
|
-
</dd>
|
|
579
|
-
</dl>
|
|
580
|
-
</dd>
|
|
581
|
-
</dl>
|
|
582
|
-
|
|
583
|
-
</dd>
|
|
584
|
-
</dl>
|
|
585
|
-
</details>
|
|
586
|
-
|
|
587
533
|
## EmpathicVoice Tools
|
|
588
534
|
|
|
589
535
|
<details><summary><code>client.empathicVoice.tools.<a href="/src/api/resources/empathicVoice/resources/tools/client/Client.ts">listTools</a>({ ...params }) -> core.Page<Hume.ReturnUserDefinedTool | undefined></code></summary>
|
|
@@ -16,4 +16,3 @@ var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
|
16
16
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
17
|
__exportStar(require("./resources"), exports);
|
|
18
18
|
__exportStar(require("./types"), exports);
|
|
19
|
-
__exportStar(require("./client"), exports);
|
|
@@ -7,7 +7,7 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const AssistantEnd: core.serialization.ObjectSchema<serializers.empathicVoice.AssistantEnd.Raw, Hume.empathicVoice.AssistantEnd>;
|
|
8
8
|
export declare namespace AssistantEnd {
|
|
9
9
|
interface Raw {
|
|
10
|
-
custom_session_id?: string | null;
|
|
11
10
|
type: "assistant_end";
|
|
11
|
+
custom_session_id?: string | null;
|
|
12
12
|
}
|
|
13
13
|
}
|