hume 0.12.2 → 0.13.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.mock/definition/empathic-voice/__package__.yml +667 -712
- package/.mock/definition/empathic-voice/chat.yml +29 -23
- package/.mock/definition/empathic-voice/chatWebhooks.yml +3 -3
- package/.mock/definition/tts/__package__.yml +70 -87
- package/.mock/fern.config.json +1 -1
- package/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/index.js +0 -15
- package/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/api/resources/empathicVoice/errors/index.js +0 -1
- package/api/resources/empathicVoice/types/AssistantEnd.d.ts +2 -2
- package/api/resources/empathicVoice/types/AssistantInput.d.ts +2 -2
- package/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/api/resources/empathicVoice/types/AudioInput.d.ts +6 -6
- package/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/api/resources/empathicVoice/types/ReturnPrompt.d.ts +2 -2
- package/api/resources/empathicVoice/types/SessionSettings.d.ts +29 -29
- package/api/resources/empathicVoice/types/Tool.d.ts +6 -6
- package/api/resources/empathicVoice/types/ToolCallMessage.d.ts +6 -6
- package/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/api/resources/empathicVoice/types/UserInput.d.ts +2 -2
- package/api/resources/empathicVoice/types/UserInterruption.d.ts +4 -4
- package/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/api/resources/empathicVoice/types/index.js +16 -17
- package/api/resources/tts/client/Client.d.ts +21 -23
- package/api/resources/tts/client/Client.js +50 -58
- package/api/resources/tts/client/index.d.ts +0 -1
- package/api/resources/tts/client/index.js +0 -15
- package/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/api/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/index.js +0 -15
- package/dist/api/resources/empathicVoice/errors/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/errors/index.js +0 -1
- package/dist/api/resources/empathicVoice/types/AssistantEnd.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/AssistantInput.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/AudioInput.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
- package/dist/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/Context.d.ts +8 -14
- package/dist/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
- package/dist/api/resources/empathicVoice/types/ContextType.js +1 -2
- package/dist/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
- package/dist/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
- package/dist/api/resources/empathicVoice/types/ReturnPrompt.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/SessionSettings.d.ts +29 -29
- package/dist/api/resources/empathicVoice/types/Tool.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/ToolCallMessage.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
- package/dist/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/UserInput.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/UserInterruption.d.ts +4 -4
- package/dist/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
- package/dist/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
- package/dist/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
- package/dist/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
- package/dist/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
- package/dist/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
- package/dist/api/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/api/resources/empathicVoice/types/index.js +16 -17
- package/dist/api/resources/tts/client/Client.d.ts +21 -23
- package/dist/api/resources/tts/client/Client.js +50 -58
- package/dist/api/resources/tts/client/index.d.ts +0 -1
- package/dist/api/resources/tts/client/index.js +0 -15
- package/dist/api/resources/tts/types/PostedTts.d.ts +8 -8
- package/dist/api/resources/tts/types/PostedUtterance.d.ts +6 -6
- package/dist/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
- package/dist/api/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/api/resources/tts/types/Snippet.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/index.js +0 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/dist/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/dist/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/dist/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/dist/serialization/resources/empathicVoice/types/index.js +16 -17
- package/dist/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/dist/serialization/resources/tts/types/PostedTts.js +3 -3
- package/dist/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/dist/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/dist/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/dist/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/dist/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/dist/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/dist/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/dist/serialization/resources/tts/types/Snippet.js +3 -3
- package/dist/version.d.ts +1 -1
- package/dist/version.js +1 -1
- package/package.json +1 -1
- package/reference.md +24 -78
- package/serialization/resources/empathicVoice/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/index.js +0 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
- package/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
- package/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
- package/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/Context.js +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ContextType.js +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
- package/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
- package/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
- package/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
- package/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
- package/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/Tool.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
- package/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
- package/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
- package/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInput.js +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
- package/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
- package/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
- package/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
- package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
- package/serialization/resources/empathicVoice/types/index.d.ts +16 -17
- package/serialization/resources/empathicVoice/types/index.js +16 -17
- package/serialization/resources/tts/types/PostedTts.d.ts +3 -3
- package/serialization/resources/tts/types/PostedTts.js +3 -3
- package/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
- package/serialization/resources/tts/types/PostedUtterance.js +2 -2
- package/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
- package/serialization/resources/tts/types/ReturnGeneration.js +3 -3
- package/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
- package/serialization/resources/tts/types/ReturnTts.js +1 -1
- package/serialization/resources/tts/types/Snippet.d.ts +3 -3
- package/serialization/resources/tts/types/Snippet.js +3 -3
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/api/resources/tts/client/requests/index.d.ts +0 -1
- package/api/resources/tts/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
- package/dist/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/api/resources/empathicVoice/client/requests/index.js +0 -2
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
- package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
- package/dist/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
- package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
- package/dist/api/resources/tts/client/requests/index.d.ts +0 -1
- package/dist/api/resources/tts/client/requests/index.js +0 -2
- package/dist/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/index.js +0 -17
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/dist/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/dist/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
- package/serialization/resources/empathicVoice/client/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/index.js +0 -17
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
- package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
- package/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
- package/serialization/resources/empathicVoice/client/requests/index.js +0 -5
- package/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
- package/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
package/reference.md
CHANGED
|
@@ -33,28 +33,26 @@ The response includes the base64-encoded audio and metadata in JSON format.
|
|
|
33
33
|
|
|
34
34
|
```typescript
|
|
35
35
|
await client.tts.synthesizeJson({
|
|
36
|
-
|
|
36
|
+
utterances: [
|
|
37
|
+
{
|
|
38
|
+
text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
|
|
39
|
+
description:
|
|
40
|
+
"Middle-aged masculine voice with a clear, rhythmic Scots lilt, rounded vowels, and a warm, steady tone with an articulate, academic quality.",
|
|
41
|
+
},
|
|
42
|
+
],
|
|
43
|
+
context: {
|
|
37
44
|
utterances: [
|
|
38
45
|
{
|
|
39
|
-
text: "
|
|
46
|
+
text: "How can people see beauty so differently?",
|
|
40
47
|
description:
|
|
41
|
-
"
|
|
48
|
+
"A curious student with a clear and respectful tone, seeking clarification on Hume's ideas with a straightforward question.",
|
|
42
49
|
},
|
|
43
50
|
],
|
|
44
|
-
context: {
|
|
45
|
-
utterances: [
|
|
46
|
-
{
|
|
47
|
-
text: "How can people see beauty so differently?",
|
|
48
|
-
description:
|
|
49
|
-
"A curious student with a clear and respectful tone, seeking clarification on Hume's ideas with a straightforward question.",
|
|
50
|
-
},
|
|
51
|
-
],
|
|
52
|
-
},
|
|
53
|
-
format: {
|
|
54
|
-
type: "mp3",
|
|
55
|
-
},
|
|
56
|
-
numGenerations: 1,
|
|
57
51
|
},
|
|
52
|
+
format: {
|
|
53
|
+
type: "mp3",
|
|
54
|
+
},
|
|
55
|
+
numGenerations: 1,
|
|
58
56
|
});
|
|
59
57
|
```
|
|
60
58
|
|
|
@@ -71,7 +69,7 @@ await client.tts.synthesizeJson({
|
|
|
71
69
|
<dl>
|
|
72
70
|
<dd>
|
|
73
71
|
|
|
74
|
-
**request:** `Hume.
|
|
72
|
+
**request:** `Hume.PostedTts`
|
|
75
73
|
|
|
76
74
|
</dd>
|
|
77
75
|
</dl>
|
|
@@ -170,7 +168,7 @@ await client.tts.synthesizeFile({
|
|
|
170
168
|
</dl>
|
|
171
169
|
</details>
|
|
172
170
|
|
|
173
|
-
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">
|
|
171
|
+
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">synthesizeJsonStreaming</a>({ ...params }) -> core.Stream<Hume.SnippetAudioChunk></code></summary>
|
|
174
172
|
<dl>
|
|
175
173
|
<dd>
|
|
176
174
|
|
|
@@ -184,6 +182,8 @@ await client.tts.synthesizeFile({
|
|
|
184
182
|
|
|
185
183
|
Streams synthesized speech using the specified voice. If no voice is provided, a novel voice will be generated dynamically. Optionally, additional context can be included to influence the speech's style and prosody.
|
|
186
184
|
|
|
185
|
+
The response is a stream of JSON objects including audio encoded in base64.
|
|
186
|
+
|
|
187
187
|
</dd>
|
|
188
188
|
</dl>
|
|
189
189
|
</dd>
|
|
@@ -198,7 +198,7 @@ Streams synthesized speech using the specified voice. If no voice is provided, a
|
|
|
198
198
|
<dd>
|
|
199
199
|
|
|
200
200
|
```typescript
|
|
201
|
-
await client.tts.
|
|
201
|
+
const response = await client.tts.synthesizeJsonStreaming({
|
|
202
202
|
utterances: [
|
|
203
203
|
{
|
|
204
204
|
text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
|
|
@@ -209,6 +209,9 @@ await client.tts.synthesizeFileStreaming({
|
|
|
209
209
|
},
|
|
210
210
|
],
|
|
211
211
|
});
|
|
212
|
+
for await (const item of response) {
|
|
213
|
+
console.log(item);
|
|
214
|
+
}
|
|
212
215
|
```
|
|
213
216
|
|
|
214
217
|
</dd>
|
|
@@ -243,7 +246,7 @@ await client.tts.synthesizeFileStreaming({
|
|
|
243
246
|
</dl>
|
|
244
247
|
</details>
|
|
245
248
|
|
|
246
|
-
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">
|
|
249
|
+
<details><summary><code>client.tts.<a href="/src/api/resources/tts/client/Client.ts">synthesizeFileStreaming</a>({ ...params }) -> stream.Readable</code></summary>
|
|
247
250
|
<dl>
|
|
248
251
|
<dd>
|
|
249
252
|
|
|
@@ -257,8 +260,6 @@ await client.tts.synthesizeFileStreaming({
|
|
|
257
260
|
|
|
258
261
|
Streams synthesized speech using the specified voice. If no voice is provided, a novel voice will be generated dynamically. Optionally, additional context can be included to influence the speech's style and prosody.
|
|
259
262
|
|
|
260
|
-
The response is a stream of JSON objects including audio encoded in base64.
|
|
261
|
-
|
|
262
263
|
</dd>
|
|
263
264
|
</dl>
|
|
264
265
|
</dd>
|
|
@@ -273,7 +274,7 @@ The response is a stream of JSON objects including audio encoded in base64.
|
|
|
273
274
|
<dd>
|
|
274
275
|
|
|
275
276
|
```typescript
|
|
276
|
-
|
|
277
|
+
await client.tts.synthesizeFileStreaming({
|
|
277
278
|
utterances: [
|
|
278
279
|
{
|
|
279
280
|
text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
|
|
@@ -284,9 +285,6 @@ const response = await client.tts.synthesizeJsonStreaming({
|
|
|
284
285
|
},
|
|
285
286
|
],
|
|
286
287
|
});
|
|
287
|
-
for await (const item of response) {
|
|
288
|
-
console.log(item);
|
|
289
|
-
}
|
|
290
288
|
```
|
|
291
289
|
|
|
292
290
|
</dd>
|
|
@@ -532,58 +530,6 @@ await client.tts.voices.delete({
|
|
|
532
530
|
</dl>
|
|
533
531
|
</details>
|
|
534
532
|
|
|
535
|
-
## EmpathicVoice
|
|
536
|
-
|
|
537
|
-
<details><summary><code>client.empathicVoice.<a href="/src/api/resources/empathicVoice/client/Client.ts">customLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost</a>({ ...params }) -> Hume.SupportsToolUse</code></summary>
|
|
538
|
-
<dl>
|
|
539
|
-
<dd>
|
|
540
|
-
|
|
541
|
-
#### 🔌 Usage
|
|
542
|
-
|
|
543
|
-
<dl>
|
|
544
|
-
<dd>
|
|
545
|
-
|
|
546
|
-
<dl>
|
|
547
|
-
<dd>
|
|
548
|
-
|
|
549
|
-
```typescript
|
|
550
|
-
await client.empathicVoice.customLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost({
|
|
551
|
-
modelResource: "model_resource",
|
|
552
|
-
});
|
|
553
|
-
```
|
|
554
|
-
|
|
555
|
-
</dd>
|
|
556
|
-
</dl>
|
|
557
|
-
</dd>
|
|
558
|
-
</dl>
|
|
559
|
-
|
|
560
|
-
#### ⚙️ Parameters
|
|
561
|
-
|
|
562
|
-
<dl>
|
|
563
|
-
<dd>
|
|
564
|
-
|
|
565
|
-
<dl>
|
|
566
|
-
<dd>
|
|
567
|
-
|
|
568
|
-
**request:** `Hume.empathicVoice.BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost`
|
|
569
|
-
|
|
570
|
-
</dd>
|
|
571
|
-
</dl>
|
|
572
|
-
|
|
573
|
-
<dl>
|
|
574
|
-
<dd>
|
|
575
|
-
|
|
576
|
-
**requestOptions:** `EmpathicVoice.RequestOptions`
|
|
577
|
-
|
|
578
|
-
</dd>
|
|
579
|
-
</dl>
|
|
580
|
-
</dd>
|
|
581
|
-
</dl>
|
|
582
|
-
|
|
583
|
-
</dd>
|
|
584
|
-
</dl>
|
|
585
|
-
</details>
|
|
586
|
-
|
|
587
533
|
## EmpathicVoice Tools
|
|
588
534
|
|
|
589
535
|
<details><summary><code>client.empathicVoice.tools.<a href="/src/api/resources/empathicVoice/resources/tools/client/Client.ts">listTools</a>({ ...params }) -> core.Page<Hume.ReturnUserDefinedTool | undefined></code></summary>
|
|
@@ -16,4 +16,3 @@ var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
|
16
16
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
17
|
__exportStar(require("./resources"), exports);
|
|
18
18
|
__exportStar(require("./types"), exports);
|
|
19
|
-
__exportStar(require("./client"), exports);
|
|
@@ -7,7 +7,7 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const AssistantEnd: core.serialization.ObjectSchema<serializers.empathicVoice.AssistantEnd.Raw, Hume.empathicVoice.AssistantEnd>;
|
|
8
8
|
export declare namespace AssistantEnd {
|
|
9
9
|
interface Raw {
|
|
10
|
-
custom_session_id?: string | null;
|
|
11
10
|
type: "assistant_end";
|
|
11
|
+
custom_session_id?: string | null;
|
|
12
12
|
}
|
|
13
13
|
}
|
|
@@ -39,6 +39,6 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.AssistantEnd = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.AssistantEnd = core.serialization.object({
|
|
42
|
-
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
43
42
|
type: core.serialization.stringLiteral("assistant_end"),
|
|
43
|
+
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
44
44
|
});
|
|
@@ -7,8 +7,8 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const AssistantInput: core.serialization.ObjectSchema<serializers.empathicVoice.AssistantInput.Raw, Hume.empathicVoice.AssistantInput>;
|
|
8
8
|
export declare namespace AssistantInput {
|
|
9
9
|
interface Raw {
|
|
10
|
+
type: "assistant_input";
|
|
10
11
|
custom_session_id?: string | null;
|
|
11
12
|
text: string;
|
|
12
|
-
type: "assistant_input";
|
|
13
13
|
}
|
|
14
14
|
}
|
|
@@ -39,7 +39,7 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.AssistantInput = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.AssistantInput = core.serialization.object({
|
|
42
|
+
type: core.serialization.stringLiteral("assistant_input"),
|
|
42
43
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
43
44
|
text: core.serialization.string(),
|
|
44
|
-
type: core.serialization.stringLiteral("assistant_input"),
|
|
45
45
|
});
|
|
@@ -9,11 +9,11 @@ import { Inference } from "./Inference";
|
|
|
9
9
|
export declare const AssistantMessage: core.serialization.ObjectSchema<serializers.empathicVoice.AssistantMessage.Raw, Hume.empathicVoice.AssistantMessage>;
|
|
10
10
|
export declare namespace AssistantMessage {
|
|
11
11
|
interface Raw {
|
|
12
|
+
type: "assistant_message";
|
|
12
13
|
custom_session_id?: string | null;
|
|
13
|
-
from_text: boolean;
|
|
14
14
|
id?: string | null;
|
|
15
15
|
message: ChatMessage.Raw;
|
|
16
16
|
models: Inference.Raw;
|
|
17
|
-
|
|
17
|
+
from_text: boolean;
|
|
18
18
|
}
|
|
19
19
|
}
|
|
@@ -41,10 +41,10 @@ const core = __importStar(require("../../../../core"));
|
|
|
41
41
|
const ChatMessage_1 = require("./ChatMessage");
|
|
42
42
|
const Inference_1 = require("./Inference");
|
|
43
43
|
exports.AssistantMessage = core.serialization.object({
|
|
44
|
+
type: core.serialization.stringLiteral("assistant_message"),
|
|
44
45
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
45
|
-
fromText: core.serialization.property("from_text", core.serialization.boolean()),
|
|
46
46
|
id: core.serialization.string().optional(),
|
|
47
47
|
message: ChatMessage_1.ChatMessage,
|
|
48
48
|
models: Inference_1.Inference,
|
|
49
|
-
|
|
49
|
+
fromText: core.serialization.property("from_text", core.serialization.boolean()),
|
|
50
50
|
});
|
|
@@ -8,9 +8,9 @@ import { Inference } from "./Inference";
|
|
|
8
8
|
export declare const AssistantProsody: core.serialization.ObjectSchema<serializers.empathicVoice.AssistantProsody.Raw, Hume.empathicVoice.AssistantProsody>;
|
|
9
9
|
export declare namespace AssistantProsody {
|
|
10
10
|
interface Raw {
|
|
11
|
+
type: "assistant_prosody";
|
|
11
12
|
custom_session_id?: string | null;
|
|
12
|
-
id?: string | null;
|
|
13
13
|
models: Inference.Raw;
|
|
14
|
-
|
|
14
|
+
id?: string | null;
|
|
15
15
|
}
|
|
16
16
|
}
|
|
@@ -40,8 +40,8 @@ exports.AssistantProsody = void 0;
|
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const Inference_1 = require("./Inference");
|
|
42
42
|
exports.AssistantProsody = core.serialization.object({
|
|
43
|
+
type: core.serialization.stringLiteral("assistant_prosody"),
|
|
43
44
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
44
|
-
id: core.serialization.string().optional(),
|
|
45
45
|
models: Inference_1.Inference,
|
|
46
|
-
|
|
46
|
+
id: core.serialization.string().optional(),
|
|
47
47
|
});
|
|
@@ -8,8 +8,8 @@ import { Encoding } from "./Encoding";
|
|
|
8
8
|
export declare const AudioConfiguration: core.serialization.ObjectSchema<serializers.empathicVoice.AudioConfiguration.Raw, Hume.empathicVoice.AudioConfiguration>;
|
|
9
9
|
export declare namespace AudioConfiguration {
|
|
10
10
|
interface Raw {
|
|
11
|
-
channels: number;
|
|
12
11
|
encoding: Encoding.Raw;
|
|
12
|
+
channels: number;
|
|
13
13
|
sample_rate: number;
|
|
14
14
|
}
|
|
15
15
|
}
|
|
@@ -40,7 +40,7 @@ exports.AudioConfiguration = void 0;
|
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const Encoding_1 = require("./Encoding");
|
|
42
42
|
exports.AudioConfiguration = core.serialization.object({
|
|
43
|
-
channels: core.serialization.number(),
|
|
44
43
|
encoding: Encoding_1.Encoding,
|
|
44
|
+
channels: core.serialization.number(),
|
|
45
45
|
sampleRate: core.serialization.property("sample_rate", core.serialization.number()),
|
|
46
46
|
});
|
|
@@ -7,8 +7,8 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const AudioInput: core.serialization.ObjectSchema<serializers.empathicVoice.AudioInput.Raw, Hume.empathicVoice.AudioInput>;
|
|
8
8
|
export declare namespace AudioInput {
|
|
9
9
|
interface Raw {
|
|
10
|
+
type: "audio_input";
|
|
10
11
|
custom_session_id?: string | null;
|
|
11
12
|
data: string;
|
|
12
|
-
type: "audio_input";
|
|
13
13
|
}
|
|
14
14
|
}
|
|
@@ -39,7 +39,7 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.AudioInput = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.AudioInput = core.serialization.object({
|
|
42
|
+
type: core.serialization.stringLiteral("audio_input"),
|
|
42
43
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
43
44
|
data: core.serialization.string(),
|
|
44
|
-
type: core.serialization.stringLiteral("audio_input"),
|
|
45
45
|
});
|
|
@@ -7,10 +7,10 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const AudioOutput: core.serialization.ObjectSchema<serializers.empathicVoice.AudioOutput.Raw, Hume.empathicVoice.AudioOutput>;
|
|
8
8
|
export declare namespace AudioOutput {
|
|
9
9
|
interface Raw {
|
|
10
|
+
type: "audio_output";
|
|
10
11
|
custom_session_id?: string | null;
|
|
11
|
-
data: string;
|
|
12
12
|
id: string;
|
|
13
13
|
index: number;
|
|
14
|
-
|
|
14
|
+
data: string;
|
|
15
15
|
}
|
|
16
16
|
}
|
|
@@ -39,9 +39,9 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.AudioOutput = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.AudioOutput = core.serialization.object({
|
|
42
|
+
type: core.serialization.stringLiteral("audio_output"),
|
|
42
43
|
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
43
|
-
data: core.serialization.string(),
|
|
44
44
|
id: core.serialization.string(),
|
|
45
45
|
index: core.serialization.number(),
|
|
46
|
-
|
|
46
|
+
data: core.serialization.string(),
|
|
47
47
|
});
|
|
@@ -8,7 +8,7 @@ import { BuiltInTool } from "./BuiltInTool";
|
|
|
8
8
|
export declare const BuiltinToolConfig: core.serialization.ObjectSchema<serializers.empathicVoice.BuiltinToolConfig.Raw, Hume.empathicVoice.BuiltinToolConfig>;
|
|
9
9
|
export declare namespace BuiltinToolConfig {
|
|
10
10
|
interface Raw {
|
|
11
|
-
fallback_content?: string | null;
|
|
12
11
|
name: BuiltInTool.Raw;
|
|
12
|
+
fallback_content?: string | null;
|
|
13
13
|
}
|
|
14
14
|
}
|
|
@@ -40,6 +40,6 @@ exports.BuiltinToolConfig = void 0;
|
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const BuiltInTool_1 = require("./BuiltInTool");
|
|
42
42
|
exports.BuiltinToolConfig = core.serialization.object({
|
|
43
|
-
fallbackContent: core.serialization.property("fallback_content", core.serialization.string().optional()),
|
|
44
43
|
name: BuiltInTool_1.BuiltInTool,
|
|
44
|
+
fallbackContent: core.serialization.property("fallback_content", core.serialization.string().optional()),
|
|
45
45
|
});
|
|
@@ -10,8 +10,8 @@ import { ChatMessageToolResult } from "./ChatMessageToolResult";
|
|
|
10
10
|
export declare const ChatMessage: core.serialization.ObjectSchema<serializers.empathicVoice.ChatMessage.Raw, Hume.empathicVoice.ChatMessage>;
|
|
11
11
|
export declare namespace ChatMessage {
|
|
12
12
|
interface Raw {
|
|
13
|
-
content?: string | null;
|
|
14
13
|
role: Role.Raw;
|
|
14
|
+
content?: string | null;
|
|
15
15
|
tool_call?: ToolCallMessage.Raw | null;
|
|
16
16
|
tool_result?: ChatMessageToolResult.Raw | null;
|
|
17
17
|
}
|
|
@@ -42,8 +42,8 @@ const Role_1 = require("./Role");
|
|
|
42
42
|
const ToolCallMessage_1 = require("./ToolCallMessage");
|
|
43
43
|
const ChatMessageToolResult_1 = require("./ChatMessageToolResult");
|
|
44
44
|
exports.ChatMessage = core.serialization.object({
|
|
45
|
-
content: core.serialization.string().optional(),
|
|
46
45
|
role: Role_1.Role,
|
|
46
|
+
content: core.serialization.string().optional(),
|
|
47
47
|
toolCall: core.serialization.property("tool_call", ToolCallMessage_1.ToolCallMessage.optional()),
|
|
48
48
|
toolResult: core.serialization.property("tool_result", ChatMessageToolResult_1.ChatMessageToolResult.optional()),
|
|
49
49
|
});
|
|
@@ -7,10 +7,10 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const ChatMetadata: core.serialization.ObjectSchema<serializers.empathicVoice.ChatMetadata.Raw, Hume.empathicVoice.ChatMetadata>;
|
|
8
8
|
export declare namespace ChatMetadata {
|
|
9
9
|
interface Raw {
|
|
10
|
+
type: "chat_metadata";
|
|
11
|
+
custom_session_id?: string | null;
|
|
10
12
|
chat_group_id: string;
|
|
11
13
|
chat_id: string;
|
|
12
|
-
custom_session_id?: string | null;
|
|
13
14
|
request_id?: string | null;
|
|
14
|
-
type: "chat_metadata";
|
|
15
15
|
}
|
|
16
16
|
}
|
|
@@ -39,9 +39,9 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.ChatMetadata = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.ChatMetadata = core.serialization.object({
|
|
42
|
+
type: core.serialization.stringLiteral("chat_metadata"),
|
|
43
|
+
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
42
44
|
chatGroupId: core.serialization.property("chat_group_id", core.serialization.string()),
|
|
43
45
|
chatId: core.serialization.property("chat_id", core.serialization.string()),
|
|
44
|
-
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
45
46
|
requestId: core.serialization.property("request_id", core.serialization.string().optional()),
|
|
46
|
-
type: core.serialization.stringLiteral("chat_metadata"),
|
|
47
47
|
});
|
|
@@ -8,7 +8,7 @@ import { ContextType } from "./ContextType";
|
|
|
8
8
|
export declare const Context: core.serialization.ObjectSchema<serializers.empathicVoice.Context.Raw, Hume.empathicVoice.Context>;
|
|
9
9
|
export declare namespace Context {
|
|
10
10
|
interface Raw {
|
|
11
|
-
text: string;
|
|
12
11
|
type?: ContextType.Raw | null;
|
|
12
|
+
text: string;
|
|
13
13
|
}
|
|
14
14
|
}
|
|
@@ -40,6 +40,6 @@ exports.Context = void 0;
|
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
const ContextType_1 = require("./ContextType");
|
|
42
42
|
exports.Context = core.serialization.object({
|
|
43
|
-
text: core.serialization.string(),
|
|
44
43
|
type: ContextType_1.ContextType.optional(),
|
|
44
|
+
text: core.serialization.string(),
|
|
45
45
|
});
|
|
@@ -6,5 +6,5 @@ import * as Hume from "../../../../api/index";
|
|
|
6
6
|
import * as core from "../../../../core";
|
|
7
7
|
export declare const ContextType: core.serialization.Schema<serializers.empathicVoice.ContextType.Raw, Hume.empathicVoice.ContextType>;
|
|
8
8
|
export declare namespace ContextType {
|
|
9
|
-
type Raw = "
|
|
9
|
+
type Raw = "temporary" | "persistent";
|
|
10
10
|
}
|
|
@@ -38,4 +38,4 @@ var __importStar = (this && this.__importStar) || (function () {
|
|
|
38
38
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
39
39
|
exports.ContextType = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
|
-
exports.ContextType = core.serialization.enum_(["
|
|
41
|
+
exports.ContextType = core.serialization.enum_(["temporary", "persistent"]);
|
|
@@ -6,5 +6,5 @@ import * as Hume from "../../../../api/index";
|
|
|
6
6
|
import * as core from "../../../../core";
|
|
7
7
|
export declare const LanguageModelType: core.serialization.Schema<serializers.empathicVoice.LanguageModelType.Raw, Hume.empathicVoice.LanguageModelType>;
|
|
8
8
|
export declare namespace LanguageModelType {
|
|
9
|
-
type Raw = "claude-3-7-sonnet-latest" | "claude-3-5-sonnet-latest" | "claude-3-5-haiku-latest" | "claude-3-5-sonnet-20240620" | "claude-3-opus-20240229" | "claude-3-sonnet-20240229" | "claude-3-haiku-20240307" | "claude-sonnet-4-20250514" | "us.anthropic.claude-3-5-haiku-20241022-v1:0" | "us.anthropic.claude-3-5-sonnet-20240620-v1:0" | "us.anthropic.claude-3-haiku-20240307-v1:0" | "gpt-oss-120b" | "qwen-3-235b-a22b" | "qwen-3-235b-a22b-instruct-2507" | "qwen-3-235b-a22b-thinking-2507" | "gemini-1.5-pro" | "gemini-1.5-flash" | "gemini-1.5-pro-002" | "gemini-1.5-flash-002" | "gemini-2.0-flash" | "gemini-2.5-flash" | "gemini-2.5-flash-preview-04-17" | "gpt-4-turbo" | "gpt-4-turbo-preview" | "gpt-3.5-turbo-0125" | "gpt-3.5-turbo" | "gpt-4o" | "gpt-4o-mini" | "gpt-4.1" | "gemma-7b-it" | "llama3-8b-8192" | "llama3-70b-8192" | "llama-3.1-70b-versatile" | "llama-3.3-70b-versatile" | "llama-3.1-8b-instant" | "moonshotai/kimi-k2-instruct" | "accounts/fireworks/models/mixtral-8x7b-instruct" | "accounts/fireworks/models/llama-v3p1-405b-instruct" | "accounts/fireworks/models/llama-v3p1-70b-instruct" | "accounts/fireworks/models/llama-v3p1-8b-instruct" | "sonar" | "sonar-pro" | "sambanova" | "DeepSeek-R1-Distill-Llama-70B" | "Llama-4-Maverick-17B-128E-Instruct" | "Qwen3-32B" | "ellm" | "custom-language-model" | "hume-evi-3-web-search";
|
|
9
|
+
type Raw = "claude-3-7-sonnet-latest" | "claude-3-5-sonnet-latest" | "claude-3-5-haiku-latest" | "claude-3-5-sonnet-20240620" | "claude-3-opus-20240229" | "claude-3-sonnet-20240229" | "claude-3-haiku-20240307" | "claude-sonnet-4-20250514" | "us.anthropic.claude-3-5-haiku-20241022-v1:0" | "us.anthropic.claude-3-5-sonnet-20240620-v1:0" | "us.anthropic.claude-3-haiku-20240307-v1:0" | "gpt-oss-120b" | "qwen-3-235b-a22b" | "qwen-3-235b-a22b-instruct-2507" | "qwen-3-235b-a22b-thinking-2507" | "gemini-1.5-pro" | "gemini-1.5-flash" | "gemini-1.5-pro-002" | "gemini-1.5-flash-002" | "gemini-2.0-flash" | "gemini-2.5-flash" | "gemini-2.5-flash-preview-04-17" | "gpt-4-turbo" | "gpt-4-turbo-preview" | "gpt-3.5-turbo-0125" | "gpt-3.5-turbo" | "gpt-4o" | "gpt-4o-mini" | "gpt-4.1" | "gpt-5" | "gpt-5-mini" | "gpt-5-nano" | "gemma-7b-it" | "llama3-8b-8192" | "llama3-70b-8192" | "llama-3.1-70b-versatile" | "llama-3.3-70b-versatile" | "llama-3.1-8b-instant" | "moonshotai/kimi-k2-instruct" | "accounts/fireworks/models/mixtral-8x7b-instruct" | "accounts/fireworks/models/llama-v3p1-405b-instruct" | "accounts/fireworks/models/llama-v3p1-70b-instruct" | "accounts/fireworks/models/llama-v3p1-8b-instruct" | "sonar" | "sonar-pro" | "sambanova" | "DeepSeek-R1-Distill-Llama-70B" | "Llama-4-Maverick-17B-128E-Instruct" | "Qwen3-32B" | "ellm" | "custom-language-model" | "hume-evi-3-web-search";
|
|
10
10
|
}
|
|
@@ -7,7 +7,7 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const PauseAssistantMessage: core.serialization.ObjectSchema<serializers.empathicVoice.PauseAssistantMessage.Raw, Hume.empathicVoice.PauseAssistantMessage>;
|
|
8
8
|
export declare namespace PauseAssistantMessage {
|
|
9
9
|
interface Raw {
|
|
10
|
-
custom_session_id?: string | null;
|
|
11
10
|
type: "pause_assistant_message";
|
|
11
|
+
custom_session_id?: string | null;
|
|
12
12
|
}
|
|
13
13
|
}
|
|
@@ -39,6 +39,6 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.PauseAssistantMessage = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.PauseAssistantMessage = core.serialization.object({
|
|
42
|
-
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
43
42
|
type: core.serialization.stringLiteral("pause_assistant_message"),
|
|
43
|
+
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
44
44
|
});
|
|
@@ -7,7 +7,7 @@ import * as core from "../../../../core";
|
|
|
7
7
|
export declare const ResumeAssistantMessage: core.serialization.ObjectSchema<serializers.empathicVoice.ResumeAssistantMessage.Raw, Hume.empathicVoice.ResumeAssistantMessage>;
|
|
8
8
|
export declare namespace ResumeAssistantMessage {
|
|
9
9
|
interface Raw {
|
|
10
|
-
custom_session_id?: string | null;
|
|
11
10
|
type: "resume_assistant_message";
|
|
11
|
+
custom_session_id?: string | null;
|
|
12
12
|
}
|
|
13
13
|
}
|
|
@@ -39,6 +39,6 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
39
39
|
exports.ResumeAssistantMessage = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
41
|
exports.ResumeAssistantMessage = core.serialization.object({
|
|
42
|
-
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
43
42
|
type: core.serialization.stringLiteral("resume_assistant_message"),
|
|
43
|
+
customSessionId: core.serialization.property("custom_session_id", core.serialization.string().optional()),
|
|
44
44
|
});
|
|
@@ -4,34 +4,34 @@
|
|
|
4
4
|
import * as serializers from "../../../index";
|
|
5
5
|
import * as Hume from "../../../../api/index";
|
|
6
6
|
import * as core from "../../../../core";
|
|
7
|
+
import { ReturnUserDefinedTool } from "./ReturnUserDefinedTool";
|
|
7
8
|
import { ReturnLanguageModel } from "./ReturnLanguageModel";
|
|
8
9
|
import { ReturnBuiltinTool } from "./ReturnBuiltinTool";
|
|
9
10
|
import { ReturnTimeoutSpecs } from "./ReturnTimeoutSpecs";
|
|
11
|
+
import { ReturnNudgeSpec } from "./ReturnNudgeSpec";
|
|
10
12
|
import { ReturnEventMessageSpecs } from "./ReturnEventMessageSpecs";
|
|
11
13
|
import { ReturnEllmModel } from "./ReturnEllmModel";
|
|
12
|
-
import { ReturnWebhookSpec } from "./ReturnWebhookSpec";
|
|
13
|
-
import { ReturnNudgeSpec } from "./ReturnNudgeSpec";
|
|
14
14
|
import { ReturnPrompt } from "./ReturnPrompt";
|
|
15
|
-
import {
|
|
15
|
+
import { ReturnWebhookSpec } from "./ReturnWebhookSpec";
|
|
16
16
|
export declare const ReturnConfig: core.serialization.ObjectSchema<serializers.empathicVoice.ReturnConfig.Raw, Hume.empathicVoice.ReturnConfig>;
|
|
17
17
|
export declare namespace ReturnConfig {
|
|
18
18
|
interface Raw {
|
|
19
19
|
name?: string | null;
|
|
20
20
|
id?: string | null;
|
|
21
21
|
version?: number | null;
|
|
22
|
+
tools?: (ReturnUserDefinedTool.Raw | null | undefined)[] | null;
|
|
23
|
+
version_description?: string | null;
|
|
22
24
|
language_model?: ReturnLanguageModel.Raw | null;
|
|
23
25
|
builtin_tools?: (ReturnBuiltinTool.Raw | null | undefined)[] | null;
|
|
24
26
|
evi_version?: string | null;
|
|
25
27
|
timeouts?: ReturnTimeoutSpecs.Raw | null;
|
|
28
|
+
nudges?: ReturnNudgeSpec.Raw | null;
|
|
26
29
|
event_messages?: ReturnEventMessageSpecs.Raw | null;
|
|
27
30
|
ellm_model?: ReturnEllmModel.Raw | null;
|
|
31
|
+
voice?: unknown | null;
|
|
32
|
+
prompt?: ReturnPrompt.Raw | null;
|
|
28
33
|
webhooks?: (ReturnWebhookSpec.Raw | null | undefined)[] | null;
|
|
29
|
-
version_description?: string | null;
|
|
30
34
|
created_on?: number | null;
|
|
31
35
|
modified_on?: number | null;
|
|
32
|
-
nudges?: ReturnNudgeSpec.Raw | null;
|
|
33
|
-
voice?: unknown | null;
|
|
34
|
-
prompt?: ReturnPrompt.Raw | null;
|
|
35
|
-
tools?: (ReturnUserDefinedTool.Raw | null | undefined)[] | null;
|
|
36
36
|
}
|
|
37
37
|
}
|
|
@@ -38,31 +38,31 @@ var __importStar = (this && this.__importStar) || (function () {
|
|
|
38
38
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
39
39
|
exports.ReturnConfig = void 0;
|
|
40
40
|
const core = __importStar(require("../../../../core"));
|
|
41
|
+
const ReturnUserDefinedTool_1 = require("./ReturnUserDefinedTool");
|
|
41
42
|
const ReturnLanguageModel_1 = require("./ReturnLanguageModel");
|
|
42
43
|
const ReturnBuiltinTool_1 = require("./ReturnBuiltinTool");
|
|
43
44
|
const ReturnTimeoutSpecs_1 = require("./ReturnTimeoutSpecs");
|
|
45
|
+
const ReturnNudgeSpec_1 = require("./ReturnNudgeSpec");
|
|
44
46
|
const ReturnEventMessageSpecs_1 = require("./ReturnEventMessageSpecs");
|
|
45
47
|
const ReturnEllmModel_1 = require("./ReturnEllmModel");
|
|
46
|
-
const ReturnWebhookSpec_1 = require("./ReturnWebhookSpec");
|
|
47
|
-
const ReturnNudgeSpec_1 = require("./ReturnNudgeSpec");
|
|
48
48
|
const ReturnPrompt_1 = require("./ReturnPrompt");
|
|
49
|
-
const
|
|
49
|
+
const ReturnWebhookSpec_1 = require("./ReturnWebhookSpec");
|
|
50
50
|
exports.ReturnConfig = core.serialization.object({
|
|
51
51
|
name: core.serialization.string().optional(),
|
|
52
52
|
id: core.serialization.string().optional(),
|
|
53
53
|
version: core.serialization.number().optional(),
|
|
54
|
+
tools: core.serialization.list(ReturnUserDefinedTool_1.ReturnUserDefinedTool.optional()).optional(),
|
|
55
|
+
versionDescription: core.serialization.property("version_description", core.serialization.string().optional()),
|
|
54
56
|
languageModel: core.serialization.property("language_model", ReturnLanguageModel_1.ReturnLanguageModel.optional()),
|
|
55
57
|
builtinTools: core.serialization.property("builtin_tools", core.serialization.list(ReturnBuiltinTool_1.ReturnBuiltinTool.optional()).optional()),
|
|
56
58
|
eviVersion: core.serialization.property("evi_version", core.serialization.string().optional()),
|
|
57
59
|
timeouts: ReturnTimeoutSpecs_1.ReturnTimeoutSpecs.optional(),
|
|
60
|
+
nudges: ReturnNudgeSpec_1.ReturnNudgeSpec.optional(),
|
|
58
61
|
eventMessages: core.serialization.property("event_messages", ReturnEventMessageSpecs_1.ReturnEventMessageSpecs.optional()),
|
|
59
62
|
ellmModel: core.serialization.property("ellm_model", ReturnEllmModel_1.ReturnEllmModel.optional()),
|
|
63
|
+
voice: core.serialization.unknown().optional(),
|
|
64
|
+
prompt: ReturnPrompt_1.ReturnPrompt.optional(),
|
|
60
65
|
webhooks: core.serialization.list(ReturnWebhookSpec_1.ReturnWebhookSpec.optional()).optional(),
|
|
61
|
-
versionDescription: core.serialization.property("version_description", core.serialization.string().optional()),
|
|
62
66
|
createdOn: core.serialization.property("created_on", core.serialization.number().optional()),
|
|
63
67
|
modifiedOn: core.serialization.property("modified_on", core.serialization.number().optional()),
|
|
64
|
-
nudges: ReturnNudgeSpec_1.ReturnNudgeSpec.optional(),
|
|
65
|
-
voice: core.serialization.unknown().optional(),
|
|
66
|
-
prompt: ReturnPrompt_1.ReturnPrompt.optional(),
|
|
67
|
-
tools: core.serialization.list(ReturnUserDefinedTool_1.ReturnUserDefinedTool.optional()).optional(),
|
|
68
68
|
});
|
|
@@ -12,8 +12,8 @@ export declare namespace ReturnPrompt {
|
|
|
12
12
|
id: string;
|
|
13
13
|
text: string;
|
|
14
14
|
version: number;
|
|
15
|
-
version_type: ReturnPromptVersionType.Raw;
|
|
16
15
|
version_description?: string | null;
|
|
16
|
+
version_type: ReturnPromptVersionType.Raw;
|
|
17
17
|
created_on: number;
|
|
18
18
|
modified_on: number;
|
|
19
19
|
}
|
|
@@ -44,8 +44,8 @@ exports.ReturnPrompt = core.serialization.object({
|
|
|
44
44
|
id: core.serialization.string(),
|
|
45
45
|
text: core.serialization.string(),
|
|
46
46
|
version: core.serialization.number(),
|
|
47
|
-
versionType: core.serialization.property("version_type", ReturnPromptVersionType_1.ReturnPromptVersionType),
|
|
48
47
|
versionDescription: core.serialization.property("version_description", core.serialization.string().optional()),
|
|
48
|
+
versionType: core.serialization.property("version_type", ReturnPromptVersionType_1.ReturnPromptVersionType),
|
|
49
49
|
createdOn: core.serialization.property("created_on", core.serialization.number()),
|
|
50
50
|
modifiedOn: core.serialization.property("modified_on", core.serialization.number()),
|
|
51
51
|
});
|