hume 0.12.2 → 0.13.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (289) hide show
  1. package/.mock/definition/empathic-voice/__package__.yml +667 -712
  2. package/.mock/definition/empathic-voice/chat.yml +29 -23
  3. package/.mock/definition/empathic-voice/chatWebhooks.yml +3 -3
  4. package/.mock/definition/tts/__package__.yml +70 -87
  5. package/.mock/fern.config.json +1 -1
  6. package/api/resources/empathicVoice/client/index.d.ts +0 -1
  7. package/api/resources/empathicVoice/client/index.js +0 -15
  8. package/api/resources/empathicVoice/errors/index.d.ts +0 -1
  9. package/api/resources/empathicVoice/errors/index.js +0 -1
  10. package/api/resources/empathicVoice/types/AssistantEnd.d.ts +2 -2
  11. package/api/resources/empathicVoice/types/AssistantInput.d.ts +2 -2
  12. package/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
  13. package/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
  14. package/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
  15. package/api/resources/empathicVoice/types/AudioInput.d.ts +6 -6
  16. package/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
  17. package/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
  18. package/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
  19. package/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
  20. package/api/resources/empathicVoice/types/Context.d.ts +8 -14
  21. package/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
  22. package/api/resources/empathicVoice/types/ContextType.js +1 -2
  23. package/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
  24. package/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
  25. package/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +2 -2
  26. package/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
  27. package/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
  28. package/api/resources/empathicVoice/types/ReturnPrompt.d.ts +2 -2
  29. package/api/resources/empathicVoice/types/SessionSettings.d.ts +29 -29
  30. package/api/resources/empathicVoice/types/Tool.d.ts +6 -6
  31. package/api/resources/empathicVoice/types/ToolCallMessage.d.ts +6 -6
  32. package/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
  33. package/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
  34. package/api/resources/empathicVoice/types/UserInput.d.ts +2 -2
  35. package/api/resources/empathicVoice/types/UserInterruption.d.ts +4 -4
  36. package/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
  37. package/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
  38. package/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
  39. package/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
  40. package/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
  41. package/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
  42. package/api/resources/empathicVoice/types/index.d.ts +16 -17
  43. package/api/resources/empathicVoice/types/index.js +16 -17
  44. package/api/resources/tts/client/Client.d.ts +21 -23
  45. package/api/resources/tts/client/Client.js +50 -58
  46. package/api/resources/tts/client/index.d.ts +0 -1
  47. package/api/resources/tts/client/index.js +0 -15
  48. package/api/resources/tts/types/PostedTts.d.ts +8 -8
  49. package/api/resources/tts/types/PostedUtterance.d.ts +6 -6
  50. package/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
  51. package/api/resources/tts/types/ReturnTts.d.ts +1 -1
  52. package/api/resources/tts/types/Snippet.d.ts +6 -6
  53. package/dist/api/resources/empathicVoice/client/index.d.ts +0 -1
  54. package/dist/api/resources/empathicVoice/client/index.js +0 -15
  55. package/dist/api/resources/empathicVoice/errors/index.d.ts +0 -1
  56. package/dist/api/resources/empathicVoice/errors/index.js +0 -1
  57. package/dist/api/resources/empathicVoice/types/AssistantEnd.d.ts +2 -2
  58. package/dist/api/resources/empathicVoice/types/AssistantInput.d.ts +2 -2
  59. package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +8 -8
  60. package/dist/api/resources/empathicVoice/types/AssistantProsody.d.ts +6 -6
  61. package/dist/api/resources/empathicVoice/types/AudioConfiguration.d.ts +2 -2
  62. package/dist/api/resources/empathicVoice/types/AudioInput.d.ts +6 -6
  63. package/dist/api/resources/empathicVoice/types/AudioOutput.d.ts +4 -4
  64. package/dist/api/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
  65. package/dist/api/resources/empathicVoice/types/ChatMessage.d.ts +2 -2
  66. package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +8 -8
  67. package/dist/api/resources/empathicVoice/types/Context.d.ts +8 -14
  68. package/dist/api/resources/empathicVoice/types/ContextType.d.ts +2 -3
  69. package/dist/api/resources/empathicVoice/types/ContextType.js +1 -2
  70. package/dist/api/resources/empathicVoice/types/LanguageModelType.d.ts +4 -1
  71. package/dist/api/resources/empathicVoice/types/LanguageModelType.js +3 -0
  72. package/dist/api/resources/empathicVoice/types/PauseAssistantMessage.d.ts +2 -2
  73. package/dist/api/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +2 -2
  74. package/dist/api/resources/empathicVoice/types/ReturnConfig.d.ts +7 -7
  75. package/dist/api/resources/empathicVoice/types/ReturnPrompt.d.ts +2 -2
  76. package/dist/api/resources/empathicVoice/types/SessionSettings.d.ts +29 -29
  77. package/dist/api/resources/empathicVoice/types/Tool.d.ts +6 -6
  78. package/dist/api/resources/empathicVoice/types/ToolCallMessage.d.ts +6 -6
  79. package/dist/api/resources/empathicVoice/types/ToolErrorMessage.d.ts +16 -16
  80. package/dist/api/resources/empathicVoice/types/ToolResponseMessage.d.ts +8 -8
  81. package/dist/api/resources/empathicVoice/types/UserInput.d.ts +2 -2
  82. package/dist/api/resources/empathicVoice/types/UserInterruption.d.ts +4 -4
  83. package/dist/api/resources/empathicVoice/types/UserMessage.d.ts +12 -12
  84. package/dist/api/resources/empathicVoice/types/VoiceProvider.d.ts +2 -1
  85. package/dist/api/resources/empathicVoice/types/VoiceProvider.js +1 -0
  86. package/dist/api/resources/empathicVoice/types/WebSocketError.d.ts +10 -10
  87. package/dist/api/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +8 -8
  88. package/dist/api/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +6 -6
  89. package/dist/api/resources/empathicVoice/types/index.d.ts +16 -17
  90. package/dist/api/resources/empathicVoice/types/index.js +16 -17
  91. package/dist/api/resources/tts/client/Client.d.ts +21 -23
  92. package/dist/api/resources/tts/client/Client.js +50 -58
  93. package/dist/api/resources/tts/client/index.d.ts +0 -1
  94. package/dist/api/resources/tts/client/index.js +0 -15
  95. package/dist/api/resources/tts/types/PostedTts.d.ts +8 -8
  96. package/dist/api/resources/tts/types/PostedUtterance.d.ts +6 -6
  97. package/dist/api/resources/tts/types/ReturnGeneration.d.ts +5 -5
  98. package/dist/api/resources/tts/types/ReturnTts.d.ts +1 -1
  99. package/dist/api/resources/tts/types/Snippet.d.ts +6 -6
  100. package/dist/serialization/resources/empathicVoice/index.d.ts +0 -1
  101. package/dist/serialization/resources/empathicVoice/index.js +0 -1
  102. package/dist/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
  103. package/dist/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
  104. package/dist/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
  105. package/dist/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
  106. package/dist/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
  107. package/dist/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
  108. package/dist/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
  109. package/dist/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
  110. package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
  111. package/dist/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
  112. package/dist/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
  113. package/dist/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
  114. package/dist/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
  115. package/dist/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
  116. package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
  117. package/dist/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
  118. package/dist/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
  119. package/dist/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
  120. package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
  121. package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
  122. package/dist/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
  123. package/dist/serialization/resources/empathicVoice/types/Context.js +1 -1
  124. package/dist/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
  125. package/dist/serialization/resources/empathicVoice/types/ContextType.js +1 -1
  126. package/dist/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
  127. package/dist/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
  128. package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
  129. package/dist/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
  130. package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
  131. package/dist/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
  132. package/dist/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
  133. package/dist/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
  134. package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
  135. package/dist/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
  136. package/dist/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
  137. package/dist/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
  138. package/dist/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
  139. package/dist/serialization/resources/empathicVoice/types/Tool.js +3 -3
  140. package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
  141. package/dist/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
  142. package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
  143. package/dist/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
  144. package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
  145. package/dist/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
  146. package/dist/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
  147. package/dist/serialization/resources/empathicVoice/types/UserInput.js +1 -1
  148. package/dist/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
  149. package/dist/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
  150. package/dist/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
  151. package/dist/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
  152. package/dist/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
  153. package/dist/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
  154. package/dist/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
  155. package/dist/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
  156. package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
  157. package/dist/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
  158. package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
  159. package/dist/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
  160. package/dist/serialization/resources/empathicVoice/types/index.d.ts +16 -17
  161. package/dist/serialization/resources/empathicVoice/types/index.js +16 -17
  162. package/dist/serialization/resources/tts/types/PostedTts.d.ts +3 -3
  163. package/dist/serialization/resources/tts/types/PostedTts.js +3 -3
  164. package/dist/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
  165. package/dist/serialization/resources/tts/types/PostedUtterance.js +2 -2
  166. package/dist/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
  167. package/dist/serialization/resources/tts/types/ReturnGeneration.js +3 -3
  168. package/dist/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
  169. package/dist/serialization/resources/tts/types/ReturnTts.js +1 -1
  170. package/dist/serialization/resources/tts/types/Snippet.d.ts +3 -3
  171. package/dist/serialization/resources/tts/types/Snippet.js +3 -3
  172. package/dist/version.d.ts +1 -1
  173. package/dist/version.js +1 -1
  174. package/package.json +1 -1
  175. package/reference.md +24 -78
  176. package/serialization/resources/empathicVoice/index.d.ts +0 -1
  177. package/serialization/resources/empathicVoice/index.js +0 -1
  178. package/serialization/resources/empathicVoice/types/AssistantEnd.d.ts +1 -1
  179. package/serialization/resources/empathicVoice/types/AssistantEnd.js +1 -1
  180. package/serialization/resources/empathicVoice/types/AssistantInput.d.ts +1 -1
  181. package/serialization/resources/empathicVoice/types/AssistantInput.js +1 -1
  182. package/serialization/resources/empathicVoice/types/AssistantMessage.d.ts +2 -2
  183. package/serialization/resources/empathicVoice/types/AssistantMessage.js +2 -2
  184. package/serialization/resources/empathicVoice/types/AssistantProsody.d.ts +2 -2
  185. package/serialization/resources/empathicVoice/types/AssistantProsody.js +2 -2
  186. package/serialization/resources/empathicVoice/types/AudioConfiguration.d.ts +1 -1
  187. package/serialization/resources/empathicVoice/types/AudioConfiguration.js +1 -1
  188. package/serialization/resources/empathicVoice/types/AudioInput.d.ts +1 -1
  189. package/serialization/resources/empathicVoice/types/AudioInput.js +1 -1
  190. package/serialization/resources/empathicVoice/types/AudioOutput.d.ts +2 -2
  191. package/serialization/resources/empathicVoice/types/AudioOutput.js +2 -2
  192. package/serialization/resources/empathicVoice/types/BuiltinToolConfig.d.ts +1 -1
  193. package/serialization/resources/empathicVoice/types/BuiltinToolConfig.js +1 -1
  194. package/serialization/resources/empathicVoice/types/ChatMessage.d.ts +1 -1
  195. package/serialization/resources/empathicVoice/types/ChatMessage.js +1 -1
  196. package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +2 -2
  197. package/serialization/resources/empathicVoice/types/ChatMetadata.js +2 -2
  198. package/serialization/resources/empathicVoice/types/Context.d.ts +1 -1
  199. package/serialization/resources/empathicVoice/types/Context.js +1 -1
  200. package/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
  201. package/serialization/resources/empathicVoice/types/ContextType.js +1 -1
  202. package/serialization/resources/empathicVoice/types/LanguageModelType.d.ts +1 -1
  203. package/serialization/resources/empathicVoice/types/LanguageModelType.js +3 -0
  204. package/serialization/resources/empathicVoice/types/PauseAssistantMessage.d.ts +1 -1
  205. package/serialization/resources/empathicVoice/types/PauseAssistantMessage.js +1 -1
  206. package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.d.ts +1 -1
  207. package/serialization/resources/empathicVoice/types/ResumeAssistantMessage.js +1 -1
  208. package/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +8 -8
  209. package/serialization/resources/empathicVoice/types/ReturnConfig.js +8 -8
  210. package/serialization/resources/empathicVoice/types/ReturnPrompt.d.ts +1 -1
  211. package/serialization/resources/empathicVoice/types/ReturnPrompt.js +1 -1
  212. package/serialization/resources/empathicVoice/types/SessionSettings.d.ts +8 -8
  213. package/serialization/resources/empathicVoice/types/SessionSettings.js +8 -8
  214. package/serialization/resources/empathicVoice/types/Tool.d.ts +3 -3
  215. package/serialization/resources/empathicVoice/types/Tool.js +3 -3
  216. package/serialization/resources/empathicVoice/types/ToolCallMessage.d.ts +3 -3
  217. package/serialization/resources/empathicVoice/types/ToolCallMessage.js +3 -3
  218. package/serialization/resources/empathicVoice/types/ToolErrorMessage.d.ts +6 -6
  219. package/serialization/resources/empathicVoice/types/ToolErrorMessage.js +6 -6
  220. package/serialization/resources/empathicVoice/types/ToolResponseMessage.d.ts +2 -2
  221. package/serialization/resources/empathicVoice/types/ToolResponseMessage.js +2 -2
  222. package/serialization/resources/empathicVoice/types/UserInput.d.ts +1 -1
  223. package/serialization/resources/empathicVoice/types/UserInput.js +1 -1
  224. package/serialization/resources/empathicVoice/types/UserInterruption.d.ts +1 -1
  225. package/serialization/resources/empathicVoice/types/UserInterruption.js +1 -1
  226. package/serialization/resources/empathicVoice/types/UserMessage.d.ts +3 -3
  227. package/serialization/resources/empathicVoice/types/UserMessage.js +3 -3
  228. package/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
  229. package/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
  230. package/serialization/resources/empathicVoice/types/WebSocketError.d.ts +3 -3
  231. package/serialization/resources/empathicVoice/types/WebSocketError.js +3 -3
  232. package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.d.ts +4 -4
  233. package/serialization/resources/empathicVoice/types/WebhookEventChatEnded.js +4 -4
  234. package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.d.ts +3 -3
  235. package/serialization/resources/empathicVoice/types/WebhookEventChatStarted.js +3 -3
  236. package/serialization/resources/empathicVoice/types/index.d.ts +16 -17
  237. package/serialization/resources/empathicVoice/types/index.js +16 -17
  238. package/serialization/resources/tts/types/PostedTts.d.ts +3 -3
  239. package/serialization/resources/tts/types/PostedTts.js +3 -3
  240. package/serialization/resources/tts/types/PostedUtterance.d.ts +2 -2
  241. package/serialization/resources/tts/types/PostedUtterance.js +2 -2
  242. package/serialization/resources/tts/types/ReturnGeneration.d.ts +3 -3
  243. package/serialization/resources/tts/types/ReturnGeneration.js +3 -3
  244. package/serialization/resources/tts/types/ReturnTts.d.ts +1 -1
  245. package/serialization/resources/tts/types/ReturnTts.js +1 -1
  246. package/serialization/resources/tts/types/Snippet.d.ts +3 -3
  247. package/serialization/resources/tts/types/Snippet.js +3 -3
  248. package/version.d.ts +1 -1
  249. package/version.js +1 -1
  250. package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
  251. package/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
  252. package/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
  253. package/api/resources/empathicVoice/client/requests/index.js +0 -2
  254. package/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
  255. package/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
  256. package/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
  257. package/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
  258. package/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
  259. package/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
  260. package/api/resources/tts/client/requests/index.d.ts +0 -1
  261. package/api/resources/tts/client/requests/index.js +0 -2
  262. package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
  263. package/dist/api/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -5
  264. package/dist/api/resources/empathicVoice/client/requests/index.d.ts +0 -1
  265. package/dist/api/resources/empathicVoice/client/requests/index.js +0 -2
  266. package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.d.ts +0 -9
  267. package/dist/api/resources/empathicVoice/errors/UnprocessableEntityError.js +0 -52
  268. package/dist/api/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -7
  269. package/dist/api/resources/empathicVoice/types/SupportsToolUse.js +0 -5
  270. package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.d.ts +0 -36
  271. package/dist/api/resources/tts/client/requests/SynthesizeJsonRequest.js +0 -5
  272. package/dist/api/resources/tts/client/requests/index.d.ts +0 -1
  273. package/dist/api/resources/tts/client/requests/index.js +0 -2
  274. package/dist/serialization/resources/empathicVoice/client/index.d.ts +0 -1
  275. package/dist/serialization/resources/empathicVoice/client/index.js +0 -17
  276. package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
  277. package/dist/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
  278. package/dist/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
  279. package/dist/serialization/resources/empathicVoice/client/requests/index.js +0 -5
  280. package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
  281. package/dist/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
  282. package/serialization/resources/empathicVoice/client/index.d.ts +0 -1
  283. package/serialization/resources/empathicVoice/client/index.js +0 -17
  284. package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.d.ts +0 -12
  285. package/serialization/resources/empathicVoice/client/requests/BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost.js +0 -43
  286. package/serialization/resources/empathicVoice/client/requests/index.d.ts +0 -1
  287. package/serialization/resources/empathicVoice/client/requests/index.js +0 -5
  288. package/serialization/resources/empathicVoice/types/SupportsToolUse.d.ts +0 -13
  289. package/serialization/resources/empathicVoice/types/SupportsToolUse.js +0 -44
@@ -1,73 +1,20 @@
1
1
  errors:
2
- UnprocessableEntityError:
3
- status-code: 422
4
- type: HTTPValidationError
5
- docs: Validation Error
6
- examples:
7
- - value: {}
8
2
  BadRequestError:
9
3
  status-code: 400
10
4
  type: ErrorResponse
11
5
  docs: Bad Request
12
6
  examples:
13
7
  - value: {}
14
- service:
15
- auth: false
16
- base-path: ''
17
- endpoints:
18
- custom_language_model_supports_tool_use_v0_evi_custom_language_model_supports_tool_use_post:
19
- path: /v0/evi/custom_language_model_supports_tool_use
20
- method: POST
21
- auth: true
22
- source:
23
- openapi: evi-openapi.json
24
- display-name: Custom Language Model Supports Tool Use
25
- request:
26
- name: >-
27
- BodyCustomLanguageModelSupportsToolUseV0EviCustomLanguageModelSupportsToolUsePost
28
- body:
29
- properties:
30
- model_resource: string
31
- content-type: application/json
32
- response:
33
- docs: Successful Response
34
- type: SupportsToolUse
35
- status-code: 200
36
- errors:
37
- - UnprocessableEntityError
38
- examples:
39
- - request:
40
- model_resource: model_resource
41
- response:
42
- body:
43
- model_resource: model_resource
44
- supports_tool_use: true
45
- source:
46
- openapi: evi-openapi.json
47
8
  types:
48
- AssistantEnd:
49
- docs: When provided, the output is an assistant end message.
50
- properties:
51
- custom_session_id:
52
- type: optional<string>
53
- docs: >-
54
- Used to manage conversational state, correlate frontend and backend
55
- data, and persist conversations across EVI sessions.
56
- type:
57
- type: literal<"assistant_end">
58
- docs: >-
59
- The type of message sent through the socket; for an Assistant End
60
- message, this must be `assistant_end`.
61
-
62
-
63
- This message indicates the conclusion of the assistant’s response,
64
- signaling that the assistant has finished speaking for the current
65
- conversational turn.
66
- source:
67
- openapi: evi-asyncapi.json
68
9
  AssistantInput:
69
10
  docs: When provided, the input is spoken by EVI.
70
11
  properties:
12
+ type:
13
+ type: literal<"assistant_input">
14
+ docs: >-
15
+ The type of message sent through the socket; must be `assistant_input`
16
+ for our server to correctly identify and process it as an Assistant
17
+ Input message.
71
18
  custom_session_id:
72
19
  type: optional<string>
73
20
  docs: >-
@@ -86,84 +33,16 @@ types:
86
33
  the context of the conversation. The synthesized audio is streamed
87
34
  back to the user as an [Assistant
88
35
  Message](/reference/empathic-voice-interface-evi/chat/chat#receive.AssistantMessage.type).
89
- type:
90
- type: literal<"assistant_input">
91
- docs: >-
92
- The type of message sent through the socket; must be `assistant_input`
93
- for our server to correctly identify and process it as an Assistant
94
- Input message.
95
- source:
96
- openapi: evi-asyncapi.json
97
- AssistantMessage:
98
- docs: When provided, the output is an assistant message.
99
- properties:
100
- custom_session_id:
101
- type: optional<string>
102
- docs: >-
103
- Used to manage conversational state, correlate frontend and backend
104
- data, and persist conversations across EVI sessions.
105
- from_text:
106
- type: boolean
107
- docs: >-
108
- Indicates if this message was inserted into the conversation as text
109
- from an [Assistant Input
110
- message](/reference/empathic-voice-interface-evi/chat/chat#send.AssistantInput.text).
111
- id:
112
- type: optional<string>
113
- docs: >-
114
- ID of the assistant message. Allows the Assistant Message to be
115
- tracked and referenced.
116
- message:
117
- type: ChatMessage
118
- docs: Transcript of the message.
119
- models:
120
- type: Inference
121
- docs: Inference model results.
122
- type:
123
- type: literal<"assistant_message">
124
- docs: >-
125
- The type of message sent through the socket; for an Assistant Message,
126
- this must be `assistant_message`.
127
-
128
-
129
- This message contains both a transcript of the assistant’s response
130
- and the expression measurement predictions of the assistant’s audio
131
- output.
132
- source:
133
- openapi: evi-asyncapi.json
134
- AssistantProsody:
135
- docs: When provided, the output is an Assistant Prosody message.
136
- properties:
137
- custom_session_id:
138
- type: optional<string>
139
- docs: >-
140
- Used to manage conversational state, correlate frontend and backend
141
- data, and persist conversations across EVI sessions.
142
- id:
143
- type: optional<string>
144
- docs: Unique identifier for the segment.
145
- models:
146
- type: Inference
147
- docs: Inference model results.
148
- type:
149
- type: literal<"assistant_prosody">
150
- docs: >-
151
- The type of message sent through the socket; for an Assistant Prosody
152
- message, this must be `assistant_PROSODY`.
153
-
154
-
155
- This message the expression measurement predictions of the assistant's
156
- audio output.
157
36
  source:
158
37
  openapi: evi-asyncapi.json
159
38
  AudioConfiguration:
160
39
  properties:
161
- channels:
162
- type: integer
163
- docs: Number of audio channels.
164
40
  encoding:
165
41
  type: Encoding
166
42
  docs: Encoding format of the audio input, such as `linear16`.
43
+ channels:
44
+ type: integer
45
+ docs: Number of audio channels.
167
46
  sample_rate:
168
47
  type: integer
169
48
  docs: >-
@@ -174,6 +53,17 @@ types:
174
53
  AudioInput:
175
54
  docs: When provided, the input is audio.
176
55
  properties:
56
+ type:
57
+ type: literal<"audio_input">
58
+ docs: >-
59
+ The type of message sent through the socket; must be `audio_input` for
60
+ our server to correctly identify and process it as an Audio Input
61
+ message.
62
+
63
+
64
+ This message is used for sending audio input data to EVI for
65
+ processing and expression measurement. Audio data should be sent as a
66
+ continuous stream, encoded in Base64.
177
67
  custom_session_id:
178
68
  type: optional<string>
179
69
  docs: >-
@@ -196,48 +86,6 @@ types:
196
86
 
197
87
  Hume recommends streaming audio with a buffer window of 20
198
88
  milliseconds (ms), or 100 milliseconds (ms) for web applications.
199
- type:
200
- type: literal<"audio_input">
201
- docs: >-
202
- The type of message sent through the socket; must be `audio_input` for
203
- our server to correctly identify and process it as an Audio Input
204
- message.
205
-
206
-
207
- This message is used for sending audio input data to EVI for
208
- processing and expression measurement. Audio data should be sent as a
209
- continuous stream, encoded in Base64.
210
- source:
211
- openapi: evi-asyncapi.json
212
- AudioOutput:
213
- docs: >-
214
- The type of message sent through the socket; for an Audio Output message,
215
- this must be `audio_output`.
216
- properties:
217
- custom_session_id:
218
- type: optional<string>
219
- docs: >-
220
- Used to manage conversational state, correlate frontend and backend
221
- data, and persist conversations across EVI sessions.
222
- data:
223
- type: string
224
- docs: >-
225
- Base64 encoded audio output. This encoded audio is transmitted to the
226
- client, where it can be decoded and played back as part of the user
227
- interaction.
228
- id:
229
- type: string
230
- docs: >-
231
- ID of the audio output. Allows the Audio Output message to be tracked
232
- and referenced.
233
- index:
234
- type: integer
235
- docs: Index of the chunk of audio relative to the whole audio segment.
236
- type:
237
- type: literal<"audio_output">
238
- docs: >-
239
- The type of message sent through the socket; for an Audio Output
240
- message, this must be `audio_output`.
241
89
  source:
242
90
  openapi: evi-asyncapi.json
243
91
  BuiltInTool:
@@ -251,87 +99,31 @@ types:
251
99
  openapi: evi-asyncapi.json
252
100
  BuiltinToolConfig:
253
101
  properties:
102
+ name:
103
+ type: BuiltInTool
254
104
  fallback_content:
255
105
  type: optional<string>
256
106
  docs: >-
257
107
  Optional text passed to the supplemental LLM if the tool call fails.
258
108
  The LLM then uses this text to generate a response back to the user,
259
109
  ensuring continuity in the conversation.
260
- name:
261
- type: BuiltInTool
262
- source:
263
- openapi: evi-asyncapi.json
264
- ChatMessageToolResult:
265
- discriminated: false
266
- docs: Function call response from client.
267
- union:
268
- - type: ToolResponseMessage
269
- - type: ToolErrorMessage
270
- source:
271
- openapi: evi-asyncapi.json
272
- inline: true
273
- ChatMessage:
274
- properties:
275
- content:
276
- type: optional<string>
277
- docs: Transcript of the message.
278
- role:
279
- type: Role
280
- docs: Role of who is providing the message.
281
- tool_call:
282
- type: optional<ToolCallMessage>
283
- docs: Function call name and arguments.
284
- tool_result:
285
- type: optional<ChatMessageToolResult>
286
- docs: Function call response from client.
287
110
  source:
288
111
  openapi: evi-asyncapi.json
289
- ChatMetadata:
290
- docs: When provided, the output is a chat metadata message.
112
+ Context:
291
113
  properties:
292
- chat_group_id:
293
- type: string
114
+ type:
115
+ type: optional<ContextType>
294
116
  docs: >-
295
- ID of the Chat Group.
296
-
297
-
298
- Used to resume a Chat when passed in the
299
- [resumed_chat_group_id](/reference/empathic-voice-interface-evi/chat/chat#request.query.resumed_chat_group_id)
300
- query parameter of a subsequent connection request. This allows EVI to
301
- continue the conversation from where it left off within the Chat
302
- Group.
117
+ The persistence level of the injected context. Specifies how long the
118
+ injected context will remain active in the session.
303
119
 
304
120
 
305
- Learn more about [supporting chat
306
- resumability](/docs/empathic-voice-interface-evi/faq#does-evi-support-chat-resumability)
307
- from the EVI FAQ.
308
- chat_id:
309
- type: string
310
- docs: >-
311
- ID of the Chat session. Allows the Chat session to be tracked and
312
- referenced.
313
- custom_session_id:
314
- type: optional<string>
315
- docs: >-
316
- Used to manage conversational state, correlate frontend and backend
317
- data, and persist conversations across EVI sessions.
318
- request_id:
319
- type: optional<string>
320
- docs: ID of the initiating request.
321
- type:
322
- type: literal<"chat_metadata">
323
- docs: >-
324
- The type of message sent through the socket; for a Chat Metadata
325
- message, this must be `chat_metadata`.
121
+ - **Temporary**: Context that is only applied to the following
122
+ assistant response.
326
123
 
327
124
 
328
- The Chat Metadata message is the first message you receive after
329
- establishing a connection with EVI and contains important identifiers
330
- for the current Chat session.
331
- source:
332
- openapi: evi-asyncapi.json
333
- Context:
334
- properties:
125
+ - **Persistent**: Context that is applied to all subsequent assistant
126
+ responses for the remainder of the Chat.
335
127
  text:
336
128
  type: string
337
129
  docs: >-
@@ -340,163 +132,29 @@ types:
340
132
  conversation.
341
133
 
342
134
 
343
- This text will be appended to the end of user messages based on the
344
- chosen persistence level. For example, if you want to remind EVI of
345
- its role as a helpful weather assistant, the context you insert will
346
- be appended to the end of user messages as `{Context: You are a
347
- helpful weather assistant}`.
348
- type:
349
- type: optional<ContextType>
350
- docs: >-
351
- The persistence level of the injected context. Specifies how long the
352
- injected context will remain active in the session.
353
-
354
-
355
- There are three possible context types:
356
-
357
-
358
- - **Persistent**: The context is appended to all user messages for the
359
- duration of the session.
360
-
361
-
362
- - **Temporary**: The context is appended only to the next user
363
- message.
364
-
365
- - **Editable**: The original context is updated to reflect the new context.
366
-
367
- If the type is not specified, it will default to `temporary`.
135
+ This text will be appended to the end of
136
+ [user_messages](/reference/speech-to-speech-evi/chat#receive.UserMessage.message.content)
137
+ based on the chosen persistence level. For example, if you want to
138
+ remind EVI of its role as a helpful weather assistant, the context you
139
+ insert will be appended to the end of user messages as `{Context: You
140
+ are a helpful weather assistant}`.
368
141
  source:
369
142
  openapi: evi-asyncapi.json
370
143
  ContextType:
371
144
  enum:
372
- - editable
373
- - persistent
374
145
  - temporary
146
+ - persistent
375
147
  source:
376
148
  openapi: evi-asyncapi.json
377
- EmotionScores:
149
+ Encoding:
150
+ type: literal<"linear16">
151
+ ErrorLevel:
152
+ type: literal<"warn">
153
+ PauseAssistantMessage:
154
+ docs: >-
155
+ Pause responses from EVI. Chat history is still saved and sent after
156
+ resuming.
378
157
  properties:
379
- Admiration: double
380
- Adoration: double
381
- Aesthetic Appreciation: double
382
- Amusement: double
383
- Anger: double
384
- Anxiety: double
385
- Awe: double
386
- Awkwardness: double
387
- Boredom: double
388
- Calmness: double
389
- Concentration: double
390
- Confusion: double
391
- Contemplation: double
392
- Contempt: double
393
- Contentment: double
394
- Craving: double
395
- Desire: double
396
- Determination: double
397
- Disappointment: double
398
- Disgust: double
399
- Distress: double
400
- Doubt: double
401
- Ecstasy: double
402
- Embarrassment: double
403
- Empathic Pain: double
404
- Entrancement: double
405
- Envy: double
406
- Excitement: double
407
- Fear: double
408
- Guilt: double
409
- Horror: double
410
- Interest: double
411
- Joy: double
412
- Love: double
413
- Nostalgia: double
414
- Pain: double
415
- Pride: double
416
- Realization: double
417
- Relief: double
418
- Romance: double
419
- Sadness: double
420
- Satisfaction: double
421
- Shame: double
422
- Surprise (negative): double
423
- Surprise (positive): double
424
- Sympathy: double
425
- Tiredness: double
426
- Triumph: double
427
- source:
428
- openapi: evi-asyncapi.json
429
- Encoding:
430
- type: literal<"linear16">
431
- WebSocketError:
432
- docs: When provided, the output is an error message.
433
- properties:
434
- code:
435
- type: string
436
- docs: Error code. Identifies the type of error encountered.
437
- custom_session_id:
438
- type: optional<string>
439
- docs: >-
440
- Used to manage conversational state, correlate frontend and backend
441
- data, and persist conversations across EVI sessions.
442
- message:
443
- type: string
444
- docs: Detailed description of the error.
445
- request_id:
446
- type: optional<string>
447
- docs: ID of the initiating request.
448
- slug:
449
- type: string
450
- docs: >-
451
- Short, human-readable identifier and description for the error. See a
452
- complete list of error slugs on the [Errors
453
- page](/docs/resources/errors).
454
- type:
455
- type: literal<"error">
456
- docs: >-
457
- The type of message sent through the socket; for a Web Socket Error
458
- message, this must be `error`.
459
-
460
-
461
- This message indicates a disruption in the WebSocket connection, such
462
- as an unexpected disconnection, protocol error, or data transmission
463
- issue.
464
- source:
465
- openapi: evi-asyncapi.json
466
- ErrorLevel:
467
- type: literal<"warn">
468
- Inference:
469
- properties:
470
- prosody:
471
- type: optional<ProsodyInference>
472
- docs: >-
473
- Prosody model inference results.
474
-
475
-
476
- EVI uses the prosody model to measure 48 emotions related to speech
477
- and vocal characteristics within a given expression.
478
- source:
479
- openapi: evi-asyncapi.json
480
- MillisecondInterval:
481
- properties:
482
- begin:
483
- type: integer
484
- docs: Start time of the interval in milliseconds.
485
- end:
486
- type: integer
487
- docs: End time of the interval in milliseconds.
488
- source:
489
- openapi: evi-asyncapi.json
490
- PauseAssistantMessage:
491
- docs: >-
492
- Pause responses from EVI. Chat history is still saved and sent after
493
- resuming.
494
- properties:
495
- custom_session_id:
496
- type: optional<string>
497
- docs: >-
498
- Used to manage conversational state, correlate frontend and backend
499
- data, and persist conversations across EVI sessions.
500
158
  type:
501
159
  type: literal<"pause_assistant_message">
502
160
  docs: >-
@@ -510,24 +168,11 @@ types:
510
168
  message](/reference/empathic-voice-interface-evi/chat/chat#send.ResumeAssistantMessage.type)
511
169
  is sent. When paused, EVI won’t respond, but transcriptions of your
512
170
  audio inputs will still be recorded.
513
- source:
514
- openapi: evi-asyncapi.json
515
- ProsodyInference:
516
- properties:
517
- scores:
518
- type: EmotionScores
171
+ custom_session_id:
172
+ type: optional<string>
519
173
  docs: >-
520
- The confidence scores for 48 emotions within the detected expression
521
- of an audio sample.
522
-
523
-
524
- Scores typically range from 0 to 1, with higher values indicating a
525
- stronger confidence level in the measured attribute.
526
-
527
-
528
- See our guide on [interpreting expression measurement
529
- results](/docs/expression-measurement/faq#how-do-i-interpret-my-results)
530
- to learn more.
174
+ Used to manage conversational state, correlate frontend and backend
175
+ data, and persist conversations across EVI sessions.
531
176
  source:
532
177
  openapi: evi-asyncapi.json
533
178
  ResumeAssistantMessage:
@@ -535,11 +180,6 @@ types:
535
180
  Resume responses from EVI. Chat history sent while paused will now be
536
181
  sent.
537
182
  properties:
538
- custom_session_id:
539
- type: optional<string>
540
- docs: >-
541
- Used to manage conversational state, correlate frontend and backend
542
- data, and persist conversations across EVI sessions.
543
183
  type:
544
184
  type: literal<"resume_assistant_message">
545
185
  docs: >-
@@ -554,15 +194,11 @@ types:
554
194
  then send a `resume_assistant_message`, EVI will respond to the second
555
195
  question and have added the first question to its conversation
556
196
  context.)
557
- source:
558
- openapi: evi-asyncapi.json
559
- Role:
560
- enum:
561
- - assistant
562
- - system
563
- - user
564
- - all
565
- - tool
197
+ custom_session_id:
198
+ type: optional<string>
199
+ docs: >-
200
+ Used to manage conversational state, correlate frontend and backend
201
+ data, and persist conversations across EVI sessions.
566
202
  source:
567
203
  openapi: evi-asyncapi.json
568
204
  SessionSettingsVariablesValue:
@@ -577,48 +213,22 @@ types:
577
213
  SessionSettings:
578
214
  docs: Settings for this chat session.
579
215
  properties:
580
- audio:
581
- type: optional<AudioConfiguration>
582
- docs: >-
583
- Configuration details for the audio input used during the session.
584
- Ensures the audio is being correctly set up for processing.
585
-
586
-
587
- This optional field is only required when the audio input is encoded
588
- in PCM Linear 16 (16-bit, little-endian, signed PCM WAV data). For
589
- detailed instructions on how to configure session settings for PCM
590
- Linear 16 audio, please refer to the [Session Settings
591
- guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
592
- builtin_tools:
593
- type: optional<list<BuiltinToolConfig>>
594
- docs: >-
595
- List of built-in tools to enable for the session.
596
-
597
-
598
- Tools are resources used by EVI to perform various tasks, such as
599
- searching the web or calling external APIs. Built-in tools, like web
600
- search, are natively integrated, while user-defined tools are created
601
- and invoked by the user. To learn more, see our [Tool Use
602
- Guide](/docs/empathic-voice-interface-evi/features/tool-use).
603
-
604
-
605
- Currently, the only built-in tool Hume provides is **Web Search**.
606
- When enabled, Web Search equips EVI with the ability to search the web
607
- for up-to-date information.
608
- context:
609
- type: optional<Context>
216
+ type:
217
+ type: literal<"session_settings">
610
218
  docs: >-
611
- Allows developers to inject additional context into the conversation,
612
- which is appended to the end of user messages for the session.
219
+ The type of message sent through the socket; must be
220
+ `session_settings` for our server to correctly identify and process it
221
+ as a Session Settings message.
613
222
 
614
223
 
615
- When included in a Session Settings message, the provided context can
616
- be used to remind the LLM of its role in every user message, prevent
617
- it from forgetting important details, or add new relevant information
618
- to the conversation.
224
+ Session settings are temporary and apply only to the current Chat
225
+ session. These settings can be adjusted dynamically based on the
226
+ requirements of each session to ensure optimal performance and user
227
+ experience.
619
228
 
620
229
 
621
- Set to `null` to disable context injection.
230
+ For more information, please refer to the [Session Settings
231
+ guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
622
232
  custom_session_id:
623
233
  type: optional<string>
624
234
  docs: >-
@@ -637,17 +247,6 @@ types:
637
247
  language
638
248
  model](/docs/empathic-voice-interface-evi/guides/custom-language-model)
639
249
  with EVI to learn more.
640
- language_model_api_key:
641
- type: optional<string>
642
- docs: >-
643
- Third party API key for the supplemental language model.
644
-
645
-
646
- When provided, EVI will use this key instead of Hume’s API key for the
647
- supplemental LLM. This allows you to bypass rate limits and utilize
648
- your own API key as needed.
649
- metadata:
650
- type: optional<map<string, unknown>>
651
250
  system_prompt:
652
251
  type: optional<string>
653
252
  docs: >-
@@ -670,6 +269,41 @@ types:
670
269
 
671
270
  For help writing a system prompt, see our [Prompting
672
271
  Guide](/docs/empathic-voice-interface-evi/guides/prompting).
272
+ context:
273
+ type: optional<Context>
274
+ docs: >-
275
+ Field for injecting additional context into the conversation, which is
276
+ appended to the end of user messages for the session.
277
+
278
+
279
+ When included in a Session Settings message, the provided context can
280
+ be used to remind the LLM of its role in every user message, prevent
281
+ it from forgetting important details, or add new relevant information
282
+ to the conversation.
283
+
284
+
285
+ Set to `null` to clear injected context.
286
+ audio:
287
+ type: optional<AudioConfiguration>
288
+ docs: >-
289
+ Configuration details for the audio input used during the session.
290
+ Ensures the audio is being correctly set up for processing.
291
+
292
+
293
+ This optional field is only required when the audio input is encoded
294
+ in PCM Linear 16 (16-bit, little-endian, signed PCM WAV data). For
295
+ detailed instructions on how to configure session settings for PCM
296
+ Linear 16 audio, please refer to the [Session Settings
297
+ guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
298
+ language_model_api_key:
299
+ type: optional<string>
300
+ docs: >-
301
+ Third party API key for the supplemental language model.
302
+
303
+
304
+ When provided, EVI will use this key instead of Hume’s API key for the
305
+ supplemental LLM. This allows you to bypass rate limits and utilize
306
+ your own API key as needed.
673
307
  tools:
674
308
  type: optional<list<Tool>>
675
309
  docs: >-
@@ -681,22 +315,24 @@ types:
681
315
  search, are natively integrated, while user-defined tools are created
682
316
  and invoked by the user. To learn more, see our [Tool Use
683
317
  Guide](/docs/empathic-voice-interface-evi/features/tool-use).
684
- type:
685
- type: literal<"session_settings">
318
+ builtin_tools:
319
+ type: optional<list<BuiltinToolConfig>>
686
320
  docs: >-
687
- The type of message sent through the socket; must be
688
- `session_settings` for our server to correctly identify and process it
689
- as a Session Settings message.
321
+ List of built-in tools to enable for the session.
690
322
 
691
323
 
692
- Session settings are temporary and apply only to the current Chat
693
- session. These settings can be adjusted dynamically based on the
694
- requirements of each session to ensure optimal performance and user
695
- experience.
324
+ Tools are resources used by EVI to perform various tasks, such as
325
+ searching the web or calling external APIs. Built-in tools, like web
326
+ search, are natively integrated, while user-defined tools are created
327
+ and invoked by the user. To learn more, see our [Tool Use
328
+ Guide](/docs/empathic-voice-interface-evi/features/tool-use).
696
329
 
697
330
 
698
- For more information, please refer to the [Session Settings
699
- guide](/docs/empathic-voice-interface-evi/configuration/session-settings).
331
+ Currently, the only built-in tool Hume provides is **Web Search**.
332
+ When enabled, Web Search equips EVI with the ability to search the web
333
+ for up-to-date information.
334
+ metadata:
335
+ type: optional<map<string, unknown>>
700
336
  variables:
701
337
  type: optional<map<string, SessionSettingsVariablesValue>>
702
338
  docs: >-
@@ -719,17 +355,9 @@ types:
719
355
  openapi: evi-asyncapi.json
720
356
  Tool:
721
357
  properties:
722
- description:
723
- type: optional<string>
724
- docs: >-
725
- An optional description of what the tool does, used by the
726
- supplemental LLM to choose when and how to call the function.
727
- fallback_content:
728
- type: optional<string>
729
- docs: >-
730
- Optional text passed to the supplemental LLM if the tool call fails.
731
- The LLM then uses this text to generate a response back to the user,
732
- ensuring continuity in the conversation.
358
+ type:
359
+ type: ToolType
360
+ docs: Type of tool. Set to `function` for user-defined tools.
733
361
  name:
734
362
  type: string
735
363
  docs: Name of the user-defined tool to be enabled.
@@ -743,90 +371,92 @@ types:
743
371
  including the expected data type and description for each input field.
744
372
  Structured as a JSON schema, this format ensures the tool receives
745
373
  data in the expected format.
746
- type:
747
- type: ToolType
748
- docs: Type of tool. Set to `function` for user-defined tools.
374
+ description:
375
+ type: optional<string>
376
+ docs: >-
377
+ An optional description of what the tool does, used by the
378
+ supplemental LLM to choose when and how to call the function.
379
+ fallback_content:
380
+ type: optional<string>
381
+ docs: >-
382
+ Optional text passed to the supplemental LLM if the tool call fails.
383
+ The LLM then uses this text to generate a response back to the user,
384
+ ensuring continuity in the conversation.
749
385
  source:
750
386
  openapi: evi-asyncapi.json
751
- ToolCallMessage:
752
- docs: When provided, the output is a tool call.
387
+ ToolErrorMessage:
388
+ docs: When provided, the output is a function call error.
753
389
  properties:
390
+ type:
391
+ type: literal<"tool_error">
392
+ docs: >-
393
+ The type of message sent through the socket; for a Tool Error message,
394
+ this must be `tool_error`.
395
+
396
+
397
+ Upon receiving a [Tool Call
398
+ message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
399
+ and failing to invoke the function, this message is sent to notify EVI
400
+ of the tool's failure.
754
401
  custom_session_id:
755
402
  type: optional<string>
756
403
  docs: >-
757
404
  Used to manage conversational state, correlate frontend and backend
758
405
  data, and persist conversations across EVI sessions.
759
- name:
760
- type: string
761
- docs: Name of the tool called.
762
- parameters:
763
- type: string
764
- docs: >-
765
- Parameters of the tool.
766
-
767
-
768
- These parameters define the inputs needed for the tool’s execution,
769
- including the expected data type and description for each input field.
770
- Structured as a stringified JSON schema, this format ensures the tool
771
- receives data in the expected format.
772
- response_required:
773
- type: boolean
774
- docs: >-
775
- Indicates whether a response to the tool call is required from the
776
- developer, either in the form of a [Tool Response
777
- message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolResponseMessage.type)
778
- or a [Tool Error
779
- message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolErrorMessage.type).
780
- tool_call_id:
781
- type: string
782
- docs: >-
783
- The unique identifier for a specific tool call instance.
784
-
785
-
786
- This ID is used to track the request and response of a particular tool
787
- invocation, ensuring that the correct response is linked to the
788
- appropriate request.
789
406
  tool_type:
790
407
  type: optional<ToolType>
791
408
  docs: >-
792
409
  Type of tool called. Either `builtin` for natively implemented tools,
793
410
  like web search, or `function` for user-defined tools.
794
- type:
795
- type: literal<"tool_call">
411
+ tool_call_id:
412
+ type: string
796
413
  docs: >-
797
- The type of message sent through the socket; for a Tool Call message,
798
- this must be `tool_call`.
414
+ The unique identifier for a specific tool call instance.
799
415
 
800
416
 
801
- This message indicates that the supplemental LLM has detected a need
802
- to invoke the specified tool.
803
- source:
804
- openapi: evi-asyncapi.json
805
- ToolErrorMessage:
806
- docs: When provided, the output is a function call error.
807
- properties:
808
- code:
809
- type: optional<string>
810
- docs: Error code. Identifies the type of error encountered.
417
+ This ID is used to track the request and response of a particular tool
418
+ invocation, ensuring that the Tool Error message is linked to the
419
+ appropriate tool call request. The specified `tool_call_id` must match
420
+ the one received in the [Tool Call
421
+ message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type).
811
422
  content:
812
423
  type: optional<string>
813
424
  docs: >-
814
425
  Optional text passed to the supplemental LLM in place of the tool call
815
426
  result. The LLM then uses this text to generate a response back to the
816
427
  user, ensuring continuity in the conversation if the tool errors.
817
- custom_session_id:
818
- type: optional<string>
819
- docs: >-
820
- Used to manage conversational state, correlate frontend and backend
821
- data, and persist conversations across EVI sessions.
822
428
  error:
823
429
  type: string
824
430
  docs: Error message from the tool call, not exposed to the LLM or user.
431
+ code:
432
+ type: optional<string>
433
+ docs: Error code. Identifies the type of error encountered.
825
434
  level:
826
435
  type: optional<ErrorLevel>
827
436
  docs: >-
828
437
  Indicates the severity of an error; for a Tool Error message, this
829
438
  must be `warn` to signal an unexpected event.
439
+ source:
440
+ openapi: evi-asyncapi.json
441
+ ToolResponseMessage:
442
+ docs: When provided, the output is a function call response.
443
+ properties:
444
+ type:
445
+ type: literal<"tool_response">
446
+ docs: >-
447
+ The type of message sent through the socket; for a Tool Response
448
+ message, this must be `tool_response`.
449
+
450
+
451
+ Upon receiving a [Tool Call
452
+ message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
453
+ and successfully invoking the function, this message is sent to convey
454
+ the result of the function call back to EVI.
455
+ custom_session_id:
456
+ type: optional<string>
457
+ docs: >-
458
+ Used to manage conversational state, correlate frontend and backend
459
+ data, and persist conversations across EVI sessions.
830
460
  tool_call_id:
831
461
  type: string
832
462
  docs: >-
@@ -834,41 +464,406 @@ types:
834
464
 
835
465
 
836
466
  This ID is used to track the request and response of a particular tool
837
- invocation, ensuring that the Tool Error message is linked to the
838
- appropriate tool call request. The specified `tool_call_id` must match
839
- the one received in the [Tool Call
467
+ invocation, ensuring that the correct response is linked to the
468
+ appropriate request. The specified `tool_call_id` must match the one
469
+ received in the [Tool Call
470
+ message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.tool_call_id).
471
+ content:
472
+ type: string
473
+ docs: >-
474
+ Return value of the tool call. Contains the output generated by the
475
+ tool to pass back to EVI.
476
+ tool_name:
477
+ type: optional<string>
478
+ docs: >-
479
+ Name of the tool.
480
+
481
+
482
+ Include this optional field to help the supplemental LLM identify
483
+ which tool generated the response. The specified `tool_name` must
484
+ match the one received in the [Tool Call
840
485
  message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type).
841
486
  tool_type:
842
487
  type: optional<ToolType>
843
488
  docs: >-
844
489
  Type of tool called. Either `builtin` for natively implemented tools,
845
490
  like web search, or `function` for user-defined tools.
491
+ source:
492
+ openapi: evi-asyncapi.json
493
+ ToolType:
494
+ enum:
495
+ - builtin
496
+ - function
497
+ source:
498
+ openapi: evi-asyncapi.json
499
+ UserInput:
500
+ docs: >-
501
+ User text to insert into the conversation. Text sent through a User Input
502
+ message is treated as the user's speech to EVI. EVI processes this input
503
+ and provides a corresponding response.
504
+
505
+
506
+ Expression measurement results are not available for User Input messages,
507
+ as the prosody model relies on audio input and cannot process text alone.
508
+ properties:
846
509
  type:
847
- type: literal<"tool_error">
510
+ type: literal<"user_input">
848
511
  docs: >-
849
- The type of message sent through the socket; for a Tool Error message,
850
- this must be `tool_error`.
512
+ The type of message sent through the socket; must be `user_input` for
513
+ our server to correctly identify and process it as a User Input
514
+ message.
515
+ custom_session_id:
516
+ type: optional<string>
517
+ docs: >-
518
+ Used to manage conversational state, correlate frontend and backend
519
+ data, and persist conversations across EVI sessions.
520
+ text:
521
+ type: string
522
+ docs: >-
523
+ User text to insert into the conversation. Text sent through a User
524
+ Input message is treated as the user’s speech to EVI. EVI processes
525
+ this input and provides a corresponding response.
851
526
 
852
527
 
853
- Upon receiving a [Tool Call
854
- message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
855
- and failing to invoke the function, this message is sent to notify EVI
856
- of the tool's failure.
528
+ Expression measurement results are not available for User Input
529
+ messages, as the prosody model relies on audio input and cannot
530
+ process text alone.
857
531
  source:
858
532
  openapi: evi-asyncapi.json
859
- ToolResponseMessage:
860
- docs: When provided, the output is a function call response.
533
+ AssistantEnd:
534
+ docs: When provided, the output is an assistant end message.
861
535
  properties:
862
- content:
863
- type: string
536
+ type:
537
+ type: literal<"assistant_end">
864
538
  docs: >-
865
- Return value of the tool call. Contains the output generated by the
866
- tool to pass back to EVI.
539
+ The type of message sent through the socket; for an Assistant End
540
+ message, this must be `assistant_end`.
541
+
542
+
543
+ This message indicates the conclusion of the assistant’s response,
544
+ signaling that the assistant has finished speaking for the current
545
+ conversational turn.
546
+ custom_session_id:
547
+ type: optional<string>
548
+ docs: >-
549
+ Used to manage conversational state, correlate frontend and backend
550
+ data, and persist conversations across EVI sessions.
551
+ source:
552
+ openapi: evi-asyncapi.json
553
+ AssistantMessage:
554
+ docs: When provided, the output is an assistant message.
555
+ properties:
556
+ type:
557
+ type: literal<"assistant_message">
558
+ docs: >-
559
+ The type of message sent through the socket; for an Assistant Message,
560
+ this must be `assistant_message`.
561
+
562
+
563
+ This message contains both a transcript of the assistant’s response
564
+ and the expression measurement predictions of the assistant’s audio
565
+ output.
867
566
  custom_session_id:
868
567
  type: optional<string>
869
568
  docs: >-
870
569
  Used to manage conversational state, correlate frontend and backend
871
570
  data, and persist conversations across EVI sessions.
571
+ id:
572
+ type: optional<string>
573
+ docs: >-
574
+ ID of the assistant message. Allows the Assistant Message to be
575
+ tracked and referenced.
576
+ message:
577
+ type: ChatMessage
578
+ docs: Transcript of the message.
579
+ models:
580
+ type: Inference
581
+ docs: Inference model results.
582
+ from_text:
583
+ type: boolean
584
+ docs: >-
585
+ Indicates if this message was inserted into the conversation as text
586
+ from an [Assistant Input
587
+ message](/reference/empathic-voice-interface-evi/chat/chat#send.AssistantInput.text).
588
+ source:
589
+ openapi: evi-asyncapi.json
590
+ AssistantProsody:
591
+ docs: When provided, the output is an Assistant Prosody message.
592
+ properties:
593
+ type:
594
+ type: literal<"assistant_prosody">
595
+ docs: >-
596
+ The type of message sent through the socket; for an Assistant Prosody
597
+ message, this must be `assistant_PROSODY`.
598
+
599
+
600
+ This message the expression measurement predictions of the assistant's
601
+ audio output.
602
+ custom_session_id:
603
+ type: optional<string>
604
+ docs: >-
605
+ Used to manage conversational state, correlate frontend and backend
606
+ data, and persist conversations across EVI sessions.
607
+ models:
608
+ type: Inference
609
+ docs: Inference model results.
610
+ id:
611
+ type: optional<string>
612
+ docs: Unique identifier for the segment.
613
+ source:
614
+ openapi: evi-asyncapi.json
615
+ AudioOutput:
616
+ docs: >-
617
+ The type of message sent through the socket; for an Audio Output message,
618
+ this must be `audio_output`.
619
+ properties:
620
+ type:
621
+ type: literal<"audio_output">
622
+ docs: >-
623
+ The type of message sent through the socket; for an Audio Output
624
+ message, this must be `audio_output`.
625
+ custom_session_id:
626
+ type: optional<string>
627
+ docs: >-
628
+ Used to manage conversational state, correlate frontend and backend
629
+ data, and persist conversations across EVI sessions.
630
+ id:
631
+ type: string
632
+ docs: >-
633
+ ID of the audio output. Allows the Audio Output message to be tracked
634
+ and referenced.
635
+ index:
636
+ type: integer
637
+ docs: Index of the chunk of audio relative to the whole audio segment.
638
+ data:
639
+ type: string
640
+ docs: >-
641
+ Base64 encoded audio output. This encoded audio is transmitted to the
642
+ client, where it can be decoded and played back as part of the user
643
+ interaction.
644
+ source:
645
+ openapi: evi-asyncapi.json
646
+ ChatMessageToolResult:
647
+ discriminated: false
648
+ docs: Function call response from client.
649
+ union:
650
+ - type: ToolResponseMessage
651
+ - type: ToolErrorMessage
652
+ source:
653
+ openapi: evi-asyncapi.json
654
+ inline: true
655
+ ChatMessage:
656
+ properties:
657
+ role:
658
+ type: Role
659
+ docs: Role of who is providing the message.
660
+ content:
661
+ type: optional<string>
662
+ docs: Transcript of the message.
663
+ tool_call:
664
+ type: optional<ToolCallMessage>
665
+ docs: Function call name and arguments.
666
+ tool_result:
667
+ type: optional<ChatMessageToolResult>
668
+ docs: Function call response from client.
669
+ source:
670
+ openapi: evi-asyncapi.json
671
+ ChatMetadata:
672
+ docs: When provided, the output is a chat metadata message.
673
+ properties:
674
+ type:
675
+ type: literal<"chat_metadata">
676
+ docs: >-
677
+ The type of message sent through the socket; for a Chat Metadata
678
+ message, this must be `chat_metadata`.
679
+
680
+
681
+ The Chat Metadata message is the first message you receive after
682
+ establishing a connection with EVI and contains important identifiers
683
+ for the current Chat session.
684
+ custom_session_id:
685
+ type: optional<string>
686
+ docs: >-
687
+ Used to manage conversational state, correlate frontend and backend
688
+ data, and persist conversations across EVI sessions.
689
+ chat_group_id:
690
+ type: string
691
+ docs: >-
692
+ ID of the Chat Group.
693
+
694
+
695
+ Used to resume a Chat when passed in the
696
+ [resumed_chat_group_id](/reference/empathic-voice-interface-evi/chat/chat#request.query.resumed_chat_group_id)
697
+ query parameter of a subsequent connection request. This allows EVI to
698
+ continue the conversation from where it left off within the Chat
699
+ Group.
700
+
701
+
702
+ Learn more about [supporting chat
703
+ resumability](/docs/empathic-voice-interface-evi/faq#does-evi-support-chat-resumability)
704
+ from the EVI FAQ.
705
+ chat_id:
706
+ type: string
707
+ docs: >-
708
+ ID of the Chat session. Allows the Chat session to be tracked and
709
+ referenced.
710
+ request_id:
711
+ type: optional<string>
712
+ docs: ID of the initiating request.
713
+ source:
714
+ openapi: evi-asyncapi.json
715
+ EmotionScores:
716
+ properties:
717
+ Admiration: double
718
+ Adoration: double
719
+ Aesthetic Appreciation: double
720
+ Amusement: double
721
+ Anger: double
722
+ Anxiety: double
723
+ Awe: double
724
+ Awkwardness: double
725
+ Boredom: double
726
+ Calmness: double
727
+ Concentration: double
728
+ Confusion: double
729
+ Contemplation: double
730
+ Contempt: double
731
+ Contentment: double
732
+ Craving: double
733
+ Desire: double
734
+ Determination: double
735
+ Disappointment: double
736
+ Disgust: double
737
+ Distress: double
738
+ Doubt: double
739
+ Ecstasy: double
740
+ Embarrassment: double
741
+ Empathic Pain: double
742
+ Entrancement: double
743
+ Envy: double
744
+ Excitement: double
745
+ Fear: double
746
+ Guilt: double
747
+ Horror: double
748
+ Interest: double
749
+ Joy: double
750
+ Love: double
751
+ Nostalgia: double
752
+ Pain: double
753
+ Pride: double
754
+ Realization: double
755
+ Relief: double
756
+ Romance: double
757
+ Sadness: double
758
+ Satisfaction: double
759
+ Shame: double
760
+ Surprise (negative): double
761
+ Surprise (positive): double
762
+ Sympathy: double
763
+ Tiredness: double
764
+ Triumph: double
765
+ source:
766
+ openapi: evi-asyncapi.json
767
+ WebSocketError:
768
+ docs: When provided, the output is an error message.
769
+ properties:
770
+ type:
771
+ type: literal<"error">
772
+ docs: >-
773
+ The type of message sent through the socket; for a Web Socket Error
774
+ message, this must be `error`.
775
+
776
+
777
+ This message indicates a disruption in the WebSocket connection, such
778
+ as an unexpected disconnection, protocol error, or data transmission
779
+ issue.
780
+ custom_session_id:
781
+ type: optional<string>
782
+ docs: >-
783
+ Used to manage conversational state, correlate frontend and backend
784
+ data, and persist conversations across EVI sessions.
785
+ code:
786
+ type: string
787
+ docs: Error code. Identifies the type of error encountered.
788
+ slug:
789
+ type: string
790
+ docs: >-
791
+ Short, human-readable identifier and description for the error. See a
792
+ complete list of error slugs on the [Errors
793
+ page](/docs/resources/errors).
794
+ message:
795
+ type: string
796
+ docs: Detailed description of the error.
797
+ request_id:
798
+ type: optional<string>
799
+ docs: ID of the initiating request.
800
+ source:
801
+ openapi: evi-asyncapi.json
802
+ Inference:
803
+ properties:
804
+ prosody:
805
+ type: optional<ProsodyInference>
806
+ docs: >-
807
+ Prosody model inference results.
808
+
809
+
810
+ EVI uses the prosody model to measure 48 emotions related to speech
811
+ and vocal characteristics within a given expression.
812
+ source:
813
+ openapi: evi-asyncapi.json
814
+ MillisecondInterval:
815
+ properties:
816
+ begin:
817
+ type: integer
818
+ docs: Start time of the interval in milliseconds.
819
+ end:
820
+ type: integer
821
+ docs: End time of the interval in milliseconds.
822
+ source:
823
+ openapi: evi-asyncapi.json
824
+ ProsodyInference:
825
+ properties:
826
+ scores:
827
+ type: EmotionScores
828
+ docs: >-
829
+ The confidence scores for 48 emotions within the detected expression
830
+ of an audio sample.
831
+
832
+
833
+ Scores typically range from 0 to 1, with higher values indicating a
834
+ stronger confidence level in the measured attribute.
835
+
836
+
837
+ See our guide on [interpreting expression measurement
838
+ results](/docs/expression-measurement/faq#how-do-i-interpret-my-results)
839
+ to learn more.
840
+ source:
841
+ openapi: evi-asyncapi.json
842
+ Role:
843
+ enum:
844
+ - assistant
845
+ - system
846
+ - user
847
+ - all
848
+ - tool
849
+ source:
850
+ openapi: evi-asyncapi.json
851
+ ToolCallMessage:
852
+ docs: When provided, the output is a tool call.
853
+ properties:
854
+ name:
855
+ type: string
856
+ docs: Name of the tool called.
857
+ parameters:
858
+ type: string
859
+ docs: >-
860
+ Parameters of the tool.
861
+
862
+
863
+ These parameters define the inputs needed for the tool’s execution,
864
+ including the expected data type and description for each input field.
865
+ Structured as a stringified JSON schema, this format ensures the tool
866
+ receives data in the expected format.
872
867
  tool_call_id:
873
868
  type: string
874
869
  docs: >-
@@ -877,88 +872,39 @@ types:
877
872
 
878
873
  This ID is used to track the request and response of a particular tool
879
874
  invocation, ensuring that the correct response is linked to the
880
- appropriate request. The specified `tool_call_id` must match the one
881
- received in the [Tool Call
882
- message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.tool_call_id).
883
- tool_name:
884
- type: optional<string>
885
- docs: >-
886
- Name of the tool.
887
-
888
-
889
- Include this optional field to help the supplemental LLM identify
890
- which tool generated the response. The specified `tool_name` must
891
- match the one received in the [Tool Call
892
- message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type).
893
- tool_type:
894
- type: optional<ToolType>
895
- docs: >-
896
- Type of tool called. Either `builtin` for natively implemented tools,
897
- like web search, or `function` for user-defined tools.
875
+ appropriate request.
898
876
  type:
899
- type: literal<"tool_response">
877
+ type: literal<"tool_call">
900
878
  docs: >-
901
- The type of message sent through the socket; for a Tool Response
902
- message, this must be `tool_response`.
903
-
904
-
905
- Upon receiving a [Tool Call
906
- message](/reference/empathic-voice-interface-evi/chat/chat#receive.ToolCallMessage.type)
907
- and successfully invoking the function, this message is sent to convey
908
- the result of the function call back to EVI.
909
- source:
910
- openapi: evi-asyncapi.json
911
- ToolType:
912
- enum:
913
- - builtin
914
- - function
915
- source:
916
- openapi: evi-asyncapi.json
917
- UserInput:
918
- docs: >-
919
- User text to insert into the conversation. Text sent through a User Input
920
- message is treated as the user's speech to EVI. EVI processes this input
921
- and provides a corresponding response.
879
+ The type of message sent through the socket; for a Tool Call message,
880
+ this must be `tool_call`.
922
881
 
923
882
 
924
- Expression measurement results are not available for User Input messages,
925
- as the prosody model relies on audio input and cannot process text alone.
926
- properties:
883
+ This message indicates that the supplemental LLM has detected a need
884
+ to invoke the specified tool.
927
885
  custom_session_id:
928
886
  type: optional<string>
929
887
  docs: >-
930
888
  Used to manage conversational state, correlate frontend and backend
931
889
  data, and persist conversations across EVI sessions.
932
- text:
933
- type: string
890
+ tool_type:
891
+ type: optional<ToolType>
934
892
  docs: >-
935
- User text to insert into the conversation. Text sent through a User
936
- Input message is treated as the user’s speech to EVI. EVI processes
937
- this input and provides a corresponding response.
938
-
939
-
940
- Expression measurement results are not available for User Input
941
- messages, as the prosody model relies on audio input and cannot
942
- process text alone.
943
- type:
944
- type: literal<"user_input">
893
+ Type of tool called. Either `builtin` for natively implemented tools,
894
+ like web search, or `function` for user-defined tools.
895
+ response_required:
896
+ type: boolean
945
897
  docs: >-
946
- The type of message sent through the socket; must be `user_input` for
947
- our server to correctly identify and process it as a User Input
948
- message.
898
+ Indicates whether a response to the tool call is required from the
899
+ developer, either in the form of a [Tool Response
900
+ message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolResponseMessage.type)
901
+ or a [Tool Error
902
+ message](/reference/empathic-voice-interface-evi/chat/chat#send.ToolErrorMessage.type).
949
903
  source:
950
904
  openapi: evi-asyncapi.json
951
905
  UserInterruption:
952
906
  docs: When provided, the output is an interruption.
953
907
  properties:
954
- custom_session_id:
955
- type: optional<string>
956
- docs: >-
957
- Used to manage conversational state, correlate frontend and backend
958
- data, and persist conversations across EVI sessions.
959
- time:
960
- type: integer
961
- docs: Unix timestamp of the detected user interruption.
962
908
  type:
963
909
  type: literal<"user_interruption">
964
910
  docs: >-
@@ -971,16 +917,48 @@ types:
971
917
  message to signal the interruption event. This message allows the
972
918
  system to stop the current audio playback, clear the audio queue, and
973
919
  prepare to handle new user input.
920
+ custom_session_id:
921
+ type: optional<string>
922
+ docs: >-
923
+ Used to manage conversational state, correlate frontend and backend
924
+ data, and persist conversations across EVI sessions.
925
+ time:
926
+ type: integer
927
+ docs: Unix timestamp of the detected user interruption.
974
928
  source:
975
929
  openapi: evi-asyncapi.json
976
930
  UserMessage:
977
931
  docs: When provided, the output is a user message.
978
932
  properties:
933
+ type:
934
+ type: literal<"user_message">
935
+ docs: >-
936
+ The type of message sent through the socket; for a User Message, this
937
+ must be `user_message`.
938
+
939
+
940
+ This message contains both a transcript of the user’s input and the
941
+ expression measurement predictions if the input was sent as an [Audio
942
+ Input
943
+ message](/reference/empathic-voice-interface-evi/chat/chat#send.AudioInput.type).
944
+ Expression measurement predictions are not provided for a [User Input
945
+ message](/reference/empathic-voice-interface-evi/chat/chat#send.UserInput.type),
946
+ as the prosody model relies on audio input and cannot process text
947
+ alone.
979
948
  custom_session_id:
980
949
  type: optional<string>
981
950
  docs: >-
982
951
  Used to manage conversational state, correlate frontend and backend
983
952
  data, and persist conversations across EVI sessions.
953
+ message:
954
+ type: ChatMessage
955
+ docs: Transcript of the message.
956
+ models:
957
+ type: Inference
958
+ docs: Inference model results.
959
+ time:
960
+ type: MillisecondInterval
961
+ docs: Start and End time of user message.
984
962
  from_text:
985
963
  type: boolean
986
964
  docs: >-
@@ -1005,30 +983,6 @@ types:
1005
983
  Interim transcripts are only sent when the
1006
984
  [`verbose_transcription`](/reference/empathic-voice-interface-evi/chat/chat#request.query.verbose_transcription)
1007
985
  query parameter is set to `true` in the initial handshake.
1008
- message:
1009
- type: ChatMessage
1010
- docs: Transcript of the message.
1011
- models:
1012
- type: Inference
1013
- docs: Inference model results.
1014
- time:
1015
- type: MillisecondInterval
1016
- docs: Start and End time of user message.
1017
- type:
1018
- type: literal<"user_message">
1019
- docs: >-
1020
- The type of message sent through the socket; for a User Message, this
1021
- must be `user_message`.
1022
-
1023
-
1024
- This message contains both a transcript of the user’s input and the
1025
- expression measurement predictions if the input was sent as an [Audio
1026
- Input
1027
- message](/reference/empathic-voice-interface-evi/chat/chat#send.AudioInput.type).
1028
- Expression measurement predictions are not provided for a [User Input
1029
- message](/reference/empathic-voice-interface-evi/chat/chat#send.UserInput.type),
1030
- as the prosody model relies on audio input and cannot process text
1031
- alone.
1032
986
  source:
1033
987
  openapi: evi-asyncapi.json
1034
988
  JsonMessage:
@@ -1112,6 +1066,12 @@ types:
1112
1066
  name: Gpt4OMini
1113
1067
  - value: gpt-4.1
1114
1068
  name: Gpt41
1069
+ - value: gpt-5
1070
+ name: Gpt5
1071
+ - value: gpt-5-mini
1072
+ name: Gpt5Mini
1073
+ - value: gpt-5-nano
1074
+ name: Gpt5Nano
1115
1075
  - value: gemma-7b-it
1116
1076
  name: Gemma7BIt
1117
1077
  - value: llama3-8b-8192
@@ -1166,12 +1126,6 @@ types:
1166
1126
  - CEREBRAS
1167
1127
  source:
1168
1128
  openapi: evi-openapi.json
1169
- SupportsToolUse:
1170
- properties:
1171
- model_resource: string
1172
- supports_tool_use: boolean
1173
- source:
1174
- openapi: evi-openapi.json
1175
1129
  ValidationErrorLocItem:
1176
1130
  discriminated: false
1177
1131
  union:
@@ -1188,29 +1142,20 @@ types:
1188
1142
  type: string
1189
1143
  source:
1190
1144
  openapi: evi-openapi.json
1191
- WebhookEventBase:
1192
- docs: Represents the fields common to all webhook events.
1193
- properties:
1194
- chat_group_id:
1195
- type: string
1196
- docs: Unique ID of the **Chat Group** associated with the **Chat** session.
1197
- chat_id:
1198
- type: string
1199
- docs: Unique ID of the **Chat** session.
1200
- config_id:
1201
- type: optional<string>
1202
- docs: Unique ID of the EVI **Config** used for the session.
1203
- source:
1204
- openapi: evi-openapi.json
1205
- WebhookEvent:
1206
- discriminated: false
1207
- union:
1208
- - WebhookEventChatStarted
1209
- - WebhookEventChatEnded
1210
- source:
1211
- openapi: evi-openapi.json
1212
1145
  WebhookEventChatEnded:
1213
1146
  properties:
1147
+ event_name:
1148
+ type: optional<literal<"chat_ended">>
1149
+ docs: Always `chat_ended`.
1150
+ end_time:
1151
+ type: integer
1152
+ docs: Unix timestamp (in milliseconds) indicating when the session ended.
1153
+ duration_seconds:
1154
+ type: integer
1155
+ docs: Total duration of the session in seconds.
1156
+ end_reason:
1157
+ type: WebhookEventChatStatus
1158
+ docs: Reason for the session's termination.
1214
1159
  caller_number:
1215
1160
  type: optional<string>
1216
1161
  docs: >-
@@ -1225,18 +1170,6 @@ types:
1225
1170
  language
1226
1171
  model](/docs/empathic-voice-interface-evi/custom-language-model) in
1227
1172
  the EVI Config.
1228
- duration_seconds:
1229
- type: integer
1230
- docs: Total duration of the session in seconds.
1231
- end_reason:
1232
- type: WebhookEventChatStatus
1233
- docs: Reason for the session's termination.
1234
- end_time:
1235
- type: integer
1236
- docs: Unix timestamp (in milliseconds) indicating when the session ended.
1237
- event_name:
1238
- type: optional<literal<"chat_ended">>
1239
- docs: Always `chat_ended`.
1240
1173
  extends:
1241
1174
  - WebhookEventBase
1242
1175
  source:
@@ -1249,6 +1182,18 @@ types:
1249
1182
  openapi: evi-openapi.json
1250
1183
  WebhookEventChatStarted:
1251
1184
  properties:
1185
+ event_name:
1186
+ type: optional<literal<"chat_started">>
1187
+ docs: Always `chat_started`.
1188
+ start_time:
1189
+ type: integer
1190
+ docs: Unix timestamp (in milliseconds) indicating when the session started.
1191
+ chat_start_type:
1192
+ type: WebhookEventChatStartType
1193
+ docs: >-
1194
+ Indicates whether the chat is the first in a new Chat Group
1195
+ (`new_chat_group`) or the continuation of an existing chat group
1196
+ (`resumed_chat_group`).
1252
1197
  caller_number:
1253
1198
  type: optional<string>
1254
1199
  docs: >-
@@ -1256,12 +1201,6 @@ types:
1256
1201
  This field is included only if the Chat was created via the [Twilio
1257
1202
  phone calling](/docs/empathic-voice-interface-evi/phone-calling)
1258
1203
  integration.
1259
- chat_start_type:
1260
- type: WebhookEventChatStartType
1261
- docs: >-
1262
- Indicates whether the chat is the first in a new Chat Group
1263
- (`new_chat_group`) or the continuation of an existing chat group
1264
- (`resumed_chat_group`).
1265
1204
  custom_session_id:
1266
1205
  type: optional<string>
1267
1206
  docs: >-
@@ -1269,12 +1208,6 @@ types:
1269
1208
  language
1270
1209
  model](/docs/empathic-voice-interface-evi/custom-language-model) in
1271
1210
  the EVI Config.
1272
- event_name:
1273
- type: optional<literal<"chat_started">>
1274
- docs: Always `chat_started`.
1275
- start_time:
1276
- type: integer
1277
- docs: Unix timestamp (in milliseconds) indicating when the session started.
1278
1211
  extends:
1279
1212
  - WebhookEventBase
1280
1213
  source:
@@ -1290,6 +1223,27 @@ types:
1290
1223
  - ERROR
1291
1224
  source:
1292
1225
  openapi: evi-openapi.json
1226
+ WebhookEvent:
1227
+ discriminated: false
1228
+ union:
1229
+ - WebhookEventChatStarted
1230
+ - WebhookEventChatEnded
1231
+ source:
1232
+ openapi: evi-openapi.json
1233
+ WebhookEventBase:
1234
+ docs: Represents the fields common to all webhook events.
1235
+ properties:
1236
+ chat_group_id:
1237
+ type: string
1238
+ docs: Unique ID of the **Chat Group** associated with the **Chat** session.
1239
+ chat_id:
1240
+ type: string
1241
+ docs: Unique ID of the **Chat** session.
1242
+ config_id:
1243
+ type: optional<string>
1244
+ docs: Unique ID of the EVI **Config** used for the session.
1245
+ source:
1246
+ openapi: evi-openapi.json
1293
1247
  ErrorResponse:
1294
1248
  properties:
1295
1249
  error: optional<string>
@@ -1483,14 +1437,14 @@ types:
1483
1437
  Version numbers are integer values representing different iterations
1484
1438
  of the Prompt. Each update to the Prompt increments its version
1485
1439
  number.
1440
+ version_description:
1441
+ type: optional<string>
1442
+ docs: An optional description of the Prompt version.
1486
1443
  version_type:
1487
1444
  type: ReturnPromptVersionType
1488
1445
  docs: >-
1489
1446
  Versioning method for a Prompt. Either `FIXED` for using a fixed
1490
1447
  version number or `LATEST` for auto-updating to the latest version.
1491
- version_description:
1492
- type: optional<string>
1493
- docs: An optional description of the Prompt version.
1494
1448
  created_on:
1495
1449
  type: long
1496
1450
  docs: >-
@@ -1556,6 +1510,12 @@ types:
1556
1510
  Version numbers are integer values representing different iterations
1557
1511
  of the Config. Each update to the Config increments its version
1558
1512
  number.
1513
+ tools:
1514
+ type: optional<list<optional<ReturnUserDefinedTool>>>
1515
+ docs: List of user-defined tools associated with this Config.
1516
+ version_description:
1517
+ type: optional<string>
1518
+ docs: An optional description of the Config version.
1559
1519
  language_model:
1560
1520
  type: optional<ReturnLanguageModel>
1561
1521
  docs: >-
@@ -1581,6 +1541,7 @@ types:
1581
1541
  3](/docs/speech-to-speech-evi/configuration/evi-version#migrating-to-evi-3)
1582
1542
  before then.
1583
1543
  timeouts: optional<ReturnTimeoutSpecs>
1544
+ nudges: optional<ReturnNudgeSpec>
1584
1545
  event_messages: optional<ReturnEventMessageSpecs>
1585
1546
  ellm_model:
1586
1547
  type: optional<ReturnEllmModel>
@@ -1592,12 +1553,11 @@ types:
1592
1553
  model that takes into account both expression measures and language.
1593
1554
  The eLLM generates short, empathic language responses and guides
1594
1555
  text-to-speech (TTS) prosody.
1556
+ voice: optional<unknown>
1557
+ prompt: optional<ReturnPrompt>
1595
1558
  webhooks:
1596
1559
  type: optional<list<optional<ReturnWebhookSpec>>>
1597
1560
  docs: Map of webhooks associated with this config.
1598
- version_description:
1599
- type: optional<string>
1600
- docs: An optional description of the Config version.
1601
1561
  created_on:
1602
1562
  type: optional<long>
1603
1563
  docs: >-
@@ -1608,12 +1568,6 @@ types:
1608
1568
  docs: >-
1609
1569
  Time at which the Config was last modified. Measured in seconds since
1610
1570
  the Unix epoch.
1611
- nudges: optional<ReturnNudgeSpec>
1612
- voice: optional<unknown>
1613
- prompt: optional<ReturnPrompt>
1614
- tools:
1615
- type: optional<list<optional<ReturnUserDefinedTool>>>
1616
- docs: List of user-defined tools associated with this Config.
1617
1571
  source:
1618
1572
  openapi: evi-openapi.json
1619
1573
  ReturnPagedChatsPaginationDirection:
@@ -2980,6 +2934,7 @@ types:
2980
2934
  enum:
2981
2935
  - HUME_AI
2982
2936
  - CUSTOM_VOICE
2937
+ - OCTAVE_COMBINED
2983
2938
  source:
2984
2939
  openapi: evi-openapi.json
2985
2940
  VoiceId: