hume 0.13.1 → 0.13.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (65) hide show
  1. package/.mock/definition/empathic-voice/__package__.yml +39 -30
  2. package/.mock/definition/tts/__package__.yml +85 -47
  3. package/.mock/definition/tts/voices.yml +9 -9
  4. package/api/resources/empathicVoice/resources/chat/client/Client.js +3 -0
  5. package/api/resources/empathicVoice/types/ContextType.d.ts +2 -2
  6. package/api/resources/empathicVoice/types/ContextType.js +1 -1
  7. package/api/resources/empathicVoice/types/JsonMessage.d.ts +1 -1
  8. package/api/resources/empathicVoice/types/ReturnConfig.d.ts +2 -1
  9. package/api/resources/empathicVoice/types/ReturnVoice.d.ts +12 -0
  10. package/api/resources/empathicVoice/types/ReturnVoice.js +5 -0
  11. package/api/resources/empathicVoice/types/VoiceProvider.d.ts +1 -2
  12. package/api/resources/empathicVoice/types/VoiceProvider.js +0 -1
  13. package/api/resources/empathicVoice/types/index.d.ts +4 -3
  14. package/api/resources/empathicVoice/types/index.js +4 -3
  15. package/api/resources/tts/client/Client.d.ts +5 -5
  16. package/api/resources/tts/client/Client.js +5 -5
  17. package/api/resources/tts/types/SnippetAudioChunk.d.ts +20 -0
  18. package/dist/api/resources/empathicVoice/resources/chat/client/Client.js +3 -0
  19. package/dist/api/resources/empathicVoice/types/ContextType.d.ts +2 -2
  20. package/dist/api/resources/empathicVoice/types/ContextType.js +1 -1
  21. package/dist/api/resources/empathicVoice/types/JsonMessage.d.ts +1 -1
  22. package/dist/api/resources/empathicVoice/types/ReturnConfig.d.ts +2 -1
  23. package/dist/api/resources/empathicVoice/types/ReturnVoice.d.ts +12 -0
  24. package/dist/api/resources/empathicVoice/types/ReturnVoice.js +5 -0
  25. package/dist/api/resources/empathicVoice/types/VoiceProvider.d.ts +1 -2
  26. package/dist/api/resources/empathicVoice/types/VoiceProvider.js +0 -1
  27. package/dist/api/resources/empathicVoice/types/index.d.ts +4 -3
  28. package/dist/api/resources/empathicVoice/types/index.js +4 -3
  29. package/dist/api/resources/tts/client/Client.d.ts +5 -5
  30. package/dist/api/resources/tts/client/Client.js +5 -5
  31. package/dist/api/resources/tts/types/SnippetAudioChunk.d.ts +20 -0
  32. package/dist/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
  33. package/dist/serialization/resources/empathicVoice/types/ContextType.js +1 -1
  34. package/dist/serialization/resources/empathicVoice/types/JsonMessage.d.ts +2 -2
  35. package/dist/serialization/resources/empathicVoice/types/JsonMessage.js +2 -2
  36. package/dist/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +2 -1
  37. package/dist/serialization/resources/empathicVoice/types/ReturnConfig.js +2 -1
  38. package/dist/serialization/resources/empathicVoice/types/ReturnVoice.d.ts +15 -0
  39. package/dist/serialization/resources/empathicVoice/types/ReturnVoice.js +46 -0
  40. package/dist/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
  41. package/dist/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
  42. package/dist/serialization/resources/empathicVoice/types/index.d.ts +4 -3
  43. package/dist/serialization/resources/empathicVoice/types/index.js +4 -3
  44. package/dist/serialization/resources/tts/types/SnippetAudioChunk.d.ts +12 -0
  45. package/dist/serialization/resources/tts/types/SnippetAudioChunk.js +14 -1
  46. package/dist/version.d.ts +1 -1
  47. package/dist/version.js +1 -1
  48. package/package.json +1 -1
  49. package/reference.md +14 -14
  50. package/serialization/resources/empathicVoice/types/ContextType.d.ts +1 -1
  51. package/serialization/resources/empathicVoice/types/ContextType.js +1 -1
  52. package/serialization/resources/empathicVoice/types/JsonMessage.d.ts +2 -2
  53. package/serialization/resources/empathicVoice/types/JsonMessage.js +2 -2
  54. package/serialization/resources/empathicVoice/types/ReturnConfig.d.ts +2 -1
  55. package/serialization/resources/empathicVoice/types/ReturnConfig.js +2 -1
  56. package/serialization/resources/empathicVoice/types/ReturnVoice.d.ts +15 -0
  57. package/serialization/resources/empathicVoice/types/ReturnVoice.js +46 -0
  58. package/serialization/resources/empathicVoice/types/VoiceProvider.d.ts +1 -1
  59. package/serialization/resources/empathicVoice/types/VoiceProvider.js +1 -1
  60. package/serialization/resources/empathicVoice/types/index.d.ts +4 -3
  61. package/serialization/resources/empathicVoice/types/index.js +4 -3
  62. package/serialization/resources/tts/types/SnippetAudioChunk.d.ts +12 -0
  63. package/serialization/resources/tts/types/SnippetAudioChunk.js +14 -1
  64. package/version.d.ts +1 -1
  65. package/version.js +1 -1
@@ -142,8 +142,8 @@ types:
142
142
  openapi: evi-asyncapi.json
143
143
  ContextType:
144
144
  enum:
145
- - temporary
146
145
  - persistent
146
+ - temporary
147
147
  source:
148
148
  openapi: evi-asyncapi.json
149
149
  Encoding:
@@ -990,6 +990,7 @@ types:
990
990
  union:
991
991
  - type: AssistantEnd
992
992
  - type: AssistantMessage
993
+ - type: AssistantProsody
993
994
  - type: ChatMetadata
994
995
  - type: WebSocketError
995
996
  - type: UserInterruption
@@ -997,7 +998,6 @@ types:
997
998
  - type: ToolCallMessage
998
999
  - type: ToolResponseMessage
999
1000
  - type: ToolErrorMessage
1000
- - type: AssistantProsody
1001
1001
  source:
1002
1002
  openapi: evi-asyncapi.json
1003
1003
  HTTPValidationError:
@@ -1142,6 +1142,26 @@ types:
1142
1142
  type: string
1143
1143
  source:
1144
1144
  openapi: evi-openapi.json
1145
+ VoiceId:
1146
+ properties:
1147
+ id:
1148
+ type: string
1149
+ docs: ID of the voice in the `Voice Library`.
1150
+ provider:
1151
+ type: optional<VoiceProvider>
1152
+ docs: Model provider associated with this Voice ID.
1153
+ source:
1154
+ openapi: evi-openapi.json
1155
+ VoiceName:
1156
+ properties:
1157
+ name:
1158
+ type: string
1159
+ docs: Name of the voice in the `Voice Library`.
1160
+ provider:
1161
+ type: optional<VoiceProvider>
1162
+ docs: Model provider associated with this Voice Name.
1163
+ source:
1164
+ openapi: evi-openapi.json
1145
1165
  WebhookEventChatEnded:
1146
1166
  properties:
1147
1167
  event_name:
@@ -1551,7 +1571,9 @@ types:
1551
1571
  model that takes into account both expression measures and language.
1552
1572
  The eLLM generates short, empathic language responses and guides
1553
1573
  text-to-speech (TTS) prosody.
1554
- voice: optional<unknown>
1574
+ voice:
1575
+ type: optional<ReturnVoice>
1576
+ docs: A voice specification associated with this Config.
1555
1577
  prompt: optional<ReturnPrompt>
1556
1578
  webhooks:
1557
1579
  type: optional<list<optional<ReturnWebhookSpec>>>
@@ -2928,33 +2950,6 @@ types:
2928
2950
  minutes).
2929
2951
  source:
2930
2952
  openapi: evi-openapi.json
2931
- VoiceProvider:
2932
- enum:
2933
- - HUME_AI
2934
- - CUSTOM_VOICE
2935
- - OCTAVE_COMBINED
2936
- source:
2937
- openapi: evi-openapi.json
2938
- VoiceId:
2939
- properties:
2940
- id:
2941
- type: string
2942
- docs: ID of the voice in the `Voice Library`.
2943
- provider:
2944
- type: optional<VoiceProvider>
2945
- docs: Model provider associated with this Voice ID.
2946
- source:
2947
- openapi: evi-openapi.json
2948
- VoiceName:
2949
- properties:
2950
- name:
2951
- type: string
2952
- docs: Name of the voice in the `Voice Library`.
2953
- provider:
2954
- type: optional<VoiceProvider>
2955
- docs: Model provider associated with this Voice Name.
2956
- source:
2957
- openapi: evi-openapi.json
2958
2953
  VoiceRef:
2959
2954
  discriminated: false
2960
2955
  union:
@@ -2962,3 +2957,17 @@ types:
2962
2957
  - type: VoiceName
2963
2958
  source:
2964
2959
  openapi: evi-openapi.json
2960
+ ReturnVoice:
2961
+ docs: An Octave voice available for text-to-speech
2962
+ properties:
2963
+ id: optional<string>
2964
+ name: optional<string>
2965
+ provider: optional<VoiceProvider>
2966
+ source:
2967
+ openapi: evi-openapi.json
2968
+ VoiceProvider:
2969
+ enum:
2970
+ - HUME_AI
2971
+ - CUSTOM_VOICE
2972
+ source:
2973
+ openapi: evi-openapi.json
@@ -29,7 +29,7 @@ service:
29
29
  The response includes the base64-encoded audio and metadata in JSON
30
30
  format.
31
31
  source:
32
- openapi: tts-openapi.yml
32
+ openapi: tts-openapi.json
33
33
  display-name: Text-to-speech (Json)
34
34
  request:
35
35
  body:
@@ -43,14 +43,6 @@ service:
43
43
  - UnprocessableEntityError
44
44
  examples:
45
45
  - request:
46
- utterances:
47
- - text: >-
48
- Beauty is no quality in things themselves: It exists merely in
49
- the mind which contemplates them.
50
- description: >-
51
- Middle-aged masculine voice with a clear, rhythmic Scots lilt,
52
- rounded vowels, and a warm, steady tone with an articulate,
53
- academic quality.
54
46
  context:
55
47
  utterances:
56
48
  - text: How can people see beauty so differently?
@@ -61,16 +53,24 @@ service:
61
53
  format:
62
54
  type: mp3
63
55
  num_generations: 1
56
+ utterances:
57
+ - text: >-
58
+ Beauty is no quality in things themselves: It exists merely in
59
+ the mind which contemplates them.
60
+ description: >-
61
+ Middle-aged masculine voice with a clear, rhythmic Scots lilt,
62
+ rounded vowels, and a warm, steady tone with an articulate,
63
+ academic quality.
64
64
  response:
65
65
  body:
66
66
  generations:
67
- - generation_id: 795c949a-1510-4a80-9646-7d0863b023ab
67
+ - audio: //PExAA0DDYRvkpNfhv3JI5JZ...etc.
68
68
  duration: 7.44225
69
- file_size: 120192
70
69
  encoding:
71
70
  format: mp3
72
71
  sample_rate: 48000
73
- audio: //PExAA0DDYRvkpNfhv3JI5JZ...etc.
72
+ file_size: 120192
73
+ generation_id: 795c949a-1510-4a80-9646-7d0863b023ab
74
74
  snippets:
75
75
  - - audio: //PExAA0DDYRvkpNfhv3JI5JZ...etc.
76
76
  generation_id: 795c949a-1510-4a80-9646-7d0863b023ab
@@ -93,7 +93,7 @@ service:
93
93
 
94
94
  The response contains the generated audio file in the requested format.
95
95
  source:
96
- openapi: tts-openapi.yml
96
+ openapi: tts-openapi.json
97
97
  display-name: Text-to-speech (File)
98
98
  request:
99
99
  body:
@@ -107,6 +107,11 @@ service:
107
107
  - UnprocessableEntityError
108
108
  examples:
109
109
  - request:
110
+ context:
111
+ generation_id: 09ad914d-8e7f-40f8-a279-e34f07f7dab2
112
+ format:
113
+ type: mp3
114
+ num_generations: 1
110
115
  utterances:
111
116
  - text: >-
112
117
  Beauty is no quality in things themselves: It exists merely in
@@ -115,11 +120,6 @@ service:
115
120
  Middle-aged masculine voice with a clear, rhythmic Scots lilt,
116
121
  rounded vowels, and a warm, steady tone with an articulate,
117
122
  academic quality.
118
- context:
119
- generation_id: 09ad914d-8e7f-40f8-a279-e34f07f7dab2
120
- format:
121
- type: mp3
122
- num_generations: 1
123
123
  synthesize-json-streaming:
124
124
  path: /v0/tts/stream/json
125
125
  method: POST
@@ -134,7 +134,7 @@ service:
134
134
  The response is a stream of JSON objects including audio encoded in
135
135
  base64.
136
136
  source:
137
- openapi: tts-openapi.yml
137
+ openapi: tts-openapi.json
138
138
  display-name: Text-to-speech (Streamed JSON)
139
139
  request:
140
140
  body:
@@ -165,7 +165,7 @@ service:
165
165
  additional context can be included to influence the speech's style and
166
166
  prosody.
167
167
  source:
168
- openapi: tts-openapi.yml
168
+ openapi: tts-openapi.json
169
169
  display-name: Text-to-speech (Streamed File)
170
170
  request:
171
171
  body:
@@ -187,7 +187,7 @@ service:
187
187
  name: Male English Actor
188
188
  provider: HUME_AI
189
189
  source:
190
- openapi: tts-openapi.yml
190
+ openapi: tts-openapi.json
191
191
  types:
192
192
  PostedContextWithGenerationId:
193
193
  properties:
@@ -198,13 +198,13 @@ types:
198
198
  consistent speech style and prosody across multiple requests.
199
199
  Including context may increase audio generation times.
200
200
  source:
201
- openapi: tts-openapi.yml
201
+ openapi: tts-openapi.json
202
202
  PostedContextWithUtterances:
203
203
  properties:
204
204
  utterances:
205
205
  type: list<PostedUtterance>
206
206
  source:
207
- openapi: tts-openapi.yml
207
+ openapi: tts-openapi.json
208
208
  AudioEncoding:
209
209
  docs: >-
210
210
  Encoding information about the generated audio, including the `format` and
@@ -219,14 +219,14 @@ types:
219
219
  The sample rate (`Hz`) of the generated audio. The default sample rate
220
220
  is `48000 Hz`.
221
221
  source:
222
- openapi: tts-openapi.yml
222
+ openapi: tts-openapi.json
223
223
  AudioFormatType:
224
224
  enum:
225
225
  - mp3
226
226
  - pcm
227
227
  - wav
228
228
  source:
229
- openapi: tts-openapi.yml
229
+ openapi: tts-openapi.json
230
230
  ReturnGeneration:
231
231
  properties:
232
232
  generation_id:
@@ -256,17 +256,17 @@ types:
256
256
  optimized for speech delivery.
257
257
  type: list<list<Snippet>>
258
258
  source:
259
- openapi: tts-openapi.yml
259
+ openapi: tts-openapi.json
260
260
  HTTPValidationError:
261
261
  properties:
262
262
  detail:
263
263
  type: optional<list<ValidationError>>
264
264
  source:
265
- openapi: tts-openapi.yml
265
+ openapi: tts-openapi.json
266
266
  FormatMp3:
267
267
  properties: {}
268
268
  source:
269
- openapi: tts-openapi.yml
269
+ openapi: tts-openapi.json
270
270
  PostedContext:
271
271
  discriminated: false
272
272
  docs: >-
@@ -277,7 +277,7 @@ types:
277
277
  - type: PostedContextWithGenerationId
278
278
  - type: PostedContextWithUtterances
279
279
  source:
280
- openapi: tts-openapi.yml
280
+ openapi: tts-openapi.json
281
281
  inline: true
282
282
  Format:
283
283
  discriminant: type
@@ -291,7 +291,7 @@ types:
291
291
  wav:
292
292
  type: FormatWav
293
293
  source:
294
- openapi: tts-openapi.yml
294
+ openapi: tts-openapi.json
295
295
  PostedTts:
296
296
  properties:
297
297
  context:
@@ -373,7 +373,7 @@ types:
373
373
  must be `1` or omitted).
374
374
  default: true
375
375
  source:
376
- openapi: tts-openapi.yml
376
+ openapi: tts-openapi.json
377
377
  ReturnTts:
378
378
  properties:
379
379
  request_id:
@@ -385,7 +385,7 @@ types:
385
385
  generations:
386
386
  type: list<ReturnGeneration>
387
387
  source:
388
- openapi: tts-openapi.yml
388
+ openapi: tts-openapi.json
389
389
  ReturnVoice:
390
390
  docs: An Octave voice available for text-to-speech
391
391
  properties:
@@ -404,11 +404,11 @@ types:
404
404
  Voices created through this endpoint will always have the provider set
405
405
  to `CUSTOM_VOICE`, indicating a custom voice stored in your account.
406
406
  source:
407
- openapi: tts-openapi.yml
407
+ openapi: tts-openapi.json
408
408
  FormatPcm:
409
409
  properties: {}
410
410
  source:
411
- openapi: tts-openapi.yml
411
+ openapi: tts-openapi.json
412
412
  Snippet:
413
413
  properties:
414
414
  id:
@@ -434,11 +434,49 @@ types:
434
434
  The segmented audio output in the requested format, encoded as a
435
435
  base64 string.
436
436
  source:
437
- openapi: tts-openapi.yml
437
+ openapi: tts-openapi.json
438
438
  SnippetAudioChunk:
439
- properties: {}
439
+ properties:
440
+ generation_id:
441
+ type: string
442
+ docs: >-
443
+ The generation ID of the parent snippet that this chunk corresponds
444
+ to.
445
+ snippet_id:
446
+ type: string
447
+ docs: The ID of the parent snippet that this chunk corresponds to.
448
+ text:
449
+ type: string
450
+ docs: The text of the parent snippet that this chunk corresponds to.
451
+ transcribed_text:
452
+ type: optional<string>
453
+ docs: >-
454
+ The transcribed text of the generated audio of the parent snippet that
455
+ this chunk corresponds to. It is only present if `instant_mode` is set
456
+ to `false`.
457
+ chunk_index:
458
+ type: integer
459
+ docs: The index of the audio chunk in the snippet.
460
+ audio:
461
+ type: string
462
+ docs: The generated audio output chunk in the requested format.
463
+ audio_format:
464
+ type: AudioFormatType
465
+ docs: The generated audio output format.
466
+ is_last_chunk:
467
+ type: boolean
468
+ docs: >-
469
+ Whether or not this is the last chunk streamed back from the decoder
470
+ for one input snippet.
471
+ utterance_index:
472
+ type: optional<integer>
473
+ docs: >-
474
+ The index of the utterance in the request that the parent snippet of
475
+ this chunk corresponds to.
476
+ snippet:
477
+ type: Snippet
440
478
  source:
441
- openapi: tts-openapi.yml
479
+ openapi: tts-openapi.json
442
480
  PostedUtterance:
443
481
  properties:
444
482
  text:
@@ -492,14 +530,14 @@ types:
492
530
  min: 0
493
531
  max: 5
494
532
  source:
495
- openapi: tts-openapi.yml
533
+ openapi: tts-openapi.json
496
534
  ValidationErrorLocItem:
497
535
  discriminated: false
498
536
  union:
499
537
  - string
500
538
  - integer
501
539
  source:
502
- openapi: tts-openapi.yml
540
+ openapi: tts-openapi.json
503
541
  inline: true
504
542
  ValidationError:
505
543
  properties:
@@ -508,7 +546,7 @@ types:
508
546
  msg: string
509
547
  type: string
510
548
  source:
511
- openapi: tts-openapi.yml
549
+ openapi: tts-openapi.json
512
550
  PostedUtteranceVoiceWithId:
513
551
  properties:
514
552
  id:
@@ -537,7 +575,7 @@ types:
537
575
  users. In contrast, your custom voices are private and accessible only
538
576
  via requests authenticated with your API key.
539
577
  source:
540
- openapi: tts-openapi.yml
578
+ openapi: tts-openapi.json
541
579
  PostedUtteranceVoiceWithName:
542
580
  properties:
543
581
  name:
@@ -566,31 +604,31 @@ types:
566
604
  users. In contrast, your custom voices are private and accessible only
567
605
  via requests authenticated with your API key.
568
606
  source:
569
- openapi: tts-openapi.yml
607
+ openapi: tts-openapi.json
570
608
  VoiceProvider:
571
609
  enum:
572
610
  - HUME_AI
573
611
  - CUSTOM_VOICE
574
612
  source:
575
- openapi: tts-openapi.yml
613
+ openapi: tts-openapi.json
576
614
  PostedUtteranceVoice:
577
615
  discriminated: false
578
616
  union:
579
617
  - type: PostedUtteranceVoiceWithId
580
618
  - type: PostedUtteranceVoiceWithName
581
619
  source:
582
- openapi: tts-openapi.yml
620
+ openapi: tts-openapi.json
583
621
  FormatWav:
584
622
  properties: {}
585
623
  source:
586
- openapi: tts-openapi.yml
624
+ openapi: tts-openapi.json
587
625
  ErrorResponse:
588
626
  properties:
589
627
  error: optional<string>
590
628
  message: optional<string>
591
629
  code: optional<string>
592
630
  source:
593
- openapi: tts-openapi.yml
631
+ openapi: tts-openapi.json
594
632
  ReturnPagedVoices:
595
633
  docs: A paginated list Octave voices available for text-to-speech
596
634
  properties:
@@ -619,4 +657,4 @@ types:
619
657
  List of voices returned for the specified `page_number` and
620
658
  `page_size`.
621
659
  source:
622
- openapi: tts-openapi.yml
660
+ openapi: tts-openapi.json
@@ -15,7 +15,7 @@ service:
15
15
  offset: $request.page_number
16
16
  results: $response.voices_page
17
17
  source:
18
- openapi: tts-openapi.yml
18
+ openapi: tts-openapi.json
19
19
  display-name: List voices
20
20
  request:
21
21
  name: VoicesListRequest
@@ -70,11 +70,11 @@ service:
70
70
  page_size: 10
71
71
  total_pages: 1
72
72
  voices_page:
73
- - name: David Hume
74
- id: c42352c0-4566-455d-b180-0f654b65b525
73
+ - id: c42352c0-4566-455d-b180-0f654b65b525
74
+ name: David Hume
75
75
  provider: CUSTOM_VOICE
76
- - name: Goliath Hume
77
- id: d87352b0-26a3-4b11-081b-d157a5674d19
76
+ - id: d87352b0-26a3-4b11-081b-d157a5674d19
77
+ name: Goliath Hume
78
78
  provider: CUSTOM_VOICE
79
79
  create:
80
80
  path: /v0/tts/voices
@@ -89,7 +89,7 @@ service:
89
89
  ensuring consistent speech style and prosody. For more details on voice
90
90
  creation, see the [Voices Guide](/docs/text-to-speech-tts/voices).
91
91
  source:
92
- openapi: tts-openapi.yml
92
+ openapi: tts-openapi.json
93
93
  display-name: Create voice
94
94
  request:
95
95
  name: PostedVoice
@@ -117,8 +117,8 @@ service:
117
117
  name: David Hume
118
118
  response:
119
119
  body:
120
- name: David Hume
121
120
  id: c42352c0-4566-455d-b180-0f654b65b525
121
+ name: David Hume
122
122
  provider: CUSTOM_VOICE
123
123
  delete:
124
124
  path: /v0/tts/voices
@@ -126,7 +126,7 @@ service:
126
126
  auth: true
127
127
  docs: Deletes a previously generated custom voice.
128
128
  source:
129
- openapi: tts-openapi.yml
129
+ openapi: tts-openapi.json
130
130
  display-name: Delete voice
131
131
  request:
132
132
  name: VoicesDeleteRequest
@@ -140,4 +140,4 @@ service:
140
140
  - query-parameters:
141
141
  name: David Hume
142
142
  source:
143
- openapi: tts-openapi.yml
143
+ openapi: tts-openapi.json
@@ -82,6 +82,9 @@ class Chat {
82
82
  if (args.verboseTranscription !== null) {
83
83
  queryParams["verbose_transcription"] = args.verboseTranscription ? "true" : "false";
84
84
  }
85
+ if (args.voiceId !== null && args.voiceId !== undefined && args.voiceId !== "") {
86
+ queryParams["voice_id"] = args.voiceId;
87
+ }
85
88
  if (args.queryParams !== null && args.queryParams !== undefined) {
86
89
  for (const [name, value] of Object.entries(args.queryParams)) {
87
90
  queryParams[name] = value;
@@ -1,8 +1,8 @@
1
1
  /**
2
2
  * This file was auto-generated by Fern from our API Definition.
3
3
  */
4
- export type ContextType = "temporary" | "persistent";
4
+ export type ContextType = "persistent" | "temporary";
5
5
  export declare const ContextType: {
6
- readonly Temporary: "temporary";
7
6
  readonly Persistent: "persistent";
7
+ readonly Temporary: "temporary";
8
8
  };
@@ -5,6 +5,6 @@
5
5
  Object.defineProperty(exports, "__esModule", { value: true });
6
6
  exports.ContextType = void 0;
7
7
  exports.ContextType = {
8
- Temporary: "temporary",
9
8
  Persistent: "persistent",
9
+ Temporary: "temporary",
10
10
  };
@@ -2,4 +2,4 @@
2
2
  * This file was auto-generated by Fern from our API Definition.
3
3
  */
4
4
  import * as Hume from "../../../index";
5
- export type JsonMessage = Hume.empathicVoice.AssistantEnd | Hume.empathicVoice.AssistantMessage | Hume.empathicVoice.ChatMetadata | Hume.empathicVoice.WebSocketError | Hume.empathicVoice.UserInterruption | Hume.empathicVoice.UserMessage | Hume.empathicVoice.ToolCallMessage | Hume.empathicVoice.ToolResponseMessage | Hume.empathicVoice.ToolErrorMessage | Hume.empathicVoice.AssistantProsody;
5
+ export type JsonMessage = Hume.empathicVoice.AssistantEnd | Hume.empathicVoice.AssistantMessage | Hume.empathicVoice.AssistantProsody | Hume.empathicVoice.ChatMetadata | Hume.empathicVoice.WebSocketError | Hume.empathicVoice.UserInterruption | Hume.empathicVoice.UserMessage | Hume.empathicVoice.ToolCallMessage | Hume.empathicVoice.ToolResponseMessage | Hume.empathicVoice.ToolErrorMessage;
@@ -45,7 +45,8 @@ export interface ReturnConfig {
45
45
  * Hume's eLLM (empathic Large Language Model) is a multimodal language model that takes into account both expression measures and language. The eLLM generates short, empathic language responses and guides text-to-speech (TTS) prosody.
46
46
  */
47
47
  ellmModel?: Hume.empathicVoice.ReturnEllmModel;
48
- voice?: unknown;
48
+ /** A voice specification associated with this Config. */
49
+ voice?: Hume.empathicVoice.ReturnVoice;
49
50
  prompt?: Hume.empathicVoice.ReturnPrompt;
50
51
  /** Map of webhooks associated with this config. */
51
52
  webhooks?: (Hume.empathicVoice.ReturnWebhookSpec | undefined)[];
@@ -0,0 +1,12 @@
1
+ /**
2
+ * This file was auto-generated by Fern from our API Definition.
3
+ */
4
+ import * as Hume from "../../../index";
5
+ /**
6
+ * An Octave voice available for text-to-speech
7
+ */
8
+ export interface ReturnVoice {
9
+ id?: string;
10
+ name?: string;
11
+ provider?: Hume.empathicVoice.VoiceProvider;
12
+ }
@@ -0,0 +1,5 @@
1
+ "use strict";
2
+ /**
3
+ * This file was auto-generated by Fern from our API Definition.
4
+ */
5
+ Object.defineProperty(exports, "__esModule", { value: true });
@@ -1,9 +1,8 @@
1
1
  /**
2
2
  * This file was auto-generated by Fern from our API Definition.
3
3
  */
4
- export type VoiceProvider = "HUME_AI" | "CUSTOM_VOICE" | "OCTAVE_COMBINED";
4
+ export type VoiceProvider = "HUME_AI" | "CUSTOM_VOICE";
5
5
  export declare const VoiceProvider: {
6
6
  readonly HumeAi: "HUME_AI";
7
7
  readonly CustomVoice: "CUSTOM_VOICE";
8
- readonly OctaveCombined: "OCTAVE_COMBINED";
9
8
  };
@@ -7,5 +7,4 @@ exports.VoiceProvider = void 0;
7
7
  exports.VoiceProvider = {
8
8
  HumeAi: "HUME_AI",
9
9
  CustomVoice: "CUSTOM_VOICE",
10
- OctaveCombined: "OCTAVE_COMBINED",
11
10
  };
@@ -38,6 +38,8 @@ export * from "./LanguageModelType";
38
38
  export * from "./ModelProviderEnum";
39
39
  export * from "./ValidationErrorLocItem";
40
40
  export * from "./ValidationError";
41
+ export * from "./VoiceId";
42
+ export * from "./VoiceName";
41
43
  export * from "./WebhookEventChatEnded";
42
44
  export * from "./WebhookEventChatStartType";
43
45
  export * from "./WebhookEventChatStarted";
@@ -102,7 +104,6 @@ export * from "./PostedEventMessageSpec";
102
104
  export * from "./PostedTimeoutSpec";
103
105
  export * from "./ReturnEventMessageSpec";
104
106
  export * from "./ReturnTimeoutSpec";
105
- export * from "./VoiceProvider";
106
- export * from "./VoiceId";
107
- export * from "./VoiceName";
108
107
  export * from "./VoiceRef";
108
+ export * from "./ReturnVoice";
109
+ export * from "./VoiceProvider";
@@ -54,6 +54,8 @@ __exportStar(require("./LanguageModelType"), exports);
54
54
  __exportStar(require("./ModelProviderEnum"), exports);
55
55
  __exportStar(require("./ValidationErrorLocItem"), exports);
56
56
  __exportStar(require("./ValidationError"), exports);
57
+ __exportStar(require("./VoiceId"), exports);
58
+ __exportStar(require("./VoiceName"), exports);
57
59
  __exportStar(require("./WebhookEventChatEnded"), exports);
58
60
  __exportStar(require("./WebhookEventChatStartType"), exports);
59
61
  __exportStar(require("./WebhookEventChatStarted"), exports);
@@ -118,7 +120,6 @@ __exportStar(require("./PostedEventMessageSpec"), exports);
118
120
  __exportStar(require("./PostedTimeoutSpec"), exports);
119
121
  __exportStar(require("./ReturnEventMessageSpec"), exports);
120
122
  __exportStar(require("./ReturnTimeoutSpec"), exports);
121
- __exportStar(require("./VoiceProvider"), exports);
122
- __exportStar(require("./VoiceId"), exports);
123
- __exportStar(require("./VoiceName"), exports);
124
123
  __exportStar(require("./VoiceRef"), exports);
124
+ __exportStar(require("./ReturnVoice"), exports);
125
+ __exportStar(require("./VoiceProvider"), exports);
@@ -44,10 +44,6 @@ export declare class Tts {
44
44
  *
45
45
  * @example
46
46
  * await client.tts.synthesizeJson({
47
- * utterances: [{
48
- * text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
49
- * description: "Middle-aged masculine voice with a clear, rhythmic Scots lilt, rounded vowels, and a warm, steady tone with an articulate, academic quality."
50
- * }],
51
47
  * context: {
52
48
  * utterances: [{
53
49
  * text: "How can people see beauty so differently?",
@@ -57,7 +53,11 @@ export declare class Tts {
57
53
  * format: {
58
54
  * type: "mp3"
59
55
  * },
60
- * numGenerations: 1
56
+ * numGenerations: 1,
57
+ * utterances: [{
58
+ * text: "Beauty is no quality in things themselves: It exists merely in the mind which contemplates them.",
59
+ * description: "Middle-aged masculine voice with a clear, rhythmic Scots lilt, rounded vowels, and a warm, steady tone with an articulate, academic quality."
60
+ * }]
61
61
  * })
62
62
  */
63
63
  synthesizeJson(request: Hume.tts.PostedTts, requestOptions?: Tts.RequestOptions): core.HttpResponsePromise<Hume.tts.ReturnTts>;