hume 0.9.1 → 0.9.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.mock/definition/api.yml +6 -6
- package/.mock/definition/empathic-voice/__package__.yml +2810 -2703
- package/.mock/definition/empathic-voice/chat.yml +143 -143
- package/.mock/definition/empathic-voice/chatGroups.yml +579 -508
- package/.mock/definition/empathic-voice/chats.yml +490 -449
- package/.mock/definition/empathic-voice/configs.yml +913 -871
- package/.mock/definition/empathic-voice/customVoices.yml +255 -234
- package/.mock/definition/empathic-voice/prompts.yml +523 -526
- package/.mock/definition/empathic-voice/tools.yml +588 -588
- package/.mock/definition/expression-measurement/batch/__package__.yml +1758 -1758
- package/.mock/definition/expression-measurement/stream/__package__.yml +486 -485
- package/.mock/fern.config.json +3 -3
- package/.prettierignore +1 -0
- package/api/resources/empathicVoice/resources/chat/client/Socket.d.ts +4 -1
- package/api/resources/empathicVoice/resources/chat/client/Socket.js +6 -1
- package/api/resources/empathicVoice/resources/chatGroups/client/Client.d.ts +11 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/Client.js +77 -3
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsGetAudioRequest.d.ts +25 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/index.d.ts +1 -0
- package/api/resources/empathicVoice/resources/chats/client/Client.d.ts +10 -0
- package/api/resources/empathicVoice/resources/chats/client/Client.js +63 -2
- package/api/resources/empathicVoice/resources/configs/client/Client.js +9 -9
- package/api/resources/empathicVoice/resources/configs/client/requests/ConfigsListConfigsRequest.d.ts +1 -1
- package/api/resources/empathicVoice/resources/customVoices/client/Client.d.ts +2 -7
- package/api/resources/empathicVoice/resources/customVoices/client/Client.js +8 -13
- package/api/resources/empathicVoice/resources/customVoices/client/requests/CustomVoicesListCustomVoicesRequest.d.ts +1 -1
- package/api/resources/empathicVoice/resources/customVoices/client/requests/PostedCustomVoiceName.d.ts +0 -6
- package/api/resources/empathicVoice/resources/prompts/client/Client.js +9 -9
- package/api/resources/empathicVoice/resources/prompts/client/requests/PromptsListPromptsRequest.d.ts +2 -2
- package/api/resources/empathicVoice/resources/tools/client/Client.js +9 -9
- package/api/resources/empathicVoice/resources/tools/client/requests/ToolsListToolsRequest.d.ts +1 -1
- package/api/resources/empathicVoice/types/AudioOutput.d.ts +1 -1
- package/api/resources/empathicVoice/types/PostedCustomVoice.d.ts +2 -2
- package/api/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.d.ts +3 -2
- package/api/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.js +2 -1
- package/api/resources/empathicVoice/types/PostedCustomVoiceParameters.d.ts +51 -9
- package/api/resources/empathicVoice/types/PostedVoice.d.ts +1 -1
- package/api/resources/empathicVoice/types/ReturnCustomVoice.d.ts +2 -2
- package/api/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.d.ts +3 -2
- package/api/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.js +2 -1
- package/api/resources/empathicVoice/types/ReturnCustomVoiceParameters.d.ts +51 -9
- package/api/resources/empathicVoice/types/ReturnVoice.d.ts +1 -1
- package/api/resources/empathicVoice/types/UserInput.d.ts +3 -1
- package/api/resources/empathicVoice/types/VoiceNameEnum.d.ts +2 -1
- package/api/resources/empathicVoice/types/VoiceNameEnum.js +1 -0
- package/api/resources/empathicVoice/types/index.d.ts +0 -2
- package/api/resources/empathicVoice/types/index.js +0 -2
- package/api/resources/expressionMeasurement/resources/batch/client/Client.js +6 -6
- package/api/resources/index.d.ts +1 -1
- package/api/resources/index.js +2 -2
- package/dist/api/resources/empathicVoice/resources/chat/client/Socket.d.ts +4 -1
- package/dist/api/resources/empathicVoice/resources/chat/client/Socket.js +6 -1
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/Client.d.ts +11 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/Client.js +77 -3
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsGetAudioRequest.d.ts +25 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/index.d.ts +1 -0
- package/dist/api/resources/empathicVoice/resources/chats/client/Client.d.ts +10 -0
- package/dist/api/resources/empathicVoice/resources/chats/client/Client.js +63 -2
- package/dist/api/resources/empathicVoice/resources/configs/client/Client.js +9 -9
- package/dist/api/resources/empathicVoice/resources/configs/client/requests/ConfigsListConfigsRequest.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/customVoices/client/Client.d.ts +2 -7
- package/dist/api/resources/empathicVoice/resources/customVoices/client/Client.js +8 -13
- package/dist/api/resources/empathicVoice/resources/customVoices/client/requests/CustomVoicesListCustomVoicesRequest.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/customVoices/client/requests/PostedCustomVoiceName.d.ts +0 -6
- package/dist/api/resources/empathicVoice/resources/prompts/client/Client.js +9 -9
- package/dist/api/resources/empathicVoice/resources/prompts/client/requests/PromptsListPromptsRequest.d.ts +2 -2
- package/dist/api/resources/empathicVoice/resources/tools/client/Client.js +9 -9
- package/dist/api/resources/empathicVoice/resources/tools/client/requests/ToolsListToolsRequest.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/AudioOutput.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/PostedCustomVoice.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.d.ts +3 -2
- package/dist/api/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.js +2 -1
- package/dist/api/resources/empathicVoice/types/PostedCustomVoiceParameters.d.ts +51 -9
- package/dist/api/resources/empathicVoice/types/PostedVoice.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ReturnCustomVoice.d.ts +2 -2
- package/dist/api/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.d.ts +3 -2
- package/dist/api/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.js +2 -1
- package/dist/api/resources/empathicVoice/types/ReturnCustomVoiceParameters.d.ts +51 -9
- package/dist/api/resources/empathicVoice/types/ReturnVoice.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/UserInput.d.ts +3 -1
- package/dist/api/resources/empathicVoice/types/VoiceNameEnum.d.ts +2 -1
- package/dist/api/resources/empathicVoice/types/VoiceNameEnum.js +1 -0
- package/dist/api/resources/empathicVoice/types/index.d.ts +0 -2
- package/dist/api/resources/empathicVoice/types/index.js +0 -2
- package/dist/api/resources/expressionMeasurement/resources/batch/client/Client.js +6 -6
- package/dist/api/resources/index.d.ts +1 -1
- package/dist/api/resources/index.js +2 -2
- package/dist/serialization/resources/empathicVoice/types/PostedCustomVoice.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedCustomVoice.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/PostedCustomVoiceParameters.d.ts +9 -2
- package/dist/serialization/resources/empathicVoice/types/PostedCustomVoiceParameters.js +9 -2
- package/dist/serialization/resources/empathicVoice/types/ReturnCustomVoice.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnCustomVoice.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.js +1 -1
- package/dist/serialization/resources/empathicVoice/types/ReturnCustomVoiceParameters.d.ts +9 -2
- package/dist/serialization/resources/empathicVoice/types/ReturnCustomVoiceParameters.js +9 -2
- package/dist/serialization/resources/empathicVoice/types/VoiceNameEnum.d.ts +1 -1
- package/dist/serialization/resources/empathicVoice/types/VoiceNameEnum.js +1 -0
- package/dist/serialization/resources/empathicVoice/types/index.d.ts +0 -2
- package/dist/serialization/resources/empathicVoice/types/index.js +0 -2
- package/dist/serialization/resources/index.d.ts +1 -1
- package/dist/serialization/resources/index.js +2 -2
- package/dist/version.d.ts +1 -1
- package/dist/version.js +1 -1
- package/dist/wrapper/HumeClient.d.ts +7 -3
- package/dist/wrapper/HumeClient.js +2 -2
- package/package.json +1 -1
- package/reference.md +573 -607
- package/serialization/resources/empathicVoice/types/PostedCustomVoice.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedCustomVoice.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/PostedCustomVoiceBaseVoice.js +1 -1
- package/serialization/resources/empathicVoice/types/PostedCustomVoiceParameters.d.ts +9 -2
- package/serialization/resources/empathicVoice/types/PostedCustomVoiceParameters.js +9 -2
- package/serialization/resources/empathicVoice/types/ReturnCustomVoice.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ReturnCustomVoice.js +1 -1
- package/serialization/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/ReturnCustomVoiceBaseVoice.js +1 -1
- package/serialization/resources/empathicVoice/types/ReturnCustomVoiceParameters.d.ts +9 -2
- package/serialization/resources/empathicVoice/types/ReturnCustomVoiceParameters.js +9 -2
- package/serialization/resources/empathicVoice/types/VoiceNameEnum.d.ts +1 -1
- package/serialization/resources/empathicVoice/types/VoiceNameEnum.js +1 -0
- package/serialization/resources/empathicVoice/types/index.d.ts +0 -2
- package/serialization/resources/empathicVoice/types/index.js +0 -2
- package/serialization/resources/index.d.ts +1 -1
- package/serialization/resources/index.js +2 -2
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/wrapper/HumeClient.d.ts +7 -3
- package/wrapper/HumeClient.js +2 -2
- package/api/resources/empathicVoice/types/ExtendedVoiceArgs.d.ts +0 -9
- package/api/resources/empathicVoice/types/VoiceArgs.d.ts +0 -13
- package/dist/api/resources/empathicVoice/types/ExtendedVoiceArgs.d.ts +0 -9
- package/dist/api/resources/empathicVoice/types/ExtendedVoiceArgs.js +0 -5
- package/dist/api/resources/empathicVoice/types/VoiceArgs.d.ts +0 -13
- package/dist/api/resources/empathicVoice/types/VoiceArgs.js +0 -5
- package/dist/serialization/resources/empathicVoice/types/ExtendedVoiceArgs.d.ts +0 -15
- package/dist/serialization/resources/empathicVoice/types/ExtendedVoiceArgs.js +0 -36
- package/dist/serialization/resources/empathicVoice/types/VoiceArgs.d.ts +0 -19
- package/dist/serialization/resources/empathicVoice/types/VoiceArgs.js +0 -40
- package/serialization/resources/empathicVoice/types/ExtendedVoiceArgs.d.ts +0 -15
- package/serialization/resources/empathicVoice/types/ExtendedVoiceArgs.js +0 -36
- package/serialization/resources/empathicVoice/types/VoiceArgs.d.ts +0 -19
- package/serialization/resources/empathicVoice/types/VoiceArgs.js +0 -40
- /package/api/resources/empathicVoice/{types/ExtendedVoiceArgs.js → resources/chatGroups/client/requests/ChatGroupsGetAudioRequest.js} +0 -0
- /package/{api/resources/empathicVoice/types/VoiceArgs.js → dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsGetAudioRequest.js} +0 -0
|
@@ -1,528 +1,529 @@
|
|
|
1
1
|
channel:
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
2
|
+
path: /v0/stream/models
|
|
3
|
+
auth: false
|
|
4
|
+
display-name: Stream
|
|
5
|
+
headers:
|
|
6
|
+
X-Hume-Api-Key:
|
|
7
|
+
type: string
|
|
8
|
+
name: humeApiKey
|
|
9
|
+
messages:
|
|
10
|
+
subscribe:
|
|
11
|
+
origin: server
|
|
12
|
+
body: SubscribeEvent
|
|
13
|
+
publish:
|
|
14
|
+
origin: client
|
|
15
|
+
body:
|
|
16
|
+
type: StreamModelsEndpointPayload
|
|
17
|
+
docs: Models endpoint payload
|
|
18
|
+
examples:
|
|
19
|
+
- messages:
|
|
20
|
+
- type: publish
|
|
21
|
+
body: {}
|
|
22
|
+
- type: subscribe
|
|
23
|
+
body: {}
|
|
23
24
|
types:
|
|
24
|
-
|
|
25
|
+
StreamModelPredictionsJobDetails:
|
|
26
|
+
docs: >
|
|
27
|
+
If the job_details flag was set in the request, details about the current
|
|
28
|
+
streaming job will be returned in the response body.
|
|
29
|
+
properties:
|
|
30
|
+
job_id:
|
|
31
|
+
type: optional<string>
|
|
32
|
+
docs: ID of the current streaming job.
|
|
33
|
+
source:
|
|
34
|
+
openapi: streaming-asyncapi.yml
|
|
35
|
+
StreamModelPredictionsBurstPredictionsItem:
|
|
36
|
+
properties:
|
|
37
|
+
time: optional<TimeRange>
|
|
38
|
+
emotions: optional<EmotionEmbedding>
|
|
39
|
+
source:
|
|
40
|
+
openapi: streaming-asyncapi.yml
|
|
41
|
+
StreamModelPredictionsBurst:
|
|
42
|
+
docs: Response for the vocal burst emotion model.
|
|
43
|
+
properties:
|
|
44
|
+
predictions: optional<list<StreamModelPredictionsBurstPredictionsItem>>
|
|
45
|
+
source:
|
|
46
|
+
openapi: streaming-asyncapi.yml
|
|
47
|
+
StreamModelPredictionsFacePredictionsItem:
|
|
48
|
+
properties:
|
|
49
|
+
frame:
|
|
50
|
+
type: optional<double>
|
|
51
|
+
docs: Frame number
|
|
52
|
+
time:
|
|
53
|
+
type: optional<double>
|
|
54
|
+
docs: Time in seconds when face detection occurred.
|
|
55
|
+
bbox: optional<StreamBoundingBox>
|
|
56
|
+
prob:
|
|
57
|
+
type: optional<double>
|
|
58
|
+
docs: The predicted probability that a detected face was actually a face.
|
|
59
|
+
face_id:
|
|
60
|
+
type: optional<string>
|
|
61
|
+
docs: >-
|
|
62
|
+
Identifier for a face. Not that this defaults to `unknown` unless face
|
|
63
|
+
identification is enabled in the face model configuration.
|
|
64
|
+
emotions: optional<EmotionEmbedding>
|
|
65
|
+
facs: optional<EmotionEmbedding>
|
|
66
|
+
descriptions: optional<EmotionEmbedding>
|
|
67
|
+
source:
|
|
68
|
+
openapi: streaming-asyncapi.yml
|
|
69
|
+
StreamModelPredictionsFace:
|
|
70
|
+
docs: Response for the facial expression emotion model.
|
|
71
|
+
properties:
|
|
72
|
+
predictions: optional<list<StreamModelPredictionsFacePredictionsItem>>
|
|
73
|
+
source:
|
|
74
|
+
openapi: streaming-asyncapi.yml
|
|
75
|
+
StreamModelPredictionsFacemeshPredictionsItem:
|
|
76
|
+
properties:
|
|
77
|
+
emotions: optional<EmotionEmbedding>
|
|
78
|
+
source:
|
|
79
|
+
openapi: streaming-asyncapi.yml
|
|
80
|
+
StreamModelPredictionsFacemesh:
|
|
81
|
+
docs: Response for the facemesh emotion model.
|
|
82
|
+
properties:
|
|
83
|
+
predictions: optional<list<StreamModelPredictionsFacemeshPredictionsItem>>
|
|
84
|
+
source:
|
|
85
|
+
openapi: streaming-asyncapi.yml
|
|
86
|
+
StreamModelPredictionsLanguagePredictionsItem:
|
|
87
|
+
properties:
|
|
88
|
+
text:
|
|
89
|
+
type: optional<string>
|
|
90
|
+
docs: A segment of text (like a word or a sentence).
|
|
91
|
+
position: optional<TextPosition>
|
|
92
|
+
emotions: optional<EmotionEmbedding>
|
|
93
|
+
sentiment: optional<Sentiment>
|
|
94
|
+
toxicity: optional<Toxicity>
|
|
95
|
+
source:
|
|
96
|
+
openapi: streaming-asyncapi.yml
|
|
97
|
+
StreamModelPredictionsLanguage:
|
|
98
|
+
docs: Response for the language emotion model.
|
|
99
|
+
properties:
|
|
100
|
+
predictions: optional<list<StreamModelPredictionsLanguagePredictionsItem>>
|
|
101
|
+
source:
|
|
102
|
+
openapi: streaming-asyncapi.yml
|
|
103
|
+
StreamModelPredictionsProsodyPredictionsItem:
|
|
104
|
+
properties:
|
|
105
|
+
time: optional<TimeRange>
|
|
106
|
+
emotions: optional<EmotionEmbedding>
|
|
107
|
+
source:
|
|
108
|
+
openapi: streaming-asyncapi.yml
|
|
109
|
+
StreamModelPredictionsProsody:
|
|
110
|
+
docs: Response for the speech prosody emotion model.
|
|
111
|
+
properties:
|
|
112
|
+
predictions: optional<list<StreamModelPredictionsProsodyPredictionsItem>>
|
|
113
|
+
source:
|
|
114
|
+
openapi: streaming-asyncapi.yml
|
|
115
|
+
StreamModelPredictions:
|
|
116
|
+
docs: Model predictions
|
|
117
|
+
properties:
|
|
118
|
+
payload_id:
|
|
119
|
+
type: optional<string>
|
|
25
120
|
docs: >
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
properties:
|
|
36
|
-
time: optional<TimeRange>
|
|
37
|
-
emotions: optional<EmotionEmbedding>
|
|
38
|
-
source:
|
|
39
|
-
openapi: streaming-asyncapi.yml
|
|
40
|
-
StreamModelPredictionsBurst:
|
|
121
|
+
If a payload ID was passed in the request, the same payload ID will be
|
|
122
|
+
sent back in the response body.
|
|
123
|
+
job_details:
|
|
124
|
+
type: optional<StreamModelPredictionsJobDetails>
|
|
125
|
+
docs: >
|
|
126
|
+
If the job_details flag was set in the request, details about the
|
|
127
|
+
current streaming job will be returned in the response body.
|
|
128
|
+
burst:
|
|
129
|
+
type: optional<StreamModelPredictionsBurst>
|
|
41
130
|
docs: Response for the vocal burst emotion model.
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
source:
|
|
45
|
-
openapi: streaming-asyncapi.yml
|
|
46
|
-
StreamModelPredictionsFacePredictionsItem:
|
|
47
|
-
properties:
|
|
48
|
-
frame:
|
|
49
|
-
type: optional<double>
|
|
50
|
-
docs: Frame number
|
|
51
|
-
time:
|
|
52
|
-
type: optional<double>
|
|
53
|
-
docs: Time in seconds when face detection occurred.
|
|
54
|
-
bbox: optional<StreamBoundingBox>
|
|
55
|
-
prob:
|
|
56
|
-
type: optional<double>
|
|
57
|
-
docs: The predicted probability that a detected face was actually a face.
|
|
58
|
-
face_id:
|
|
59
|
-
type: optional<string>
|
|
60
|
-
docs: >-
|
|
61
|
-
Identifier for a face. Not that this defaults to `unknown` unless face
|
|
62
|
-
identification is enabled in the face model configuration.
|
|
63
|
-
emotions: optional<EmotionEmbedding>
|
|
64
|
-
facs: optional<EmotionEmbedding>
|
|
65
|
-
descriptions: optional<EmotionEmbedding>
|
|
66
|
-
source:
|
|
67
|
-
openapi: streaming-asyncapi.yml
|
|
68
|
-
StreamModelPredictionsFace:
|
|
131
|
+
face:
|
|
132
|
+
type: optional<StreamModelPredictionsFace>
|
|
69
133
|
docs: Response for the facial expression emotion model.
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
source:
|
|
73
|
-
openapi: streaming-asyncapi.yml
|
|
74
|
-
StreamModelPredictionsFacemeshPredictionsItem:
|
|
75
|
-
properties:
|
|
76
|
-
emotions: optional<EmotionEmbedding>
|
|
77
|
-
source:
|
|
78
|
-
openapi: streaming-asyncapi.yml
|
|
79
|
-
StreamModelPredictionsFacemesh:
|
|
134
|
+
facemesh:
|
|
135
|
+
type: optional<StreamModelPredictionsFacemesh>
|
|
80
136
|
docs: Response for the facemesh emotion model.
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
source:
|
|
84
|
-
openapi: streaming-asyncapi.yml
|
|
85
|
-
StreamModelPredictionsLanguagePredictionsItem:
|
|
86
|
-
properties:
|
|
87
|
-
text:
|
|
88
|
-
type: optional<string>
|
|
89
|
-
docs: A segment of text (like a word or a sentence).
|
|
90
|
-
position: optional<TextPosition>
|
|
91
|
-
emotions: optional<EmotionEmbedding>
|
|
92
|
-
sentiment: optional<Sentiment>
|
|
93
|
-
toxicity: optional<Toxicity>
|
|
94
|
-
source:
|
|
95
|
-
openapi: streaming-asyncapi.yml
|
|
96
|
-
StreamModelPredictionsLanguage:
|
|
137
|
+
language:
|
|
138
|
+
type: optional<StreamModelPredictionsLanguage>
|
|
97
139
|
docs: Response for the language emotion model.
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
source:
|
|
101
|
-
openapi: streaming-asyncapi.yml
|
|
102
|
-
StreamModelPredictionsProsodyPredictionsItem:
|
|
103
|
-
properties:
|
|
104
|
-
time: optional<TimeRange>
|
|
105
|
-
emotions: optional<EmotionEmbedding>
|
|
106
|
-
source:
|
|
107
|
-
openapi: streaming-asyncapi.yml
|
|
108
|
-
StreamModelPredictionsProsody:
|
|
140
|
+
prosody:
|
|
141
|
+
type: optional<StreamModelPredictionsProsody>
|
|
109
142
|
docs: Response for the speech prosody emotion model.
|
|
110
|
-
|
|
111
|
-
|
|
112
|
-
|
|
113
|
-
|
|
114
|
-
|
|
115
|
-
|
|
116
|
-
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
126
|
-
|
|
127
|
-
|
|
128
|
-
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
133
|
-
facemesh:
|
|
134
|
-
type: optional<StreamModelPredictionsFacemesh>
|
|
135
|
-
docs: Response for the facemesh emotion model.
|
|
136
|
-
language:
|
|
137
|
-
type: optional<StreamModelPredictionsLanguage>
|
|
138
|
-
docs: Response for the language emotion model.
|
|
139
|
-
prosody:
|
|
140
|
-
type: optional<StreamModelPredictionsProsody>
|
|
141
|
-
docs: Response for the speech prosody emotion model.
|
|
142
|
-
source:
|
|
143
|
-
openapi: streaming-asyncapi.yml
|
|
144
|
-
JobDetails:
|
|
143
|
+
source:
|
|
144
|
+
openapi: streaming-asyncapi.yml
|
|
145
|
+
JobDetails:
|
|
146
|
+
docs: >
|
|
147
|
+
If the job_details flag was set in the request, details about the current
|
|
148
|
+
streaming job will be returned in the response body.
|
|
149
|
+
properties:
|
|
150
|
+
job_id:
|
|
151
|
+
type: optional<string>
|
|
152
|
+
docs: ID of the current streaming job.
|
|
153
|
+
source:
|
|
154
|
+
openapi: streaming-asyncapi.yml
|
|
155
|
+
StreamErrorMessage:
|
|
156
|
+
docs: Error message
|
|
157
|
+
properties:
|
|
158
|
+
error:
|
|
159
|
+
type: optional<string>
|
|
160
|
+
docs: Error message text.
|
|
161
|
+
code:
|
|
162
|
+
type: optional<string>
|
|
163
|
+
docs: Unique identifier for the error.
|
|
164
|
+
payload_id:
|
|
165
|
+
type: optional<string>
|
|
145
166
|
docs: >
|
|
146
|
-
|
|
147
|
-
|
|
148
|
-
|
|
149
|
-
|
|
150
|
-
type: optional<string>
|
|
151
|
-
docs: ID of the current streaming job.
|
|
152
|
-
source:
|
|
153
|
-
openapi: streaming-asyncapi.yml
|
|
154
|
-
StreamErrorMessage:
|
|
155
|
-
docs: Error message
|
|
156
|
-
properties:
|
|
157
|
-
error:
|
|
158
|
-
type: optional<string>
|
|
159
|
-
docs: Error message text.
|
|
160
|
-
code:
|
|
161
|
-
type: optional<string>
|
|
162
|
-
docs: Unique identifier for the error.
|
|
163
|
-
payload_id:
|
|
164
|
-
type: optional<string>
|
|
165
|
-
docs: >
|
|
166
|
-
If a payload ID was passed in the request, the same payload ID will be
|
|
167
|
-
sent back in the response body.
|
|
168
|
-
job_details:
|
|
169
|
-
type: optional<JobDetails>
|
|
170
|
-
docs: >
|
|
171
|
-
If the job_details flag was set in the request, details about the
|
|
172
|
-
current streaming job will be returned in the response body.
|
|
173
|
-
source:
|
|
174
|
-
openapi: streaming-asyncapi.yml
|
|
175
|
-
StreamWarningMessageJobDetails:
|
|
167
|
+
If a payload ID was passed in the request, the same payload ID will be
|
|
168
|
+
sent back in the response body.
|
|
169
|
+
job_details:
|
|
170
|
+
type: optional<JobDetails>
|
|
176
171
|
docs: >
|
|
177
|
-
|
|
178
|
-
|
|
179
|
-
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
|
|
192
|
-
|
|
193
|
-
|
|
194
|
-
|
|
195
|
-
|
|
196
|
-
|
|
197
|
-
|
|
198
|
-
|
|
199
|
-
|
|
200
|
-
|
|
201
|
-
|
|
202
|
-
If the job_details flag was set in the request, details about the
|
|
203
|
-
current streaming job will be returned in the response body.
|
|
204
|
-
source:
|
|
205
|
-
openapi: streaming-asyncapi.yml
|
|
206
|
-
SubscribeEvent:
|
|
207
|
-
discriminated: false
|
|
208
|
-
union:
|
|
209
|
-
- type: StreamModelPredictions
|
|
210
|
-
docs: Model predictions
|
|
211
|
-
- type: StreamErrorMessage
|
|
212
|
-
docs: Error message
|
|
213
|
-
- type: StreamWarningMessage
|
|
214
|
-
docs: Warning message
|
|
215
|
-
source:
|
|
216
|
-
openapi: streaming-asyncapi.yml
|
|
217
|
-
StreamFace:
|
|
172
|
+
If the job_details flag was set in the request, details about the
|
|
173
|
+
current streaming job will be returned in the response body.
|
|
174
|
+
source:
|
|
175
|
+
openapi: streaming-asyncapi.yml
|
|
176
|
+
StreamWarningMessageJobDetails:
|
|
177
|
+
docs: >
|
|
178
|
+
If the job_details flag was set in the request, details about the current
|
|
179
|
+
streaming job will be returned in the response body.
|
|
180
|
+
properties:
|
|
181
|
+
job_id:
|
|
182
|
+
type: optional<string>
|
|
183
|
+
docs: ID of the current streaming job.
|
|
184
|
+
source:
|
|
185
|
+
openapi: streaming-asyncapi.yml
|
|
186
|
+
StreamWarningMessage:
|
|
187
|
+
docs: Warning message
|
|
188
|
+
properties:
|
|
189
|
+
warning:
|
|
190
|
+
type: optional<string>
|
|
191
|
+
docs: Warning message text.
|
|
192
|
+
code:
|
|
193
|
+
type: optional<string>
|
|
194
|
+
docs: Unique identifier for the error.
|
|
195
|
+
payload_id:
|
|
196
|
+
type: optional<string>
|
|
218
197
|
docs: >
|
|
219
|
-
|
|
198
|
+
If a payload ID was passed in the request, the same payload ID will be
|
|
199
|
+
sent back in the response body.
|
|
200
|
+
job_details:
|
|
201
|
+
type: optional<StreamWarningMessageJobDetails>
|
|
202
|
+
docs: >
|
|
203
|
+
If the job_details flag was set in the request, details about the
|
|
204
|
+
current streaming job will be returned in the response body.
|
|
205
|
+
source:
|
|
206
|
+
openapi: streaming-asyncapi.yml
|
|
207
|
+
SubscribeEvent:
|
|
208
|
+
discriminated: false
|
|
209
|
+
union:
|
|
210
|
+
- type: StreamModelPredictions
|
|
211
|
+
docs: Model predictions
|
|
212
|
+
- type: StreamErrorMessage
|
|
213
|
+
docs: Error message
|
|
214
|
+
- type: StreamWarningMessage
|
|
215
|
+
docs: Warning message
|
|
216
|
+
source:
|
|
217
|
+
openapi: streaming-asyncapi.yml
|
|
218
|
+
StreamFace:
|
|
219
|
+
docs: >
|
|
220
|
+
Configuration for the facial expression emotion model.
|
|
220
221
|
|
|
221
222
|
|
|
222
|
-
|
|
223
|
-
|
|
224
|
-
|
|
225
|
-
|
|
226
|
-
|
|
227
|
-
|
|
228
|
-
|
|
229
|
-
|
|
230
|
-
|
|
231
|
-
|
|
232
|
-
|
|
233
|
-
|
|
234
|
-
|
|
235
|
-
|
|
236
|
-
|
|
237
|
-
|
|
238
|
-
|
|
239
|
-
|
|
240
|
-
|
|
241
|
-
|
|
242
|
-
|
|
243
|
-
|
|
244
|
-
|
|
245
|
-
|
|
246
|
-
|
|
247
|
-
|
|
248
|
-
|
|
249
|
-
|
|
250
|
-
|
|
251
|
-
|
|
252
|
-
|
|
253
|
-
|
|
254
|
-
|
|
255
|
-
|
|
256
|
-
|
|
257
|
-
|
|
258
|
-
|
|
259
|
-
docs: >
|
|
260
|
-
Minimum bounding box side length in pixels to treat as a face. Faces
|
|
261
|
-
detected with a bounding box side length in pixels less than this
|
|
262
|
-
threshold will be omitted from the response.
|
|
263
|
-
default: 3
|
|
264
|
-
source:
|
|
265
|
-
openapi: streaming-asyncapi.yml
|
|
266
|
-
StreamLanguage:
|
|
267
|
-
docs: Configuration for the language emotion model.
|
|
268
|
-
properties:
|
|
269
|
-
sentiment:
|
|
270
|
-
type: optional<map<string, unknown>>
|
|
271
|
-
docs: >-
|
|
272
|
-
Configuration for sentiment predictions. If missing or null, no
|
|
273
|
-
sentiment predictions will be generated.
|
|
274
|
-
toxicity:
|
|
275
|
-
type: optional<map<string, unknown>>
|
|
276
|
-
docs: >-
|
|
277
|
-
Configuration for toxicity predictions. If missing or null, no
|
|
278
|
-
toxicity predictions will be generated.
|
|
279
|
-
granularity:
|
|
280
|
-
type: optional<string>
|
|
281
|
-
docs: >-
|
|
282
|
-
The granularity at which to generate predictions. Values are `word`,
|
|
283
|
-
`sentence`, `utterance`, or `passage`. To get a single prediction for
|
|
284
|
-
the entire text of your streaming payload use `passage`. Default value
|
|
285
|
-
is `word`.
|
|
286
|
-
source:
|
|
287
|
-
openapi: streaming-asyncapi.yml
|
|
288
|
-
Config:
|
|
223
|
+
Note: Using the `reset_stream` parameter does not have any effect on face
|
|
224
|
+
identification. A single face identifier cache is maintained over a full
|
|
225
|
+
session whether `reset_stream` is used or not.
|
|
226
|
+
properties:
|
|
227
|
+
facs:
|
|
228
|
+
type: optional<map<string, unknown>>
|
|
229
|
+
docs: >-
|
|
230
|
+
Configuration for FACS predictions. If missing or null, no FACS
|
|
231
|
+
predictions will be generated.
|
|
232
|
+
descriptions:
|
|
233
|
+
type: optional<map<string, unknown>>
|
|
234
|
+
docs: >-
|
|
235
|
+
Configuration for Descriptions predictions. If missing or null, no
|
|
236
|
+
Descriptions predictions will be generated.
|
|
237
|
+
identify_faces:
|
|
238
|
+
type: optional<boolean>
|
|
239
|
+
docs: >
|
|
240
|
+
Whether to return identifiers for faces across frames. If true, unique
|
|
241
|
+
identifiers will be assigned to face bounding boxes to differentiate
|
|
242
|
+
different faces. If false, all faces will be tagged with an "unknown"
|
|
243
|
+
ID.
|
|
244
|
+
default: false
|
|
245
|
+
fps_pred:
|
|
246
|
+
type: optional<double>
|
|
247
|
+
docs: >
|
|
248
|
+
Number of frames per second to process. Other frames will be omitted
|
|
249
|
+
from the response.
|
|
250
|
+
default: 3
|
|
251
|
+
prob_threshold:
|
|
252
|
+
type: optional<double>
|
|
253
|
+
docs: >
|
|
254
|
+
Face detection probability threshold. Faces detected with a
|
|
255
|
+
probability less than this threshold will be omitted from the
|
|
256
|
+
response.
|
|
257
|
+
default: 3
|
|
258
|
+
min_face_size:
|
|
259
|
+
type: optional<double>
|
|
289
260
|
docs: >
|
|
290
|
-
|
|
291
|
-
|
|
292
|
-
|
|
293
|
-
|
|
294
|
-
|
|
295
|
-
|
|
296
|
-
|
|
261
|
+
Minimum bounding box side length in pixels to treat as a face. Faces
|
|
262
|
+
detected with a bounding box side length in pixels less than this
|
|
263
|
+
threshold will be omitted from the response.
|
|
264
|
+
default: 3
|
|
265
|
+
source:
|
|
266
|
+
openapi: streaming-asyncapi.yml
|
|
267
|
+
StreamLanguage:
|
|
268
|
+
docs: Configuration for the language emotion model.
|
|
269
|
+
properties:
|
|
270
|
+
sentiment:
|
|
271
|
+
type: optional<map<string, unknown>>
|
|
272
|
+
docs: >-
|
|
273
|
+
Configuration for sentiment predictions. If missing or null, no
|
|
274
|
+
sentiment predictions will be generated.
|
|
275
|
+
toxicity:
|
|
276
|
+
type: optional<map<string, unknown>>
|
|
277
|
+
docs: >-
|
|
278
|
+
Configuration for toxicity predictions. If missing or null, no
|
|
279
|
+
toxicity predictions will be generated.
|
|
280
|
+
granularity:
|
|
281
|
+
type: optional<string>
|
|
282
|
+
docs: >-
|
|
283
|
+
The granularity at which to generate predictions. Values are `word`,
|
|
284
|
+
`sentence`, `utterance`, or `passage`. To get a single prediction for
|
|
285
|
+
the entire text of your streaming payload use `passage`. Default value
|
|
286
|
+
is `word`.
|
|
287
|
+
source:
|
|
288
|
+
openapi: streaming-asyncapi.yml
|
|
289
|
+
Config:
|
|
290
|
+
docs: >
|
|
291
|
+
Configuration used to specify which models should be used and with what
|
|
292
|
+
settings.
|
|
293
|
+
properties:
|
|
294
|
+
burst:
|
|
295
|
+
type: optional<map<string, unknown>>
|
|
296
|
+
docs: |
|
|
297
|
+
Configuration for the vocal burst emotion model.
|
|
297
298
|
|
|
298
|
-
|
|
299
|
+
Note: Model configuration is not currently available in streaming.
|
|
299
300
|
|
|
300
|
-
|
|
301
|
-
|
|
302
|
-
|
|
303
|
-
|
|
304
|
-
|
|
301
|
+
Please use the default configuration by passing an empty object `{}`.
|
|
302
|
+
face:
|
|
303
|
+
type: optional<StreamFace>
|
|
304
|
+
docs: >
|
|
305
|
+
Configuration for the facial expression emotion model.
|
|
305
306
|
|
|
306
307
|
|
|
307
|
-
|
|
308
|
-
|
|
309
|
-
|
|
310
|
-
|
|
311
|
-
|
|
312
|
-
|
|
313
|
-
|
|
308
|
+
Note: Using the `reset_stream` parameter does not have any effect on
|
|
309
|
+
face identification. A single face identifier cache is maintained over
|
|
310
|
+
a full session whether `reset_stream` is used or not.
|
|
311
|
+
facemesh:
|
|
312
|
+
type: optional<map<string, unknown>>
|
|
313
|
+
docs: |
|
|
314
|
+
Configuration for the facemesh emotion model.
|
|
314
315
|
|
|
315
|
-
|
|
316
|
+
Note: Model configuration is not currently available in streaming.
|
|
316
317
|
|
|
317
|
-
|
|
318
|
-
|
|
319
|
-
|
|
320
|
-
|
|
321
|
-
|
|
322
|
-
|
|
323
|
-
|
|
324
|
-
|
|
318
|
+
Please use the default configuration by passing an empty object `{}`.
|
|
319
|
+
language:
|
|
320
|
+
type: optional<StreamLanguage>
|
|
321
|
+
docs: Configuration for the language emotion model.
|
|
322
|
+
prosody:
|
|
323
|
+
type: optional<map<string, unknown>>
|
|
324
|
+
docs: |
|
|
325
|
+
Configuration for the speech prosody emotion model.
|
|
325
326
|
|
|
326
|
-
|
|
327
|
+
Note: Model configuration is not currently available in streaming.
|
|
327
328
|
|
|
328
|
-
|
|
329
|
-
|
|
330
|
-
|
|
331
|
-
|
|
332
|
-
|
|
333
|
-
|
|
334
|
-
|
|
335
|
-
|
|
336
|
-
|
|
337
|
-
|
|
338
|
-
|
|
339
|
-
|
|
340
|
-
|
|
341
|
-
|
|
342
|
-
|
|
343
|
-
|
|
344
|
-
|
|
329
|
+
Please use the default configuration by passing an empty object `{}`.
|
|
330
|
+
source:
|
|
331
|
+
openapi: streaming-asyncapi.yml
|
|
332
|
+
StreamModelsEndpointPayload:
|
|
333
|
+
docs: Models endpoint payload
|
|
334
|
+
properties:
|
|
335
|
+
data:
|
|
336
|
+
type: optional<string>
|
|
337
|
+
models:
|
|
338
|
+
type: optional<Config>
|
|
339
|
+
docs: >
|
|
340
|
+
Configuration used to specify which models should be used and with
|
|
341
|
+
what settings.
|
|
342
|
+
stream_window_ms:
|
|
343
|
+
type: optional<double>
|
|
344
|
+
docs: >
|
|
345
|
+
Length in milliseconds of streaming sliding window.
|
|
345
346
|
|
|
346
347
|
|
|
347
|
-
|
|
348
|
-
|
|
348
|
+
Extending the length of this window will prepend media context from
|
|
349
|
+
past payloads into the current payload.
|
|
349
350
|
|
|
350
351
|
|
|
351
|
-
|
|
352
|
-
|
|
353
|
-
|
|
354
|
-
|
|
352
|
+
For example, if on the first payload you send 500ms of data and on the
|
|
353
|
+
second payload you send an additional 500ms of data, a window of at
|
|
354
|
+
least 1000ms will allow the model to process all 1000ms of stream
|
|
355
|
+
data.
|
|
355
356
|
|
|
356
357
|
|
|
357
|
-
|
|
358
|
-
|
|
358
|
+
A window of 600ms would append the full 500ms of the second payload to
|
|
359
|
+
the last 100ms of the first payload.
|
|
359
360
|
|
|
360
361
|
|
|
361
|
-
|
|
362
|
-
|
|
363
|
-
|
|
364
|
-
|
|
365
|
-
|
|
366
|
-
|
|
367
|
-
|
|
368
|
-
|
|
369
|
-
|
|
370
|
-
|
|
371
|
-
|
|
372
|
-
|
|
362
|
+
Note: This feature is currently only supported for audio data and
|
|
363
|
+
audio models. For other file types and models this parameter will be
|
|
364
|
+
ignored.
|
|
365
|
+
default: 5000
|
|
366
|
+
validation:
|
|
367
|
+
min: 500
|
|
368
|
+
max: 10000
|
|
369
|
+
reset_stream:
|
|
370
|
+
type: optional<boolean>
|
|
371
|
+
docs: >
|
|
372
|
+
Whether to reset the streaming sliding window before processing the
|
|
373
|
+
current payload.
|
|
373
374
|
|
|
374
375
|
|
|
375
|
-
|
|
376
|
-
|
|
376
|
+
If this parameter is set to `true` then past context will be deleted
|
|
377
|
+
before processing the current payload.
|
|
377
378
|
|
|
378
379
|
|
|
379
|
-
|
|
380
|
-
|
|
381
|
-
|
|
382
|
-
|
|
383
|
-
|
|
384
|
-
|
|
385
|
-
|
|
386
|
-
|
|
380
|
+
Use reset_stream when one audio file is done being processed and you
|
|
381
|
+
do not want context to leak across files.
|
|
382
|
+
default: false
|
|
383
|
+
raw_text:
|
|
384
|
+
type: optional<boolean>
|
|
385
|
+
docs: >
|
|
386
|
+
Set to `true` to enable the data parameter to be parsed as raw text
|
|
387
|
+
rather than base64 encoded bytes.
|
|
387
388
|
|
|
388
|
-
|
|
389
|
-
|
|
390
|
-
|
|
391
|
-
|
|
392
|
-
|
|
393
|
-
|
|
394
|
-
|
|
395
|
-
|
|
389
|
+
This parameter is useful if you want to send text to be processed by
|
|
390
|
+
the language model, but it cannot be used with other file types like
|
|
391
|
+
audio, image, or video.
|
|
392
|
+
default: false
|
|
393
|
+
job_details:
|
|
394
|
+
type: optional<boolean>
|
|
395
|
+
docs: >
|
|
396
|
+
Set to `true` to get details about the job.
|
|
396
397
|
|
|
397
398
|
|
|
398
|
-
|
|
399
|
-
|
|
400
|
-
|
|
399
|
+
This parameter can be set in the same payload as data or it can be set
|
|
400
|
+
without data and models configuration to get the job details between
|
|
401
|
+
payloads.
|
|
401
402
|
|
|
402
403
|
|
|
403
|
-
|
|
404
|
-
|
|
405
|
-
|
|
406
|
-
|
|
407
|
-
|
|
408
|
-
|
|
409
|
-
|
|
404
|
+
This parameter is useful to get the unique job ID.
|
|
405
|
+
default: false
|
|
406
|
+
payload_id:
|
|
407
|
+
type: optional<string>
|
|
408
|
+
docs: >
|
|
409
|
+
Pass an arbitrary string as the payload ID and get it back at the top
|
|
410
|
+
level of the socket response.
|
|
410
411
|
|
|
411
412
|
|
|
412
|
-
|
|
413
|
-
|
|
414
|
-
|
|
415
|
-
|
|
416
|
-
|
|
417
|
-
|
|
418
|
-
|
|
419
|
-
|
|
420
|
-
|
|
421
|
-
|
|
422
|
-
|
|
423
|
-
|
|
424
|
-
|
|
425
|
-
|
|
426
|
-
|
|
427
|
-
|
|
428
|
-
|
|
429
|
-
|
|
430
|
-
|
|
431
|
-
|
|
432
|
-
|
|
433
|
-
|
|
434
|
-
|
|
435
|
-
|
|
436
|
-
|
|
437
|
-
|
|
438
|
-
|
|
439
|
-
|
|
440
|
-
|
|
441
|
-
|
|
442
|
-
|
|
443
|
-
|
|
444
|
-
|
|
445
|
-
|
|
446
|
-
|
|
447
|
-
|
|
448
|
-
|
|
449
|
-
|
|
450
|
-
|
|
451
|
-
|
|
452
|
-
|
|
453
|
-
|
|
454
|
-
|
|
455
|
-
|
|
456
|
-
|
|
457
|
-
|
|
458
|
-
|
|
459
|
-
|
|
460
|
-
|
|
461
|
-
|
|
462
|
-
|
|
463
|
-
|
|
464
|
-
|
|
465
|
-
|
|
466
|
-
|
|
467
|
-
|
|
468
|
-
|
|
469
|
-
|
|
470
|
-
|
|
471
|
-
|
|
472
|
-
|
|
473
|
-
|
|
474
|
-
|
|
475
|
-
|
|
476
|
-
|
|
477
|
-
|
|
478
|
-
|
|
479
|
-
|
|
480
|
-
|
|
481
|
-
|
|
482
|
-
|
|
483
|
-
|
|
484
|
-
|
|
485
|
-
|
|
486
|
-
|
|
487
|
-
|
|
488
|
-
|
|
489
|
-
|
|
490
|
-
|
|
491
|
-
|
|
492
|
-
|
|
493
|
-
|
|
494
|
-
|
|
495
|
-
|
|
496
|
-
|
|
497
|
-
|
|
498
|
-
|
|
499
|
-
|
|
500
|
-
|
|
501
|
-
|
|
502
|
-
|
|
413
|
+
This can be useful if you have multiple requests running
|
|
414
|
+
asynchronously and want to disambiguate responses as they are
|
|
415
|
+
received.
|
|
416
|
+
source:
|
|
417
|
+
openapi: streaming-asyncapi.yml
|
|
418
|
+
EmotionEmbeddingItem:
|
|
419
|
+
properties:
|
|
420
|
+
name:
|
|
421
|
+
type: optional<string>
|
|
422
|
+
docs: Name of the emotion being expressed.
|
|
423
|
+
score:
|
|
424
|
+
type: optional<double>
|
|
425
|
+
docs: Embedding value for the emotion being expressed.
|
|
426
|
+
source:
|
|
427
|
+
openapi: streaming-asyncapi.yml
|
|
428
|
+
EmotionEmbedding:
|
|
429
|
+
docs: A high-dimensional embedding in emotion space.
|
|
430
|
+
type: list<EmotionEmbeddingItem>
|
|
431
|
+
StreamBoundingBox:
|
|
432
|
+
docs: A bounding box around a face.
|
|
433
|
+
properties:
|
|
434
|
+
x:
|
|
435
|
+
type: optional<double>
|
|
436
|
+
docs: x-coordinate of bounding box top left corner.
|
|
437
|
+
validation:
|
|
438
|
+
min: 0
|
|
439
|
+
'y':
|
|
440
|
+
type: optional<double>
|
|
441
|
+
docs: y-coordinate of bounding box top left corner.
|
|
442
|
+
validation:
|
|
443
|
+
min: 0
|
|
444
|
+
w:
|
|
445
|
+
type: optional<double>
|
|
446
|
+
docs: Bounding box width.
|
|
447
|
+
validation:
|
|
448
|
+
min: 0
|
|
449
|
+
h:
|
|
450
|
+
type: optional<double>
|
|
451
|
+
docs: Bounding box height.
|
|
452
|
+
validation:
|
|
453
|
+
min: 0
|
|
454
|
+
source:
|
|
455
|
+
openapi: streaming-asyncapi.yml
|
|
456
|
+
TimeRange:
|
|
457
|
+
docs: A time range with a beginning and end, measured in seconds.
|
|
458
|
+
properties:
|
|
459
|
+
begin:
|
|
460
|
+
type: optional<double>
|
|
461
|
+
docs: Beginning of time range in seconds.
|
|
462
|
+
validation:
|
|
463
|
+
min: 0
|
|
464
|
+
end:
|
|
465
|
+
type: optional<double>
|
|
466
|
+
docs: End of time range in seconds.
|
|
467
|
+
validation:
|
|
468
|
+
min: 0
|
|
469
|
+
source:
|
|
470
|
+
openapi: streaming-asyncapi.yml
|
|
471
|
+
TextPosition:
|
|
472
|
+
docs: >
|
|
473
|
+
Position of a segment of text within a larger document, measured in
|
|
474
|
+
characters. Uses zero-based indexing. The beginning index is inclusive and
|
|
475
|
+
the end index is exclusive.
|
|
476
|
+
properties:
|
|
477
|
+
begin:
|
|
478
|
+
type: optional<double>
|
|
479
|
+
docs: The index of the first character in the text segment, inclusive.
|
|
480
|
+
validation:
|
|
481
|
+
min: 0
|
|
482
|
+
end:
|
|
483
|
+
type: optional<double>
|
|
484
|
+
docs: The index of the last character in the text segment, exclusive.
|
|
485
|
+
validation:
|
|
486
|
+
min: 0
|
|
487
|
+
source:
|
|
488
|
+
openapi: streaming-asyncapi.yml
|
|
489
|
+
SentimentItem:
|
|
490
|
+
properties:
|
|
491
|
+
name:
|
|
492
|
+
type: optional<string>
|
|
493
|
+
docs: Level of sentiment, ranging from 1 (negative) to 9 (positive)
|
|
494
|
+
score:
|
|
495
|
+
type: optional<double>
|
|
496
|
+
docs: Prediction for this level of sentiment
|
|
497
|
+
source:
|
|
498
|
+
openapi: streaming-asyncapi.yml
|
|
499
|
+
Sentiment:
|
|
500
|
+
docs: >-
|
|
501
|
+
Sentiment predictions returned as a distribution. This model predicts the
|
|
502
|
+
probability that a given text could be interpreted as having each
|
|
503
|
+
sentiment level from 1 (negative) to 9 (positive).
|
|
503
504
|
|
|
504
505
|
|
|
505
|
-
|
|
506
|
-
|
|
507
|
-
|
|
508
|
-
|
|
509
|
-
|
|
510
|
-
|
|
511
|
-
|
|
512
|
-
|
|
513
|
-
|
|
514
|
-
|
|
515
|
-
|
|
516
|
-
|
|
517
|
-
|
|
518
|
-
|
|
519
|
-
|
|
520
|
-
|
|
521
|
-
|
|
522
|
-
|
|
523
|
-
|
|
524
|
-
|
|
525
|
-
|
|
526
|
-
|
|
527
|
-
|
|
528
|
-
|
|
506
|
+
Compared to returning one estimate of sentiment, this enables a more
|
|
507
|
+
nuanced analysis of a text's meaning. For example, a text with very
|
|
508
|
+
neutral sentiment would have an average rating of 5. But also a text that
|
|
509
|
+
could be interpreted as having very positive sentiment or very negative
|
|
510
|
+
sentiment would also have an average rating of 5. The average sentiment is
|
|
511
|
+
less informative than the distribution over sentiment, so this API returns
|
|
512
|
+
a value for each sentiment level.
|
|
513
|
+
type: list<SentimentItem>
|
|
514
|
+
ToxicityItem:
|
|
515
|
+
properties:
|
|
516
|
+
name:
|
|
517
|
+
type: optional<string>
|
|
518
|
+
docs: Category of toxicity.
|
|
519
|
+
score:
|
|
520
|
+
type: optional<double>
|
|
521
|
+
docs: Prediction for this category of toxicity
|
|
522
|
+
source:
|
|
523
|
+
openapi: streaming-asyncapi.yml
|
|
524
|
+
Toxicity:
|
|
525
|
+
docs: >-
|
|
526
|
+
Toxicity predictions returned as probabilities that the text can be
|
|
527
|
+
classified into the following categories: toxic, severe_toxic, obscene,
|
|
528
|
+
threat, insult, and identity_hate.
|
|
529
|
+
type: list<ToxicityItem>
|