hume 0.6.0 → 0.6.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.mock/definition/api.yml +12 -0
- package/.mock/definition/custom-models/__package__.yml +1392 -0
- package/.mock/definition/custom-models/datasets.yml +406 -0
- package/.mock/definition/custom-models/files.yml +263 -0
- package/.mock/definition/custom-models/jobs.yml +44 -0
- package/.mock/definition/custom-models/models.yml +303 -0
- package/.mock/definition/empathic-voice/__package__.yml +896 -0
- package/.mock/definition/empathic-voice/chat.yml +59 -0
- package/.mock/definition/empathic-voice/chatGroups.yml +80 -0
- package/.mock/definition/empathic-voice/chats.yml +86 -0
- package/.mock/definition/empathic-voice/configs.yml +413 -0
- package/.mock/definition/empathic-voice/prompts.yml +265 -0
- package/.mock/definition/empathic-voice/tools.yml +398 -0
- package/.mock/definition/expression-measurement/__package__.yml +1122 -0
- package/.mock/definition/expression-measurement/batch.yml +352 -0
- package/.mock/definition/expression-measurement/stream.yml +362 -0
- package/.mock/fern.config.json +4 -0
- package/Client.d.ts +3 -3
- package/Client.js +6 -6
- package/api/resources/customModels/resources/datasets/client/Client.js +9 -9
- package/api/resources/customModels/resources/files/client/Client.js +7 -7
- package/api/resources/customModels/resources/jobs/client/Client.js +2 -2
- package/api/resources/customModels/resources/models/client/Client.js +6 -6
- package/api/resources/empathicVoice/client/Client.d.ts +3 -0
- package/api/resources/empathicVoice/client/Client.js +5 -0
- package/api/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +1 -1
- package/api/resources/empathicVoice/resources/chatGroups/client/Client.d.ts +42 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/Client.js +178 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/index.d.ts +1 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/index.js +17 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.d.ts +21 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.js +5 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.d.ts +17 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.js +5 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/index.d.ts +2 -0
- package/api/resources/empathicVoice/resources/chatGroups/client/requests/index.js +2 -0
- package/api/resources/empathicVoice/resources/chatGroups/index.d.ts +1 -0
- package/api/resources/empathicVoice/resources/chatGroups/index.js +17 -0
- package/api/resources/empathicVoice/resources/chats/client/Client.js +2 -2
- package/api/resources/empathicVoice/resources/configs/client/Client.js +9 -9
- package/api/resources/empathicVoice/resources/index.d.ts +2 -0
- package/api/resources/empathicVoice/resources/index.js +3 -1
- package/api/resources/empathicVoice/resources/prompts/client/Client.js +9 -9
- package/api/resources/empathicVoice/resources/tools/client/Client.js +9 -9
- package/api/resources/empathicVoice/types/AssistantMessage.d.ts +1 -1
- package/api/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
- package/api/resources/empathicVoice/types/ChatMetadata.js +5 -0
- package/api/resources/empathicVoice/types/ReturnChat.d.ts +2 -0
- package/api/resources/empathicVoice/types/ReturnChatEvent.d.ts +2 -0
- package/api/resources/empathicVoice/types/ReturnChatGroup.d.ts +19 -0
- package/api/resources/empathicVoice/types/ReturnChatGroup.js +5 -0
- package/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +24 -0
- package/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +5 -0
- package/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +19 -0
- package/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +5 -0
- package/api/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +2 -0
- package/api/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
- package/api/resources/empathicVoice/types/ReturnPagedChatGroups.js +5 -0
- package/api/resources/empathicVoice/types/index.d.ts +5 -0
- package/api/resources/empathicVoice/types/index.js +5 -0
- package/api/resources/expressionMeasurement/resources/batch/client/Client.js +6 -6
- package/api/resources/index.d.ts +1 -1
- package/api/resources/index.js +2 -2
- package/core/form-data-utils/FormDataWrapper.js +2 -1
- package/dist/Client.d.ts +3 -3
- package/dist/Client.js +6 -6
- package/dist/api/resources/customModels/resources/datasets/client/Client.js +9 -9
- package/dist/api/resources/customModels/resources/files/client/Client.js +7 -7
- package/dist/api/resources/customModels/resources/jobs/client/Client.js +2 -2
- package/dist/api/resources/customModels/resources/models/client/Client.js +6 -6
- package/dist/api/resources/empathicVoice/client/Client.d.ts +3 -0
- package/dist/api/resources/empathicVoice/client/Client.js +5 -0
- package/dist/api/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +1 -1
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/Client.d.ts +42 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/Client.js +178 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/index.d.ts +1 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/index.js +17 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.d.ts +21 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.js +5 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.d.ts +17 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.js +5 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/index.d.ts +2 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/index.js +2 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/index.d.ts +1 -0
- package/dist/api/resources/empathicVoice/resources/chatGroups/index.js +17 -0
- package/dist/api/resources/empathicVoice/resources/chats/client/Client.js +2 -2
- package/dist/api/resources/empathicVoice/resources/configs/client/Client.js +9 -9
- package/dist/api/resources/empathicVoice/resources/index.d.ts +2 -0
- package/dist/api/resources/empathicVoice/resources/index.js +3 -1
- package/dist/api/resources/empathicVoice/resources/prompts/client/Client.js +9 -9
- package/dist/api/resources/empathicVoice/resources/tools/client/Client.js +9 -9
- package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +1 -1
- package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
- package/dist/api/resources/empathicVoice/types/ChatMetadata.js +5 -0
- package/dist/api/resources/empathicVoice/types/ReturnChat.d.ts +2 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatEvent.d.ts +2 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatGroup.d.ts +19 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatGroup.js +5 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +24 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +5 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +19 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +5 -0
- package/dist/api/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +2 -0
- package/dist/api/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
- package/dist/api/resources/empathicVoice/types/ReturnPagedChatGroups.js +5 -0
- package/dist/api/resources/empathicVoice/types/index.d.ts +5 -0
- package/dist/api/resources/empathicVoice/types/index.js +5 -0
- package/dist/api/resources/expressionMeasurement/resources/batch/client/Client.js +6 -6
- package/dist/api/resources/index.d.ts +1 -1
- package/dist/api/resources/index.js +2 -2
- package/dist/core/form-data-utils/FormDataWrapper.js +2 -1
- package/dist/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +2 -1
- package/dist/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.js +2 -0
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
- package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +36 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChat.d.ts +1 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChat.js +1 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatEvent.d.ts +1 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatEvent.js +1 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatGroup.d.ts +17 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatGroup.js +38 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +20 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +41 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +17 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +38 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +1 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.js +1 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
- package/dist/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.js +36 -0
- package/dist/serialization/resources/empathicVoice/types/index.d.ts +5 -0
- package/dist/serialization/resources/empathicVoice/types/index.js +5 -0
- package/dist/serialization/resources/index.d.ts +1 -1
- package/dist/serialization/resources/index.js +2 -2
- package/package.json +1 -1
- package/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +2 -1
- package/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.js +2 -0
- package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
- package/serialization/resources/empathicVoice/types/ChatMetadata.js +36 -0
- package/serialization/resources/empathicVoice/types/ReturnChat.d.ts +1 -0
- package/serialization/resources/empathicVoice/types/ReturnChat.js +1 -0
- package/serialization/resources/empathicVoice/types/ReturnChatEvent.d.ts +1 -0
- package/serialization/resources/empathicVoice/types/ReturnChatEvent.js +1 -0
- package/serialization/resources/empathicVoice/types/ReturnChatGroup.d.ts +17 -0
- package/serialization/resources/empathicVoice/types/ReturnChatGroup.js +38 -0
- package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +20 -0
- package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +41 -0
- package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +17 -0
- package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +38 -0
- package/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +1 -0
- package/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.js +1 -0
- package/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
- package/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.js +36 -0
- package/serialization/resources/empathicVoice/types/index.d.ts +5 -0
- package/serialization/resources/empathicVoice/types/index.js +5 -0
- package/serialization/resources/index.d.ts +1 -1
- package/serialization/resources/index.js +2 -2
|
@@ -0,0 +1,1122 @@
|
|
|
1
|
+
types:
|
|
2
|
+
Alternative: literal<"language_only">
|
|
3
|
+
Bcp47Tag:
|
|
4
|
+
enum:
|
|
5
|
+
- zh
|
|
6
|
+
- da
|
|
7
|
+
- nl
|
|
8
|
+
- en
|
|
9
|
+
- value: en-AU
|
|
10
|
+
name: EnAu
|
|
11
|
+
- value: en-IN
|
|
12
|
+
name: EnIn
|
|
13
|
+
- value: en-NZ
|
|
14
|
+
name: EnNz
|
|
15
|
+
- value: en-GB
|
|
16
|
+
name: EnGb
|
|
17
|
+
- fr
|
|
18
|
+
- value: fr-CA
|
|
19
|
+
name: FrCa
|
|
20
|
+
- de
|
|
21
|
+
- hi
|
|
22
|
+
- value: hi-Latn
|
|
23
|
+
name: HiLatn
|
|
24
|
+
- id
|
|
25
|
+
- it
|
|
26
|
+
- ja
|
|
27
|
+
- ko
|
|
28
|
+
- 'no'
|
|
29
|
+
- pl
|
|
30
|
+
- pt
|
|
31
|
+
- value: pt-BR
|
|
32
|
+
name: PtBr
|
|
33
|
+
- value: pt-PT
|
|
34
|
+
name: PtPt
|
|
35
|
+
- ru
|
|
36
|
+
- es
|
|
37
|
+
- value: es-419
|
|
38
|
+
name: Es419
|
|
39
|
+
- sv
|
|
40
|
+
- ta
|
|
41
|
+
- tr
|
|
42
|
+
- uk
|
|
43
|
+
BoundingBox:
|
|
44
|
+
docs: A bounding box around a face.
|
|
45
|
+
properties:
|
|
46
|
+
x:
|
|
47
|
+
type: optional<double>
|
|
48
|
+
docs: x-coordinate of bounding box top left corner.
|
|
49
|
+
validation:
|
|
50
|
+
min: 0
|
|
51
|
+
'y':
|
|
52
|
+
type: optional<double>
|
|
53
|
+
docs: y-coordinate of bounding box top left corner.
|
|
54
|
+
validation:
|
|
55
|
+
min: 0
|
|
56
|
+
w:
|
|
57
|
+
type: optional<double>
|
|
58
|
+
docs: Bounding box width.
|
|
59
|
+
validation:
|
|
60
|
+
min: 0
|
|
61
|
+
h:
|
|
62
|
+
type: optional<double>
|
|
63
|
+
docs: Bounding box height.
|
|
64
|
+
validation:
|
|
65
|
+
min: 0
|
|
66
|
+
BurstPrediction:
|
|
67
|
+
properties:
|
|
68
|
+
time: TimeInterval
|
|
69
|
+
emotions:
|
|
70
|
+
docs: A high-dimensional embedding in emotion space.
|
|
71
|
+
type: list<EmotionScore>
|
|
72
|
+
descriptions:
|
|
73
|
+
docs: Modality-specific descriptive features and their scores.
|
|
74
|
+
type: list<DescriptionsScore>
|
|
75
|
+
Classification: map<string, unknown>
|
|
76
|
+
CompletedEmbeddingGeneration:
|
|
77
|
+
properties:
|
|
78
|
+
created_timestamp_ms:
|
|
79
|
+
type: integer
|
|
80
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
81
|
+
started_timestamp_ms:
|
|
82
|
+
type: integer
|
|
83
|
+
docs: When this job started (Unix timestamp in milliseconds).
|
|
84
|
+
ended_timestamp_ms:
|
|
85
|
+
type: integer
|
|
86
|
+
docs: When this job ended (Unix timestamp in milliseconds).
|
|
87
|
+
CompletedInference:
|
|
88
|
+
properties:
|
|
89
|
+
created_timestamp_ms:
|
|
90
|
+
type: integer
|
|
91
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
92
|
+
started_timestamp_ms:
|
|
93
|
+
type: integer
|
|
94
|
+
docs: When this job started (Unix timestamp in milliseconds).
|
|
95
|
+
ended_timestamp_ms:
|
|
96
|
+
type: integer
|
|
97
|
+
docs: When this job ended (Unix timestamp in milliseconds).
|
|
98
|
+
num_predictions:
|
|
99
|
+
type: integer
|
|
100
|
+
docs: The number of predictions that were generated by this job.
|
|
101
|
+
num_errors:
|
|
102
|
+
type: integer
|
|
103
|
+
docs: The number of errors that occurred while running this job.
|
|
104
|
+
CompletedTlInference:
|
|
105
|
+
properties:
|
|
106
|
+
created_timestamp_ms:
|
|
107
|
+
type: integer
|
|
108
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
109
|
+
started_timestamp_ms:
|
|
110
|
+
type: integer
|
|
111
|
+
docs: When this job started (Unix timestamp in milliseconds).
|
|
112
|
+
ended_timestamp_ms:
|
|
113
|
+
type: integer
|
|
114
|
+
docs: When this job ended (Unix timestamp in milliseconds).
|
|
115
|
+
num_predictions:
|
|
116
|
+
type: integer
|
|
117
|
+
docs: The number of predictions that were generated by this job.
|
|
118
|
+
num_errors:
|
|
119
|
+
type: integer
|
|
120
|
+
docs: The number of errors that occurred while running this job.
|
|
121
|
+
CompletedTraining:
|
|
122
|
+
properties:
|
|
123
|
+
created_timestamp_ms:
|
|
124
|
+
type: integer
|
|
125
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
126
|
+
started_timestamp_ms:
|
|
127
|
+
type: integer
|
|
128
|
+
docs: When this job started (Unix timestamp in milliseconds).
|
|
129
|
+
ended_timestamp_ms:
|
|
130
|
+
type: integer
|
|
131
|
+
docs: When this job ended (Unix timestamp in milliseconds).
|
|
132
|
+
custom_model: TrainingCustomModel
|
|
133
|
+
alternatives: optional<map<string, TrainingCustomModel>>
|
|
134
|
+
CustomModelPrediction:
|
|
135
|
+
properties:
|
|
136
|
+
output: map<string, double>
|
|
137
|
+
error: string
|
|
138
|
+
task_type: string
|
|
139
|
+
CustomModelRequest:
|
|
140
|
+
properties:
|
|
141
|
+
name: string
|
|
142
|
+
description: optional<string>
|
|
143
|
+
tags: optional<list<Tag>>
|
|
144
|
+
Dataset:
|
|
145
|
+
discriminated: false
|
|
146
|
+
union:
|
|
147
|
+
- DatasetId
|
|
148
|
+
- DatasetVersionId
|
|
149
|
+
DatasetId:
|
|
150
|
+
properties:
|
|
151
|
+
id:
|
|
152
|
+
type: string
|
|
153
|
+
validation:
|
|
154
|
+
format: uuid
|
|
155
|
+
DatasetVersionId:
|
|
156
|
+
properties:
|
|
157
|
+
version_id:
|
|
158
|
+
type: string
|
|
159
|
+
validation:
|
|
160
|
+
format: uuid
|
|
161
|
+
DescriptionsScore:
|
|
162
|
+
properties:
|
|
163
|
+
name:
|
|
164
|
+
type: string
|
|
165
|
+
docs: Name of the descriptive feature being expressed.
|
|
166
|
+
score:
|
|
167
|
+
type: string
|
|
168
|
+
docs: Embedding value for the descriptive feature being expressed.
|
|
169
|
+
Direction:
|
|
170
|
+
enum:
|
|
171
|
+
- asc
|
|
172
|
+
- desc
|
|
173
|
+
EmbeddingGenerationBaseRequest:
|
|
174
|
+
properties:
|
|
175
|
+
registry_file_details:
|
|
176
|
+
type: optional<list<RegistryFileDetail>>
|
|
177
|
+
docs: File ID and File URL pairs for an asset registry file
|
|
178
|
+
EmotionScore:
|
|
179
|
+
properties:
|
|
180
|
+
name:
|
|
181
|
+
type: string
|
|
182
|
+
docs: Name of the emotion being expressed.
|
|
183
|
+
score:
|
|
184
|
+
type: double
|
|
185
|
+
docs: Embedding value for the emotion being expressed.
|
|
186
|
+
Error:
|
|
187
|
+
properties:
|
|
188
|
+
message:
|
|
189
|
+
type: string
|
|
190
|
+
docs: An error message.
|
|
191
|
+
file:
|
|
192
|
+
type: string
|
|
193
|
+
docs: A file path relative to the top level source URL or file.
|
|
194
|
+
EvaluationArgs:
|
|
195
|
+
properties:
|
|
196
|
+
validation: optional<ValidationArgs>
|
|
197
|
+
Face:
|
|
198
|
+
properties:
|
|
199
|
+
fps_pred:
|
|
200
|
+
type: optional<double>
|
|
201
|
+
docs: >-
|
|
202
|
+
Number of frames per second to process. Other frames will be omitted
|
|
203
|
+
from the response. Set to `0` to process every frame.
|
|
204
|
+
default: 3
|
|
205
|
+
prob_threshold:
|
|
206
|
+
type: optional<double>
|
|
207
|
+
docs: >-
|
|
208
|
+
Face detection probability threshold. Faces detected with a
|
|
209
|
+
probability less than this threshold will be omitted from the
|
|
210
|
+
response.
|
|
211
|
+
default: 0.99
|
|
212
|
+
validation:
|
|
213
|
+
min: 0
|
|
214
|
+
max: 1
|
|
215
|
+
identify_faces:
|
|
216
|
+
type: optional<boolean>
|
|
217
|
+
docs: >-
|
|
218
|
+
Whether to return identifiers for faces across frames. If `true`,
|
|
219
|
+
unique identifiers will be assigned to face bounding boxes to
|
|
220
|
+
differentiate different faces. If `false`, all faces will be tagged
|
|
221
|
+
with an `unknown` ID.
|
|
222
|
+
min_face_size:
|
|
223
|
+
type: optional<integer>
|
|
224
|
+
docs: >-
|
|
225
|
+
Minimum bounding box side length in pixels to treat as a face. Faces
|
|
226
|
+
detected with a bounding box side length in pixels less than this
|
|
227
|
+
threshold will be omitted from the response.
|
|
228
|
+
default: 60
|
|
229
|
+
validation:
|
|
230
|
+
min: 10
|
|
231
|
+
facs: optional<Unconfigurable>
|
|
232
|
+
descriptions: optional<Unconfigurable>
|
|
233
|
+
save_faces:
|
|
234
|
+
type: optional<boolean>
|
|
235
|
+
docs: >-
|
|
236
|
+
Whether to extract and save the detected faces in the artifacts zip
|
|
237
|
+
created by each job.
|
|
238
|
+
FacePrediction:
|
|
239
|
+
properties:
|
|
240
|
+
frame:
|
|
241
|
+
type: integer
|
|
242
|
+
docs: Frame number
|
|
243
|
+
time:
|
|
244
|
+
type: double
|
|
245
|
+
docs: Time in seconds when face detection occurred.
|
|
246
|
+
prob:
|
|
247
|
+
type: double
|
|
248
|
+
docs: The predicted probability that a detected face was actually a face.
|
|
249
|
+
box: BoundingBox
|
|
250
|
+
emotions:
|
|
251
|
+
docs: A high-dimensional embedding in emotion space.
|
|
252
|
+
type: list<EmotionScore>
|
|
253
|
+
facs:
|
|
254
|
+
type: optional<list<FacsScore>>
|
|
255
|
+
docs: FACS 2.0 features and their scores.
|
|
256
|
+
descriptions:
|
|
257
|
+
type: optional<list<DescriptionsScore>>
|
|
258
|
+
docs: Modality-specific descriptive features and their scores.
|
|
259
|
+
FacemeshPrediction:
|
|
260
|
+
properties:
|
|
261
|
+
emotions:
|
|
262
|
+
docs: A high-dimensional embedding in emotion space.
|
|
263
|
+
type: list<EmotionScore>
|
|
264
|
+
FacsScore:
|
|
265
|
+
properties:
|
|
266
|
+
name:
|
|
267
|
+
type: string
|
|
268
|
+
docs: Name of the FACS 2.0 feature being expressed.
|
|
269
|
+
score:
|
|
270
|
+
type: string
|
|
271
|
+
docs: Embedding value for the FACS 2.0 feature being expressed.
|
|
272
|
+
Failed:
|
|
273
|
+
properties:
|
|
274
|
+
created_timestamp_ms:
|
|
275
|
+
type: integer
|
|
276
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
277
|
+
started_timestamp_ms:
|
|
278
|
+
type: integer
|
|
279
|
+
docs: When this job started (Unix timestamp in milliseconds).
|
|
280
|
+
ended_timestamp_ms:
|
|
281
|
+
type: integer
|
|
282
|
+
docs: When this job ended (Unix timestamp in milliseconds).
|
|
283
|
+
message:
|
|
284
|
+
type: string
|
|
285
|
+
docs: An error message.
|
|
286
|
+
File:
|
|
287
|
+
properties:
|
|
288
|
+
filename:
|
|
289
|
+
type: optional<string>
|
|
290
|
+
docs: The name of the file.
|
|
291
|
+
content_type:
|
|
292
|
+
type: optional<string>
|
|
293
|
+
docs: The content type of the file.
|
|
294
|
+
md5sum:
|
|
295
|
+
type: string
|
|
296
|
+
docs: The MD5 checksum of the file.
|
|
297
|
+
Granularity:
|
|
298
|
+
enum:
|
|
299
|
+
- word
|
|
300
|
+
- sentence
|
|
301
|
+
- utterance
|
|
302
|
+
- conversational_turn
|
|
303
|
+
docs: >-
|
|
304
|
+
The granularity at which to generate predictions. `utterance` corresponds
|
|
305
|
+
to a natural pause or break in conversation, while `conversational_turn`
|
|
306
|
+
corresponds to a change in speaker.
|
|
307
|
+
GroupedPredictionsBurstPrediction:
|
|
308
|
+
properties:
|
|
309
|
+
id:
|
|
310
|
+
type: string
|
|
311
|
+
docs: >-
|
|
312
|
+
An automatically generated label to identify individuals in your media
|
|
313
|
+
file. Will be `unknown` if you have chosen to disable identification,
|
|
314
|
+
or if the model is unable to distinguish between individuals.
|
|
315
|
+
predictions: list<BurstPrediction>
|
|
316
|
+
GroupedPredictionsFacePrediction:
|
|
317
|
+
properties:
|
|
318
|
+
id:
|
|
319
|
+
type: string
|
|
320
|
+
docs: >-
|
|
321
|
+
An automatically generated label to identify individuals in your media
|
|
322
|
+
file. Will be `unknown` if you have chosen to disable identification,
|
|
323
|
+
or if the model is unable to distinguish between individuals.
|
|
324
|
+
predictions: list<FacePrediction>
|
|
325
|
+
GroupedPredictionsFacemeshPrediction:
|
|
326
|
+
properties:
|
|
327
|
+
id:
|
|
328
|
+
type: string
|
|
329
|
+
docs: >-
|
|
330
|
+
An automatically generated label to identify individuals in your media
|
|
331
|
+
file. Will be `unknown` if you have chosen to disable identification,
|
|
332
|
+
or if the model is unable to distinguish between individuals.
|
|
333
|
+
predictions: list<FacemeshPrediction>
|
|
334
|
+
GroupedPredictionsLanguagePrediction:
|
|
335
|
+
properties:
|
|
336
|
+
id:
|
|
337
|
+
type: string
|
|
338
|
+
docs: >-
|
|
339
|
+
An automatically generated label to identify individuals in your media
|
|
340
|
+
file. Will be `unknown` if you have chosen to disable identification,
|
|
341
|
+
or if the model is unable to distinguish between individuals.
|
|
342
|
+
predictions: list<LanguagePrediction>
|
|
343
|
+
GroupedPredictionsNerPrediction:
|
|
344
|
+
properties:
|
|
345
|
+
id:
|
|
346
|
+
type: string
|
|
347
|
+
docs: >-
|
|
348
|
+
An automatically generated label to identify individuals in your media
|
|
349
|
+
file. Will be `unknown` if you have chosen to disable identification,
|
|
350
|
+
or if the model is unable to distinguish between individuals.
|
|
351
|
+
predictions: list<NerPrediction>
|
|
352
|
+
GroupedPredictionsProsodyPrediction:
|
|
353
|
+
properties:
|
|
354
|
+
id:
|
|
355
|
+
type: string
|
|
356
|
+
docs: >-
|
|
357
|
+
An automatically generated label to identify individuals in your media
|
|
358
|
+
file. Will be `unknown` if you have chosen to disable identification,
|
|
359
|
+
or if the model is unable to distinguish between individuals.
|
|
360
|
+
predictions: list<ProsodyPrediction>
|
|
361
|
+
InProgress:
|
|
362
|
+
properties:
|
|
363
|
+
created_timestamp_ms:
|
|
364
|
+
type: integer
|
|
365
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
366
|
+
started_timestamp_ms:
|
|
367
|
+
type: integer
|
|
368
|
+
docs: When this job started (Unix timestamp in milliseconds).
|
|
369
|
+
InferenceBaseRequest:
|
|
370
|
+
properties:
|
|
371
|
+
models: optional<Models>
|
|
372
|
+
transcription: optional<Transcription>
|
|
373
|
+
urls:
|
|
374
|
+
type: optional<list<string>>
|
|
375
|
+
docs: >-
|
|
376
|
+
URLs to the media files to be processed. Each must be a valid public
|
|
377
|
+
URL to a media file (see recommended input filetypes) or an archive
|
|
378
|
+
(`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`) of media files.
|
|
379
|
+
|
|
380
|
+
|
|
381
|
+
If you wish to supply more than 100 URLs, consider providing them as
|
|
382
|
+
an archive (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`).
|
|
383
|
+
registry_files:
|
|
384
|
+
type: optional<list<string>>
|
|
385
|
+
docs: List of File IDs corresponding to the files in the asset registry.
|
|
386
|
+
text:
|
|
387
|
+
type: optional<list<string>>
|
|
388
|
+
docs: Text to supply directly to our language and NER models.
|
|
389
|
+
callback_url:
|
|
390
|
+
type: optional<string>
|
|
391
|
+
docs: >-
|
|
392
|
+
If provided, a `POST` request will be made to the URL with the
|
|
393
|
+
generated predictions on completion or the error message on failure.
|
|
394
|
+
validation:
|
|
395
|
+
format: url
|
|
396
|
+
notify:
|
|
397
|
+
type: optional<boolean>
|
|
398
|
+
docs: >-
|
|
399
|
+
Whether to send an email notification to the user upon job
|
|
400
|
+
completion/failure.
|
|
401
|
+
InferencePrediction:
|
|
402
|
+
properties:
|
|
403
|
+
file:
|
|
404
|
+
type: string
|
|
405
|
+
docs: A file path relative to the top level source URL or file.
|
|
406
|
+
models: ModelsPredictions
|
|
407
|
+
InferenceRequest:
|
|
408
|
+
properties:
|
|
409
|
+
models: optional<Models>
|
|
410
|
+
transcription: optional<Transcription>
|
|
411
|
+
urls:
|
|
412
|
+
type: optional<list<string>>
|
|
413
|
+
docs: >-
|
|
414
|
+
URLs to the media files to be processed. Each must be a valid public
|
|
415
|
+
URL to a media file (see recommended input filetypes) or an archive
|
|
416
|
+
(`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`) of media files.
|
|
417
|
+
|
|
418
|
+
|
|
419
|
+
If you wish to supply more than 100 URLs, consider providing them as
|
|
420
|
+
an archive (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`).
|
|
421
|
+
registry_files:
|
|
422
|
+
type: optional<list<string>>
|
|
423
|
+
docs: List of File IDs corresponding to the files in the asset registry.
|
|
424
|
+
text:
|
|
425
|
+
type: optional<list<string>>
|
|
426
|
+
docs: Text to supply directly to our language and NER models.
|
|
427
|
+
callback_url:
|
|
428
|
+
type: optional<string>
|
|
429
|
+
docs: >-
|
|
430
|
+
If provided, a `POST` request will be made to the URL with the
|
|
431
|
+
generated predictions on completion or the error message on failure.
|
|
432
|
+
validation:
|
|
433
|
+
format: url
|
|
434
|
+
notify:
|
|
435
|
+
type: optional<boolean>
|
|
436
|
+
docs: >-
|
|
437
|
+
Whether to send an email notification to the user upon job
|
|
438
|
+
completion/failure.
|
|
439
|
+
files: list<File>
|
|
440
|
+
InferenceResults:
|
|
441
|
+
properties:
|
|
442
|
+
predictions: list<InferencePrediction>
|
|
443
|
+
errors: list<Error>
|
|
444
|
+
InferenceSourcePredictResult:
|
|
445
|
+
properties:
|
|
446
|
+
source: Source
|
|
447
|
+
results: optional<InferenceResults>
|
|
448
|
+
error:
|
|
449
|
+
type: optional<string>
|
|
450
|
+
docs: An error message.
|
|
451
|
+
JobEmbeddingGeneration:
|
|
452
|
+
properties:
|
|
453
|
+
job_id:
|
|
454
|
+
type: string
|
|
455
|
+
docs: The ID associated with this job.
|
|
456
|
+
validation:
|
|
457
|
+
format: uuid
|
|
458
|
+
user_id:
|
|
459
|
+
type: string
|
|
460
|
+
validation:
|
|
461
|
+
format: uuid
|
|
462
|
+
request: EmbeddingGenerationBaseRequest
|
|
463
|
+
state: StateEmbeddingGeneration
|
|
464
|
+
JobInference:
|
|
465
|
+
properties:
|
|
466
|
+
job_id:
|
|
467
|
+
type: string
|
|
468
|
+
docs: The ID associated with this job.
|
|
469
|
+
validation:
|
|
470
|
+
format: uuid
|
|
471
|
+
user_id:
|
|
472
|
+
type: string
|
|
473
|
+
validation:
|
|
474
|
+
format: uuid
|
|
475
|
+
request: InferenceRequest
|
|
476
|
+
state: StateInference
|
|
477
|
+
JobTlInference:
|
|
478
|
+
properties:
|
|
479
|
+
job_id:
|
|
480
|
+
type: string
|
|
481
|
+
docs: The ID associated with this job.
|
|
482
|
+
validation:
|
|
483
|
+
format: uuid
|
|
484
|
+
user_id:
|
|
485
|
+
type: string
|
|
486
|
+
validation:
|
|
487
|
+
format: uuid
|
|
488
|
+
request: TlInferenceBaseRequest
|
|
489
|
+
state: StateTlInference
|
|
490
|
+
JobTraining:
|
|
491
|
+
properties:
|
|
492
|
+
job_id:
|
|
493
|
+
type: string
|
|
494
|
+
docs: The ID associated with this job.
|
|
495
|
+
validation:
|
|
496
|
+
format: uuid
|
|
497
|
+
user_id:
|
|
498
|
+
type: string
|
|
499
|
+
validation:
|
|
500
|
+
format: uuid
|
|
501
|
+
request: TrainingBaseRequest
|
|
502
|
+
state: StateTraining
|
|
503
|
+
JobId:
|
|
504
|
+
properties:
|
|
505
|
+
job_id:
|
|
506
|
+
type: string
|
|
507
|
+
docs: The ID of the started job.
|
|
508
|
+
validation:
|
|
509
|
+
format: uuid
|
|
510
|
+
Language:
|
|
511
|
+
properties:
|
|
512
|
+
granularity: optional<Granularity>
|
|
513
|
+
sentiment: optional<Unconfigurable>
|
|
514
|
+
toxicity: optional<Unconfigurable>
|
|
515
|
+
identify_speakers:
|
|
516
|
+
type: optional<boolean>
|
|
517
|
+
docs: >-
|
|
518
|
+
Whether to return identifiers for speakers over time. If `true`,
|
|
519
|
+
unique identifiers will be assigned to spoken words to differentiate
|
|
520
|
+
different speakers. If `false`, all speakers will be tagged with an
|
|
521
|
+
`unknown` ID.
|
|
522
|
+
LanguagePrediction:
|
|
523
|
+
properties:
|
|
524
|
+
text:
|
|
525
|
+
type: string
|
|
526
|
+
docs: A segment of text (like a word or a sentence).
|
|
527
|
+
position: PositionInterval
|
|
528
|
+
time: optional<TimeInterval>
|
|
529
|
+
confidence:
|
|
530
|
+
type: optional<double>
|
|
531
|
+
docs: >-
|
|
532
|
+
Value between `0.0` and `1.0` that indicates our transcription model's
|
|
533
|
+
relative confidence in this text.
|
|
534
|
+
speaker_confidence:
|
|
535
|
+
type: optional<double>
|
|
536
|
+
docs: >-
|
|
537
|
+
Value between `0.0` and `1.0` that indicates our transcription model's
|
|
538
|
+
relative confidence that this text was spoken by this speaker.
|
|
539
|
+
emotions:
|
|
540
|
+
docs: A high-dimensional embedding in emotion space.
|
|
541
|
+
type: list<EmotionScore>
|
|
542
|
+
sentiment:
|
|
543
|
+
type: optional<list<SentimentScore>>
|
|
544
|
+
docs: >-
|
|
545
|
+
Sentiment predictions returned as a distribution. This model predicts
|
|
546
|
+
the probability that a given text could be interpreted as having each
|
|
547
|
+
sentiment level from `1` (negative) to `9` (positive).
|
|
548
|
+
|
|
549
|
+
|
|
550
|
+
Compared to returning one estimate of sentiment, this enables a more
|
|
551
|
+
nuanced analysis of a text's meaning. For example, a text with very
|
|
552
|
+
neutral sentiment would have an average rating of `5`. But also a text
|
|
553
|
+
that could be interpreted as having very positive sentiment or very
|
|
554
|
+
negative sentiment would also have an average rating of `5`. The
|
|
555
|
+
average sentiment is less informative than the distribution over
|
|
556
|
+
sentiment, so this API returns a value for each sentiment level.
|
|
557
|
+
toxicity:
|
|
558
|
+
type: optional<list<ToxicityScore>>
|
|
559
|
+
docs: >-
|
|
560
|
+
Toxicity predictions returned as probabilities that the text can be
|
|
561
|
+
classified into the following categories: `toxic`, `severe_toxic`,
|
|
562
|
+
`obscene`, `threat`, `insult`, and `identity_hate`.
|
|
563
|
+
Models:
|
|
564
|
+
properties:
|
|
565
|
+
face: optional<Face>
|
|
566
|
+
burst: optional<Unconfigurable>
|
|
567
|
+
prosody: optional<Prosody>
|
|
568
|
+
language: optional<Language>
|
|
569
|
+
ner: optional<Ner>
|
|
570
|
+
facemesh: optional<Unconfigurable>
|
|
571
|
+
ModelsPredictions:
|
|
572
|
+
properties:
|
|
573
|
+
face: optional<PredictionsOptionalNullFacePrediction>
|
|
574
|
+
burst: optional<PredictionsOptionalNullBurstPrediction>
|
|
575
|
+
prosody: optional<PredictionsOptionalTranscriptionMetadataProsodyPrediction>
|
|
576
|
+
language: optional<PredictionsOptionalTranscriptionMetadataLanguagePrediction>
|
|
577
|
+
ner: optional<PredictionsOptionalTranscriptionMetadataNerPrediction>
|
|
578
|
+
facemesh: optional<PredictionsOptionalNullFacemeshPrediction>
|
|
579
|
+
Ner:
|
|
580
|
+
properties:
|
|
581
|
+
identify_speakers:
|
|
582
|
+
type: optional<boolean>
|
|
583
|
+
docs: >-
|
|
584
|
+
Whether to return identifiers for speakers over time. If `true`,
|
|
585
|
+
unique identifiers will be assigned to spoken words to differentiate
|
|
586
|
+
different speakers. If `false`, all speakers will be tagged with an
|
|
587
|
+
`unknown` ID.
|
|
588
|
+
NerPrediction:
|
|
589
|
+
properties:
|
|
590
|
+
entity:
|
|
591
|
+
type: string
|
|
592
|
+
docs: The recognized topic or entity.
|
|
593
|
+
position: PositionInterval
|
|
594
|
+
entity_confidence:
|
|
595
|
+
type: double
|
|
596
|
+
docs: Our NER model's relative confidence in the recognized topic or entity.
|
|
597
|
+
support:
|
|
598
|
+
type: double
|
|
599
|
+
docs: A measure of how often the entity is linked to by other entities.
|
|
600
|
+
uri:
|
|
601
|
+
type: string
|
|
602
|
+
docs: >-
|
|
603
|
+
A URL which provides more information about the recognized topic or
|
|
604
|
+
entity.
|
|
605
|
+
link_word:
|
|
606
|
+
type: string
|
|
607
|
+
docs: The specific word to which the emotion predictions are linked.
|
|
608
|
+
time: optional<TimeInterval>
|
|
609
|
+
confidence:
|
|
610
|
+
type: optional<double>
|
|
611
|
+
docs: >-
|
|
612
|
+
Value between `0.0` and `1.0` that indicates our transcription model's
|
|
613
|
+
relative confidence in this text.
|
|
614
|
+
speaker_confidence:
|
|
615
|
+
type: optional<double>
|
|
616
|
+
docs: >-
|
|
617
|
+
Value between `0.0` and `1.0` that indicates our transcription model's
|
|
618
|
+
relative confidence that this text was spoken by this speaker.
|
|
619
|
+
emotions:
|
|
620
|
+
docs: A high-dimensional embedding in emotion space.
|
|
621
|
+
type: list<EmotionScore>
|
|
622
|
+
'Null':
|
|
623
|
+
docs: No associated metadata for this model. Value will be `null`.
|
|
624
|
+
type: map<string, unknown>
|
|
625
|
+
PositionInterval:
|
|
626
|
+
docs: >-
|
|
627
|
+
Position of a segment of text within a larger document, measured in
|
|
628
|
+
characters. Uses zero-based indexing. The beginning index is inclusive and
|
|
629
|
+
the end index is exclusive.
|
|
630
|
+
properties:
|
|
631
|
+
begin:
|
|
632
|
+
type: integer
|
|
633
|
+
docs: The index of the first character in the text segment, inclusive.
|
|
634
|
+
end:
|
|
635
|
+
type: integer
|
|
636
|
+
docs: The index of the last character in the text segment, exclusive.
|
|
637
|
+
PredictionsOptionalNullBurstPrediction:
|
|
638
|
+
properties:
|
|
639
|
+
metadata: optional<Null>
|
|
640
|
+
grouped_predictions: list<GroupedPredictionsBurstPrediction>
|
|
641
|
+
PredictionsOptionalNullFacePrediction:
|
|
642
|
+
properties:
|
|
643
|
+
metadata: optional<Null>
|
|
644
|
+
grouped_predictions: list<GroupedPredictionsFacePrediction>
|
|
645
|
+
PredictionsOptionalNullFacemeshPrediction:
|
|
646
|
+
properties:
|
|
647
|
+
metadata: optional<Null>
|
|
648
|
+
grouped_predictions: list<GroupedPredictionsFacemeshPrediction>
|
|
649
|
+
PredictionsOptionalTranscriptionMetadataLanguagePrediction:
|
|
650
|
+
properties:
|
|
651
|
+
metadata: optional<TranscriptionMetadata>
|
|
652
|
+
grouped_predictions: list<GroupedPredictionsLanguagePrediction>
|
|
653
|
+
PredictionsOptionalTranscriptionMetadataNerPrediction:
|
|
654
|
+
properties:
|
|
655
|
+
metadata: optional<TranscriptionMetadata>
|
|
656
|
+
grouped_predictions: list<GroupedPredictionsNerPrediction>
|
|
657
|
+
PredictionsOptionalTranscriptionMetadataProsodyPrediction:
|
|
658
|
+
properties:
|
|
659
|
+
metadata: optional<TranscriptionMetadata>
|
|
660
|
+
grouped_predictions: list<GroupedPredictionsProsodyPrediction>
|
|
661
|
+
Prosody:
|
|
662
|
+
properties:
|
|
663
|
+
granularity: optional<Granularity>
|
|
664
|
+
window: optional<Window>
|
|
665
|
+
identify_speakers:
|
|
666
|
+
type: optional<boolean>
|
|
667
|
+
docs: >-
|
|
668
|
+
Whether to return identifiers for speakers over time. If `true`,
|
|
669
|
+
unique identifiers will be assigned to spoken words to differentiate
|
|
670
|
+
different speakers. If `false`, all speakers will be tagged with an
|
|
671
|
+
`unknown` ID.
|
|
672
|
+
ProsodyPrediction:
|
|
673
|
+
properties:
|
|
674
|
+
text:
|
|
675
|
+
type: optional<string>
|
|
676
|
+
docs: A segment of text (like a word or a sentence).
|
|
677
|
+
time: TimeInterval
|
|
678
|
+
confidence:
|
|
679
|
+
type: optional<double>
|
|
680
|
+
docs: >-
|
|
681
|
+
Value between `0.0` and `1.0` that indicates our transcription model's
|
|
682
|
+
relative confidence in this text.
|
|
683
|
+
speaker_confidence:
|
|
684
|
+
type: optional<double>
|
|
685
|
+
docs: >-
|
|
686
|
+
Value between `0.0` and `1.0` that indicates our transcription model's
|
|
687
|
+
relative confidence that this text was spoken by this speaker.
|
|
688
|
+
emotions:
|
|
689
|
+
docs: A high-dimensional embedding in emotion space.
|
|
690
|
+
type: list<EmotionScore>
|
|
691
|
+
Queued:
|
|
692
|
+
properties:
|
|
693
|
+
created_timestamp_ms:
|
|
694
|
+
type: integer
|
|
695
|
+
docs: When this job was created (Unix timestamp in milliseconds).
|
|
696
|
+
RegistryFileDetail:
|
|
697
|
+
properties:
|
|
698
|
+
file_id:
|
|
699
|
+
type: string
|
|
700
|
+
docs: File ID in the Asset Registry
|
|
701
|
+
file_url:
|
|
702
|
+
type: string
|
|
703
|
+
docs: URL to the file in the Asset Registry
|
|
704
|
+
Regression: map<string, unknown>
|
|
705
|
+
SentimentScore:
|
|
706
|
+
properties:
|
|
707
|
+
name:
|
|
708
|
+
type: string
|
|
709
|
+
docs: Level of sentiment, ranging from `1` (negative) to `9` (positive)
|
|
710
|
+
score:
|
|
711
|
+
type: string
|
|
712
|
+
docs: Prediction for this level of sentiment
|
|
713
|
+
SortBy:
|
|
714
|
+
enum:
|
|
715
|
+
- created
|
|
716
|
+
- started
|
|
717
|
+
- ended
|
|
718
|
+
Source:
|
|
719
|
+
discriminant: type
|
|
720
|
+
base-properties: {}
|
|
721
|
+
union:
|
|
722
|
+
url: SourceUrl
|
|
723
|
+
file: SourceFile
|
|
724
|
+
text: SourceTextSource
|
|
725
|
+
SourceFile:
|
|
726
|
+
properties: {}
|
|
727
|
+
extends:
|
|
728
|
+
- File
|
|
729
|
+
SourceTextSource:
|
|
730
|
+
properties: {}
|
|
731
|
+
SourceUrl:
|
|
732
|
+
properties: {}
|
|
733
|
+
extends:
|
|
734
|
+
- Url
|
|
735
|
+
Url:
|
|
736
|
+
properties:
|
|
737
|
+
url:
|
|
738
|
+
type: string
|
|
739
|
+
docs: The URL of the source media file.
|
|
740
|
+
StateEmbeddingGeneration:
|
|
741
|
+
discriminant: status
|
|
742
|
+
base-properties: {}
|
|
743
|
+
union:
|
|
744
|
+
QUEUED: StateEmbeddingGenerationQueued
|
|
745
|
+
IN_PROGRESS: StateEmbeddingGenerationInProgress
|
|
746
|
+
COMPLETED: StateEmbeddingGenerationCompletedEmbeddingGeneration
|
|
747
|
+
FAILED: StateEmbeddingGenerationFailed
|
|
748
|
+
StateEmbeddingGenerationCompletedEmbeddingGeneration:
|
|
749
|
+
properties: {}
|
|
750
|
+
extends:
|
|
751
|
+
- CompletedEmbeddingGeneration
|
|
752
|
+
StateEmbeddingGenerationFailed:
|
|
753
|
+
properties: {}
|
|
754
|
+
extends:
|
|
755
|
+
- Failed
|
|
756
|
+
StateEmbeddingGenerationInProgress:
|
|
757
|
+
properties: {}
|
|
758
|
+
extends:
|
|
759
|
+
- InProgress
|
|
760
|
+
StateEmbeddingGenerationQueued:
|
|
761
|
+
properties: {}
|
|
762
|
+
extends:
|
|
763
|
+
- Queued
|
|
764
|
+
StateInference:
|
|
765
|
+
discriminant: status
|
|
766
|
+
base-properties: {}
|
|
767
|
+
union:
|
|
768
|
+
QUEUED: StateInferenceQueued
|
|
769
|
+
IN_PROGRESS: StateInferenceInProgress
|
|
770
|
+
COMPLETED: StateInferenceCompletedInference
|
|
771
|
+
FAILED: StateInferenceFailed
|
|
772
|
+
StateInferenceCompletedInference:
|
|
773
|
+
properties: {}
|
|
774
|
+
extends:
|
|
775
|
+
- CompletedInference
|
|
776
|
+
StateInferenceFailed:
|
|
777
|
+
properties: {}
|
|
778
|
+
extends:
|
|
779
|
+
- Failed
|
|
780
|
+
StateInferenceInProgress:
|
|
781
|
+
properties: {}
|
|
782
|
+
extends:
|
|
783
|
+
- InProgress
|
|
784
|
+
StateInferenceQueued:
|
|
785
|
+
properties: {}
|
|
786
|
+
extends:
|
|
787
|
+
- Queued
|
|
788
|
+
StateTlInference:
|
|
789
|
+
discriminant: status
|
|
790
|
+
base-properties: {}
|
|
791
|
+
union:
|
|
792
|
+
QUEUED: StateTlInferenceQueued
|
|
793
|
+
IN_PROGRESS: StateTlInferenceInProgress
|
|
794
|
+
COMPLETED: StateTlInferenceCompletedTlInference
|
|
795
|
+
FAILED: StateTlInferenceFailed
|
|
796
|
+
StateTlInferenceCompletedTlInference:
|
|
797
|
+
properties: {}
|
|
798
|
+
extends:
|
|
799
|
+
- CompletedTlInference
|
|
800
|
+
StateTlInferenceFailed:
|
|
801
|
+
properties: {}
|
|
802
|
+
extends:
|
|
803
|
+
- Failed
|
|
804
|
+
StateTlInferenceInProgress:
|
|
805
|
+
properties: {}
|
|
806
|
+
extends:
|
|
807
|
+
- InProgress
|
|
808
|
+
StateTlInferenceQueued:
|
|
809
|
+
properties: {}
|
|
810
|
+
extends:
|
|
811
|
+
- Queued
|
|
812
|
+
StateTraining:
|
|
813
|
+
discriminant: status
|
|
814
|
+
base-properties: {}
|
|
815
|
+
union:
|
|
816
|
+
QUEUED: StateTrainingQueued
|
|
817
|
+
IN_PROGRESS: StateTrainingInProgress
|
|
818
|
+
COMPLETED: StateTrainingCompletedTraining
|
|
819
|
+
FAILED: StateTrainingFailed
|
|
820
|
+
StateTrainingCompletedTraining:
|
|
821
|
+
properties: {}
|
|
822
|
+
extends:
|
|
823
|
+
- CompletedTraining
|
|
824
|
+
StateTrainingFailed:
|
|
825
|
+
properties: {}
|
|
826
|
+
extends:
|
|
827
|
+
- Failed
|
|
828
|
+
StateTrainingInProgress:
|
|
829
|
+
properties: {}
|
|
830
|
+
extends:
|
|
831
|
+
- InProgress
|
|
832
|
+
StateTrainingQueued:
|
|
833
|
+
properties: {}
|
|
834
|
+
extends:
|
|
835
|
+
- Queued
|
|
836
|
+
Status:
|
|
837
|
+
enum:
|
|
838
|
+
- QUEUED
|
|
839
|
+
- IN_PROGRESS
|
|
840
|
+
- COMPLETED
|
|
841
|
+
- FAILED
|
|
842
|
+
TlInferencePrediction:
|
|
843
|
+
properties:
|
|
844
|
+
file:
|
|
845
|
+
type: string
|
|
846
|
+
docs: A file path relative to the top level source URL or file.
|
|
847
|
+
file_type: string
|
|
848
|
+
custom_models: map<string, CustomModelPrediction>
|
|
849
|
+
TlInferenceResults:
|
|
850
|
+
properties:
|
|
851
|
+
predictions: list<TlInferencePrediction>
|
|
852
|
+
errors: list<Error>
|
|
853
|
+
TlInferenceSourcePredictResult:
|
|
854
|
+
properties:
|
|
855
|
+
source: Source
|
|
856
|
+
results: optional<TlInferenceResults>
|
|
857
|
+
error:
|
|
858
|
+
type: optional<string>
|
|
859
|
+
docs: An error message.
|
|
860
|
+
Tag:
|
|
861
|
+
properties:
|
|
862
|
+
key: string
|
|
863
|
+
value: string
|
|
864
|
+
Target:
|
|
865
|
+
discriminated: false
|
|
866
|
+
union:
|
|
867
|
+
- integer
|
|
868
|
+
- double
|
|
869
|
+
- string
|
|
870
|
+
Task:
|
|
871
|
+
discriminant: type
|
|
872
|
+
base-properties: {}
|
|
873
|
+
union:
|
|
874
|
+
classification: TaskClassification
|
|
875
|
+
regression: TaskRegression
|
|
876
|
+
TaskClassification:
|
|
877
|
+
properties: {}
|
|
878
|
+
TaskRegression:
|
|
879
|
+
properties: {}
|
|
880
|
+
TextSource: map<string, unknown>
|
|
881
|
+
TimeInterval:
|
|
882
|
+
docs: A time range with a beginning and end, measured in seconds.
|
|
883
|
+
properties:
|
|
884
|
+
begin:
|
|
885
|
+
type: double
|
|
886
|
+
docs: Beginning of time range in seconds.
|
|
887
|
+
end:
|
|
888
|
+
type: double
|
|
889
|
+
docs: End of time range in seconds.
|
|
890
|
+
TlInferenceBaseRequest:
|
|
891
|
+
properties:
|
|
892
|
+
custom_model: CustomModel
|
|
893
|
+
urls:
|
|
894
|
+
type: optional<list<string>>
|
|
895
|
+
docs: >-
|
|
896
|
+
URLs to the media files to be processed. Each must be a valid public
|
|
897
|
+
URL to a media file (see recommended input filetypes) or an archive
|
|
898
|
+
(`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`) of media files.
|
|
899
|
+
|
|
900
|
+
|
|
901
|
+
If you wish to supply more than 100 URLs, consider providing them as
|
|
902
|
+
an archive (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`).
|
|
903
|
+
registry_files:
|
|
904
|
+
type: optional<list<string>>
|
|
905
|
+
docs: List of File IDs corresponding to the files in the asset registry.
|
|
906
|
+
callback_url:
|
|
907
|
+
type: optional<string>
|
|
908
|
+
docs: >-
|
|
909
|
+
If provided, a `POST` request will be made to the URL with the
|
|
910
|
+
generated predictions on completion or the error message on failure.
|
|
911
|
+
validation:
|
|
912
|
+
format: url
|
|
913
|
+
notify:
|
|
914
|
+
type: optional<boolean>
|
|
915
|
+
docs: >-
|
|
916
|
+
Whether to send an email notification to the user upon job
|
|
917
|
+
completion/failure.
|
|
918
|
+
CustomModel:
|
|
919
|
+
discriminated: false
|
|
920
|
+
union:
|
|
921
|
+
- CustomModelId
|
|
922
|
+
- CustomModelVersionId
|
|
923
|
+
CustomModelId:
|
|
924
|
+
properties:
|
|
925
|
+
id: string
|
|
926
|
+
CustomModelVersionId:
|
|
927
|
+
properties:
|
|
928
|
+
version_id: string
|
|
929
|
+
ToxicityScore:
|
|
930
|
+
properties:
|
|
931
|
+
name:
|
|
932
|
+
type: string
|
|
933
|
+
docs: Category of toxicity.
|
|
934
|
+
score:
|
|
935
|
+
type: string
|
|
936
|
+
docs: Prediction for this category of toxicity
|
|
937
|
+
TrainingBaseRequest:
|
|
938
|
+
properties:
|
|
939
|
+
custom_model: CustomModelRequest
|
|
940
|
+
dataset: Dataset
|
|
941
|
+
target_feature:
|
|
942
|
+
type: optional<string>
|
|
943
|
+
default: label
|
|
944
|
+
task: optional<Task>
|
|
945
|
+
evaluation: optional<EvaluationArgs>
|
|
946
|
+
alternatives: optional<list<Alternative>>
|
|
947
|
+
callback_url:
|
|
948
|
+
type: optional<string>
|
|
949
|
+
validation:
|
|
950
|
+
format: url
|
|
951
|
+
notify: optional<boolean>
|
|
952
|
+
TrainingCustomModel:
|
|
953
|
+
properties:
|
|
954
|
+
id: string
|
|
955
|
+
version_id: optional<string>
|
|
956
|
+
Transcription:
|
|
957
|
+
properties:
|
|
958
|
+
language: optional<Bcp47Tag>
|
|
959
|
+
identify_speakers:
|
|
960
|
+
type: optional<boolean>
|
|
961
|
+
docs: >-
|
|
962
|
+
Whether to return identifiers for speakers over time. If `true`,
|
|
963
|
+
unique identifiers will be assigned to spoken words to differentiate
|
|
964
|
+
different speakers. If `false`, all speakers will be tagged with an
|
|
965
|
+
`unknown` ID.
|
|
966
|
+
confidence_threshold:
|
|
967
|
+
type: optional<double>
|
|
968
|
+
docs: >-
|
|
969
|
+
Transcript confidence threshold. Transcripts generated with a
|
|
970
|
+
confidence less than this threshold will be considered invalid and not
|
|
971
|
+
used as an input for model inference.
|
|
972
|
+
default: 0.5
|
|
973
|
+
validation:
|
|
974
|
+
min: 0
|
|
975
|
+
max: 1
|
|
976
|
+
TranscriptionMetadata:
|
|
977
|
+
docs: Transcription metadata for your media file.
|
|
978
|
+
properties:
|
|
979
|
+
confidence:
|
|
980
|
+
type: double
|
|
981
|
+
docs: >-
|
|
982
|
+
Value between `0.0` and `1.0` indicating our transcription model's
|
|
983
|
+
relative confidence in the transcription of your media file.
|
|
984
|
+
detected_language: optional<Bcp47Tag>
|
|
985
|
+
Type:
|
|
986
|
+
enum:
|
|
987
|
+
- EMBEDDING_GENERATION
|
|
988
|
+
- INFERENCE
|
|
989
|
+
- TL_INFERENCE
|
|
990
|
+
- TRAINING
|
|
991
|
+
Unconfigurable:
|
|
992
|
+
docs: >-
|
|
993
|
+
To include predictions for this model type, set this field to `{}`. It is
|
|
994
|
+
currently not configurable further.
|
|
995
|
+
type: map<string, unknown>
|
|
996
|
+
UnionJob:
|
|
997
|
+
discriminant: type
|
|
998
|
+
base-properties: {}
|
|
999
|
+
union:
|
|
1000
|
+
EMBEDDING_GENERATION: UnionJobJobEmbeddingGeneration
|
|
1001
|
+
INFERENCE: UnionJobJobInference
|
|
1002
|
+
TL_INFERENCE: UnionJobJobTlInference
|
|
1003
|
+
TRAINING: UnionJobJobTraining
|
|
1004
|
+
UnionJobJobEmbeddingGeneration:
|
|
1005
|
+
properties: {}
|
|
1006
|
+
extends:
|
|
1007
|
+
- JobEmbeddingGeneration
|
|
1008
|
+
UnionJobJobInference:
|
|
1009
|
+
properties: {}
|
|
1010
|
+
extends:
|
|
1011
|
+
- JobInference
|
|
1012
|
+
UnionJobJobTlInference:
|
|
1013
|
+
properties: {}
|
|
1014
|
+
extends:
|
|
1015
|
+
- JobTlInference
|
|
1016
|
+
UnionJobJobTraining:
|
|
1017
|
+
properties: {}
|
|
1018
|
+
extends:
|
|
1019
|
+
- JobTraining
|
|
1020
|
+
UnionPredictResult:
|
|
1021
|
+
discriminated: false
|
|
1022
|
+
union:
|
|
1023
|
+
- InferenceSourcePredictResult
|
|
1024
|
+
- TlInferenceSourcePredictResult
|
|
1025
|
+
ValidationArgs:
|
|
1026
|
+
properties:
|
|
1027
|
+
positive_label: optional<Target>
|
|
1028
|
+
When:
|
|
1029
|
+
enum:
|
|
1030
|
+
- created_before
|
|
1031
|
+
- created_after
|
|
1032
|
+
Window:
|
|
1033
|
+
properties:
|
|
1034
|
+
length:
|
|
1035
|
+
type: optional<double>
|
|
1036
|
+
docs: The length of the sliding window.
|
|
1037
|
+
default: 4
|
|
1038
|
+
validation:
|
|
1039
|
+
min: 0.5
|
|
1040
|
+
step:
|
|
1041
|
+
type: optional<double>
|
|
1042
|
+
docs: The step size of the sliding window.
|
|
1043
|
+
default: 1
|
|
1044
|
+
validation:
|
|
1045
|
+
min: 0.5
|
|
1046
|
+
EmotionEmbeddingItem:
|
|
1047
|
+
properties:
|
|
1048
|
+
name:
|
|
1049
|
+
type: optional<string>
|
|
1050
|
+
docs: Name of the emotion being expressed.
|
|
1051
|
+
score:
|
|
1052
|
+
type: optional<double>
|
|
1053
|
+
docs: Embedding value for the emotion being expressed.
|
|
1054
|
+
EmotionEmbedding:
|
|
1055
|
+
docs: A high-dimensional embedding in emotion space.
|
|
1056
|
+
type: list<EmotionEmbeddingItem>
|
|
1057
|
+
TimeRange:
|
|
1058
|
+
docs: A time range with a beginning and end, measured in seconds.
|
|
1059
|
+
properties:
|
|
1060
|
+
begin:
|
|
1061
|
+
type: optional<double>
|
|
1062
|
+
docs: Beginning of time range in seconds.
|
|
1063
|
+
validation:
|
|
1064
|
+
min: 0
|
|
1065
|
+
end:
|
|
1066
|
+
type: optional<double>
|
|
1067
|
+
docs: End of time range in seconds.
|
|
1068
|
+
validation:
|
|
1069
|
+
min: 0
|
|
1070
|
+
TextPosition:
|
|
1071
|
+
docs: >
|
|
1072
|
+
Position of a segment of text within a larger document, measured in
|
|
1073
|
+
characters. Uses zero-based indexing. The beginning index is inclusive and
|
|
1074
|
+
the end index is exclusive.
|
|
1075
|
+
properties:
|
|
1076
|
+
begin:
|
|
1077
|
+
type: optional<double>
|
|
1078
|
+
docs: The index of the first character in the text segment, inclusive.
|
|
1079
|
+
validation:
|
|
1080
|
+
min: 0
|
|
1081
|
+
end:
|
|
1082
|
+
type: optional<double>
|
|
1083
|
+
docs: The index of the last character in the text segment, exclusive.
|
|
1084
|
+
validation:
|
|
1085
|
+
min: 0
|
|
1086
|
+
SentimentItem:
|
|
1087
|
+
properties:
|
|
1088
|
+
name:
|
|
1089
|
+
type: optional<string>
|
|
1090
|
+
docs: Level of sentiment, ranging from 1 (negative) to 9 (positive)
|
|
1091
|
+
score:
|
|
1092
|
+
type: optional<double>
|
|
1093
|
+
docs: Prediction for this level of sentiment
|
|
1094
|
+
Sentiment:
|
|
1095
|
+
docs: >-
|
|
1096
|
+
Sentiment predictions returned as a distribution. This model predicts the
|
|
1097
|
+
probability that a given text could be interpreted as having each
|
|
1098
|
+
sentiment level from 1 (negative) to 9 (positive).
|
|
1099
|
+
|
|
1100
|
+
|
|
1101
|
+
Compared to returning one estimate of sentiment, this enables a more
|
|
1102
|
+
nuanced analysis of a text's meaning. For example, a text with very
|
|
1103
|
+
neutral sentiment would have an average rating of 5. But also a text that
|
|
1104
|
+
could be interpreted as having very positive sentiment or very negative
|
|
1105
|
+
sentiment would also have an average rating of 5. The average sentiment is
|
|
1106
|
+
less informative than the distribution over sentiment, so this API returns
|
|
1107
|
+
a value for each sentiment level.
|
|
1108
|
+
type: list<SentimentItem>
|
|
1109
|
+
ToxicityItem:
|
|
1110
|
+
properties:
|
|
1111
|
+
name:
|
|
1112
|
+
type: optional<string>
|
|
1113
|
+
docs: Category of toxicity.
|
|
1114
|
+
score:
|
|
1115
|
+
type: optional<double>
|
|
1116
|
+
docs: Prediction for this category of toxicity
|
|
1117
|
+
Toxicity:
|
|
1118
|
+
docs: >-
|
|
1119
|
+
Toxicity predictions returned as probabilities that the text can be
|
|
1120
|
+
classified into the following categories: toxic, severe_toxic, obscene,
|
|
1121
|
+
threat, insult, and identity_hate.
|
|
1122
|
+
type: list<ToxicityItem>
|