hume 0.6.0 → 0.6.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (156) hide show
  1. package/.mock/definition/api.yml +12 -0
  2. package/.mock/definition/custom-models/__package__.yml +1392 -0
  3. package/.mock/definition/custom-models/datasets.yml +406 -0
  4. package/.mock/definition/custom-models/files.yml +263 -0
  5. package/.mock/definition/custom-models/jobs.yml +44 -0
  6. package/.mock/definition/custom-models/models.yml +303 -0
  7. package/.mock/definition/empathic-voice/__package__.yml +896 -0
  8. package/.mock/definition/empathic-voice/chat.yml +59 -0
  9. package/.mock/definition/empathic-voice/chatGroups.yml +80 -0
  10. package/.mock/definition/empathic-voice/chats.yml +86 -0
  11. package/.mock/definition/empathic-voice/configs.yml +413 -0
  12. package/.mock/definition/empathic-voice/prompts.yml +265 -0
  13. package/.mock/definition/empathic-voice/tools.yml +398 -0
  14. package/.mock/definition/expression-measurement/__package__.yml +1122 -0
  15. package/.mock/definition/expression-measurement/batch.yml +352 -0
  16. package/.mock/definition/expression-measurement/stream.yml +362 -0
  17. package/.mock/fern.config.json +4 -0
  18. package/Client.d.ts +3 -3
  19. package/Client.js +6 -6
  20. package/api/resources/customModels/resources/datasets/client/Client.js +9 -9
  21. package/api/resources/customModels/resources/files/client/Client.js +7 -7
  22. package/api/resources/customModels/resources/jobs/client/Client.js +2 -2
  23. package/api/resources/customModels/resources/models/client/Client.js +6 -6
  24. package/api/resources/empathicVoice/client/Client.d.ts +3 -0
  25. package/api/resources/empathicVoice/client/Client.js +5 -0
  26. package/api/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +1 -1
  27. package/api/resources/empathicVoice/resources/chatGroups/client/Client.d.ts +42 -0
  28. package/api/resources/empathicVoice/resources/chatGroups/client/Client.js +178 -0
  29. package/api/resources/empathicVoice/resources/chatGroups/client/index.d.ts +1 -0
  30. package/api/resources/empathicVoice/resources/chatGroups/client/index.js +17 -0
  31. package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.d.ts +21 -0
  32. package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.js +5 -0
  33. package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.d.ts +17 -0
  34. package/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.js +5 -0
  35. package/api/resources/empathicVoice/resources/chatGroups/client/requests/index.d.ts +2 -0
  36. package/api/resources/empathicVoice/resources/chatGroups/client/requests/index.js +2 -0
  37. package/api/resources/empathicVoice/resources/chatGroups/index.d.ts +1 -0
  38. package/api/resources/empathicVoice/resources/chatGroups/index.js +17 -0
  39. package/api/resources/empathicVoice/resources/chats/client/Client.js +2 -2
  40. package/api/resources/empathicVoice/resources/configs/client/Client.js +9 -9
  41. package/api/resources/empathicVoice/resources/index.d.ts +2 -0
  42. package/api/resources/empathicVoice/resources/index.js +3 -1
  43. package/api/resources/empathicVoice/resources/prompts/client/Client.js +9 -9
  44. package/api/resources/empathicVoice/resources/tools/client/Client.js +9 -9
  45. package/api/resources/empathicVoice/types/AssistantMessage.d.ts +1 -1
  46. package/api/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
  47. package/api/resources/empathicVoice/types/ChatMetadata.js +5 -0
  48. package/api/resources/empathicVoice/types/ReturnChat.d.ts +2 -0
  49. package/api/resources/empathicVoice/types/ReturnChatEvent.d.ts +2 -0
  50. package/api/resources/empathicVoice/types/ReturnChatGroup.d.ts +19 -0
  51. package/api/resources/empathicVoice/types/ReturnChatGroup.js +5 -0
  52. package/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +24 -0
  53. package/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +5 -0
  54. package/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +19 -0
  55. package/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +5 -0
  56. package/api/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +2 -0
  57. package/api/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
  58. package/api/resources/empathicVoice/types/ReturnPagedChatGroups.js +5 -0
  59. package/api/resources/empathicVoice/types/index.d.ts +5 -0
  60. package/api/resources/empathicVoice/types/index.js +5 -0
  61. package/api/resources/expressionMeasurement/resources/batch/client/Client.js +6 -6
  62. package/api/resources/index.d.ts +1 -1
  63. package/api/resources/index.js +2 -2
  64. package/core/form-data-utils/FormDataWrapper.js +2 -1
  65. package/dist/Client.d.ts +3 -3
  66. package/dist/Client.js +6 -6
  67. package/dist/api/resources/customModels/resources/datasets/client/Client.js +9 -9
  68. package/dist/api/resources/customModels/resources/files/client/Client.js +7 -7
  69. package/dist/api/resources/customModels/resources/jobs/client/Client.js +2 -2
  70. package/dist/api/resources/customModels/resources/models/client/Client.js +6 -6
  71. package/dist/api/resources/empathicVoice/client/Client.d.ts +3 -0
  72. package/dist/api/resources/empathicVoice/client/Client.js +5 -0
  73. package/dist/api/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +1 -1
  74. package/dist/api/resources/empathicVoice/resources/chatGroups/client/Client.d.ts +42 -0
  75. package/dist/api/resources/empathicVoice/resources/chatGroups/client/Client.js +178 -0
  76. package/dist/api/resources/empathicVoice/resources/chatGroups/client/index.d.ts +1 -0
  77. package/dist/api/resources/empathicVoice/resources/chatGroups/client/index.js +17 -0
  78. package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.d.ts +21 -0
  79. package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupEventsRequest.js +5 -0
  80. package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.d.ts +17 -0
  81. package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/ChatGroupsListChatGroupsRequest.js +5 -0
  82. package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/index.d.ts +2 -0
  83. package/dist/api/resources/empathicVoice/resources/chatGroups/client/requests/index.js +2 -0
  84. package/dist/api/resources/empathicVoice/resources/chatGroups/index.d.ts +1 -0
  85. package/dist/api/resources/empathicVoice/resources/chatGroups/index.js +17 -0
  86. package/dist/api/resources/empathicVoice/resources/chats/client/Client.js +2 -2
  87. package/dist/api/resources/empathicVoice/resources/configs/client/Client.js +9 -9
  88. package/dist/api/resources/empathicVoice/resources/index.d.ts +2 -0
  89. package/dist/api/resources/empathicVoice/resources/index.js +3 -1
  90. package/dist/api/resources/empathicVoice/resources/prompts/client/Client.js +9 -9
  91. package/dist/api/resources/empathicVoice/resources/tools/client/Client.js +9 -9
  92. package/dist/api/resources/empathicVoice/types/AssistantMessage.d.ts +1 -1
  93. package/dist/api/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
  94. package/dist/api/resources/empathicVoice/types/ChatMetadata.js +5 -0
  95. package/dist/api/resources/empathicVoice/types/ReturnChat.d.ts +2 -0
  96. package/dist/api/resources/empathicVoice/types/ReturnChatEvent.d.ts +2 -0
  97. package/dist/api/resources/empathicVoice/types/ReturnChatGroup.d.ts +19 -0
  98. package/dist/api/resources/empathicVoice/types/ReturnChatGroup.js +5 -0
  99. package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +24 -0
  100. package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +5 -0
  101. package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +19 -0
  102. package/dist/api/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +5 -0
  103. package/dist/api/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +2 -0
  104. package/dist/api/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
  105. package/dist/api/resources/empathicVoice/types/ReturnPagedChatGroups.js +5 -0
  106. package/dist/api/resources/empathicVoice/types/index.d.ts +5 -0
  107. package/dist/api/resources/empathicVoice/types/index.js +5 -0
  108. package/dist/api/resources/expressionMeasurement/resources/batch/client/Client.js +6 -6
  109. package/dist/api/resources/index.d.ts +1 -1
  110. package/dist/api/resources/index.js +2 -2
  111. package/dist/core/form-data-utils/FormDataWrapper.js +2 -1
  112. package/dist/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +2 -1
  113. package/dist/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.js +2 -0
  114. package/dist/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
  115. package/dist/serialization/resources/empathicVoice/types/ChatMetadata.js +36 -0
  116. package/dist/serialization/resources/empathicVoice/types/ReturnChat.d.ts +1 -0
  117. package/dist/serialization/resources/empathicVoice/types/ReturnChat.js +1 -0
  118. package/dist/serialization/resources/empathicVoice/types/ReturnChatEvent.d.ts +1 -0
  119. package/dist/serialization/resources/empathicVoice/types/ReturnChatEvent.js +1 -0
  120. package/dist/serialization/resources/empathicVoice/types/ReturnChatGroup.d.ts +17 -0
  121. package/dist/serialization/resources/empathicVoice/types/ReturnChatGroup.js +38 -0
  122. package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +20 -0
  123. package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +41 -0
  124. package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +17 -0
  125. package/dist/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +38 -0
  126. package/dist/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +1 -0
  127. package/dist/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.js +1 -0
  128. package/dist/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
  129. package/dist/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.js +36 -0
  130. package/dist/serialization/resources/empathicVoice/types/index.d.ts +5 -0
  131. package/dist/serialization/resources/empathicVoice/types/index.js +5 -0
  132. package/dist/serialization/resources/index.d.ts +1 -1
  133. package/dist/serialization/resources/index.js +2 -2
  134. package/package.json +1 -1
  135. package/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.d.ts +2 -1
  136. package/serialization/resources/empathicVoice/resources/chat/types/SubscribeEvent.js +2 -0
  137. package/serialization/resources/empathicVoice/types/ChatMetadata.d.ts +15 -0
  138. package/serialization/resources/empathicVoice/types/ChatMetadata.js +36 -0
  139. package/serialization/resources/empathicVoice/types/ReturnChat.d.ts +1 -0
  140. package/serialization/resources/empathicVoice/types/ReturnChat.js +1 -0
  141. package/serialization/resources/empathicVoice/types/ReturnChatEvent.d.ts +1 -0
  142. package/serialization/resources/empathicVoice/types/ReturnChatEvent.js +1 -0
  143. package/serialization/resources/empathicVoice/types/ReturnChatGroup.d.ts +17 -0
  144. package/serialization/resources/empathicVoice/types/ReturnChatGroup.js +38 -0
  145. package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.d.ts +20 -0
  146. package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedChats.js +41 -0
  147. package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.d.ts +17 -0
  148. package/serialization/resources/empathicVoice/types/ReturnChatGroupPagedEvents.js +38 -0
  149. package/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.d.ts +1 -0
  150. package/serialization/resources/empathicVoice/types/ReturnChatPagedEvents.js +1 -0
  151. package/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.d.ts +15 -0
  152. package/serialization/resources/empathicVoice/types/ReturnPagedChatGroups.js +36 -0
  153. package/serialization/resources/empathicVoice/types/index.d.ts +5 -0
  154. package/serialization/resources/empathicVoice/types/index.js +5 -0
  155. package/serialization/resources/index.d.ts +1 -1
  156. package/serialization/resources/index.js +2 -2
@@ -0,0 +1,1122 @@
1
+ types:
2
+ Alternative: literal<"language_only">
3
+ Bcp47Tag:
4
+ enum:
5
+ - zh
6
+ - da
7
+ - nl
8
+ - en
9
+ - value: en-AU
10
+ name: EnAu
11
+ - value: en-IN
12
+ name: EnIn
13
+ - value: en-NZ
14
+ name: EnNz
15
+ - value: en-GB
16
+ name: EnGb
17
+ - fr
18
+ - value: fr-CA
19
+ name: FrCa
20
+ - de
21
+ - hi
22
+ - value: hi-Latn
23
+ name: HiLatn
24
+ - id
25
+ - it
26
+ - ja
27
+ - ko
28
+ - 'no'
29
+ - pl
30
+ - pt
31
+ - value: pt-BR
32
+ name: PtBr
33
+ - value: pt-PT
34
+ name: PtPt
35
+ - ru
36
+ - es
37
+ - value: es-419
38
+ name: Es419
39
+ - sv
40
+ - ta
41
+ - tr
42
+ - uk
43
+ BoundingBox:
44
+ docs: A bounding box around a face.
45
+ properties:
46
+ x:
47
+ type: optional<double>
48
+ docs: x-coordinate of bounding box top left corner.
49
+ validation:
50
+ min: 0
51
+ 'y':
52
+ type: optional<double>
53
+ docs: y-coordinate of bounding box top left corner.
54
+ validation:
55
+ min: 0
56
+ w:
57
+ type: optional<double>
58
+ docs: Bounding box width.
59
+ validation:
60
+ min: 0
61
+ h:
62
+ type: optional<double>
63
+ docs: Bounding box height.
64
+ validation:
65
+ min: 0
66
+ BurstPrediction:
67
+ properties:
68
+ time: TimeInterval
69
+ emotions:
70
+ docs: A high-dimensional embedding in emotion space.
71
+ type: list<EmotionScore>
72
+ descriptions:
73
+ docs: Modality-specific descriptive features and their scores.
74
+ type: list<DescriptionsScore>
75
+ Classification: map<string, unknown>
76
+ CompletedEmbeddingGeneration:
77
+ properties:
78
+ created_timestamp_ms:
79
+ type: integer
80
+ docs: When this job was created (Unix timestamp in milliseconds).
81
+ started_timestamp_ms:
82
+ type: integer
83
+ docs: When this job started (Unix timestamp in milliseconds).
84
+ ended_timestamp_ms:
85
+ type: integer
86
+ docs: When this job ended (Unix timestamp in milliseconds).
87
+ CompletedInference:
88
+ properties:
89
+ created_timestamp_ms:
90
+ type: integer
91
+ docs: When this job was created (Unix timestamp in milliseconds).
92
+ started_timestamp_ms:
93
+ type: integer
94
+ docs: When this job started (Unix timestamp in milliseconds).
95
+ ended_timestamp_ms:
96
+ type: integer
97
+ docs: When this job ended (Unix timestamp in milliseconds).
98
+ num_predictions:
99
+ type: integer
100
+ docs: The number of predictions that were generated by this job.
101
+ num_errors:
102
+ type: integer
103
+ docs: The number of errors that occurred while running this job.
104
+ CompletedTlInference:
105
+ properties:
106
+ created_timestamp_ms:
107
+ type: integer
108
+ docs: When this job was created (Unix timestamp in milliseconds).
109
+ started_timestamp_ms:
110
+ type: integer
111
+ docs: When this job started (Unix timestamp in milliseconds).
112
+ ended_timestamp_ms:
113
+ type: integer
114
+ docs: When this job ended (Unix timestamp in milliseconds).
115
+ num_predictions:
116
+ type: integer
117
+ docs: The number of predictions that were generated by this job.
118
+ num_errors:
119
+ type: integer
120
+ docs: The number of errors that occurred while running this job.
121
+ CompletedTraining:
122
+ properties:
123
+ created_timestamp_ms:
124
+ type: integer
125
+ docs: When this job was created (Unix timestamp in milliseconds).
126
+ started_timestamp_ms:
127
+ type: integer
128
+ docs: When this job started (Unix timestamp in milliseconds).
129
+ ended_timestamp_ms:
130
+ type: integer
131
+ docs: When this job ended (Unix timestamp in milliseconds).
132
+ custom_model: TrainingCustomModel
133
+ alternatives: optional<map<string, TrainingCustomModel>>
134
+ CustomModelPrediction:
135
+ properties:
136
+ output: map<string, double>
137
+ error: string
138
+ task_type: string
139
+ CustomModelRequest:
140
+ properties:
141
+ name: string
142
+ description: optional<string>
143
+ tags: optional<list<Tag>>
144
+ Dataset:
145
+ discriminated: false
146
+ union:
147
+ - DatasetId
148
+ - DatasetVersionId
149
+ DatasetId:
150
+ properties:
151
+ id:
152
+ type: string
153
+ validation:
154
+ format: uuid
155
+ DatasetVersionId:
156
+ properties:
157
+ version_id:
158
+ type: string
159
+ validation:
160
+ format: uuid
161
+ DescriptionsScore:
162
+ properties:
163
+ name:
164
+ type: string
165
+ docs: Name of the descriptive feature being expressed.
166
+ score:
167
+ type: string
168
+ docs: Embedding value for the descriptive feature being expressed.
169
+ Direction:
170
+ enum:
171
+ - asc
172
+ - desc
173
+ EmbeddingGenerationBaseRequest:
174
+ properties:
175
+ registry_file_details:
176
+ type: optional<list<RegistryFileDetail>>
177
+ docs: File ID and File URL pairs for an asset registry file
178
+ EmotionScore:
179
+ properties:
180
+ name:
181
+ type: string
182
+ docs: Name of the emotion being expressed.
183
+ score:
184
+ type: double
185
+ docs: Embedding value for the emotion being expressed.
186
+ Error:
187
+ properties:
188
+ message:
189
+ type: string
190
+ docs: An error message.
191
+ file:
192
+ type: string
193
+ docs: A file path relative to the top level source URL or file.
194
+ EvaluationArgs:
195
+ properties:
196
+ validation: optional<ValidationArgs>
197
+ Face:
198
+ properties:
199
+ fps_pred:
200
+ type: optional<double>
201
+ docs: >-
202
+ Number of frames per second to process. Other frames will be omitted
203
+ from the response. Set to `0` to process every frame.
204
+ default: 3
205
+ prob_threshold:
206
+ type: optional<double>
207
+ docs: >-
208
+ Face detection probability threshold. Faces detected with a
209
+ probability less than this threshold will be omitted from the
210
+ response.
211
+ default: 0.99
212
+ validation:
213
+ min: 0
214
+ max: 1
215
+ identify_faces:
216
+ type: optional<boolean>
217
+ docs: >-
218
+ Whether to return identifiers for faces across frames. If `true`,
219
+ unique identifiers will be assigned to face bounding boxes to
220
+ differentiate different faces. If `false`, all faces will be tagged
221
+ with an `unknown` ID.
222
+ min_face_size:
223
+ type: optional<integer>
224
+ docs: >-
225
+ Minimum bounding box side length in pixels to treat as a face. Faces
226
+ detected with a bounding box side length in pixels less than this
227
+ threshold will be omitted from the response.
228
+ default: 60
229
+ validation:
230
+ min: 10
231
+ facs: optional<Unconfigurable>
232
+ descriptions: optional<Unconfigurable>
233
+ save_faces:
234
+ type: optional<boolean>
235
+ docs: >-
236
+ Whether to extract and save the detected faces in the artifacts zip
237
+ created by each job.
238
+ FacePrediction:
239
+ properties:
240
+ frame:
241
+ type: integer
242
+ docs: Frame number
243
+ time:
244
+ type: double
245
+ docs: Time in seconds when face detection occurred.
246
+ prob:
247
+ type: double
248
+ docs: The predicted probability that a detected face was actually a face.
249
+ box: BoundingBox
250
+ emotions:
251
+ docs: A high-dimensional embedding in emotion space.
252
+ type: list<EmotionScore>
253
+ facs:
254
+ type: optional<list<FacsScore>>
255
+ docs: FACS 2.0 features and their scores.
256
+ descriptions:
257
+ type: optional<list<DescriptionsScore>>
258
+ docs: Modality-specific descriptive features and their scores.
259
+ FacemeshPrediction:
260
+ properties:
261
+ emotions:
262
+ docs: A high-dimensional embedding in emotion space.
263
+ type: list<EmotionScore>
264
+ FacsScore:
265
+ properties:
266
+ name:
267
+ type: string
268
+ docs: Name of the FACS 2.0 feature being expressed.
269
+ score:
270
+ type: string
271
+ docs: Embedding value for the FACS 2.0 feature being expressed.
272
+ Failed:
273
+ properties:
274
+ created_timestamp_ms:
275
+ type: integer
276
+ docs: When this job was created (Unix timestamp in milliseconds).
277
+ started_timestamp_ms:
278
+ type: integer
279
+ docs: When this job started (Unix timestamp in milliseconds).
280
+ ended_timestamp_ms:
281
+ type: integer
282
+ docs: When this job ended (Unix timestamp in milliseconds).
283
+ message:
284
+ type: string
285
+ docs: An error message.
286
+ File:
287
+ properties:
288
+ filename:
289
+ type: optional<string>
290
+ docs: The name of the file.
291
+ content_type:
292
+ type: optional<string>
293
+ docs: The content type of the file.
294
+ md5sum:
295
+ type: string
296
+ docs: The MD5 checksum of the file.
297
+ Granularity:
298
+ enum:
299
+ - word
300
+ - sentence
301
+ - utterance
302
+ - conversational_turn
303
+ docs: >-
304
+ The granularity at which to generate predictions. `utterance` corresponds
305
+ to a natural pause or break in conversation, while `conversational_turn`
306
+ corresponds to a change in speaker.
307
+ GroupedPredictionsBurstPrediction:
308
+ properties:
309
+ id:
310
+ type: string
311
+ docs: >-
312
+ An automatically generated label to identify individuals in your media
313
+ file. Will be `unknown` if you have chosen to disable identification,
314
+ or if the model is unable to distinguish between individuals.
315
+ predictions: list<BurstPrediction>
316
+ GroupedPredictionsFacePrediction:
317
+ properties:
318
+ id:
319
+ type: string
320
+ docs: >-
321
+ An automatically generated label to identify individuals in your media
322
+ file. Will be `unknown` if you have chosen to disable identification,
323
+ or if the model is unable to distinguish between individuals.
324
+ predictions: list<FacePrediction>
325
+ GroupedPredictionsFacemeshPrediction:
326
+ properties:
327
+ id:
328
+ type: string
329
+ docs: >-
330
+ An automatically generated label to identify individuals in your media
331
+ file. Will be `unknown` if you have chosen to disable identification,
332
+ or if the model is unable to distinguish between individuals.
333
+ predictions: list<FacemeshPrediction>
334
+ GroupedPredictionsLanguagePrediction:
335
+ properties:
336
+ id:
337
+ type: string
338
+ docs: >-
339
+ An automatically generated label to identify individuals in your media
340
+ file. Will be `unknown` if you have chosen to disable identification,
341
+ or if the model is unable to distinguish between individuals.
342
+ predictions: list<LanguagePrediction>
343
+ GroupedPredictionsNerPrediction:
344
+ properties:
345
+ id:
346
+ type: string
347
+ docs: >-
348
+ An automatically generated label to identify individuals in your media
349
+ file. Will be `unknown` if you have chosen to disable identification,
350
+ or if the model is unable to distinguish between individuals.
351
+ predictions: list<NerPrediction>
352
+ GroupedPredictionsProsodyPrediction:
353
+ properties:
354
+ id:
355
+ type: string
356
+ docs: >-
357
+ An automatically generated label to identify individuals in your media
358
+ file. Will be `unknown` if you have chosen to disable identification,
359
+ or if the model is unable to distinguish between individuals.
360
+ predictions: list<ProsodyPrediction>
361
+ InProgress:
362
+ properties:
363
+ created_timestamp_ms:
364
+ type: integer
365
+ docs: When this job was created (Unix timestamp in milliseconds).
366
+ started_timestamp_ms:
367
+ type: integer
368
+ docs: When this job started (Unix timestamp in milliseconds).
369
+ InferenceBaseRequest:
370
+ properties:
371
+ models: optional<Models>
372
+ transcription: optional<Transcription>
373
+ urls:
374
+ type: optional<list<string>>
375
+ docs: >-
376
+ URLs to the media files to be processed. Each must be a valid public
377
+ URL to a media file (see recommended input filetypes) or an archive
378
+ (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`) of media files.
379
+
380
+
381
+ If you wish to supply more than 100 URLs, consider providing them as
382
+ an archive (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`).
383
+ registry_files:
384
+ type: optional<list<string>>
385
+ docs: List of File IDs corresponding to the files in the asset registry.
386
+ text:
387
+ type: optional<list<string>>
388
+ docs: Text to supply directly to our language and NER models.
389
+ callback_url:
390
+ type: optional<string>
391
+ docs: >-
392
+ If provided, a `POST` request will be made to the URL with the
393
+ generated predictions on completion or the error message on failure.
394
+ validation:
395
+ format: url
396
+ notify:
397
+ type: optional<boolean>
398
+ docs: >-
399
+ Whether to send an email notification to the user upon job
400
+ completion/failure.
401
+ InferencePrediction:
402
+ properties:
403
+ file:
404
+ type: string
405
+ docs: A file path relative to the top level source URL or file.
406
+ models: ModelsPredictions
407
+ InferenceRequest:
408
+ properties:
409
+ models: optional<Models>
410
+ transcription: optional<Transcription>
411
+ urls:
412
+ type: optional<list<string>>
413
+ docs: >-
414
+ URLs to the media files to be processed. Each must be a valid public
415
+ URL to a media file (see recommended input filetypes) or an archive
416
+ (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`) of media files.
417
+
418
+
419
+ If you wish to supply more than 100 URLs, consider providing them as
420
+ an archive (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`).
421
+ registry_files:
422
+ type: optional<list<string>>
423
+ docs: List of File IDs corresponding to the files in the asset registry.
424
+ text:
425
+ type: optional<list<string>>
426
+ docs: Text to supply directly to our language and NER models.
427
+ callback_url:
428
+ type: optional<string>
429
+ docs: >-
430
+ If provided, a `POST` request will be made to the URL with the
431
+ generated predictions on completion or the error message on failure.
432
+ validation:
433
+ format: url
434
+ notify:
435
+ type: optional<boolean>
436
+ docs: >-
437
+ Whether to send an email notification to the user upon job
438
+ completion/failure.
439
+ files: list<File>
440
+ InferenceResults:
441
+ properties:
442
+ predictions: list<InferencePrediction>
443
+ errors: list<Error>
444
+ InferenceSourcePredictResult:
445
+ properties:
446
+ source: Source
447
+ results: optional<InferenceResults>
448
+ error:
449
+ type: optional<string>
450
+ docs: An error message.
451
+ JobEmbeddingGeneration:
452
+ properties:
453
+ job_id:
454
+ type: string
455
+ docs: The ID associated with this job.
456
+ validation:
457
+ format: uuid
458
+ user_id:
459
+ type: string
460
+ validation:
461
+ format: uuid
462
+ request: EmbeddingGenerationBaseRequest
463
+ state: StateEmbeddingGeneration
464
+ JobInference:
465
+ properties:
466
+ job_id:
467
+ type: string
468
+ docs: The ID associated with this job.
469
+ validation:
470
+ format: uuid
471
+ user_id:
472
+ type: string
473
+ validation:
474
+ format: uuid
475
+ request: InferenceRequest
476
+ state: StateInference
477
+ JobTlInference:
478
+ properties:
479
+ job_id:
480
+ type: string
481
+ docs: The ID associated with this job.
482
+ validation:
483
+ format: uuid
484
+ user_id:
485
+ type: string
486
+ validation:
487
+ format: uuid
488
+ request: TlInferenceBaseRequest
489
+ state: StateTlInference
490
+ JobTraining:
491
+ properties:
492
+ job_id:
493
+ type: string
494
+ docs: The ID associated with this job.
495
+ validation:
496
+ format: uuid
497
+ user_id:
498
+ type: string
499
+ validation:
500
+ format: uuid
501
+ request: TrainingBaseRequest
502
+ state: StateTraining
503
+ JobId:
504
+ properties:
505
+ job_id:
506
+ type: string
507
+ docs: The ID of the started job.
508
+ validation:
509
+ format: uuid
510
+ Language:
511
+ properties:
512
+ granularity: optional<Granularity>
513
+ sentiment: optional<Unconfigurable>
514
+ toxicity: optional<Unconfigurable>
515
+ identify_speakers:
516
+ type: optional<boolean>
517
+ docs: >-
518
+ Whether to return identifiers for speakers over time. If `true`,
519
+ unique identifiers will be assigned to spoken words to differentiate
520
+ different speakers. If `false`, all speakers will be tagged with an
521
+ `unknown` ID.
522
+ LanguagePrediction:
523
+ properties:
524
+ text:
525
+ type: string
526
+ docs: A segment of text (like a word or a sentence).
527
+ position: PositionInterval
528
+ time: optional<TimeInterval>
529
+ confidence:
530
+ type: optional<double>
531
+ docs: >-
532
+ Value between `0.0` and `1.0` that indicates our transcription model's
533
+ relative confidence in this text.
534
+ speaker_confidence:
535
+ type: optional<double>
536
+ docs: >-
537
+ Value between `0.0` and `1.0` that indicates our transcription model's
538
+ relative confidence that this text was spoken by this speaker.
539
+ emotions:
540
+ docs: A high-dimensional embedding in emotion space.
541
+ type: list<EmotionScore>
542
+ sentiment:
543
+ type: optional<list<SentimentScore>>
544
+ docs: >-
545
+ Sentiment predictions returned as a distribution. This model predicts
546
+ the probability that a given text could be interpreted as having each
547
+ sentiment level from `1` (negative) to `9` (positive).
548
+
549
+
550
+ Compared to returning one estimate of sentiment, this enables a more
551
+ nuanced analysis of a text's meaning. For example, a text with very
552
+ neutral sentiment would have an average rating of `5`. But also a text
553
+ that could be interpreted as having very positive sentiment or very
554
+ negative sentiment would also have an average rating of `5`. The
555
+ average sentiment is less informative than the distribution over
556
+ sentiment, so this API returns a value for each sentiment level.
557
+ toxicity:
558
+ type: optional<list<ToxicityScore>>
559
+ docs: >-
560
+ Toxicity predictions returned as probabilities that the text can be
561
+ classified into the following categories: `toxic`, `severe_toxic`,
562
+ `obscene`, `threat`, `insult`, and `identity_hate`.
563
+ Models:
564
+ properties:
565
+ face: optional<Face>
566
+ burst: optional<Unconfigurable>
567
+ prosody: optional<Prosody>
568
+ language: optional<Language>
569
+ ner: optional<Ner>
570
+ facemesh: optional<Unconfigurable>
571
+ ModelsPredictions:
572
+ properties:
573
+ face: optional<PredictionsOptionalNullFacePrediction>
574
+ burst: optional<PredictionsOptionalNullBurstPrediction>
575
+ prosody: optional<PredictionsOptionalTranscriptionMetadataProsodyPrediction>
576
+ language: optional<PredictionsOptionalTranscriptionMetadataLanguagePrediction>
577
+ ner: optional<PredictionsOptionalTranscriptionMetadataNerPrediction>
578
+ facemesh: optional<PredictionsOptionalNullFacemeshPrediction>
579
+ Ner:
580
+ properties:
581
+ identify_speakers:
582
+ type: optional<boolean>
583
+ docs: >-
584
+ Whether to return identifiers for speakers over time. If `true`,
585
+ unique identifiers will be assigned to spoken words to differentiate
586
+ different speakers. If `false`, all speakers will be tagged with an
587
+ `unknown` ID.
588
+ NerPrediction:
589
+ properties:
590
+ entity:
591
+ type: string
592
+ docs: The recognized topic or entity.
593
+ position: PositionInterval
594
+ entity_confidence:
595
+ type: double
596
+ docs: Our NER model's relative confidence in the recognized topic or entity.
597
+ support:
598
+ type: double
599
+ docs: A measure of how often the entity is linked to by other entities.
600
+ uri:
601
+ type: string
602
+ docs: >-
603
+ A URL which provides more information about the recognized topic or
604
+ entity.
605
+ link_word:
606
+ type: string
607
+ docs: The specific word to which the emotion predictions are linked.
608
+ time: optional<TimeInterval>
609
+ confidence:
610
+ type: optional<double>
611
+ docs: >-
612
+ Value between `0.0` and `1.0` that indicates our transcription model's
613
+ relative confidence in this text.
614
+ speaker_confidence:
615
+ type: optional<double>
616
+ docs: >-
617
+ Value between `0.0` and `1.0` that indicates our transcription model's
618
+ relative confidence that this text was spoken by this speaker.
619
+ emotions:
620
+ docs: A high-dimensional embedding in emotion space.
621
+ type: list<EmotionScore>
622
+ 'Null':
623
+ docs: No associated metadata for this model. Value will be `null`.
624
+ type: map<string, unknown>
625
+ PositionInterval:
626
+ docs: >-
627
+ Position of a segment of text within a larger document, measured in
628
+ characters. Uses zero-based indexing. The beginning index is inclusive and
629
+ the end index is exclusive.
630
+ properties:
631
+ begin:
632
+ type: integer
633
+ docs: The index of the first character in the text segment, inclusive.
634
+ end:
635
+ type: integer
636
+ docs: The index of the last character in the text segment, exclusive.
637
+ PredictionsOptionalNullBurstPrediction:
638
+ properties:
639
+ metadata: optional<Null>
640
+ grouped_predictions: list<GroupedPredictionsBurstPrediction>
641
+ PredictionsOptionalNullFacePrediction:
642
+ properties:
643
+ metadata: optional<Null>
644
+ grouped_predictions: list<GroupedPredictionsFacePrediction>
645
+ PredictionsOptionalNullFacemeshPrediction:
646
+ properties:
647
+ metadata: optional<Null>
648
+ grouped_predictions: list<GroupedPredictionsFacemeshPrediction>
649
+ PredictionsOptionalTranscriptionMetadataLanguagePrediction:
650
+ properties:
651
+ metadata: optional<TranscriptionMetadata>
652
+ grouped_predictions: list<GroupedPredictionsLanguagePrediction>
653
+ PredictionsOptionalTranscriptionMetadataNerPrediction:
654
+ properties:
655
+ metadata: optional<TranscriptionMetadata>
656
+ grouped_predictions: list<GroupedPredictionsNerPrediction>
657
+ PredictionsOptionalTranscriptionMetadataProsodyPrediction:
658
+ properties:
659
+ metadata: optional<TranscriptionMetadata>
660
+ grouped_predictions: list<GroupedPredictionsProsodyPrediction>
661
+ Prosody:
662
+ properties:
663
+ granularity: optional<Granularity>
664
+ window: optional<Window>
665
+ identify_speakers:
666
+ type: optional<boolean>
667
+ docs: >-
668
+ Whether to return identifiers for speakers over time. If `true`,
669
+ unique identifiers will be assigned to spoken words to differentiate
670
+ different speakers. If `false`, all speakers will be tagged with an
671
+ `unknown` ID.
672
+ ProsodyPrediction:
673
+ properties:
674
+ text:
675
+ type: optional<string>
676
+ docs: A segment of text (like a word or a sentence).
677
+ time: TimeInterval
678
+ confidence:
679
+ type: optional<double>
680
+ docs: >-
681
+ Value between `0.0` and `1.0` that indicates our transcription model's
682
+ relative confidence in this text.
683
+ speaker_confidence:
684
+ type: optional<double>
685
+ docs: >-
686
+ Value between `0.0` and `1.0` that indicates our transcription model's
687
+ relative confidence that this text was spoken by this speaker.
688
+ emotions:
689
+ docs: A high-dimensional embedding in emotion space.
690
+ type: list<EmotionScore>
691
+ Queued:
692
+ properties:
693
+ created_timestamp_ms:
694
+ type: integer
695
+ docs: When this job was created (Unix timestamp in milliseconds).
696
+ RegistryFileDetail:
697
+ properties:
698
+ file_id:
699
+ type: string
700
+ docs: File ID in the Asset Registry
701
+ file_url:
702
+ type: string
703
+ docs: URL to the file in the Asset Registry
704
+ Regression: map<string, unknown>
705
+ SentimentScore:
706
+ properties:
707
+ name:
708
+ type: string
709
+ docs: Level of sentiment, ranging from `1` (negative) to `9` (positive)
710
+ score:
711
+ type: string
712
+ docs: Prediction for this level of sentiment
713
+ SortBy:
714
+ enum:
715
+ - created
716
+ - started
717
+ - ended
718
+ Source:
719
+ discriminant: type
720
+ base-properties: {}
721
+ union:
722
+ url: SourceUrl
723
+ file: SourceFile
724
+ text: SourceTextSource
725
+ SourceFile:
726
+ properties: {}
727
+ extends:
728
+ - File
729
+ SourceTextSource:
730
+ properties: {}
731
+ SourceUrl:
732
+ properties: {}
733
+ extends:
734
+ - Url
735
+ Url:
736
+ properties:
737
+ url:
738
+ type: string
739
+ docs: The URL of the source media file.
740
+ StateEmbeddingGeneration:
741
+ discriminant: status
742
+ base-properties: {}
743
+ union:
744
+ QUEUED: StateEmbeddingGenerationQueued
745
+ IN_PROGRESS: StateEmbeddingGenerationInProgress
746
+ COMPLETED: StateEmbeddingGenerationCompletedEmbeddingGeneration
747
+ FAILED: StateEmbeddingGenerationFailed
748
+ StateEmbeddingGenerationCompletedEmbeddingGeneration:
749
+ properties: {}
750
+ extends:
751
+ - CompletedEmbeddingGeneration
752
+ StateEmbeddingGenerationFailed:
753
+ properties: {}
754
+ extends:
755
+ - Failed
756
+ StateEmbeddingGenerationInProgress:
757
+ properties: {}
758
+ extends:
759
+ - InProgress
760
+ StateEmbeddingGenerationQueued:
761
+ properties: {}
762
+ extends:
763
+ - Queued
764
+ StateInference:
765
+ discriminant: status
766
+ base-properties: {}
767
+ union:
768
+ QUEUED: StateInferenceQueued
769
+ IN_PROGRESS: StateInferenceInProgress
770
+ COMPLETED: StateInferenceCompletedInference
771
+ FAILED: StateInferenceFailed
772
+ StateInferenceCompletedInference:
773
+ properties: {}
774
+ extends:
775
+ - CompletedInference
776
+ StateInferenceFailed:
777
+ properties: {}
778
+ extends:
779
+ - Failed
780
+ StateInferenceInProgress:
781
+ properties: {}
782
+ extends:
783
+ - InProgress
784
+ StateInferenceQueued:
785
+ properties: {}
786
+ extends:
787
+ - Queued
788
+ StateTlInference:
789
+ discriminant: status
790
+ base-properties: {}
791
+ union:
792
+ QUEUED: StateTlInferenceQueued
793
+ IN_PROGRESS: StateTlInferenceInProgress
794
+ COMPLETED: StateTlInferenceCompletedTlInference
795
+ FAILED: StateTlInferenceFailed
796
+ StateTlInferenceCompletedTlInference:
797
+ properties: {}
798
+ extends:
799
+ - CompletedTlInference
800
+ StateTlInferenceFailed:
801
+ properties: {}
802
+ extends:
803
+ - Failed
804
+ StateTlInferenceInProgress:
805
+ properties: {}
806
+ extends:
807
+ - InProgress
808
+ StateTlInferenceQueued:
809
+ properties: {}
810
+ extends:
811
+ - Queued
812
+ StateTraining:
813
+ discriminant: status
814
+ base-properties: {}
815
+ union:
816
+ QUEUED: StateTrainingQueued
817
+ IN_PROGRESS: StateTrainingInProgress
818
+ COMPLETED: StateTrainingCompletedTraining
819
+ FAILED: StateTrainingFailed
820
+ StateTrainingCompletedTraining:
821
+ properties: {}
822
+ extends:
823
+ - CompletedTraining
824
+ StateTrainingFailed:
825
+ properties: {}
826
+ extends:
827
+ - Failed
828
+ StateTrainingInProgress:
829
+ properties: {}
830
+ extends:
831
+ - InProgress
832
+ StateTrainingQueued:
833
+ properties: {}
834
+ extends:
835
+ - Queued
836
+ Status:
837
+ enum:
838
+ - QUEUED
839
+ - IN_PROGRESS
840
+ - COMPLETED
841
+ - FAILED
842
+ TlInferencePrediction:
843
+ properties:
844
+ file:
845
+ type: string
846
+ docs: A file path relative to the top level source URL or file.
847
+ file_type: string
848
+ custom_models: map<string, CustomModelPrediction>
849
+ TlInferenceResults:
850
+ properties:
851
+ predictions: list<TlInferencePrediction>
852
+ errors: list<Error>
853
+ TlInferenceSourcePredictResult:
854
+ properties:
855
+ source: Source
856
+ results: optional<TlInferenceResults>
857
+ error:
858
+ type: optional<string>
859
+ docs: An error message.
860
+ Tag:
861
+ properties:
862
+ key: string
863
+ value: string
864
+ Target:
865
+ discriminated: false
866
+ union:
867
+ - integer
868
+ - double
869
+ - string
870
+ Task:
871
+ discriminant: type
872
+ base-properties: {}
873
+ union:
874
+ classification: TaskClassification
875
+ regression: TaskRegression
876
+ TaskClassification:
877
+ properties: {}
878
+ TaskRegression:
879
+ properties: {}
880
+ TextSource: map<string, unknown>
881
+ TimeInterval:
882
+ docs: A time range with a beginning and end, measured in seconds.
883
+ properties:
884
+ begin:
885
+ type: double
886
+ docs: Beginning of time range in seconds.
887
+ end:
888
+ type: double
889
+ docs: End of time range in seconds.
890
+ TlInferenceBaseRequest:
891
+ properties:
892
+ custom_model: CustomModel
893
+ urls:
894
+ type: optional<list<string>>
895
+ docs: >-
896
+ URLs to the media files to be processed. Each must be a valid public
897
+ URL to a media file (see recommended input filetypes) or an archive
898
+ (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`) of media files.
899
+
900
+
901
+ If you wish to supply more than 100 URLs, consider providing them as
902
+ an archive (`.zip`, `.tar.gz`, `.tar.bz2`, `.tar.xz`).
903
+ registry_files:
904
+ type: optional<list<string>>
905
+ docs: List of File IDs corresponding to the files in the asset registry.
906
+ callback_url:
907
+ type: optional<string>
908
+ docs: >-
909
+ If provided, a `POST` request will be made to the URL with the
910
+ generated predictions on completion or the error message on failure.
911
+ validation:
912
+ format: url
913
+ notify:
914
+ type: optional<boolean>
915
+ docs: >-
916
+ Whether to send an email notification to the user upon job
917
+ completion/failure.
918
+ CustomModel:
919
+ discriminated: false
920
+ union:
921
+ - CustomModelId
922
+ - CustomModelVersionId
923
+ CustomModelId:
924
+ properties:
925
+ id: string
926
+ CustomModelVersionId:
927
+ properties:
928
+ version_id: string
929
+ ToxicityScore:
930
+ properties:
931
+ name:
932
+ type: string
933
+ docs: Category of toxicity.
934
+ score:
935
+ type: string
936
+ docs: Prediction for this category of toxicity
937
+ TrainingBaseRequest:
938
+ properties:
939
+ custom_model: CustomModelRequest
940
+ dataset: Dataset
941
+ target_feature:
942
+ type: optional<string>
943
+ default: label
944
+ task: optional<Task>
945
+ evaluation: optional<EvaluationArgs>
946
+ alternatives: optional<list<Alternative>>
947
+ callback_url:
948
+ type: optional<string>
949
+ validation:
950
+ format: url
951
+ notify: optional<boolean>
952
+ TrainingCustomModel:
953
+ properties:
954
+ id: string
955
+ version_id: optional<string>
956
+ Transcription:
957
+ properties:
958
+ language: optional<Bcp47Tag>
959
+ identify_speakers:
960
+ type: optional<boolean>
961
+ docs: >-
962
+ Whether to return identifiers for speakers over time. If `true`,
963
+ unique identifiers will be assigned to spoken words to differentiate
964
+ different speakers. If `false`, all speakers will be tagged with an
965
+ `unknown` ID.
966
+ confidence_threshold:
967
+ type: optional<double>
968
+ docs: >-
969
+ Transcript confidence threshold. Transcripts generated with a
970
+ confidence less than this threshold will be considered invalid and not
971
+ used as an input for model inference.
972
+ default: 0.5
973
+ validation:
974
+ min: 0
975
+ max: 1
976
+ TranscriptionMetadata:
977
+ docs: Transcription metadata for your media file.
978
+ properties:
979
+ confidence:
980
+ type: double
981
+ docs: >-
982
+ Value between `0.0` and `1.0` indicating our transcription model's
983
+ relative confidence in the transcription of your media file.
984
+ detected_language: optional<Bcp47Tag>
985
+ Type:
986
+ enum:
987
+ - EMBEDDING_GENERATION
988
+ - INFERENCE
989
+ - TL_INFERENCE
990
+ - TRAINING
991
+ Unconfigurable:
992
+ docs: >-
993
+ To include predictions for this model type, set this field to `{}`. It is
994
+ currently not configurable further.
995
+ type: map<string, unknown>
996
+ UnionJob:
997
+ discriminant: type
998
+ base-properties: {}
999
+ union:
1000
+ EMBEDDING_GENERATION: UnionJobJobEmbeddingGeneration
1001
+ INFERENCE: UnionJobJobInference
1002
+ TL_INFERENCE: UnionJobJobTlInference
1003
+ TRAINING: UnionJobJobTraining
1004
+ UnionJobJobEmbeddingGeneration:
1005
+ properties: {}
1006
+ extends:
1007
+ - JobEmbeddingGeneration
1008
+ UnionJobJobInference:
1009
+ properties: {}
1010
+ extends:
1011
+ - JobInference
1012
+ UnionJobJobTlInference:
1013
+ properties: {}
1014
+ extends:
1015
+ - JobTlInference
1016
+ UnionJobJobTraining:
1017
+ properties: {}
1018
+ extends:
1019
+ - JobTraining
1020
+ UnionPredictResult:
1021
+ discriminated: false
1022
+ union:
1023
+ - InferenceSourcePredictResult
1024
+ - TlInferenceSourcePredictResult
1025
+ ValidationArgs:
1026
+ properties:
1027
+ positive_label: optional<Target>
1028
+ When:
1029
+ enum:
1030
+ - created_before
1031
+ - created_after
1032
+ Window:
1033
+ properties:
1034
+ length:
1035
+ type: optional<double>
1036
+ docs: The length of the sliding window.
1037
+ default: 4
1038
+ validation:
1039
+ min: 0.5
1040
+ step:
1041
+ type: optional<double>
1042
+ docs: The step size of the sliding window.
1043
+ default: 1
1044
+ validation:
1045
+ min: 0.5
1046
+ EmotionEmbeddingItem:
1047
+ properties:
1048
+ name:
1049
+ type: optional<string>
1050
+ docs: Name of the emotion being expressed.
1051
+ score:
1052
+ type: optional<double>
1053
+ docs: Embedding value for the emotion being expressed.
1054
+ EmotionEmbedding:
1055
+ docs: A high-dimensional embedding in emotion space.
1056
+ type: list<EmotionEmbeddingItem>
1057
+ TimeRange:
1058
+ docs: A time range with a beginning and end, measured in seconds.
1059
+ properties:
1060
+ begin:
1061
+ type: optional<double>
1062
+ docs: Beginning of time range in seconds.
1063
+ validation:
1064
+ min: 0
1065
+ end:
1066
+ type: optional<double>
1067
+ docs: End of time range in seconds.
1068
+ validation:
1069
+ min: 0
1070
+ TextPosition:
1071
+ docs: >
1072
+ Position of a segment of text within a larger document, measured in
1073
+ characters. Uses zero-based indexing. The beginning index is inclusive and
1074
+ the end index is exclusive.
1075
+ properties:
1076
+ begin:
1077
+ type: optional<double>
1078
+ docs: The index of the first character in the text segment, inclusive.
1079
+ validation:
1080
+ min: 0
1081
+ end:
1082
+ type: optional<double>
1083
+ docs: The index of the last character in the text segment, exclusive.
1084
+ validation:
1085
+ min: 0
1086
+ SentimentItem:
1087
+ properties:
1088
+ name:
1089
+ type: optional<string>
1090
+ docs: Level of sentiment, ranging from 1 (negative) to 9 (positive)
1091
+ score:
1092
+ type: optional<double>
1093
+ docs: Prediction for this level of sentiment
1094
+ Sentiment:
1095
+ docs: >-
1096
+ Sentiment predictions returned as a distribution. This model predicts the
1097
+ probability that a given text could be interpreted as having each
1098
+ sentiment level from 1 (negative) to 9 (positive).
1099
+
1100
+
1101
+ Compared to returning one estimate of sentiment, this enables a more
1102
+ nuanced analysis of a text's meaning. For example, a text with very
1103
+ neutral sentiment would have an average rating of 5. But also a text that
1104
+ could be interpreted as having very positive sentiment or very negative
1105
+ sentiment would also have an average rating of 5. The average sentiment is
1106
+ less informative than the distribution over sentiment, so this API returns
1107
+ a value for each sentiment level.
1108
+ type: list<SentimentItem>
1109
+ ToxicityItem:
1110
+ properties:
1111
+ name:
1112
+ type: optional<string>
1113
+ docs: Category of toxicity.
1114
+ score:
1115
+ type: optional<double>
1116
+ docs: Prediction for this category of toxicity
1117
+ Toxicity:
1118
+ docs: >-
1119
+ Toxicity predictions returned as probabilities that the text can be
1120
+ classified into the following categories: toxic, severe_toxic, obscene,
1121
+ threat, insult, and identity_hate.
1122
+ type: list<ToxicityItem>