@alicloud/aimiaobi20230801 1.36.0 → 1.36.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/client.js +36 -0
- package/dist/client.js.map +1 -1
- package/dist/models/AsyncCreateClipsTaskRequest.d.ts +66 -0
- package/dist/models/AsyncCreateClipsTaskRequest.js +49 -1
- package/dist/models/AsyncCreateClipsTaskRequest.js.map +1 -1
- package/dist/models/AsyncCreateClipsTaskShrinkRequest.d.ts +14 -0
- package/dist/models/AsyncCreateClipsTaskShrinkRequest.js +12 -0
- package/dist/models/AsyncCreateClipsTaskShrinkRequest.js.map +1 -1
- package/dist/models/AsyncUploadVideoRequest.d.ts +57 -0
- package/dist/models/AsyncUploadVideoRequest.js +59 -1
- package/dist/models/AsyncUploadVideoRequest.js.map +1 -1
- package/dist/models/AsyncUploadVideoShrinkRequest.d.ts +12 -0
- package/dist/models/AsyncUploadVideoShrinkRequest.js +8 -0
- package/dist/models/AsyncUploadVideoShrinkRequest.js.map +1 -1
- package/dist/models/GetAutoClipsTaskInfoResponseBody.d.ts +66 -0
- package/dist/models/GetAutoClipsTaskInfoResponseBody.js +49 -1
- package/dist/models/GetAutoClipsTaskInfoResponseBody.js.map +1 -1
- package/dist/models/model.d.ts +4 -0
- package/dist/models/model.js +42 -34
- package/dist/models/model.js.map +1 -1
- package/package.json +1 -1
- package/src/client.ts +48 -0
- package/src/models/AsyncCreateClipsTaskRequest.ts +105 -0
- package/src/models/AsyncCreateClipsTaskShrinkRequest.ts +26 -0
- package/src/models/AsyncUploadVideoRequest.ts +97 -0
- package/src/models/AsyncUploadVideoShrinkRequest.ts +20 -0
- package/src/models/GetAutoClipsTaskInfoResponseBody.ts +105 -0
- package/src/models/model.ts +4 -0
|
@@ -4,13 +4,25 @@ import * as $dara from '@darabonba/typescript';
|
|
|
4
4
|
|
|
5
5
|
export class AsyncUploadVideoShrinkRequest extends $dara.Model {
|
|
6
6
|
anlysisPrompt?: string;
|
|
7
|
+
/**
|
|
8
|
+
* @example
|
|
9
|
+
* 0.7
|
|
10
|
+
*/
|
|
11
|
+
faceIdentitySimilarityMinScore?: number;
|
|
7
12
|
referenceVideoShrink?: string;
|
|
13
|
+
removeSubtitle?: boolean;
|
|
8
14
|
/**
|
|
9
15
|
* @remarks
|
|
10
16
|
* This parameter is required.
|
|
11
17
|
*/
|
|
12
18
|
sourceVideosShrink?: string;
|
|
13
19
|
splitInterval?: number;
|
|
20
|
+
videoRolesShrink?: string;
|
|
21
|
+
/**
|
|
22
|
+
* @example
|
|
23
|
+
* 2
|
|
24
|
+
*/
|
|
25
|
+
videoShotFaceIdentityCount?: number;
|
|
14
26
|
/**
|
|
15
27
|
* @remarks
|
|
16
28
|
* This parameter is required.
|
|
@@ -22,9 +34,13 @@ export class AsyncUploadVideoShrinkRequest extends $dara.Model {
|
|
|
22
34
|
static names(): { [key: string]: string } {
|
|
23
35
|
return {
|
|
24
36
|
anlysisPrompt: 'AnlysisPrompt',
|
|
37
|
+
faceIdentitySimilarityMinScore: 'FaceIdentitySimilarityMinScore',
|
|
25
38
|
referenceVideoShrink: 'ReferenceVideo',
|
|
39
|
+
removeSubtitle: 'RemoveSubtitle',
|
|
26
40
|
sourceVideosShrink: 'SourceVideos',
|
|
27
41
|
splitInterval: 'SplitInterval',
|
|
42
|
+
videoRolesShrink: 'VideoRoles',
|
|
43
|
+
videoShotFaceIdentityCount: 'VideoShotFaceIdentityCount',
|
|
28
44
|
workspaceId: 'WorkspaceId',
|
|
29
45
|
};
|
|
30
46
|
}
|
|
@@ -32,9 +48,13 @@ export class AsyncUploadVideoShrinkRequest extends $dara.Model {
|
|
|
32
48
|
static types(): { [key: string]: any } {
|
|
33
49
|
return {
|
|
34
50
|
anlysisPrompt: 'string',
|
|
51
|
+
faceIdentitySimilarityMinScore: 'number',
|
|
35
52
|
referenceVideoShrink: 'string',
|
|
53
|
+
removeSubtitle: 'boolean',
|
|
36
54
|
sourceVideosShrink: 'string',
|
|
37
55
|
splitInterval: 'number',
|
|
56
|
+
videoRolesShrink: 'string',
|
|
57
|
+
videoShotFaceIdentityCount: 'number',
|
|
38
58
|
workspaceId: 'string',
|
|
39
59
|
};
|
|
40
60
|
}
|
|
@@ -67,6 +67,82 @@ export class GetAutoClipsTaskInfoResponseBodyDataColorWords extends $dara.Model
|
|
|
67
67
|
}
|
|
68
68
|
}
|
|
69
69
|
|
|
70
|
+
export class GetAutoClipsTaskInfoResponseBodyDataStickers extends $dara.Model {
|
|
71
|
+
/**
|
|
72
|
+
* @example
|
|
73
|
+
* 10
|
|
74
|
+
*/
|
|
75
|
+
duration?: number;
|
|
76
|
+
/**
|
|
77
|
+
* @example
|
|
78
|
+
* 8
|
|
79
|
+
*/
|
|
80
|
+
dyncFrames?: number;
|
|
81
|
+
/**
|
|
82
|
+
* @example
|
|
83
|
+
* 200
|
|
84
|
+
*/
|
|
85
|
+
height?: number;
|
|
86
|
+
/**
|
|
87
|
+
* @example
|
|
88
|
+
* 10
|
|
89
|
+
*/
|
|
90
|
+
timelineIn?: number;
|
|
91
|
+
/**
|
|
92
|
+
* @example
|
|
93
|
+
* http://xxx/xxx.gif
|
|
94
|
+
*/
|
|
95
|
+
url?: string;
|
|
96
|
+
/**
|
|
97
|
+
* @example
|
|
98
|
+
* 200
|
|
99
|
+
*/
|
|
100
|
+
width?: number;
|
|
101
|
+
/**
|
|
102
|
+
* @example
|
|
103
|
+
* 100
|
|
104
|
+
*/
|
|
105
|
+
x?: number;
|
|
106
|
+
/**
|
|
107
|
+
* @example
|
|
108
|
+
* 100
|
|
109
|
+
*/
|
|
110
|
+
y?: number;
|
|
111
|
+
static names(): { [key: string]: string } {
|
|
112
|
+
return {
|
|
113
|
+
duration: 'Duration',
|
|
114
|
+
dyncFrames: 'DyncFrames',
|
|
115
|
+
height: 'Height',
|
|
116
|
+
timelineIn: 'TimelineIn',
|
|
117
|
+
url: 'Url',
|
|
118
|
+
width: 'Width',
|
|
119
|
+
x: 'X',
|
|
120
|
+
y: 'Y',
|
|
121
|
+
};
|
|
122
|
+
}
|
|
123
|
+
|
|
124
|
+
static types(): { [key: string]: any } {
|
|
125
|
+
return {
|
|
126
|
+
duration: 'number',
|
|
127
|
+
dyncFrames: 'number',
|
|
128
|
+
height: 'number',
|
|
129
|
+
timelineIn: 'number',
|
|
130
|
+
url: 'string',
|
|
131
|
+
width: 'number',
|
|
132
|
+
x: 'number',
|
|
133
|
+
y: 'number',
|
|
134
|
+
};
|
|
135
|
+
}
|
|
136
|
+
|
|
137
|
+
validate() {
|
|
138
|
+
super.validate();
|
|
139
|
+
}
|
|
140
|
+
|
|
141
|
+
constructor(map?: { [key: string]: any }) {
|
|
142
|
+
super(map);
|
|
143
|
+
}
|
|
144
|
+
}
|
|
145
|
+
|
|
70
146
|
export class GetAutoClipsTaskInfoResponseBodyDataTimelinesClips extends $dara.Model {
|
|
71
147
|
/**
|
|
72
148
|
* @example
|
|
@@ -168,8 +244,21 @@ export class GetAutoClipsTaskInfoResponseBodyDataTimelines extends $dara.Model {
|
|
|
168
244
|
}
|
|
169
245
|
|
|
170
246
|
export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
247
|
+
closeMusic?: boolean;
|
|
248
|
+
closeSubtitle?: boolean;
|
|
249
|
+
closeVoice?: boolean;
|
|
171
250
|
colorWords?: GetAutoClipsTaskInfoResponseBodyDataColorWords[];
|
|
172
251
|
content?: string;
|
|
252
|
+
/**
|
|
253
|
+
* @example
|
|
254
|
+
* http://xxx/xxx.mp4
|
|
255
|
+
*/
|
|
256
|
+
customVoiceUrl?: string;
|
|
257
|
+
/**
|
|
258
|
+
* @example
|
|
259
|
+
* 0
|
|
260
|
+
*/
|
|
261
|
+
customVoiceVolume?: number;
|
|
173
262
|
errorMessage?: string;
|
|
174
263
|
mediaCloudTimeline?: string;
|
|
175
264
|
musicStyle?: string;
|
|
@@ -194,6 +283,7 @@ export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
|
194
283
|
* clips
|
|
195
284
|
*/
|
|
196
285
|
step?: string;
|
|
286
|
+
stickers?: GetAutoClipsTaskInfoResponseBodyDataStickers[];
|
|
197
287
|
/**
|
|
198
288
|
* @example
|
|
199
289
|
* 5
|
|
@@ -213,8 +303,13 @@ export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
|
213
303
|
voiceVolume?: number;
|
|
214
304
|
static names(): { [key: string]: string } {
|
|
215
305
|
return {
|
|
306
|
+
closeMusic: 'CloseMusic',
|
|
307
|
+
closeSubtitle: 'CloseSubtitle',
|
|
308
|
+
closeVoice: 'CloseVoice',
|
|
216
309
|
colorWords: 'ColorWords',
|
|
217
310
|
content: 'Content',
|
|
311
|
+
customVoiceUrl: 'CustomVoiceUrl',
|
|
312
|
+
customVoiceVolume: 'CustomVoiceVolume',
|
|
218
313
|
errorMessage: 'ErrorMessage',
|
|
219
314
|
mediaCloudTimeline: 'MediaCloudTimeline',
|
|
220
315
|
musicStyle: 'MusicStyle',
|
|
@@ -223,6 +318,7 @@ export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
|
223
318
|
outputVideoUrl: 'OutputVideoUrl',
|
|
224
319
|
status: 'Status',
|
|
225
320
|
step: 'Step',
|
|
321
|
+
stickers: 'Stickers',
|
|
226
322
|
subtitleFontSize: 'SubtitleFontSize',
|
|
227
323
|
taskId: 'TaskId',
|
|
228
324
|
timelines: 'Timelines',
|
|
@@ -233,8 +329,13 @@ export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
|
233
329
|
|
|
234
330
|
static types(): { [key: string]: any } {
|
|
235
331
|
return {
|
|
332
|
+
closeMusic: 'boolean',
|
|
333
|
+
closeSubtitle: 'boolean',
|
|
334
|
+
closeVoice: 'boolean',
|
|
236
335
|
colorWords: { 'type': 'array', 'itemType': GetAutoClipsTaskInfoResponseBodyDataColorWords },
|
|
237
336
|
content: 'string',
|
|
337
|
+
customVoiceUrl: 'string',
|
|
338
|
+
customVoiceVolume: 'number',
|
|
238
339
|
errorMessage: 'string',
|
|
239
340
|
mediaCloudTimeline: 'string',
|
|
240
341
|
musicStyle: 'string',
|
|
@@ -243,6 +344,7 @@ export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
|
243
344
|
outputVideoUrl: 'string',
|
|
244
345
|
status: 'number',
|
|
245
346
|
step: 'string',
|
|
347
|
+
stickers: { 'type': 'array', 'itemType': GetAutoClipsTaskInfoResponseBodyDataStickers },
|
|
246
348
|
subtitleFontSize: 'number',
|
|
247
349
|
taskId: 'string',
|
|
248
350
|
timelines: { 'type': 'array', 'itemType': GetAutoClipsTaskInfoResponseBodyDataTimelines },
|
|
@@ -255,6 +357,9 @@ export class GetAutoClipsTaskInfoResponseBodyData extends $dara.Model {
|
|
|
255
357
|
if(Array.isArray(this.colorWords)) {
|
|
256
358
|
$dara.Model.validateArray(this.colorWords);
|
|
257
359
|
}
|
|
360
|
+
if(Array.isArray(this.stickers)) {
|
|
361
|
+
$dara.Model.validateArray(this.stickers);
|
|
362
|
+
}
|
|
258
363
|
if(Array.isArray(this.timelines)) {
|
|
259
364
|
$dara.Model.validateArray(this.timelines);
|
|
260
365
|
}
|
package/src/models/model.ts
CHANGED
|
@@ -8,6 +8,7 @@ export { AddDatasetDocumentRequestDocumentMultimodalMedias } from './AddDatasetD
|
|
|
8
8
|
export { AddDatasetDocumentRequestDocument } from './AddDatasetDocumentRequest';
|
|
9
9
|
export { AddDatasetDocumentResponseBodyData } from './AddDatasetDocumentResponseBody';
|
|
10
10
|
export { AsyncCreateClipsTaskRequestColorWords } from './AsyncCreateClipsTaskRequest';
|
|
11
|
+
export { AsyncCreateClipsTaskRequestStickers } from './AsyncCreateClipsTaskRequest';
|
|
11
12
|
export { AsyncCreateClipsTaskResponseBodyData } from './AsyncCreateClipsTaskResponseBody';
|
|
12
13
|
export { AsyncCreateClipsTimeLineResponseBodyData } from './AsyncCreateClipsTimeLineResponseBody';
|
|
13
14
|
export { AsyncEditTimelineRequestTimelinesClips } from './AsyncEditTimelineRequest';
|
|
@@ -16,6 +17,8 @@ export { AsyncEditTimelineResponseBodyData } from './AsyncEditTimelineResponseBo
|
|
|
16
17
|
export { AsyncUploadTenderDocResponseBodyData } from './AsyncUploadTenderDocResponseBody';
|
|
17
18
|
export { AsyncUploadVideoRequestReferenceVideo } from './AsyncUploadVideoRequest';
|
|
18
19
|
export { AsyncUploadVideoRequestSourceVideos } from './AsyncUploadVideoRequest';
|
|
20
|
+
export { AsyncUploadVideoRequestVideoRolesRoleUrls } from './AsyncUploadVideoRequest';
|
|
21
|
+
export { AsyncUploadVideoRequestVideoRoles } from './AsyncUploadVideoRequest';
|
|
19
22
|
export { AsyncUploadVideoResponseBodyDataVideoInfos } from './AsyncUploadVideoResponseBody';
|
|
20
23
|
export { AsyncUploadVideoResponseBodyData } from './AsyncUploadVideoResponseBody';
|
|
21
24
|
export { AsyncWritingBiddingDocResponseBodyData } from './AsyncWritingBiddingDocResponseBody';
|
|
@@ -55,6 +58,7 @@ export { GenerateViewPointResponseBodyData } from './GenerateViewPointResponseBo
|
|
|
55
58
|
export { GetAuditNotePostProcessingStatusResponseBodyData } from './GetAuditNotePostProcessingStatusResponseBody';
|
|
56
59
|
export { GetAuditNoteProcessingStatusResponseBodyData } from './GetAuditNoteProcessingStatusResponseBody';
|
|
57
60
|
export { GetAutoClipsTaskInfoResponseBodyDataColorWords } from './GetAutoClipsTaskInfoResponseBody';
|
|
61
|
+
export { GetAutoClipsTaskInfoResponseBodyDataStickers } from './GetAutoClipsTaskInfoResponseBody';
|
|
58
62
|
export { GetAutoClipsTaskInfoResponseBodyDataTimelinesClips } from './GetAutoClipsTaskInfoResponseBody';
|
|
59
63
|
export { GetAutoClipsTaskInfoResponseBodyDataTimelines } from './GetAutoClipsTaskInfoResponseBody';
|
|
60
64
|
export { GetAutoClipsTaskInfoResponseBodyData } from './GetAutoClipsTaskInfoResponseBody';
|