vargai 0.4.0-alpha14 → 0.4.0-alpha15
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json
CHANGED
|
@@ -137,6 +137,15 @@ function isOverlayLayer(layer: Layer): boolean {
|
|
|
137
137
|
return isVideoOverlayLayer(layer) || isImageOverlayLayer(layer);
|
|
138
138
|
}
|
|
139
139
|
|
|
140
|
+
function isTextOverlayLayer(layer: Layer): boolean {
|
|
141
|
+
return (
|
|
142
|
+
layer.type === "title" ||
|
|
143
|
+
layer.type === "subtitle" ||
|
|
144
|
+
layer.type === "news-title" ||
|
|
145
|
+
layer.type === "slide-in-text"
|
|
146
|
+
);
|
|
147
|
+
}
|
|
148
|
+
|
|
140
149
|
function buildBaseClipFilter(
|
|
141
150
|
clip: ProcessedClip,
|
|
142
151
|
clipIndex: number,
|
|
@@ -164,7 +173,10 @@ function buildBaseClipFilter(
|
|
|
164
173
|
let baseLabel = "";
|
|
165
174
|
let inputIdx = inputOffset;
|
|
166
175
|
|
|
167
|
-
|
|
176
|
+
// Filter out overlay layers AND text overlay layers (text will be applied after image overlays)
|
|
177
|
+
const baseLayers = clip.layers.filter(
|
|
178
|
+
(l) => l && !isOverlayLayer(l) && !isTextOverlayLayer(l),
|
|
179
|
+
);
|
|
168
180
|
|
|
169
181
|
for (let i = 0; i < baseLayers.length; i++) {
|
|
170
182
|
const layer = baseLayers[i];
|
|
@@ -201,58 +213,6 @@ function buildBaseClipFilter(
|
|
|
201
213
|
inputIdx++;
|
|
202
214
|
}
|
|
203
215
|
}
|
|
204
|
-
|
|
205
|
-
if (layer.type === "title") {
|
|
206
|
-
const titleFilter = getTitleFilter(
|
|
207
|
-
layer as TitleLayer,
|
|
208
|
-
baseLabel,
|
|
209
|
-
width,
|
|
210
|
-
height,
|
|
211
|
-
clip.duration,
|
|
212
|
-
);
|
|
213
|
-
const newLabel = `title${clipIndex}_${i}`;
|
|
214
|
-
filters.push(`${titleFilter}[${newLabel}]`);
|
|
215
|
-
baseLabel = newLabel;
|
|
216
|
-
}
|
|
217
|
-
|
|
218
|
-
if (layer.type === "subtitle") {
|
|
219
|
-
const subtitleFilter = getSubtitleFilter(
|
|
220
|
-
layer as SubtitleLayer,
|
|
221
|
-
baseLabel,
|
|
222
|
-
width,
|
|
223
|
-
height,
|
|
224
|
-
clip.duration,
|
|
225
|
-
);
|
|
226
|
-
const newLabel = `sub${clipIndex}_${i}`;
|
|
227
|
-
filters.push(`${subtitleFilter}[${newLabel}]`);
|
|
228
|
-
baseLabel = newLabel;
|
|
229
|
-
}
|
|
230
|
-
|
|
231
|
-
if (layer.type === "news-title") {
|
|
232
|
-
const newsFilter = getNewsTitleFilter(
|
|
233
|
-
layer as NewsTitleLayer,
|
|
234
|
-
baseLabel,
|
|
235
|
-
width,
|
|
236
|
-
height,
|
|
237
|
-
clip.duration,
|
|
238
|
-
);
|
|
239
|
-
const newLabel = `news${clipIndex}_${i}`;
|
|
240
|
-
filters.push(`${newsFilter}[${newLabel}]`);
|
|
241
|
-
baseLabel = newLabel;
|
|
242
|
-
}
|
|
243
|
-
|
|
244
|
-
if (layer.type === "slide-in-text") {
|
|
245
|
-
const slideFilter = getSlideInTextFilter(
|
|
246
|
-
layer as SlideInTextLayer,
|
|
247
|
-
baseLabel,
|
|
248
|
-
width,
|
|
249
|
-
height,
|
|
250
|
-
clip.duration,
|
|
251
|
-
);
|
|
252
|
-
const newLabel = `slide${clipIndex}_${i}`;
|
|
253
|
-
filters.push(`${slideFilter}[${newLabel}]`);
|
|
254
|
-
baseLabel = newLabel;
|
|
255
|
-
}
|
|
256
216
|
}
|
|
257
217
|
|
|
258
218
|
return {
|
|
@@ -358,6 +318,41 @@ function collectAudioLayers(
|
|
|
358
318
|
return audioLayers;
|
|
359
319
|
}
|
|
360
320
|
|
|
321
|
+
type TextLayer = TitleLayer | SubtitleLayer | NewsTitleLayer | SlideInTextLayer;
|
|
322
|
+
|
|
323
|
+
interface TimedTextLayer {
|
|
324
|
+
layer: TextLayer;
|
|
325
|
+
startTime: number;
|
|
326
|
+
duration: number;
|
|
327
|
+
}
|
|
328
|
+
|
|
329
|
+
function collectTextLayers(clips: ProcessedClip[]): TimedTextLayer[] {
|
|
330
|
+
const textLayers: TimedTextLayer[] = [];
|
|
331
|
+
let currentTime = 0;
|
|
332
|
+
|
|
333
|
+
for (let i = 0; i < clips.length; i++) {
|
|
334
|
+
const clip = clips[i];
|
|
335
|
+
if (!clip) continue;
|
|
336
|
+
|
|
337
|
+
for (const layer of clip.layers) {
|
|
338
|
+
if (layer && isTextOverlayLayer(layer)) {
|
|
339
|
+
textLayers.push({
|
|
340
|
+
layer: layer as TextLayer,
|
|
341
|
+
startTime: currentTime,
|
|
342
|
+
duration: clip.duration,
|
|
343
|
+
});
|
|
344
|
+
}
|
|
345
|
+
}
|
|
346
|
+
|
|
347
|
+
currentTime += clip.duration;
|
|
348
|
+
if (i < clips.length - 1) {
|
|
349
|
+
currentTime -= clip.transition.duration;
|
|
350
|
+
}
|
|
351
|
+
}
|
|
352
|
+
|
|
353
|
+
return textLayers;
|
|
354
|
+
}
|
|
355
|
+
|
|
361
356
|
function buildTransitionFilter(
|
|
362
357
|
fromLabel: string,
|
|
363
358
|
toLabel: string,
|
|
@@ -744,6 +739,67 @@ export async function editly(config: EditlyConfig): Promise<void> {
|
|
|
744
739
|
finalVideoLabel = currentBase;
|
|
745
740
|
}
|
|
746
741
|
|
|
742
|
+
const textLayers = collectTextLayers(clips);
|
|
743
|
+
if (textLayers.length > 0) {
|
|
744
|
+
let currentBase = finalVideoLabel;
|
|
745
|
+
|
|
746
|
+
for (let i = 0; i < textLayers.length; i++) {
|
|
747
|
+
const timedLayer = textLayers[i];
|
|
748
|
+
if (!timedLayer) continue;
|
|
749
|
+
|
|
750
|
+
const { layer, startTime, duration } = timedLayer;
|
|
751
|
+
const outputLabel = `vwithtext${i}`;
|
|
752
|
+
|
|
753
|
+
const timedLayerWithEnable = {
|
|
754
|
+
...layer,
|
|
755
|
+
start: layer.start ?? startTime,
|
|
756
|
+
stop: layer.stop ?? startTime + duration,
|
|
757
|
+
};
|
|
758
|
+
|
|
759
|
+
if (layer.type === "title") {
|
|
760
|
+
const titleFilter = getTitleFilter(
|
|
761
|
+
timedLayerWithEnable as TitleLayer,
|
|
762
|
+
currentBase,
|
|
763
|
+
width,
|
|
764
|
+
height,
|
|
765
|
+
totalDuration,
|
|
766
|
+
);
|
|
767
|
+
allFilters.push(`${titleFilter}[${outputLabel}]`);
|
|
768
|
+
} else if (layer.type === "subtitle") {
|
|
769
|
+
const subtitleFilter = getSubtitleFilter(
|
|
770
|
+
timedLayerWithEnable as SubtitleLayer,
|
|
771
|
+
currentBase,
|
|
772
|
+
width,
|
|
773
|
+
height,
|
|
774
|
+
totalDuration,
|
|
775
|
+
);
|
|
776
|
+
allFilters.push(`${subtitleFilter}[${outputLabel}]`);
|
|
777
|
+
} else if (layer.type === "news-title") {
|
|
778
|
+
const newsFilter = getNewsTitleFilter(
|
|
779
|
+
timedLayerWithEnable as NewsTitleLayer,
|
|
780
|
+
currentBase,
|
|
781
|
+
width,
|
|
782
|
+
height,
|
|
783
|
+
totalDuration,
|
|
784
|
+
);
|
|
785
|
+
allFilters.push(`${newsFilter}[${outputLabel}]`);
|
|
786
|
+
} else if (layer.type === "slide-in-text") {
|
|
787
|
+
const slideFilter = getSlideInTextFilter(
|
|
788
|
+
timedLayerWithEnable as SlideInTextLayer,
|
|
789
|
+
currentBase,
|
|
790
|
+
width,
|
|
791
|
+
height,
|
|
792
|
+
totalDuration,
|
|
793
|
+
);
|
|
794
|
+
allFilters.push(`${slideFilter}[${outputLabel}]`);
|
|
795
|
+
}
|
|
796
|
+
|
|
797
|
+
currentBase = outputLabel;
|
|
798
|
+
}
|
|
799
|
+
|
|
800
|
+
finalVideoLabel = currentBase;
|
|
801
|
+
}
|
|
802
|
+
|
|
747
803
|
const clipAudioLayers = collectAudioLayers(clips);
|
|
748
804
|
const videoInputCount = allInputs.length;
|
|
749
805
|
const audioFilter = buildAudioFilter(
|
|
@@ -31,7 +31,8 @@ async function loadComponent(filePath: string): Promise<VargElement> {
|
|
|
31
31
|
const resolvedPath = resolve(filePath);
|
|
32
32
|
const source = await Bun.file(resolvedPath).text();
|
|
33
33
|
|
|
34
|
-
const
|
|
34
|
+
const hasAnyImport = source.includes(" from ");
|
|
35
|
+
const hasVargaiImport =
|
|
35
36
|
source.includes("from 'vargai") ||
|
|
36
37
|
source.includes('from "vargai') ||
|
|
37
38
|
source.includes("from '@vargai") ||
|
|
@@ -40,22 +41,21 @@ async function loadComponent(filePath: string): Promise<VargElement> {
|
|
|
40
41
|
const hasJsxPragma =
|
|
41
42
|
source.includes("@jsxImportSource") || source.includes("@jsx ");
|
|
42
43
|
|
|
43
|
-
|
|
44
|
+
// file has imports (relative or absolute) - import directly to preserve paths
|
|
45
|
+
if (hasAnyImport) {
|
|
44
46
|
const mod = await import(resolvedPath);
|
|
45
47
|
return mod.default;
|
|
46
48
|
}
|
|
47
49
|
|
|
50
|
+
// no imports - inject auto-imports and jsx pragma
|
|
48
51
|
const pkgDir = new URL("../../..", import.meta.url).pathname;
|
|
49
52
|
const tmpDir = `${pkgDir}/.cache/varg-render`;
|
|
50
53
|
if (!existsSync(tmpDir)) {
|
|
51
54
|
mkdirSync(tmpDir, { recursive: true });
|
|
52
55
|
}
|
|
53
56
|
|
|
54
|
-
const prepended = hasImports
|
|
55
|
-
? `/** @jsxImportSource vargai */\n`
|
|
56
|
-
: AUTO_IMPORTS;
|
|
57
57
|
const tmpFile = `${tmpDir}/${Date.now()}.tsx`;
|
|
58
|
-
await Bun.write(tmpFile,
|
|
58
|
+
await Bun.write(tmpFile, AUTO_IMPORTS + source);
|
|
59
59
|
|
|
60
60
|
try {
|
|
61
61
|
const mod = await import(tmpFile);
|
|
@@ -70,42 +70,63 @@ export async function renderRoot(
|
|
|
70
70
|
placeholderCount.total++;
|
|
71
71
|
};
|
|
72
72
|
|
|
73
|
+
const cachedGenerateImage = options.cache
|
|
74
|
+
? withCache(generateImage, { storage: fileCache({ dir: options.cache }) })
|
|
75
|
+
: generateImage;
|
|
76
|
+
|
|
77
|
+
const cachedGenerateVideo = options.cache
|
|
78
|
+
? withCache(generateVideo, { storage: fileCache({ dir: options.cache }) })
|
|
79
|
+
: generateVideo;
|
|
80
|
+
|
|
73
81
|
const wrapGenerateImage: typeof generateImage = async (opts) => {
|
|
74
82
|
if (
|
|
75
83
|
typeof opts.model === "string" ||
|
|
76
84
|
opts.model.specificationVersion !== "v3"
|
|
77
85
|
) {
|
|
78
|
-
return
|
|
86
|
+
return cachedGenerateImage(opts);
|
|
87
|
+
}
|
|
88
|
+
|
|
89
|
+
if (mode === "preview") {
|
|
90
|
+
trackPlaceholder("image");
|
|
91
|
+
}
|
|
92
|
+
|
|
93
|
+
try {
|
|
94
|
+
return await cachedGenerateImage(opts);
|
|
95
|
+
} catch (error) {
|
|
96
|
+
if (mode === "strict") throw error;
|
|
97
|
+
trackPlaceholder("image");
|
|
98
|
+
onFallback(error as Error, String(opts.prompt));
|
|
99
|
+
const wrappedModel = wrapImageModel({
|
|
100
|
+
model: opts.model,
|
|
101
|
+
middleware: imagePlaceholderFallbackMiddleware({
|
|
102
|
+
mode: "preview",
|
|
103
|
+
onFallback: () => {},
|
|
104
|
+
}),
|
|
105
|
+
});
|
|
106
|
+
return generateImage({ ...opts, model: wrappedModel });
|
|
79
107
|
}
|
|
80
|
-
const wrappedModel = wrapImageModel({
|
|
81
|
-
model: opts.model,
|
|
82
|
-
middleware: imagePlaceholderFallbackMiddleware({
|
|
83
|
-
mode,
|
|
84
|
-
onFallback: (error, prompt) => {
|
|
85
|
-
trackPlaceholder("image");
|
|
86
|
-
onFallback(error, prompt);
|
|
87
|
-
},
|
|
88
|
-
}),
|
|
89
|
-
});
|
|
90
|
-
const result = await generateImage({ ...opts, model: wrappedModel });
|
|
91
|
-
if (mode === "preview") trackPlaceholder("image");
|
|
92
|
-
return result;
|
|
93
108
|
};
|
|
94
109
|
|
|
95
110
|
const wrapGenerateVideo: typeof generateVideo = async (opts) => {
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
|
|
108
|
-
|
|
111
|
+
if (mode === "preview") {
|
|
112
|
+
trackPlaceholder("video");
|
|
113
|
+
}
|
|
114
|
+
|
|
115
|
+
try {
|
|
116
|
+
return await cachedGenerateVideo(opts);
|
|
117
|
+
} catch (error) {
|
|
118
|
+
if (mode === "strict") throw error;
|
|
119
|
+
trackPlaceholder("video");
|
|
120
|
+
onFallback(error as Error, String(opts.prompt));
|
|
121
|
+
const wrappedModel = wrapVideoModel({
|
|
122
|
+
model: opts.model,
|
|
123
|
+
middleware: placeholderFallbackMiddleware({
|
|
124
|
+
mode: "preview",
|
|
125
|
+
onFallback: () => {},
|
|
126
|
+
}),
|
|
127
|
+
});
|
|
128
|
+
return generateVideo({ ...opts, model: wrappedModel });
|
|
129
|
+
}
|
|
109
130
|
};
|
|
110
131
|
|
|
111
132
|
const ctx: RenderContext = {
|
|
@@ -113,16 +134,8 @@ export async function renderRoot(
|
|
|
113
134
|
height: props.height ?? 1080,
|
|
114
135
|
fps: props.fps ?? 30,
|
|
115
136
|
cache: options.cache ? fileCache({ dir: options.cache }) : undefined,
|
|
116
|
-
generateImage:
|
|
117
|
-
|
|
118
|
-
storage: fileCache({ dir: options.cache }),
|
|
119
|
-
})
|
|
120
|
-
: wrapGenerateImage,
|
|
121
|
-
generateVideo: options.cache
|
|
122
|
-
? withCache(wrapGenerateVideo, {
|
|
123
|
-
storage: fileCache({ dir: options.cache }),
|
|
124
|
-
})
|
|
125
|
-
: wrapGenerateVideo,
|
|
137
|
+
generateImage: wrapGenerateImage,
|
|
138
|
+
generateVideo: wrapGenerateVideo,
|
|
126
139
|
tempFiles: [],
|
|
127
140
|
progress,
|
|
128
141
|
pending: new Map(),
|
|
@@ -288,6 +301,7 @@ export async function renderRoot(
|
|
|
288
301
|
fps: ctx.fps,
|
|
289
302
|
clips,
|
|
290
303
|
audioTracks: audioTracks.length > 0 ? audioTracks : undefined,
|
|
304
|
+
verbose: options.verbose,
|
|
291
305
|
});
|
|
292
306
|
|
|
293
307
|
completeTask(progress, editlyTaskId);
|