vargai 0.4.0-alpha14 → 0.4.0-alpha15

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -64,7 +64,7 @@
64
64
  "replicate": "^1.4.0",
65
65
  "zod": "^4.2.1"
66
66
  },
67
- "version": "0.4.0-alpha14",
67
+ "version": "0.4.0-alpha15",
68
68
  "exports": {
69
69
  ".": "./src/index.ts",
70
70
  "./ai": "./src/ai-sdk/index.ts",
@@ -137,6 +137,15 @@ function isOverlayLayer(layer: Layer): boolean {
137
137
  return isVideoOverlayLayer(layer) || isImageOverlayLayer(layer);
138
138
  }
139
139
 
140
+ function isTextOverlayLayer(layer: Layer): boolean {
141
+ return (
142
+ layer.type === "title" ||
143
+ layer.type === "subtitle" ||
144
+ layer.type === "news-title" ||
145
+ layer.type === "slide-in-text"
146
+ );
147
+ }
148
+
140
149
  function buildBaseClipFilter(
141
150
  clip: ProcessedClip,
142
151
  clipIndex: number,
@@ -164,7 +173,10 @@ function buildBaseClipFilter(
164
173
  let baseLabel = "";
165
174
  let inputIdx = inputOffset;
166
175
 
167
- const baseLayers = clip.layers.filter((l) => l && !isOverlayLayer(l));
176
+ // Filter out overlay layers AND text overlay layers (text will be applied after image overlays)
177
+ const baseLayers = clip.layers.filter(
178
+ (l) => l && !isOverlayLayer(l) && !isTextOverlayLayer(l),
179
+ );
168
180
 
169
181
  for (let i = 0; i < baseLayers.length; i++) {
170
182
  const layer = baseLayers[i];
@@ -201,58 +213,6 @@ function buildBaseClipFilter(
201
213
  inputIdx++;
202
214
  }
203
215
  }
204
-
205
- if (layer.type === "title") {
206
- const titleFilter = getTitleFilter(
207
- layer as TitleLayer,
208
- baseLabel,
209
- width,
210
- height,
211
- clip.duration,
212
- );
213
- const newLabel = `title${clipIndex}_${i}`;
214
- filters.push(`${titleFilter}[${newLabel}]`);
215
- baseLabel = newLabel;
216
- }
217
-
218
- if (layer.type === "subtitle") {
219
- const subtitleFilter = getSubtitleFilter(
220
- layer as SubtitleLayer,
221
- baseLabel,
222
- width,
223
- height,
224
- clip.duration,
225
- );
226
- const newLabel = `sub${clipIndex}_${i}`;
227
- filters.push(`${subtitleFilter}[${newLabel}]`);
228
- baseLabel = newLabel;
229
- }
230
-
231
- if (layer.type === "news-title") {
232
- const newsFilter = getNewsTitleFilter(
233
- layer as NewsTitleLayer,
234
- baseLabel,
235
- width,
236
- height,
237
- clip.duration,
238
- );
239
- const newLabel = `news${clipIndex}_${i}`;
240
- filters.push(`${newsFilter}[${newLabel}]`);
241
- baseLabel = newLabel;
242
- }
243
-
244
- if (layer.type === "slide-in-text") {
245
- const slideFilter = getSlideInTextFilter(
246
- layer as SlideInTextLayer,
247
- baseLabel,
248
- width,
249
- height,
250
- clip.duration,
251
- );
252
- const newLabel = `slide${clipIndex}_${i}`;
253
- filters.push(`${slideFilter}[${newLabel}]`);
254
- baseLabel = newLabel;
255
- }
256
216
  }
257
217
 
258
218
  return {
@@ -358,6 +318,41 @@ function collectAudioLayers(
358
318
  return audioLayers;
359
319
  }
360
320
 
321
+ type TextLayer = TitleLayer | SubtitleLayer | NewsTitleLayer | SlideInTextLayer;
322
+
323
+ interface TimedTextLayer {
324
+ layer: TextLayer;
325
+ startTime: number;
326
+ duration: number;
327
+ }
328
+
329
+ function collectTextLayers(clips: ProcessedClip[]): TimedTextLayer[] {
330
+ const textLayers: TimedTextLayer[] = [];
331
+ let currentTime = 0;
332
+
333
+ for (let i = 0; i < clips.length; i++) {
334
+ const clip = clips[i];
335
+ if (!clip) continue;
336
+
337
+ for (const layer of clip.layers) {
338
+ if (layer && isTextOverlayLayer(layer)) {
339
+ textLayers.push({
340
+ layer: layer as TextLayer,
341
+ startTime: currentTime,
342
+ duration: clip.duration,
343
+ });
344
+ }
345
+ }
346
+
347
+ currentTime += clip.duration;
348
+ if (i < clips.length - 1) {
349
+ currentTime -= clip.transition.duration;
350
+ }
351
+ }
352
+
353
+ return textLayers;
354
+ }
355
+
361
356
  function buildTransitionFilter(
362
357
  fromLabel: string,
363
358
  toLabel: string,
@@ -744,6 +739,67 @@ export async function editly(config: EditlyConfig): Promise<void> {
744
739
  finalVideoLabel = currentBase;
745
740
  }
746
741
 
742
+ const textLayers = collectTextLayers(clips);
743
+ if (textLayers.length > 0) {
744
+ let currentBase = finalVideoLabel;
745
+
746
+ for (let i = 0; i < textLayers.length; i++) {
747
+ const timedLayer = textLayers[i];
748
+ if (!timedLayer) continue;
749
+
750
+ const { layer, startTime, duration } = timedLayer;
751
+ const outputLabel = `vwithtext${i}`;
752
+
753
+ const timedLayerWithEnable = {
754
+ ...layer,
755
+ start: layer.start ?? startTime,
756
+ stop: layer.stop ?? startTime + duration,
757
+ };
758
+
759
+ if (layer.type === "title") {
760
+ const titleFilter = getTitleFilter(
761
+ timedLayerWithEnable as TitleLayer,
762
+ currentBase,
763
+ width,
764
+ height,
765
+ totalDuration,
766
+ );
767
+ allFilters.push(`${titleFilter}[${outputLabel}]`);
768
+ } else if (layer.type === "subtitle") {
769
+ const subtitleFilter = getSubtitleFilter(
770
+ timedLayerWithEnable as SubtitleLayer,
771
+ currentBase,
772
+ width,
773
+ height,
774
+ totalDuration,
775
+ );
776
+ allFilters.push(`${subtitleFilter}[${outputLabel}]`);
777
+ } else if (layer.type === "news-title") {
778
+ const newsFilter = getNewsTitleFilter(
779
+ timedLayerWithEnable as NewsTitleLayer,
780
+ currentBase,
781
+ width,
782
+ height,
783
+ totalDuration,
784
+ );
785
+ allFilters.push(`${newsFilter}[${outputLabel}]`);
786
+ } else if (layer.type === "slide-in-text") {
787
+ const slideFilter = getSlideInTextFilter(
788
+ timedLayerWithEnable as SlideInTextLayer,
789
+ currentBase,
790
+ width,
791
+ height,
792
+ totalDuration,
793
+ );
794
+ allFilters.push(`${slideFilter}[${outputLabel}]`);
795
+ }
796
+
797
+ currentBase = outputLabel;
798
+ }
799
+
800
+ finalVideoLabel = currentBase;
801
+ }
802
+
747
803
  const clipAudioLayers = collectAudioLayers(clips);
748
804
  const videoInputCount = allInputs.length;
749
805
  const audioFilter = buildAudioFilter(
@@ -31,7 +31,8 @@ async function loadComponent(filePath: string): Promise<VargElement> {
31
31
  const resolvedPath = resolve(filePath);
32
32
  const source = await Bun.file(resolvedPath).text();
33
33
 
34
- const hasImports =
34
+ const hasAnyImport = source.includes(" from ");
35
+ const hasVargaiImport =
35
36
  source.includes("from 'vargai") ||
36
37
  source.includes('from "vargai') ||
37
38
  source.includes("from '@vargai") ||
@@ -40,22 +41,21 @@ async function loadComponent(filePath: string): Promise<VargElement> {
40
41
  const hasJsxPragma =
41
42
  source.includes("@jsxImportSource") || source.includes("@jsx ");
42
43
 
43
- if (hasImports && hasJsxPragma) {
44
+ // file has imports (relative or absolute) - import directly to preserve paths
45
+ if (hasAnyImport) {
44
46
  const mod = await import(resolvedPath);
45
47
  return mod.default;
46
48
  }
47
49
 
50
+ // no imports - inject auto-imports and jsx pragma
48
51
  const pkgDir = new URL("../../..", import.meta.url).pathname;
49
52
  const tmpDir = `${pkgDir}/.cache/varg-render`;
50
53
  if (!existsSync(tmpDir)) {
51
54
  mkdirSync(tmpDir, { recursive: true });
52
55
  }
53
56
 
54
- const prepended = hasImports
55
- ? `/** @jsxImportSource vargai */\n`
56
- : AUTO_IMPORTS;
57
57
  const tmpFile = `${tmpDir}/${Date.now()}.tsx`;
58
- await Bun.write(tmpFile, prepended + source);
58
+ await Bun.write(tmpFile, AUTO_IMPORTS + source);
59
59
 
60
60
  try {
61
61
  const mod = await import(tmpFile);
@@ -70,42 +70,63 @@ export async function renderRoot(
70
70
  placeholderCount.total++;
71
71
  };
72
72
 
73
+ const cachedGenerateImage = options.cache
74
+ ? withCache(generateImage, { storage: fileCache({ dir: options.cache }) })
75
+ : generateImage;
76
+
77
+ const cachedGenerateVideo = options.cache
78
+ ? withCache(generateVideo, { storage: fileCache({ dir: options.cache }) })
79
+ : generateVideo;
80
+
73
81
  const wrapGenerateImage: typeof generateImage = async (opts) => {
74
82
  if (
75
83
  typeof opts.model === "string" ||
76
84
  opts.model.specificationVersion !== "v3"
77
85
  ) {
78
- return generateImage(opts);
86
+ return cachedGenerateImage(opts);
87
+ }
88
+
89
+ if (mode === "preview") {
90
+ trackPlaceholder("image");
91
+ }
92
+
93
+ try {
94
+ return await cachedGenerateImage(opts);
95
+ } catch (error) {
96
+ if (mode === "strict") throw error;
97
+ trackPlaceholder("image");
98
+ onFallback(error as Error, String(opts.prompt));
99
+ const wrappedModel = wrapImageModel({
100
+ model: opts.model,
101
+ middleware: imagePlaceholderFallbackMiddleware({
102
+ mode: "preview",
103
+ onFallback: () => {},
104
+ }),
105
+ });
106
+ return generateImage({ ...opts, model: wrappedModel });
79
107
  }
80
- const wrappedModel = wrapImageModel({
81
- model: opts.model,
82
- middleware: imagePlaceholderFallbackMiddleware({
83
- mode,
84
- onFallback: (error, prompt) => {
85
- trackPlaceholder("image");
86
- onFallback(error, prompt);
87
- },
88
- }),
89
- });
90
- const result = await generateImage({ ...opts, model: wrappedModel });
91
- if (mode === "preview") trackPlaceholder("image");
92
- return result;
93
108
  };
94
109
 
95
110
  const wrapGenerateVideo: typeof generateVideo = async (opts) => {
96
- const wrappedModel = wrapVideoModel({
97
- model: opts.model,
98
- middleware: placeholderFallbackMiddleware({
99
- mode,
100
- onFallback: (error, prompt) => {
101
- trackPlaceholder("video");
102
- onFallback(error, prompt);
103
- },
104
- }),
105
- });
106
- const result = await generateVideo({ ...opts, model: wrappedModel });
107
- if (mode === "preview") trackPlaceholder("video");
108
- return result;
111
+ if (mode === "preview") {
112
+ trackPlaceholder("video");
113
+ }
114
+
115
+ try {
116
+ return await cachedGenerateVideo(opts);
117
+ } catch (error) {
118
+ if (mode === "strict") throw error;
119
+ trackPlaceholder("video");
120
+ onFallback(error as Error, String(opts.prompt));
121
+ const wrappedModel = wrapVideoModel({
122
+ model: opts.model,
123
+ middleware: placeholderFallbackMiddleware({
124
+ mode: "preview",
125
+ onFallback: () => {},
126
+ }),
127
+ });
128
+ return generateVideo({ ...opts, model: wrappedModel });
129
+ }
109
130
  };
110
131
 
111
132
  const ctx: RenderContext = {
@@ -113,16 +134,8 @@ export async function renderRoot(
113
134
  height: props.height ?? 1080,
114
135
  fps: props.fps ?? 30,
115
136
  cache: options.cache ? fileCache({ dir: options.cache }) : undefined,
116
- generateImage: options.cache
117
- ? withCache(wrapGenerateImage, {
118
- storage: fileCache({ dir: options.cache }),
119
- })
120
- : wrapGenerateImage,
121
- generateVideo: options.cache
122
- ? withCache(wrapGenerateVideo, {
123
- storage: fileCache({ dir: options.cache }),
124
- })
125
- : wrapGenerateVideo,
137
+ generateImage: wrapGenerateImage,
138
+ generateVideo: wrapGenerateVideo,
126
139
  tempFiles: [],
127
140
  progress,
128
141
  pending: new Map(),
@@ -288,6 +301,7 @@ export async function renderRoot(
288
301
  fps: ctx.fps,
289
302
  clips,
290
303
  audioTracks: audioTracks.length > 0 ? audioTracks : undefined,
304
+ verbose: options.verbose,
291
305
  });
292
306
 
293
307
  completeTask(progress, editlyTaskId);
@@ -223,6 +223,7 @@ export interface RenderOptions {
223
223
  output?: string;
224
224
  cache?: string;
225
225
  quiet?: boolean;
226
+ verbose?: boolean;
226
227
  mode?: RenderMode;
227
228
  defaults?: DefaultModels;
228
229
  }