@editframe/assets 0.40.0 → 0.40.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (33) hide show
  1. package/dist/Probe.cjs +28 -9
  2. package/dist/Probe.cjs.map +1 -1
  3. package/dist/Probe.d.cts +27 -26
  4. package/dist/Probe.d.ts +27 -26
  5. package/dist/Probe.js +29 -9
  6. package/dist/Probe.js.map +1 -1
  7. package/dist/VideoRenderOptions.d.cts +48 -48
  8. package/dist/VideoRenderOptions.d.ts +48 -48
  9. package/dist/generateFragmentIndex.cjs +39 -26
  10. package/dist/generateFragmentIndex.cjs.map +1 -1
  11. package/dist/generateFragmentIndex.d.cts +5 -1
  12. package/dist/generateFragmentIndex.d.ts +5 -1
  13. package/dist/generateFragmentIndex.js +35 -27
  14. package/dist/generateFragmentIndex.js.map +1 -1
  15. package/dist/generateSingleTrack.cjs.map +1 -1
  16. package/dist/generateSingleTrack.js.map +1 -1
  17. package/dist/idempotentTask.cjs +58 -6
  18. package/dist/idempotentTask.cjs.map +1 -1
  19. package/dist/idempotentTask.js +58 -7
  20. package/dist/idempotentTask.js.map +1 -1
  21. package/dist/tasks/findOrCreateCaptions.cjs +13 -6
  22. package/dist/tasks/findOrCreateCaptions.cjs.map +1 -1
  23. package/dist/tasks/findOrCreateCaptions.js +13 -6
  24. package/dist/tasks/findOrCreateCaptions.js.map +1 -1
  25. package/dist/tasks/generateScrubTrack.cjs +1 -11
  26. package/dist/tasks/generateScrubTrack.cjs.map +1 -1
  27. package/dist/tasks/generateScrubTrack.js +1 -11
  28. package/dist/tasks/generateScrubTrack.js.map +1 -1
  29. package/dist/tasks/generateTrackFragmentIndex.cjs +22 -28
  30. package/dist/tasks/generateTrackFragmentIndex.cjs.map +1 -1
  31. package/dist/tasks/generateTrackFragmentIndex.js +22 -28
  32. package/dist/tasks/generateTrackFragmentIndex.js.map +1 -1
  33. package/package.json +1 -1
@@ -2,15 +2,15 @@ const require_rolldown_runtime = require('../_virtual/rolldown_runtime.cjs');
2
2
  const require_idempotentTask = require('../idempotentTask.cjs');
3
3
  let node_child_process = require("node:child_process");
4
4
  node_child_process = require_rolldown_runtime.__toESM(node_child_process);
5
- let node_util = require("node:util");
6
- node_util = require_rolldown_runtime.__toESM(node_util);
7
5
  let debug = require("debug");
8
6
  debug = require_rolldown_runtime.__toESM(debug);
9
7
  let node_path = require("node:path");
10
8
  node_path = require_rolldown_runtime.__toESM(node_path);
9
+ let node_util = require("node:util");
10
+ node_util = require_rolldown_runtime.__toESM(node_util);
11
11
 
12
12
  //#region src/tasks/findOrCreateCaptions.ts
13
- const execPromise = (0, node_util.promisify)(node_child_process.exec);
13
+ const execFilePromise = (0, node_util.promisify)(node_child_process.execFile);
14
14
  const log = (0, debug.default)("ef:generateCaptions");
15
15
  const convertWhisperToEditframeFormat = (whisperData) => {
16
16
  return {
@@ -27,9 +27,16 @@ const convertWhisperToEditframeFormat = (whisperData) => {
27
27
  };
28
28
  };
29
29
  const generateCaptionDataFromPath = async (absolutePath) => {
30
- const command = `whisper_timestamped --language en --efficient --output_format json ${absolutePath}`;
31
- log(`Running command: ${command}`);
32
- const { stdout } = await execPromise(command);
30
+ const args = [
31
+ "--language",
32
+ "en",
33
+ "--efficient",
34
+ "--output_format",
35
+ "json",
36
+ absolutePath
37
+ ];
38
+ log("Running whisper_timestamped", args);
39
+ const { stdout } = await execFilePromise("whisper_timestamped", args);
33
40
  try {
34
41
  const captionData = convertWhisperToEditframeFormat(JSON.parse(stdout));
35
42
  return JSON.stringify(captionData, null, 2);
@@ -1 +1 @@
1
- {"version":3,"file":"findOrCreateCaptions.cjs","names":["exec","idempotentTask"],"sources":["../../src/tasks/findOrCreateCaptions.ts"],"sourcesContent":["import { basename } from \"node:path\";\nimport { promisify } from \"node:util\";\nimport { exec } from \"node:child_process\";\n\nimport debug from \"debug\";\n\nimport { idempotentTask } from \"../idempotentTask.js\";\n\nconst execPromise = promisify(exec);\n\nconst log = debug(\"ef:generateCaptions\");\n\ninterface WhisperWord {\n text: string;\n start: number;\n end: number;\n confidence: number;\n}\n\ninterface WhisperSegment {\n text: string;\n start: number;\n end: number;\n words: WhisperWord[];\n}\n\ninterface WhisperOutput {\n segments: WhisperSegment[];\n}\n\ninterface CaptionOutput {\n segments: Array<{\n start: number;\n end: number;\n text: string;\n }>;\n word_segments: Array<{\n text: string;\n start: number;\n end: number;\n }>;\n}\n\nconst convertWhisperToEditframeFormat = (\n whisperData: WhisperOutput,\n): CaptionOutput => {\n const segments = whisperData.segments.map((segment) => ({\n start: Math.round(segment.start * 1000), // Convert to milliseconds\n end: Math.round(segment.end * 1000),\n text: segment.text.trim(),\n }));\n\n const word_segments = whisperData.segments.flatMap((segment) =>\n segment.words.map((word) => ({\n text: word.text,\n start: Math.round(word.start * 1000), // Convert to milliseconds\n end: Math.round(word.end * 1000),\n })),\n );\n\n return { segments, word_segments };\n};\n\nexport const generateCaptionDataFromPath = async (absolutePath: string) => {\n const command = `whisper_timestamped --language en --efficient --output_format json ${absolutePath}`;\n log(`Running command: ${command}`);\n const { stdout } = await execPromise(command);\n\n try {\n const whisperData = JSON.parse(stdout) as WhisperOutput;\n const captionData = convertWhisperToEditframeFormat(whisperData);\n return JSON.stringify(captionData, null, 2);\n } catch (error) {\n log(`Error parsing whisper output: ${error}`);\n throw new Error(`Failed to parse whisper_timestamped output: ${error}`);\n }\n};\n\nconst generateCaptionDataTask = idempotentTask({\n label: \"captions\",\n filename: (absolutePath) => `${basename(absolutePath)}.captions.json`,\n runner: generateCaptionDataFromPath,\n});\n\nexport const findOrCreateCaptions = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateCaptionDataTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error finding or creating captions\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;;;;;AAQA,MAAM,uCAAwBA,wBAAK;AAEnC,MAAM,yBAAY,sBAAsB;AAiCxC,MAAM,mCACJ,gBACkB;AAelB,QAAO;EAAE,UAdQ,YAAY,SAAS,KAAK,aAAa;GACtD,OAAO,KAAK,MAAM,QAAQ,QAAQ,IAAK;GACvC,KAAK,KAAK,MAAM,QAAQ,MAAM,IAAK;GACnC,MAAM,QAAQ,KAAK,MAAM;GAC1B,EAAE;EAUgB,eARG,YAAY,SAAS,SAAS,YAClD,QAAQ,MAAM,KAAK,UAAU;GAC3B,MAAM,KAAK;GACX,OAAO,KAAK,MAAM,KAAK,QAAQ,IAAK;GACpC,KAAK,KAAK,MAAM,KAAK,MAAM,IAAK;GACjC,EAAE,CACJ;EAEiC;;AAGpC,MAAa,8BAA8B,OAAO,iBAAyB;CACzE,MAAM,UAAU,sEAAsE;AACtF,KAAI,oBAAoB,UAAU;CAClC,MAAM,EAAE,WAAW,MAAM,YAAY,QAAQ;AAE7C,KAAI;EAEF,MAAM,cAAc,gCADA,KAAK,MAAM,OAAO,CAC0B;AAChE,SAAO,KAAK,UAAU,aAAa,MAAM,EAAE;UACpC,OAAO;AACd,MAAI,iCAAiC,QAAQ;AAC7C,QAAM,IAAI,MAAM,+CAA+C,QAAQ;;;AAI3E,MAAM,0BAA0BC,sCAAe;CAC7C,OAAO;CACP,WAAW,iBAAiB,2BAAY,aAAa,CAAC;CACtD,QAAQ;CACT,CAAC;AAEF,MAAa,uBAAuB,OAClC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,wBAAwB,WAAW,aAAa;UACtD,OAAO;AACd,UAAQ,MAAM,sCAAsC,MAAM;AAC1D,QAAM"}
1
+ {"version":3,"file":"findOrCreateCaptions.cjs","names":["execFile","idempotentTask"],"sources":["../../src/tasks/findOrCreateCaptions.ts"],"sourcesContent":["import { basename } from \"node:path\";\nimport { promisify } from \"node:util\";\nimport { execFile } from \"node:child_process\";\n\nimport debug from \"debug\";\n\nimport { idempotentTask } from \"../idempotentTask.js\";\n\nconst execFilePromise = promisify(execFile);\n\nconst log = debug(\"ef:generateCaptions\");\n\ninterface WhisperWord {\n text: string;\n start: number;\n end: number;\n confidence: number;\n}\n\ninterface WhisperSegment {\n text: string;\n start: number;\n end: number;\n words: WhisperWord[];\n}\n\ninterface WhisperOutput {\n segments: WhisperSegment[];\n}\n\ninterface CaptionOutput {\n segments: Array<{\n start: number;\n end: number;\n text: string;\n }>;\n word_segments: Array<{\n text: string;\n start: number;\n end: number;\n }>;\n}\n\nconst convertWhisperToEditframeFormat = (\n whisperData: WhisperOutput,\n): CaptionOutput => {\n const segments = whisperData.segments.map((segment) => ({\n start: Math.round(segment.start * 1000), // Convert to milliseconds\n end: Math.round(segment.end * 1000),\n text: segment.text.trim(),\n }));\n\n const word_segments = whisperData.segments.flatMap((segment) =>\n segment.words.map((word) => ({\n text: word.text,\n start: Math.round(word.start * 1000), // Convert to milliseconds\n end: Math.round(word.end * 1000),\n })),\n );\n\n return { segments, word_segments };\n};\n\nexport const generateCaptionDataFromPath = async (absolutePath: string) => {\n const args = [\n \"--language\",\n \"en\",\n \"--efficient\",\n \"--output_format\",\n \"json\",\n absolutePath,\n ];\n log(\"Running whisper_timestamped\", args);\n const { stdout } = await execFilePromise(\"whisper_timestamped\", args);\n\n try {\n const whisperData = JSON.parse(stdout) as WhisperOutput;\n const captionData = convertWhisperToEditframeFormat(whisperData);\n return JSON.stringify(captionData, null, 2);\n } catch (error) {\n log(`Error parsing whisper output: ${error}`);\n throw new Error(`Failed to parse whisper_timestamped output: ${error}`);\n }\n};\n\nconst generateCaptionDataTask = idempotentTask({\n label: \"captions\",\n filename: (absolutePath) => `${basename(absolutePath)}.captions.json`,\n runner: generateCaptionDataFromPath,\n});\n\nexport const findOrCreateCaptions = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateCaptionDataTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error finding or creating captions\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;;;;;AAQA,MAAM,2CAA4BA,4BAAS;AAE3C,MAAM,yBAAY,sBAAsB;AAiCxC,MAAM,mCACJ,gBACkB;AAelB,QAAO;EAAE,UAdQ,YAAY,SAAS,KAAK,aAAa;GACtD,OAAO,KAAK,MAAM,QAAQ,QAAQ,IAAK;GACvC,KAAK,KAAK,MAAM,QAAQ,MAAM,IAAK;GACnC,MAAM,QAAQ,KAAK,MAAM;GAC1B,EAAE;EAUgB,eARG,YAAY,SAAS,SAAS,YAClD,QAAQ,MAAM,KAAK,UAAU;GAC3B,MAAM,KAAK;GACX,OAAO,KAAK,MAAM,KAAK,QAAQ,IAAK;GACpC,KAAK,KAAK,MAAM,KAAK,MAAM,IAAK;GACjC,EAAE,CACJ;EAEiC;;AAGpC,MAAa,8BAA8B,OAAO,iBAAyB;CACzE,MAAM,OAAO;EACX;EACA;EACA;EACA;EACA;EACA;EACD;AACD,KAAI,+BAA+B,KAAK;CACxC,MAAM,EAAE,WAAW,MAAM,gBAAgB,uBAAuB,KAAK;AAErE,KAAI;EAEF,MAAM,cAAc,gCADA,KAAK,MAAM,OAAO,CAC0B;AAChE,SAAO,KAAK,UAAU,aAAa,MAAM,EAAE;UACpC,OAAO;AACd,MAAI,iCAAiC,QAAQ;AAC7C,QAAM,IAAI,MAAM,+CAA+C,QAAQ;;;AAI3E,MAAM,0BAA0BC,sCAAe;CAC7C,OAAO;CACP,WAAW,iBAAiB,2BAAY,aAAa,CAAC;CACtD,QAAQ;CACT,CAAC;AAEF,MAAa,uBAAuB,OAClC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,wBAAwB,WAAW,aAAa;UACtD,OAAO;AACd,UAAQ,MAAM,sCAAsC,MAAM;AAC1D,QAAM"}
@@ -1,11 +1,11 @@
1
1
  import { idempotentTask } from "../idempotentTask.js";
2
- import { exec } from "node:child_process";
3
- import { promisify } from "node:util";
2
+ import { execFile } from "node:child_process";
4
3
  import debug from "debug";
5
4
  import { basename } from "node:path";
5
+ import { promisify } from "node:util";
6
6
 
7
7
  //#region src/tasks/findOrCreateCaptions.ts
8
- const execPromise = promisify(exec);
8
+ const execFilePromise = promisify(execFile);
9
9
  const log = debug("ef:generateCaptions");
10
10
  const convertWhisperToEditframeFormat = (whisperData) => {
11
11
  return {
@@ -22,9 +22,16 @@ const convertWhisperToEditframeFormat = (whisperData) => {
22
22
  };
23
23
  };
24
24
  const generateCaptionDataFromPath = async (absolutePath) => {
25
- const command = `whisper_timestamped --language en --efficient --output_format json ${absolutePath}`;
26
- log(`Running command: ${command}`);
27
- const { stdout } = await execPromise(command);
25
+ const args = [
26
+ "--language",
27
+ "en",
28
+ "--efficient",
29
+ "--output_format",
30
+ "json",
31
+ absolutePath
32
+ ];
33
+ log("Running whisper_timestamped", args);
34
+ const { stdout } = await execFilePromise("whisper_timestamped", args);
28
35
  try {
29
36
  const captionData = convertWhisperToEditframeFormat(JSON.parse(stdout));
30
37
  return JSON.stringify(captionData, null, 2);
@@ -1 +1 @@
1
- {"version":3,"file":"findOrCreateCaptions.js","names":[],"sources":["../../src/tasks/findOrCreateCaptions.ts"],"sourcesContent":["import { basename } from \"node:path\";\nimport { promisify } from \"node:util\";\nimport { exec } from \"node:child_process\";\n\nimport debug from \"debug\";\n\nimport { idempotentTask } from \"../idempotentTask.js\";\n\nconst execPromise = promisify(exec);\n\nconst log = debug(\"ef:generateCaptions\");\n\ninterface WhisperWord {\n text: string;\n start: number;\n end: number;\n confidence: number;\n}\n\ninterface WhisperSegment {\n text: string;\n start: number;\n end: number;\n words: WhisperWord[];\n}\n\ninterface WhisperOutput {\n segments: WhisperSegment[];\n}\n\ninterface CaptionOutput {\n segments: Array<{\n start: number;\n end: number;\n text: string;\n }>;\n word_segments: Array<{\n text: string;\n start: number;\n end: number;\n }>;\n}\n\nconst convertWhisperToEditframeFormat = (\n whisperData: WhisperOutput,\n): CaptionOutput => {\n const segments = whisperData.segments.map((segment) => ({\n start: Math.round(segment.start * 1000), // Convert to milliseconds\n end: Math.round(segment.end * 1000),\n text: segment.text.trim(),\n }));\n\n const word_segments = whisperData.segments.flatMap((segment) =>\n segment.words.map((word) => ({\n text: word.text,\n start: Math.round(word.start * 1000), // Convert to milliseconds\n end: Math.round(word.end * 1000),\n })),\n );\n\n return { segments, word_segments };\n};\n\nexport const generateCaptionDataFromPath = async (absolutePath: string) => {\n const command = `whisper_timestamped --language en --efficient --output_format json ${absolutePath}`;\n log(`Running command: ${command}`);\n const { stdout } = await execPromise(command);\n\n try {\n const whisperData = JSON.parse(stdout) as WhisperOutput;\n const captionData = convertWhisperToEditframeFormat(whisperData);\n return JSON.stringify(captionData, null, 2);\n } catch (error) {\n log(`Error parsing whisper output: ${error}`);\n throw new Error(`Failed to parse whisper_timestamped output: ${error}`);\n }\n};\n\nconst generateCaptionDataTask = idempotentTask({\n label: \"captions\",\n filename: (absolutePath) => `${basename(absolutePath)}.captions.json`,\n runner: generateCaptionDataFromPath,\n});\n\nexport const findOrCreateCaptions = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateCaptionDataTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error finding or creating captions\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;AAQA,MAAM,cAAc,UAAU,KAAK;AAEnC,MAAM,MAAM,MAAM,sBAAsB;AAiCxC,MAAM,mCACJ,gBACkB;AAelB,QAAO;EAAE,UAdQ,YAAY,SAAS,KAAK,aAAa;GACtD,OAAO,KAAK,MAAM,QAAQ,QAAQ,IAAK;GACvC,KAAK,KAAK,MAAM,QAAQ,MAAM,IAAK;GACnC,MAAM,QAAQ,KAAK,MAAM;GAC1B,EAAE;EAUgB,eARG,YAAY,SAAS,SAAS,YAClD,QAAQ,MAAM,KAAK,UAAU;GAC3B,MAAM,KAAK;GACX,OAAO,KAAK,MAAM,KAAK,QAAQ,IAAK;GACpC,KAAK,KAAK,MAAM,KAAK,MAAM,IAAK;GACjC,EAAE,CACJ;EAEiC;;AAGpC,MAAa,8BAA8B,OAAO,iBAAyB;CACzE,MAAM,UAAU,sEAAsE;AACtF,KAAI,oBAAoB,UAAU;CAClC,MAAM,EAAE,WAAW,MAAM,YAAY,QAAQ;AAE7C,KAAI;EAEF,MAAM,cAAc,gCADA,KAAK,MAAM,OAAO,CAC0B;AAChE,SAAO,KAAK,UAAU,aAAa,MAAM,EAAE;UACpC,OAAO;AACd,MAAI,iCAAiC,QAAQ;AAC7C,QAAM,IAAI,MAAM,+CAA+C,QAAQ;;;AAI3E,MAAM,0BAA0B,eAAe;CAC7C,OAAO;CACP,WAAW,iBAAiB,GAAG,SAAS,aAAa,CAAC;CACtD,QAAQ;CACT,CAAC;AAEF,MAAa,uBAAuB,OAClC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,wBAAwB,WAAW,aAAa;UACtD,OAAO;AACd,UAAQ,MAAM,sCAAsC,MAAM;AAC1D,QAAM"}
1
+ {"version":3,"file":"findOrCreateCaptions.js","names":[],"sources":["../../src/tasks/findOrCreateCaptions.ts"],"sourcesContent":["import { basename } from \"node:path\";\nimport { promisify } from \"node:util\";\nimport { execFile } from \"node:child_process\";\n\nimport debug from \"debug\";\n\nimport { idempotentTask } from \"../idempotentTask.js\";\n\nconst execFilePromise = promisify(execFile);\n\nconst log = debug(\"ef:generateCaptions\");\n\ninterface WhisperWord {\n text: string;\n start: number;\n end: number;\n confidence: number;\n}\n\ninterface WhisperSegment {\n text: string;\n start: number;\n end: number;\n words: WhisperWord[];\n}\n\ninterface WhisperOutput {\n segments: WhisperSegment[];\n}\n\ninterface CaptionOutput {\n segments: Array<{\n start: number;\n end: number;\n text: string;\n }>;\n word_segments: Array<{\n text: string;\n start: number;\n end: number;\n }>;\n}\n\nconst convertWhisperToEditframeFormat = (\n whisperData: WhisperOutput,\n): CaptionOutput => {\n const segments = whisperData.segments.map((segment) => ({\n start: Math.round(segment.start * 1000), // Convert to milliseconds\n end: Math.round(segment.end * 1000),\n text: segment.text.trim(),\n }));\n\n const word_segments = whisperData.segments.flatMap((segment) =>\n segment.words.map((word) => ({\n text: word.text,\n start: Math.round(word.start * 1000), // Convert to milliseconds\n end: Math.round(word.end * 1000),\n })),\n );\n\n return { segments, word_segments };\n};\n\nexport const generateCaptionDataFromPath = async (absolutePath: string) => {\n const args = [\n \"--language\",\n \"en\",\n \"--efficient\",\n \"--output_format\",\n \"json\",\n absolutePath,\n ];\n log(\"Running whisper_timestamped\", args);\n const { stdout } = await execFilePromise(\"whisper_timestamped\", args);\n\n try {\n const whisperData = JSON.parse(stdout) as WhisperOutput;\n const captionData = convertWhisperToEditframeFormat(whisperData);\n return JSON.stringify(captionData, null, 2);\n } catch (error) {\n log(`Error parsing whisper output: ${error}`);\n throw new Error(`Failed to parse whisper_timestamped output: ${error}`);\n }\n};\n\nconst generateCaptionDataTask = idempotentTask({\n label: \"captions\",\n filename: (absolutePath) => `${basename(absolutePath)}.captions.json`,\n runner: generateCaptionDataFromPath,\n});\n\nexport const findOrCreateCaptions = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateCaptionDataTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error finding or creating captions\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;AAQA,MAAM,kBAAkB,UAAU,SAAS;AAE3C,MAAM,MAAM,MAAM,sBAAsB;AAiCxC,MAAM,mCACJ,gBACkB;AAelB,QAAO;EAAE,UAdQ,YAAY,SAAS,KAAK,aAAa;GACtD,OAAO,KAAK,MAAM,QAAQ,QAAQ,IAAK;GACvC,KAAK,KAAK,MAAM,QAAQ,MAAM,IAAK;GACnC,MAAM,QAAQ,KAAK,MAAM;GAC1B,EAAE;EAUgB,eARG,YAAY,SAAS,SAAS,YAClD,QAAQ,MAAM,KAAK,UAAU;GAC3B,MAAM,KAAK;GACX,OAAO,KAAK,MAAM,KAAK,QAAQ,IAAK;GACpC,KAAK,KAAK,MAAM,KAAK,MAAM,IAAK;GACjC,EAAE,CACJ;EAEiC;;AAGpC,MAAa,8BAA8B,OAAO,iBAAyB;CACzE,MAAM,OAAO;EACX;EACA;EACA;EACA;EACA;EACA;EACD;AACD,KAAI,+BAA+B,KAAK;CACxC,MAAM,EAAE,WAAW,MAAM,gBAAgB,uBAAuB,KAAK;AAErE,KAAI;EAEF,MAAM,cAAc,gCADA,KAAK,MAAM,OAAO,CAC0B;AAChE,SAAO,KAAK,UAAU,aAAa,MAAM,EAAE;UACpC,OAAO;AACd,MAAI,iCAAiC,QAAQ;AAC7C,QAAM,IAAI,MAAM,+CAA+C,QAAQ;;;AAI3E,MAAM,0BAA0B,eAAe;CAC7C,OAAO;CACP,WAAW,iBAAiB,GAAG,SAAS,aAAa,CAAC;CACtD,QAAQ;CACT,CAAC;AAEF,MAAa,uBAAuB,OAClC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,wBAAwB,WAAW,aAAa;UACtD,OAAO;AACd,UAAQ,MAAM,sCAAsC,MAAM;AAC1D,QAAM"}
@@ -16,17 +16,7 @@ const generateScrubTrackFromPath = async (absolutePath) => {
16
16
  const probe = await require_Probe.Probe.probePath(absolutePath);
17
17
  if (probe.videoStreams.length === 0) throw new Error("No video stream found for scrub track generation");
18
18
  const scrubStream = probe.createScrubTrackReadstream();
19
- let startTimeOffsetMs;
20
- if (probe.format.start_time && Number(probe.format.start_time) !== 0) {
21
- startTimeOffsetMs = Number(probe.format.start_time) * 1e3;
22
- log(`Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`);
23
- } else {
24
- const videoStream = probe.videoStreams[0];
25
- if (videoStream && videoStream.start_time && Number(videoStream.start_time) !== 0) {
26
- startTimeOffsetMs = Number(videoStream.start_time) * 1e3;
27
- log(`Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`);
28
- }
29
- }
19
+ const startTimeOffsetMs = probe.startTimeOffsetMs;
30
20
  const outputStream = new node_stream.PassThrough();
31
21
  const indexStream = new node_stream.PassThrough();
32
22
  scrubStream.pipe(outputStream, { end: false });
@@ -1 +1 @@
1
- {"version":3,"file":"generateScrubTrack.cjs","names":["Probe","startTimeOffsetMs: number | undefined","PassThrough","generateFragmentIndex","idempotentTask","progressTimeout: NodeJS.Timeout | null"],"sources":["../../src/tasks/generateScrubTrack.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { PassThrough } from \"node:stream\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\n\nconst log = debug(\"ef:generateScrubTrack\");\n\nexport const generateScrubTrackFromPath = async (absolutePath: string) => {\n log(`Generating scrub track for ${absolutePath}`);\n\n const probe = await Probe.probePath(absolutePath);\n\n // Check if video stream exists\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg (low-res transcoded video)\n const scrubStream = probe.createScrubTrackReadstream();\n\n // Extract timing offset from probe metadata (same logic as generateTrackFragmentIndex)\n let startTimeOffsetMs: number | undefined;\n\n // First check format-level start_time\n if (probe.format.start_time && Number(probe.format.start_time) !== 0) {\n startTimeOffsetMs = Number(probe.format.start_time) * 1000;\n log(\n `Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`,\n );\n } else {\n // Check for video stream start_time (more common)\n const videoStream = probe.videoStreams[0];\n if (\n videoStream &&\n videoStream.start_time &&\n Number(videoStream.start_time) !== 0\n ) {\n startTimeOffsetMs = Number(videoStream.start_time) * 1000;\n log(\n `Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`,\n );\n }\n }\n\n // Create a PassThrough to tee the stream\n const outputStream = new PassThrough();\n const indexStream = new PassThrough();\n\n // Pipe data but DON'T end outputStream automatically - we'll control this\n scrubStream.pipe(outputStream, { end: false });\n scrubStream.pipe(indexStream);\n\n // Track when the source stream ends (but don't end output yet)\n let sourceStreamEnded = false;\n scrubStream.on(\"end\", () => {\n sourceStreamEnded = true;\n });\n\n scrubStream.on(\"error\", (error) => {\n outputStream.destroy(error);\n indexStream.destroy(error);\n });\n\n // Generate fragment index from the scrub track stream\n // Use a special track ID to identify scrub track (e.g., -1 or \"scrub\")\n // We'll use a negative track ID to distinguish from regular tracks\n const scrubTrackId = -1;\n const trackIdMapping = { 1: scrubTrackId }; // Single track 1 -> scrub track ID\n\n const fragmentIndexPromise = generateFragmentIndex(\n indexStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n // End outputStream only after BOTH source ends AND fragment index completes\n fragmentIndexPromise\n .then(() => {\n if (sourceStreamEnded) {\n outputStream.end();\n } else {\n // If fragment index completes first, wait for stream to end\n scrubStream.once(\"end\", () => {\n outputStream.end();\n });\n }\n })\n .catch((error) => {\n outputStream.destroy(error);\n });\n\n // Return both the stream and the index\n return {\n stream: outputStream,\n fragmentIndex: fragmentIndexPromise,\n };\n};\n\nexport const generateScrubTrackTask = idempotentTask({\n label: \"scrub-track\",\n filename: (absolutePath: string) =>\n `${basename(absolutePath)}.scrub-track.mp4`,\n runner: async (absolutePath: string) => {\n const probe = await Probe.probePath(absolutePath);\n\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg\n const scrubStream = probe.createScrubTrackReadstream();\n\n // Wrap in PassThrough with timeout handling to ensure stream completes\n const finalStream = new PassThrough();\n\n // Monitor progress and extend timeout based on actual work\n let progressTimeout: NodeJS.Timeout | null = null;\n\n const resetProgressTimeout = () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n\n progressTimeout = setTimeout(() => {\n if (!finalStream.destroyed) {\n console.warn(\n `Progress timeout triggered for scrub track - no activity for 30 seconds`,\n );\n finalStream.destroy(new Error(\"Scrub track generation timeout\"));\n }\n }, 30000); // 30 second sliding timeout (longer for transcoding)\n };\n\n // Start the initial timeout\n resetProgressTimeout();\n\n // Monitor data flow to detect active work\n scrubStream.on(\"data\", () => {\n resetProgressTimeout(); // Reset timeout when we see data\n });\n\n scrubStream.on(\"end\", () => {\n resetProgressTimeout(); // Reset timeout when stream ends\n finalStream.end();\n });\n\n scrubStream.on(\"error\", (error) => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n finalStream.destroy(error);\n });\n\n // Pipe data through\n scrubStream.pipe(finalStream, { end: false });\n\n // Clean up timeout when stream ends\n finalStream.on(\"end\", () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n });\n\n return finalStream;\n },\n});\n\nexport const generateScrubTrack = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateScrubTrackTask(cacheRoot, absolutePath);\n } catch (error) {\n console.error(error);\n console.trace(\"Error generating scrub track\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;;;;;AAOA,MAAM,yBAAY,wBAAwB;AAE1C,MAAa,6BAA6B,OAAO,iBAAyB;AACxE,KAAI,8BAA8B,eAAe;CAEjD,MAAM,QAAQ,MAAMA,oBAAM,UAAU,aAAa;AAGjD,KAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;CAIrE,MAAM,cAAc,MAAM,4BAA4B;CAGtD,IAAIC;AAGJ,KAAI,MAAM,OAAO,cAAc,OAAO,MAAM,OAAO,WAAW,KAAK,GAAG;AACpE,sBAAoB,OAAO,MAAM,OAAO,WAAW,GAAG;AACtD,MACE,uCAAuC,MAAM,OAAO,WAAW,KAAK,kBAAkB,KACvF;QACI;EAEL,MAAM,cAAc,MAAM,aAAa;AACvC,MACE,eACA,YAAY,cACZ,OAAO,YAAY,WAAW,KAAK,GACnC;AACA,uBAAoB,OAAO,YAAY,WAAW,GAAG;AACrD,OACE,6CAA6C,YAAY,WAAW,KAAK,kBAAkB,KAC5F;;;CAKL,MAAM,eAAe,IAAIC,yBAAa;CACtC,MAAM,cAAc,IAAIA,yBAAa;AAGrC,aAAY,KAAK,cAAc,EAAE,KAAK,OAAO,CAAC;AAC9C,aAAY,KAAK,YAAY;CAG7B,IAAI,oBAAoB;AACxB,aAAY,GAAG,aAAa;AAC1B,sBAAoB;GACpB;AAEF,aAAY,GAAG,UAAU,UAAU;AACjC,eAAa,QAAQ,MAAM;AAC3B,cAAY,QAAQ,MAAM;GAC1B;CAKF,MAAM,eAAe;CAGrB,MAAM,uBAAuBC,oDAC3B,aACA,mBAJqB,EAAE,GAAG,cAAc,CAMzC;AAGD,sBACG,WAAW;AACV,MAAI,kBACF,cAAa,KAAK;MAGlB,aAAY,KAAK,aAAa;AAC5B,gBAAa,KAAK;IAClB;GAEJ,CACD,OAAO,UAAU;AAChB,eAAa,QAAQ,MAAM;GAC3B;AAGJ,QAAO;EACL,QAAQ;EACR,eAAe;EAChB;;AAGH,MAAa,yBAAyBC,sCAAe;CACnD,OAAO;CACP,WAAW,iBACT,2BAAY,aAAa,CAAC;CAC5B,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAMJ,oBAAM,UAAU,aAAa;AAEjD,MAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;EAIrE,MAAM,cAAc,MAAM,4BAA4B;EAGtD,MAAM,cAAc,IAAIE,yBAAa;EAGrC,IAAIG,kBAAyC;EAE7C,MAAM,6BAA6B;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAG/B,qBAAkB,iBAAiB;AACjC,QAAI,CAAC,YAAY,WAAW;AAC1B,aAAQ,KACN,0EACD;AACD,iBAAY,wBAAQ,IAAI,MAAM,iCAAiC,CAAC;;MAEjE,IAAM;;AAIX,wBAAsB;AAGtB,cAAY,GAAG,cAAc;AAC3B,yBAAsB;IACtB;AAEF,cAAY,GAAG,aAAa;AAC1B,yBAAsB;AACtB,eAAY,KAAK;IACjB;AAEF,cAAY,GAAG,UAAU,UAAU;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAE/B,eAAY,QAAQ,MAAM;IAC1B;AAGF,cAAY,KAAK,aAAa,EAAE,KAAK,OAAO,CAAC;AAG7C,cAAY,GAAG,aAAa;AAC1B,OAAI,gBACF,cAAa,gBAAgB;IAE/B;AAEF,SAAO;;CAEV,CAAC;AAEF,MAAa,qBAAqB,OAChC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,uBAAuB,WAAW,aAAa;UACrD,OAAO;AACd,UAAQ,MAAM,MAAM;AACpB,UAAQ,MAAM,gCAAgC,MAAM;AACpD,QAAM"}
1
+ {"version":3,"file":"generateScrubTrack.cjs","names":["Probe","PassThrough","generateFragmentIndex","idempotentTask","progressTimeout: NodeJS.Timeout | null"],"sources":["../../src/tasks/generateScrubTrack.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { PassThrough } from \"node:stream\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\n\nconst log = debug(\"ef:generateScrubTrack\");\n\nexport const generateScrubTrackFromPath = async (absolutePath: string) => {\n log(`Generating scrub track for ${absolutePath}`);\n\n const probe = await Probe.probePath(absolutePath);\n\n // Check if video stream exists\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg (low-res transcoded video)\n const scrubStream = probe.createScrubTrackReadstream();\n\n const startTimeOffsetMs = probe.startTimeOffsetMs;\n\n // Create a PassThrough to tee the stream\n const outputStream = new PassThrough();\n const indexStream = new PassThrough();\n\n // Pipe data but DON'T end outputStream automatically - we'll control this\n scrubStream.pipe(outputStream, { end: false });\n scrubStream.pipe(indexStream);\n\n // Track when the source stream ends (but don't end output yet)\n let sourceStreamEnded = false;\n scrubStream.on(\"end\", () => {\n sourceStreamEnded = true;\n });\n\n scrubStream.on(\"error\", (error) => {\n outputStream.destroy(error);\n indexStream.destroy(error);\n });\n\n // Generate fragment index from the scrub track stream\n // Use a special track ID to identify scrub track (e.g., -1 or \"scrub\")\n // We'll use a negative track ID to distinguish from regular tracks\n const scrubTrackId = -1;\n const trackIdMapping = { 1: scrubTrackId }; // Single track 1 -> scrub track ID\n\n const fragmentIndexPromise = generateFragmentIndex(\n indexStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n // End outputStream only after BOTH source ends AND fragment index completes\n fragmentIndexPromise\n .then(() => {\n if (sourceStreamEnded) {\n outputStream.end();\n } else {\n // If fragment index completes first, wait for stream to end\n scrubStream.once(\"end\", () => {\n outputStream.end();\n });\n }\n })\n .catch((error) => {\n outputStream.destroy(error);\n });\n\n // Return both the stream and the index\n return {\n stream: outputStream,\n fragmentIndex: fragmentIndexPromise,\n };\n};\n\nexport const generateScrubTrackTask = idempotentTask({\n label: \"scrub-track\",\n filename: (absolutePath: string) =>\n `${basename(absolutePath)}.scrub-track.mp4`,\n runner: async (absolutePath: string) => {\n const probe = await Probe.probePath(absolutePath);\n\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg\n const scrubStream = probe.createScrubTrackReadstream();\n\n // Wrap in PassThrough with timeout handling to ensure stream completes\n const finalStream = new PassThrough();\n\n // Monitor progress and extend timeout based on actual work\n let progressTimeout: NodeJS.Timeout | null = null;\n\n const resetProgressTimeout = () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n\n progressTimeout = setTimeout(() => {\n if (!finalStream.destroyed) {\n console.warn(\n `Progress timeout triggered for scrub track - no activity for 30 seconds`,\n );\n finalStream.destroy(new Error(\"Scrub track generation timeout\"));\n }\n }, 30000); // 30 second sliding timeout (longer for transcoding)\n };\n\n // Start the initial timeout\n resetProgressTimeout();\n\n // Monitor data flow to detect active work\n scrubStream.on(\"data\", () => {\n resetProgressTimeout(); // Reset timeout when we see data\n });\n\n scrubStream.on(\"end\", () => {\n resetProgressTimeout(); // Reset timeout when stream ends\n finalStream.end();\n });\n\n scrubStream.on(\"error\", (error) => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n finalStream.destroy(error);\n });\n\n // Pipe data through\n scrubStream.pipe(finalStream, { end: false });\n\n // Clean up timeout when stream ends\n finalStream.on(\"end\", () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n });\n\n return finalStream;\n },\n});\n\nexport const generateScrubTrack = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateScrubTrackTask(cacheRoot, absolutePath);\n } catch (error) {\n console.error(error);\n console.trace(\"Error generating scrub track\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;;;;;AAOA,MAAM,yBAAY,wBAAwB;AAE1C,MAAa,6BAA6B,OAAO,iBAAyB;AACxE,KAAI,8BAA8B,eAAe;CAEjD,MAAM,QAAQ,MAAMA,oBAAM,UAAU,aAAa;AAGjD,KAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;CAIrE,MAAM,cAAc,MAAM,4BAA4B;CAEtD,MAAM,oBAAoB,MAAM;CAGhC,MAAM,eAAe,IAAIC,yBAAa;CACtC,MAAM,cAAc,IAAIA,yBAAa;AAGrC,aAAY,KAAK,cAAc,EAAE,KAAK,OAAO,CAAC;AAC9C,aAAY,KAAK,YAAY;CAG7B,IAAI,oBAAoB;AACxB,aAAY,GAAG,aAAa;AAC1B,sBAAoB;GACpB;AAEF,aAAY,GAAG,UAAU,UAAU;AACjC,eAAa,QAAQ,MAAM;AAC3B,cAAY,QAAQ,MAAM;GAC1B;CAKF,MAAM,eAAe;CAGrB,MAAM,uBAAuBC,oDAC3B,aACA,mBAJqB,EAAE,GAAG,cAAc,CAMzC;AAGD,sBACG,WAAW;AACV,MAAI,kBACF,cAAa,KAAK;MAGlB,aAAY,KAAK,aAAa;AAC5B,gBAAa,KAAK;IAClB;GAEJ,CACD,OAAO,UAAU;AAChB,eAAa,QAAQ,MAAM;GAC3B;AAGJ,QAAO;EACL,QAAQ;EACR,eAAe;EAChB;;AAGH,MAAa,yBAAyBC,sCAAe;CACnD,OAAO;CACP,WAAW,iBACT,2BAAY,aAAa,CAAC;CAC5B,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAMH,oBAAM,UAAU,aAAa;AAEjD,MAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;EAIrE,MAAM,cAAc,MAAM,4BAA4B;EAGtD,MAAM,cAAc,IAAIC,yBAAa;EAGrC,IAAIG,kBAAyC;EAE7C,MAAM,6BAA6B;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAG/B,qBAAkB,iBAAiB;AACjC,QAAI,CAAC,YAAY,WAAW;AAC1B,aAAQ,KACN,0EACD;AACD,iBAAY,wBAAQ,IAAI,MAAM,iCAAiC,CAAC;;MAEjE,IAAM;;AAIX,wBAAsB;AAGtB,cAAY,GAAG,cAAc;AAC3B,yBAAsB;IACtB;AAEF,cAAY,GAAG,aAAa;AAC1B,yBAAsB;AACtB,eAAY,KAAK;IACjB;AAEF,cAAY,GAAG,UAAU,UAAU;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAE/B,eAAY,QAAQ,MAAM;IAC1B;AAGF,cAAY,KAAK,aAAa,EAAE,KAAK,OAAO,CAAC;AAG7C,cAAY,GAAG,aAAa;AAC1B,OAAI,gBACF,cAAa,gBAAgB;IAE/B;AAEF,SAAO;;CAEV,CAAC;AAEF,MAAa,qBAAqB,OAChC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,uBAAuB,WAAW,aAAa;UACrD,OAAO;AACd,UAAQ,MAAM,MAAM;AACpB,UAAQ,MAAM,gCAAgC,MAAM;AACpD,QAAM"}
@@ -12,17 +12,7 @@ const generateScrubTrackFromPath = async (absolutePath) => {
12
12
  const probe = await Probe.probePath(absolutePath);
13
13
  if (probe.videoStreams.length === 0) throw new Error("No video stream found for scrub track generation");
14
14
  const scrubStream = probe.createScrubTrackReadstream();
15
- let startTimeOffsetMs;
16
- if (probe.format.start_time && Number(probe.format.start_time) !== 0) {
17
- startTimeOffsetMs = Number(probe.format.start_time) * 1e3;
18
- log(`Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`);
19
- } else {
20
- const videoStream = probe.videoStreams[0];
21
- if (videoStream && videoStream.start_time && Number(videoStream.start_time) !== 0) {
22
- startTimeOffsetMs = Number(videoStream.start_time) * 1e3;
23
- log(`Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`);
24
- }
25
- }
15
+ const startTimeOffsetMs = probe.startTimeOffsetMs;
26
16
  const outputStream = new PassThrough();
27
17
  const indexStream = new PassThrough();
28
18
  scrubStream.pipe(outputStream, { end: false });
@@ -1 +1 @@
1
- {"version":3,"file":"generateScrubTrack.js","names":["startTimeOffsetMs: number | undefined","progressTimeout: NodeJS.Timeout | null"],"sources":["../../src/tasks/generateScrubTrack.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { PassThrough } from \"node:stream\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\n\nconst log = debug(\"ef:generateScrubTrack\");\n\nexport const generateScrubTrackFromPath = async (absolutePath: string) => {\n log(`Generating scrub track for ${absolutePath}`);\n\n const probe = await Probe.probePath(absolutePath);\n\n // Check if video stream exists\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg (low-res transcoded video)\n const scrubStream = probe.createScrubTrackReadstream();\n\n // Extract timing offset from probe metadata (same logic as generateTrackFragmentIndex)\n let startTimeOffsetMs: number | undefined;\n\n // First check format-level start_time\n if (probe.format.start_time && Number(probe.format.start_time) !== 0) {\n startTimeOffsetMs = Number(probe.format.start_time) * 1000;\n log(\n `Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`,\n );\n } else {\n // Check for video stream start_time (more common)\n const videoStream = probe.videoStreams[0];\n if (\n videoStream &&\n videoStream.start_time &&\n Number(videoStream.start_time) !== 0\n ) {\n startTimeOffsetMs = Number(videoStream.start_time) * 1000;\n log(\n `Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`,\n );\n }\n }\n\n // Create a PassThrough to tee the stream\n const outputStream = new PassThrough();\n const indexStream = new PassThrough();\n\n // Pipe data but DON'T end outputStream automatically - we'll control this\n scrubStream.pipe(outputStream, { end: false });\n scrubStream.pipe(indexStream);\n\n // Track when the source stream ends (but don't end output yet)\n let sourceStreamEnded = false;\n scrubStream.on(\"end\", () => {\n sourceStreamEnded = true;\n });\n\n scrubStream.on(\"error\", (error) => {\n outputStream.destroy(error);\n indexStream.destroy(error);\n });\n\n // Generate fragment index from the scrub track stream\n // Use a special track ID to identify scrub track (e.g., -1 or \"scrub\")\n // We'll use a negative track ID to distinguish from regular tracks\n const scrubTrackId = -1;\n const trackIdMapping = { 1: scrubTrackId }; // Single track 1 -> scrub track ID\n\n const fragmentIndexPromise = generateFragmentIndex(\n indexStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n // End outputStream only after BOTH source ends AND fragment index completes\n fragmentIndexPromise\n .then(() => {\n if (sourceStreamEnded) {\n outputStream.end();\n } else {\n // If fragment index completes first, wait for stream to end\n scrubStream.once(\"end\", () => {\n outputStream.end();\n });\n }\n })\n .catch((error) => {\n outputStream.destroy(error);\n });\n\n // Return both the stream and the index\n return {\n stream: outputStream,\n fragmentIndex: fragmentIndexPromise,\n };\n};\n\nexport const generateScrubTrackTask = idempotentTask({\n label: \"scrub-track\",\n filename: (absolutePath: string) =>\n `${basename(absolutePath)}.scrub-track.mp4`,\n runner: async (absolutePath: string) => {\n const probe = await Probe.probePath(absolutePath);\n\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg\n const scrubStream = probe.createScrubTrackReadstream();\n\n // Wrap in PassThrough with timeout handling to ensure stream completes\n const finalStream = new PassThrough();\n\n // Monitor progress and extend timeout based on actual work\n let progressTimeout: NodeJS.Timeout | null = null;\n\n const resetProgressTimeout = () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n\n progressTimeout = setTimeout(() => {\n if (!finalStream.destroyed) {\n console.warn(\n `Progress timeout triggered for scrub track - no activity for 30 seconds`,\n );\n finalStream.destroy(new Error(\"Scrub track generation timeout\"));\n }\n }, 30000); // 30 second sliding timeout (longer for transcoding)\n };\n\n // Start the initial timeout\n resetProgressTimeout();\n\n // Monitor data flow to detect active work\n scrubStream.on(\"data\", () => {\n resetProgressTimeout(); // Reset timeout when we see data\n });\n\n scrubStream.on(\"end\", () => {\n resetProgressTimeout(); // Reset timeout when stream ends\n finalStream.end();\n });\n\n scrubStream.on(\"error\", (error) => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n finalStream.destroy(error);\n });\n\n // Pipe data through\n scrubStream.pipe(finalStream, { end: false });\n\n // Clean up timeout when stream ends\n finalStream.on(\"end\", () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n });\n\n return finalStream;\n },\n});\n\nexport const generateScrubTrack = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateScrubTrackTask(cacheRoot, absolutePath);\n } catch (error) {\n console.error(error);\n console.trace(\"Error generating scrub track\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;AAOA,MAAM,MAAM,MAAM,wBAAwB;AAE1C,MAAa,6BAA6B,OAAO,iBAAyB;AACxE,KAAI,8BAA8B,eAAe;CAEjD,MAAM,QAAQ,MAAM,MAAM,UAAU,aAAa;AAGjD,KAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;CAIrE,MAAM,cAAc,MAAM,4BAA4B;CAGtD,IAAIA;AAGJ,KAAI,MAAM,OAAO,cAAc,OAAO,MAAM,OAAO,WAAW,KAAK,GAAG;AACpE,sBAAoB,OAAO,MAAM,OAAO,WAAW,GAAG;AACtD,MACE,uCAAuC,MAAM,OAAO,WAAW,KAAK,kBAAkB,KACvF;QACI;EAEL,MAAM,cAAc,MAAM,aAAa;AACvC,MACE,eACA,YAAY,cACZ,OAAO,YAAY,WAAW,KAAK,GACnC;AACA,uBAAoB,OAAO,YAAY,WAAW,GAAG;AACrD,OACE,6CAA6C,YAAY,WAAW,KAAK,kBAAkB,KAC5F;;;CAKL,MAAM,eAAe,IAAI,aAAa;CACtC,MAAM,cAAc,IAAI,aAAa;AAGrC,aAAY,KAAK,cAAc,EAAE,KAAK,OAAO,CAAC;AAC9C,aAAY,KAAK,YAAY;CAG7B,IAAI,oBAAoB;AACxB,aAAY,GAAG,aAAa;AAC1B,sBAAoB;GACpB;AAEF,aAAY,GAAG,UAAU,UAAU;AACjC,eAAa,QAAQ,MAAM;AAC3B,cAAY,QAAQ,MAAM;GAC1B;CAKF,MAAM,eAAe;CAGrB,MAAM,uBAAuB,sBAC3B,aACA,mBAJqB,EAAE,GAAG,cAAc,CAMzC;AAGD,sBACG,WAAW;AACV,MAAI,kBACF,cAAa,KAAK;MAGlB,aAAY,KAAK,aAAa;AAC5B,gBAAa,KAAK;IAClB;GAEJ,CACD,OAAO,UAAU;AAChB,eAAa,QAAQ,MAAM;GAC3B;AAGJ,QAAO;EACL,QAAQ;EACR,eAAe;EAChB;;AAGH,MAAa,yBAAyB,eAAe;CACnD,OAAO;CACP,WAAW,iBACT,GAAG,SAAS,aAAa,CAAC;CAC5B,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAM,MAAM,UAAU,aAAa;AAEjD,MAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;EAIrE,MAAM,cAAc,MAAM,4BAA4B;EAGtD,MAAM,cAAc,IAAI,aAAa;EAGrC,IAAIC,kBAAyC;EAE7C,MAAM,6BAA6B;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAG/B,qBAAkB,iBAAiB;AACjC,QAAI,CAAC,YAAY,WAAW;AAC1B,aAAQ,KACN,0EACD;AACD,iBAAY,wBAAQ,IAAI,MAAM,iCAAiC,CAAC;;MAEjE,IAAM;;AAIX,wBAAsB;AAGtB,cAAY,GAAG,cAAc;AAC3B,yBAAsB;IACtB;AAEF,cAAY,GAAG,aAAa;AAC1B,yBAAsB;AACtB,eAAY,KAAK;IACjB;AAEF,cAAY,GAAG,UAAU,UAAU;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAE/B,eAAY,QAAQ,MAAM;IAC1B;AAGF,cAAY,KAAK,aAAa,EAAE,KAAK,OAAO,CAAC;AAG7C,cAAY,GAAG,aAAa;AAC1B,OAAI,gBACF,cAAa,gBAAgB;IAE/B;AAEF,SAAO;;CAEV,CAAC;AAEF,MAAa,qBAAqB,OAChC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,uBAAuB,WAAW,aAAa;UACrD,OAAO;AACd,UAAQ,MAAM,MAAM;AACpB,UAAQ,MAAM,gCAAgC,MAAM;AACpD,QAAM"}
1
+ {"version":3,"file":"generateScrubTrack.js","names":["progressTimeout: NodeJS.Timeout | null"],"sources":["../../src/tasks/generateScrubTrack.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { PassThrough } from \"node:stream\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\n\nconst log = debug(\"ef:generateScrubTrack\");\n\nexport const generateScrubTrackFromPath = async (absolutePath: string) => {\n log(`Generating scrub track for ${absolutePath}`);\n\n const probe = await Probe.probePath(absolutePath);\n\n // Check if video stream exists\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg (low-res transcoded video)\n const scrubStream = probe.createScrubTrackReadstream();\n\n const startTimeOffsetMs = probe.startTimeOffsetMs;\n\n // Create a PassThrough to tee the stream\n const outputStream = new PassThrough();\n const indexStream = new PassThrough();\n\n // Pipe data but DON'T end outputStream automatically - we'll control this\n scrubStream.pipe(outputStream, { end: false });\n scrubStream.pipe(indexStream);\n\n // Track when the source stream ends (but don't end output yet)\n let sourceStreamEnded = false;\n scrubStream.on(\"end\", () => {\n sourceStreamEnded = true;\n });\n\n scrubStream.on(\"error\", (error) => {\n outputStream.destroy(error);\n indexStream.destroy(error);\n });\n\n // Generate fragment index from the scrub track stream\n // Use a special track ID to identify scrub track (e.g., -1 or \"scrub\")\n // We'll use a negative track ID to distinguish from regular tracks\n const scrubTrackId = -1;\n const trackIdMapping = { 1: scrubTrackId }; // Single track 1 -> scrub track ID\n\n const fragmentIndexPromise = generateFragmentIndex(\n indexStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n // End outputStream only after BOTH source ends AND fragment index completes\n fragmentIndexPromise\n .then(() => {\n if (sourceStreamEnded) {\n outputStream.end();\n } else {\n // If fragment index completes first, wait for stream to end\n scrubStream.once(\"end\", () => {\n outputStream.end();\n });\n }\n })\n .catch((error) => {\n outputStream.destroy(error);\n });\n\n // Return both the stream and the index\n return {\n stream: outputStream,\n fragmentIndex: fragmentIndexPromise,\n };\n};\n\nexport const generateScrubTrackTask = idempotentTask({\n label: \"scrub-track\",\n filename: (absolutePath: string) =>\n `${basename(absolutePath)}.scrub-track.mp4`,\n runner: async (absolutePath: string) => {\n const probe = await Probe.probePath(absolutePath);\n\n if (probe.videoStreams.length === 0) {\n throw new Error(\"No video stream found for scrub track generation\");\n }\n\n // Get the scrub track stream from FFmpeg\n const scrubStream = probe.createScrubTrackReadstream();\n\n // Wrap in PassThrough with timeout handling to ensure stream completes\n const finalStream = new PassThrough();\n\n // Monitor progress and extend timeout based on actual work\n let progressTimeout: NodeJS.Timeout | null = null;\n\n const resetProgressTimeout = () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n\n progressTimeout = setTimeout(() => {\n if (!finalStream.destroyed) {\n console.warn(\n `Progress timeout triggered for scrub track - no activity for 30 seconds`,\n );\n finalStream.destroy(new Error(\"Scrub track generation timeout\"));\n }\n }, 30000); // 30 second sliding timeout (longer for transcoding)\n };\n\n // Start the initial timeout\n resetProgressTimeout();\n\n // Monitor data flow to detect active work\n scrubStream.on(\"data\", () => {\n resetProgressTimeout(); // Reset timeout when we see data\n });\n\n scrubStream.on(\"end\", () => {\n resetProgressTimeout(); // Reset timeout when stream ends\n finalStream.end();\n });\n\n scrubStream.on(\"error\", (error) => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n finalStream.destroy(error);\n });\n\n // Pipe data through\n scrubStream.pipe(finalStream, { end: false });\n\n // Clean up timeout when stream ends\n finalStream.on(\"end\", () => {\n if (progressTimeout) {\n clearTimeout(progressTimeout);\n }\n });\n\n return finalStream;\n },\n});\n\nexport const generateScrubTrack = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateScrubTrackTask(cacheRoot, absolutePath);\n } catch (error) {\n console.error(error);\n console.trace(\"Error generating scrub track\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;AAOA,MAAM,MAAM,MAAM,wBAAwB;AAE1C,MAAa,6BAA6B,OAAO,iBAAyB;AACxE,KAAI,8BAA8B,eAAe;CAEjD,MAAM,QAAQ,MAAM,MAAM,UAAU,aAAa;AAGjD,KAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;CAIrE,MAAM,cAAc,MAAM,4BAA4B;CAEtD,MAAM,oBAAoB,MAAM;CAGhC,MAAM,eAAe,IAAI,aAAa;CACtC,MAAM,cAAc,IAAI,aAAa;AAGrC,aAAY,KAAK,cAAc,EAAE,KAAK,OAAO,CAAC;AAC9C,aAAY,KAAK,YAAY;CAG7B,IAAI,oBAAoB;AACxB,aAAY,GAAG,aAAa;AAC1B,sBAAoB;GACpB;AAEF,aAAY,GAAG,UAAU,UAAU;AACjC,eAAa,QAAQ,MAAM;AAC3B,cAAY,QAAQ,MAAM;GAC1B;CAKF,MAAM,eAAe;CAGrB,MAAM,uBAAuB,sBAC3B,aACA,mBAJqB,EAAE,GAAG,cAAc,CAMzC;AAGD,sBACG,WAAW;AACV,MAAI,kBACF,cAAa,KAAK;MAGlB,aAAY,KAAK,aAAa;AAC5B,gBAAa,KAAK;IAClB;GAEJ,CACD,OAAO,UAAU;AAChB,eAAa,QAAQ,MAAM;GAC3B;AAGJ,QAAO;EACL,QAAQ;EACR,eAAe;EAChB;;AAGH,MAAa,yBAAyB,eAAe;CACnD,OAAO;CACP,WAAW,iBACT,GAAG,SAAS,aAAa,CAAC;CAC5B,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAM,MAAM,UAAU,aAAa;AAEjD,MAAI,MAAM,aAAa,WAAW,EAChC,OAAM,IAAI,MAAM,mDAAmD;EAIrE,MAAM,cAAc,MAAM,4BAA4B;EAGtD,MAAM,cAAc,IAAI,aAAa;EAGrC,IAAIA,kBAAyC;EAE7C,MAAM,6BAA6B;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAG/B,qBAAkB,iBAAiB;AACjC,QAAI,CAAC,YAAY,WAAW;AAC1B,aAAQ,KACN,0EACD;AACD,iBAAY,wBAAQ,IAAI,MAAM,iCAAiC,CAAC;;MAEjE,IAAM;;AAIX,wBAAsB;AAGtB,cAAY,GAAG,cAAc;AAC3B,yBAAsB;IACtB;AAEF,cAAY,GAAG,aAAa;AAC1B,yBAAsB;AACtB,eAAY,KAAK;IACjB;AAEF,cAAY,GAAG,UAAU,UAAU;AACjC,OAAI,gBACF,cAAa,gBAAgB;AAE/B,eAAY,QAAQ,MAAM;IAC1B;AAGF,cAAY,KAAK,aAAa,EAAE,KAAK,OAAO,CAAC;AAG7C,cAAY,GAAG,aAAa;AAC1B,OAAI,gBACF,cAAa,gBAAgB;IAE/B;AAEF,SAAO;;CAEV,CAAC;AAEF,MAAa,qBAAqB,OAChC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,uBAAuB,WAAW,aAAa;UACrD,OAAO;AACd,UAAQ,MAAM,MAAM;AACpB,UAAQ,MAAM,gCAAgC,MAAM;AACpD,QAAM"}
@@ -11,39 +11,33 @@ node_path = require_rolldown_runtime.__toESM(node_path);
11
11
  const generateTrackFragmentIndexFromPath = async (absolutePath) => {
12
12
  const log = (0, debug.default)("ef:generateTrackFragment");
13
13
  const probe = await require_Probe.Probe.probePath(absolutePath);
14
- let startTimeOffsetMs;
15
- if (probe.format.start_time && Number(probe.format.start_time) !== 0) {
16
- startTimeOffsetMs = Number(probe.format.start_time) * 1e3;
17
- log(`Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`);
18
- } else {
19
- const videoStream = probe.streams.find((stream) => stream.codec_type === "video");
20
- if (videoStream && videoStream.start_time && Number(videoStream.start_time) !== 0) {
21
- startTimeOffsetMs = Number(videoStream.start_time) * 1e3;
22
- log(`Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`);
23
- } else log("No format/stream timing offset found - will detect from composition time");
24
- }
14
+ const startTimeOffsetMs = probe.startTimeOffsetMs;
15
+ if (startTimeOffsetMs !== void 0) log(`Extracted start_time offset: ${startTimeOffsetMs}ms`);
16
+ else log("No format/stream timing offset found - will detect from composition time");
25
17
  log(`Generating track fragment index for ${absolutePath} using single-track approach`);
26
- const trackFragmentIndexes = {};
27
- for (let streamIndex = 0; streamIndex < probe.streams.length; streamIndex++) {
28
- const stream = probe.streams[streamIndex];
29
- if (stream.codec_type !== "audio" && stream.codec_type !== "video") continue;
18
+ const trackTasks = probe.streams.map((stream, streamIndex) => {
19
+ if (stream.codec_type !== "audio" && stream.codec_type !== "video") return null;
30
20
  const trackId = streamIndex + 1;
31
21
  log(`Processing track ${trackId} (${stream.codec_type})`);
32
- const singleTrackIndexes = await require_generateFragmentIndex.generateFragmentIndex(probe.createTrackReadstream(streamIndex), startTimeOffsetMs, { 0: trackId });
33
- Object.assign(trackFragmentIndexes, singleTrackIndexes);
34
- }
35
- if (probe.videoStreams.length > 0) try {
36
- log("Generating scrub track fragment index");
37
- const scrubStream = probe.createScrubTrackReadstream();
38
- const scrubTrackId = -1;
39
- const scrubFragmentIndex = await require_generateFragmentIndex.generateFragmentIndex(scrubStream, startTimeOffsetMs, { 0: scrubTrackId });
40
- if (scrubFragmentIndex[scrubTrackId]) {
41
- trackFragmentIndexes[scrubTrackId] = scrubFragmentIndex[scrubTrackId];
22
+ return require_generateFragmentIndex.generateFragmentIndex(probe.createTrackReadstream(streamIndex), startTimeOffsetMs, { 0: trackId });
23
+ }).filter((task) => task !== null);
24
+ const scrubTask = probe.videoStreams.length > 0 ? (async () => {
25
+ try {
26
+ log("Generating scrub track fragment index");
27
+ const scrubStream = probe.createScrubTrackReadstream();
28
+ const scrubTrackId = -1;
29
+ const result = await require_generateFragmentIndex.generateFragmentIndex(scrubStream, startTimeOffsetMs, { 0: scrubTrackId });
42
30
  log("Scrub track fragment index generated successfully");
31
+ return result;
32
+ } catch (error) {
33
+ log(`Failed to generate scrub track fragment index: ${error}`);
34
+ return null;
43
35
  }
44
- } catch (error) {
45
- log(`Failed to generate scrub track fragment index: ${error}`);
46
- }
36
+ })() : Promise.resolve(null);
37
+ const [trackResults, scrubResult] = await Promise.all([Promise.all(trackTasks), scrubTask]);
38
+ const trackFragmentIndexes = {};
39
+ for (const result of trackResults) Object.assign(trackFragmentIndexes, result);
40
+ if (scrubResult) Object.assign(trackFragmentIndexes, scrubResult);
47
41
  return trackFragmentIndexes;
48
42
  };
49
43
  const generateTrackFragmentIndexTask = require_idempotentTask.idempotentTask({
@@ -1 +1 @@
1
- {"version":3,"file":"generateTrackFragmentIndex.cjs","names":["Probe","startTimeOffsetMs: number | undefined","trackFragmentIndexes: Record<number, TrackFragmentIndex>","generateFragmentIndex","idempotentTask"],"sources":["../../src/tasks/generateTrackFragmentIndex.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\nimport type { TrackFragmentIndex } from \"../Probe.js\";\n\nexport const generateTrackFragmentIndexFromPath = async (\n absolutePath: string,\n) => {\n const log = debug(\"ef:generateTrackFragment\");\n const probe = await Probe.probePath(absolutePath);\n\n // Extract timing offset from probe metadata (same logic as processISOBMFF.ts)\n let startTimeOffsetMs: number | undefined;\n\n // First check format-level start_time\n if (probe.format.start_time && Number(probe.format.start_time) !== 0) {\n startTimeOffsetMs = Number(probe.format.start_time) * 1000;\n log(\n `Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`,\n );\n } else {\n // Check for video stream start_time (more common)\n const videoStream = probe.streams.find(\n (stream) => stream.codec_type === \"video\",\n );\n if (\n videoStream &&\n videoStream.start_time &&\n Number(videoStream.start_time) !== 0\n ) {\n startTimeOffsetMs = Number(videoStream.start_time) * 1000;\n log(\n `Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`,\n );\n } else {\n log(\n \"No format/stream timing offset found - will detect from composition time\",\n );\n }\n }\n\n log(\n `Generating track fragment index for ${absolutePath} using single-track approach`,\n );\n\n // FIXED: Generate fragment indexes from individual single-track files\n // This ensures byte offsets match the actual single-track files that clients will request\n const trackFragmentIndexes: Record<number, TrackFragmentIndex> = {};\n\n // Process each audio/video stream as a separate track\n for (let streamIndex = 0; streamIndex < probe.streams.length; streamIndex++) {\n const stream = probe.streams[streamIndex]!;\n\n // Only process audio and video streams\n if (stream.codec_type !== \"audio\" && stream.codec_type !== \"video\") {\n continue;\n }\n\n const trackId = streamIndex + 1; // Convert to 1-based track ID\n log(`Processing track ${trackId} (${stream.codec_type})`);\n\n // Generate single-track file and its fragment index\n const trackStream = probe.createTrackReadstream(streamIndex);\n const trackIdMapping = { 0: trackId }; // Map single-track stream index 0 to original track ID\n\n const singleTrackIndexes = await generateFragmentIndex(\n trackStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n // Merge the single-track index into the combined result\n Object.assign(trackFragmentIndexes, singleTrackIndexes);\n }\n\n // Generate scrub track fragment index if video stream exists\n if (probe.videoStreams.length > 0) {\n try {\n log(\"Generating scrub track fragment index\");\n // Generate scrub track stream and fragment index directly (don't generate full file)\n const scrubStream = probe.createScrubTrackReadstream();\n const scrubTrackId = -1;\n const trackIdMapping = { 0: scrubTrackId }; // Map single-track stream index 0 to scrub track ID -1\n\n const scrubFragmentIndex = await generateFragmentIndex(\n scrubStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n if (scrubFragmentIndex[scrubTrackId]) {\n trackFragmentIndexes[scrubTrackId] = scrubFragmentIndex[scrubTrackId]!;\n log(\"Scrub track fragment index generated successfully\");\n }\n } catch (error) {\n log(`Failed to generate scrub track fragment index: ${error}`);\n // Don't fail the entire operation if scrub track generation fails\n }\n }\n\n return trackFragmentIndexes;\n};\n\nconst generateTrackFragmentIndexTask = idempotentTask({\n label: \"trackFragmentIndex\",\n filename: (absolutePath) => `${basename(absolutePath)}.tracks.json`,\n runner: async (absolutePath: string) => {\n const index = await generateTrackFragmentIndexFromPath(absolutePath);\n return JSON.stringify(index, null, 2);\n },\n});\n\nexport const generateTrackFragmentIndex = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateTrackFragmentIndexTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error generating track fragment index\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;;;AAOA,MAAa,qCAAqC,OAChD,iBACG;CACH,MAAM,yBAAY,2BAA2B;CAC7C,MAAM,QAAQ,MAAMA,oBAAM,UAAU,aAAa;CAGjD,IAAIC;AAGJ,KAAI,MAAM,OAAO,cAAc,OAAO,MAAM,OAAO,WAAW,KAAK,GAAG;AACpE,sBAAoB,OAAO,MAAM,OAAO,WAAW,GAAG;AACtD,MACE,uCAAuC,MAAM,OAAO,WAAW,KAAK,kBAAkB,KACvF;QACI;EAEL,MAAM,cAAc,MAAM,QAAQ,MAC/B,WAAW,OAAO,eAAe,QACnC;AACD,MACE,eACA,YAAY,cACZ,OAAO,YAAY,WAAW,KAAK,GACnC;AACA,uBAAoB,OAAO,YAAY,WAAW,GAAG;AACrD,OACE,6CAA6C,YAAY,WAAW,KAAK,kBAAkB,KAC5F;QAED,KACE,2EACD;;AAIL,KACE,uCAAuC,aAAa,8BACrD;CAID,MAAMC,uBAA2D,EAAE;AAGnE,MAAK,IAAI,cAAc,GAAG,cAAc,MAAM,QAAQ,QAAQ,eAAe;EAC3E,MAAM,SAAS,MAAM,QAAQ;AAG7B,MAAI,OAAO,eAAe,WAAW,OAAO,eAAe,QACzD;EAGF,MAAM,UAAU,cAAc;AAC9B,MAAI,oBAAoB,QAAQ,IAAI,OAAO,WAAW,GAAG;EAMzD,MAAM,qBAAqB,MAAMC,oDAHb,MAAM,sBAAsB,YAAY,EAK1D,mBAJqB,EAAE,GAAG,SAAS,CAMpC;AAGD,SAAO,OAAO,sBAAsB,mBAAmB;;AAIzD,KAAI,MAAM,aAAa,SAAS,EAC9B,KAAI;AACF,MAAI,wCAAwC;EAE5C,MAAM,cAAc,MAAM,4BAA4B;EACtD,MAAM,eAAe;EAGrB,MAAM,qBAAqB,MAAMA,oDAC/B,aACA,mBAJqB,EAAE,GAAG,cAAc,CAMzC;AAED,MAAI,mBAAmB,eAAe;AACpC,wBAAqB,gBAAgB,mBAAmB;AACxD,OAAI,oDAAoD;;UAEnD,OAAO;AACd,MAAI,kDAAkD,QAAQ;;AAKlE,QAAO;;AAGT,MAAM,iCAAiCC,sCAAe;CACpD,OAAO;CACP,WAAW,iBAAiB,2BAAY,aAAa,CAAC;CACtD,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAM,mCAAmC,aAAa;AACpE,SAAO,KAAK,UAAU,OAAO,MAAM,EAAE;;CAExC,CAAC;AAEF,MAAa,6BAA6B,OACxC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,+BAA+B,WAAW,aAAa;UAC7D,OAAO;AACd,UAAQ,MAAM,yCAAyC,MAAM;AAC7D,QAAM"}
1
+ {"version":3,"file":"generateTrackFragmentIndex.cjs","names":["Probe","generateFragmentIndex","scrubTask: Promise<Record<number, TrackFragmentIndex> | null>","trackFragmentIndexes: Record<number, TrackFragmentIndex>","idempotentTask"],"sources":["../../src/tasks/generateTrackFragmentIndex.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\nimport type { TrackFragmentIndex } from \"../Probe.js\";\n\nexport const generateTrackFragmentIndexFromPath = async (\n absolutePath: string,\n) => {\n const log = debug(\"ef:generateTrackFragment\");\n const probe = await Probe.probePath(absolutePath);\n\n const startTimeOffsetMs = probe.startTimeOffsetMs;\n if (startTimeOffsetMs !== undefined) {\n log(`Extracted start_time offset: ${startTimeOffsetMs}ms`);\n } else {\n log(\n \"No format/stream timing offset found - will detect from composition time\",\n );\n }\n\n log(\n `Generating track fragment index for ${absolutePath} using single-track approach`,\n );\n\n // Process all audio/video streams and scrub track in parallel\n const trackTasks = probe.streams\n .map((stream, streamIndex) => {\n if (stream.codec_type !== \"audio\" && stream.codec_type !== \"video\") {\n return null;\n }\n const trackId = streamIndex + 1;\n log(`Processing track ${trackId} (${stream.codec_type})`);\n const trackStream = probe.createTrackReadstream(streamIndex);\n const trackIdMapping = { 0: trackId };\n return generateFragmentIndex(\n trackStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n })\n .filter(\n (task): task is Promise<Record<number, TrackFragmentIndex>> =>\n task !== null,\n );\n\n const scrubTask: Promise<Record<number, TrackFragmentIndex> | null> =\n probe.videoStreams.length > 0\n ? (async () => {\n try {\n log(\"Generating scrub track fragment index\");\n const scrubStream = probe.createScrubTrackReadstream();\n const scrubTrackId = -1;\n const result = await generateFragmentIndex(\n scrubStream,\n startTimeOffsetMs,\n { 0: scrubTrackId },\n );\n log(\"Scrub track fragment index generated successfully\");\n return result;\n } catch (error) {\n log(`Failed to generate scrub track fragment index: ${error}`);\n return null;\n }\n })()\n : Promise.resolve(null);\n\n const [trackResults, scrubResult] = await Promise.all([\n Promise.all(trackTasks),\n scrubTask,\n ]);\n\n const trackFragmentIndexes: Record<number, TrackFragmentIndex> = {};\n for (const result of trackResults) {\n Object.assign(trackFragmentIndexes, result);\n }\n if (scrubResult) {\n Object.assign(trackFragmentIndexes, scrubResult);\n }\n\n return trackFragmentIndexes;\n};\n\nconst generateTrackFragmentIndexTask = idempotentTask({\n label: \"trackFragmentIndex\",\n filename: (absolutePath) => `${basename(absolutePath)}.tracks.json`,\n runner: async (absolutePath: string) => {\n const index = await generateTrackFragmentIndexFromPath(absolutePath);\n return JSON.stringify(index, null, 2);\n },\n});\n\nexport const generateTrackFragmentIndex = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateTrackFragmentIndexTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error generating track fragment index\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;;;;AAOA,MAAa,qCAAqC,OAChD,iBACG;CACH,MAAM,yBAAY,2BAA2B;CAC7C,MAAM,QAAQ,MAAMA,oBAAM,UAAU,aAAa;CAEjD,MAAM,oBAAoB,MAAM;AAChC,KAAI,sBAAsB,OACxB,KAAI,gCAAgC,kBAAkB,IAAI;KAE1D,KACE,2EACD;AAGH,KACE,uCAAuC,aAAa,8BACrD;CAGD,MAAM,aAAa,MAAM,QACtB,KAAK,QAAQ,gBAAgB;AAC5B,MAAI,OAAO,eAAe,WAAW,OAAO,eAAe,QACzD,QAAO;EAET,MAAM,UAAU,cAAc;AAC9B,MAAI,oBAAoB,QAAQ,IAAI,OAAO,WAAW,GAAG;AAGzD,SAAOC,oDAFa,MAAM,sBAAsB,YAAY,EAI1D,mBAHqB,EAAE,GAAG,SAAS,CAKpC;GACD,CACD,QACE,SACC,SAAS,KACZ;CAEH,MAAMC,YACJ,MAAM,aAAa,SAAS,KACvB,YAAY;AACX,MAAI;AACF,OAAI,wCAAwC;GAC5C,MAAM,cAAc,MAAM,4BAA4B;GACtD,MAAM,eAAe;GACrB,MAAM,SAAS,MAAMD,oDACnB,aACA,mBACA,EAAE,GAAG,cAAc,CACpB;AACD,OAAI,oDAAoD;AACxD,UAAO;WACA,OAAO;AACd,OAAI,kDAAkD,QAAQ;AAC9D,UAAO;;KAEP,GACJ,QAAQ,QAAQ,KAAK;CAE3B,MAAM,CAAC,cAAc,eAAe,MAAM,QAAQ,IAAI,CACpD,QAAQ,IAAI,WAAW,EACvB,UACD,CAAC;CAEF,MAAME,uBAA2D,EAAE;AACnE,MAAK,MAAM,UAAU,aACnB,QAAO,OAAO,sBAAsB,OAAO;AAE7C,KAAI,YACF,QAAO,OAAO,sBAAsB,YAAY;AAGlD,QAAO;;AAGT,MAAM,iCAAiCC,sCAAe;CACpD,OAAO;CACP,WAAW,iBAAiB,2BAAY,aAAa,CAAC;CACtD,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAM,mCAAmC,aAAa;AACpE,SAAO,KAAK,UAAU,OAAO,MAAM,EAAE;;CAExC,CAAC;AAEF,MAAa,6BAA6B,OACxC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,+BAA+B,WAAW,aAAa;UAC7D,OAAO;AACd,UAAQ,MAAM,yCAAyC,MAAM;AAC7D,QAAM"}
@@ -8,39 +8,33 @@ import { basename } from "node:path";
8
8
  const generateTrackFragmentIndexFromPath = async (absolutePath) => {
9
9
  const log = debug("ef:generateTrackFragment");
10
10
  const probe = await Probe.probePath(absolutePath);
11
- let startTimeOffsetMs;
12
- if (probe.format.start_time && Number(probe.format.start_time) !== 0) {
13
- startTimeOffsetMs = Number(probe.format.start_time) * 1e3;
14
- log(`Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`);
15
- } else {
16
- const videoStream = probe.streams.find((stream) => stream.codec_type === "video");
17
- if (videoStream && videoStream.start_time && Number(videoStream.start_time) !== 0) {
18
- startTimeOffsetMs = Number(videoStream.start_time) * 1e3;
19
- log(`Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`);
20
- } else log("No format/stream timing offset found - will detect from composition time");
21
- }
11
+ const startTimeOffsetMs = probe.startTimeOffsetMs;
12
+ if (startTimeOffsetMs !== void 0) log(`Extracted start_time offset: ${startTimeOffsetMs}ms`);
13
+ else log("No format/stream timing offset found - will detect from composition time");
22
14
  log(`Generating track fragment index for ${absolutePath} using single-track approach`);
23
- const trackFragmentIndexes = {};
24
- for (let streamIndex = 0; streamIndex < probe.streams.length; streamIndex++) {
25
- const stream = probe.streams[streamIndex];
26
- if (stream.codec_type !== "audio" && stream.codec_type !== "video") continue;
15
+ const trackTasks = probe.streams.map((stream, streamIndex) => {
16
+ if (stream.codec_type !== "audio" && stream.codec_type !== "video") return null;
27
17
  const trackId = streamIndex + 1;
28
18
  log(`Processing track ${trackId} (${stream.codec_type})`);
29
- const singleTrackIndexes = await generateFragmentIndex(probe.createTrackReadstream(streamIndex), startTimeOffsetMs, { 0: trackId });
30
- Object.assign(trackFragmentIndexes, singleTrackIndexes);
31
- }
32
- if (probe.videoStreams.length > 0) try {
33
- log("Generating scrub track fragment index");
34
- const scrubStream = probe.createScrubTrackReadstream();
35
- const scrubTrackId = -1;
36
- const scrubFragmentIndex = await generateFragmentIndex(scrubStream, startTimeOffsetMs, { 0: scrubTrackId });
37
- if (scrubFragmentIndex[scrubTrackId]) {
38
- trackFragmentIndexes[scrubTrackId] = scrubFragmentIndex[scrubTrackId];
19
+ return generateFragmentIndex(probe.createTrackReadstream(streamIndex), startTimeOffsetMs, { 0: trackId });
20
+ }).filter((task) => task !== null);
21
+ const scrubTask = probe.videoStreams.length > 0 ? (async () => {
22
+ try {
23
+ log("Generating scrub track fragment index");
24
+ const scrubStream = probe.createScrubTrackReadstream();
25
+ const scrubTrackId = -1;
26
+ const result = await generateFragmentIndex(scrubStream, startTimeOffsetMs, { 0: scrubTrackId });
39
27
  log("Scrub track fragment index generated successfully");
28
+ return result;
29
+ } catch (error) {
30
+ log(`Failed to generate scrub track fragment index: ${error}`);
31
+ return null;
40
32
  }
41
- } catch (error) {
42
- log(`Failed to generate scrub track fragment index: ${error}`);
43
- }
33
+ })() : Promise.resolve(null);
34
+ const [trackResults, scrubResult] = await Promise.all([Promise.all(trackTasks), scrubTask]);
35
+ const trackFragmentIndexes = {};
36
+ for (const result of trackResults) Object.assign(trackFragmentIndexes, result);
37
+ if (scrubResult) Object.assign(trackFragmentIndexes, scrubResult);
44
38
  return trackFragmentIndexes;
45
39
  };
46
40
  const generateTrackFragmentIndexTask = idempotentTask({
@@ -1 +1 @@
1
- {"version":3,"file":"generateTrackFragmentIndex.js","names":["startTimeOffsetMs: number | undefined","trackFragmentIndexes: Record<number, TrackFragmentIndex>"],"sources":["../../src/tasks/generateTrackFragmentIndex.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\nimport type { TrackFragmentIndex } from \"../Probe.js\";\n\nexport const generateTrackFragmentIndexFromPath = async (\n absolutePath: string,\n) => {\n const log = debug(\"ef:generateTrackFragment\");\n const probe = await Probe.probePath(absolutePath);\n\n // Extract timing offset from probe metadata (same logic as processISOBMFF.ts)\n let startTimeOffsetMs: number | undefined;\n\n // First check format-level start_time\n if (probe.format.start_time && Number(probe.format.start_time) !== 0) {\n startTimeOffsetMs = Number(probe.format.start_time) * 1000;\n log(\n `Extracted format start_time offset: ${probe.format.start_time}s (${startTimeOffsetMs}ms)`,\n );\n } else {\n // Check for video stream start_time (more common)\n const videoStream = probe.streams.find(\n (stream) => stream.codec_type === \"video\",\n );\n if (\n videoStream &&\n videoStream.start_time &&\n Number(videoStream.start_time) !== 0\n ) {\n startTimeOffsetMs = Number(videoStream.start_time) * 1000;\n log(\n `Extracted video stream start_time offset: ${videoStream.start_time}s (${startTimeOffsetMs}ms)`,\n );\n } else {\n log(\n \"No format/stream timing offset found - will detect from composition time\",\n );\n }\n }\n\n log(\n `Generating track fragment index for ${absolutePath} using single-track approach`,\n );\n\n // FIXED: Generate fragment indexes from individual single-track files\n // This ensures byte offsets match the actual single-track files that clients will request\n const trackFragmentIndexes: Record<number, TrackFragmentIndex> = {};\n\n // Process each audio/video stream as a separate track\n for (let streamIndex = 0; streamIndex < probe.streams.length; streamIndex++) {\n const stream = probe.streams[streamIndex]!;\n\n // Only process audio and video streams\n if (stream.codec_type !== \"audio\" && stream.codec_type !== \"video\") {\n continue;\n }\n\n const trackId = streamIndex + 1; // Convert to 1-based track ID\n log(`Processing track ${trackId} (${stream.codec_type})`);\n\n // Generate single-track file and its fragment index\n const trackStream = probe.createTrackReadstream(streamIndex);\n const trackIdMapping = { 0: trackId }; // Map single-track stream index 0 to original track ID\n\n const singleTrackIndexes = await generateFragmentIndex(\n trackStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n // Merge the single-track index into the combined result\n Object.assign(trackFragmentIndexes, singleTrackIndexes);\n }\n\n // Generate scrub track fragment index if video stream exists\n if (probe.videoStreams.length > 0) {\n try {\n log(\"Generating scrub track fragment index\");\n // Generate scrub track stream and fragment index directly (don't generate full file)\n const scrubStream = probe.createScrubTrackReadstream();\n const scrubTrackId = -1;\n const trackIdMapping = { 0: scrubTrackId }; // Map single-track stream index 0 to scrub track ID -1\n\n const scrubFragmentIndex = await generateFragmentIndex(\n scrubStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n\n if (scrubFragmentIndex[scrubTrackId]) {\n trackFragmentIndexes[scrubTrackId] = scrubFragmentIndex[scrubTrackId]!;\n log(\"Scrub track fragment index generated successfully\");\n }\n } catch (error) {\n log(`Failed to generate scrub track fragment index: ${error}`);\n // Don't fail the entire operation if scrub track generation fails\n }\n }\n\n return trackFragmentIndexes;\n};\n\nconst generateTrackFragmentIndexTask = idempotentTask({\n label: \"trackFragmentIndex\",\n filename: (absolutePath) => `${basename(absolutePath)}.tracks.json`,\n runner: async (absolutePath: string) => {\n const index = await generateTrackFragmentIndexFromPath(absolutePath);\n return JSON.stringify(index, null, 2);\n },\n});\n\nexport const generateTrackFragmentIndex = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateTrackFragmentIndexTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error generating track fragment index\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;AAOA,MAAa,qCAAqC,OAChD,iBACG;CACH,MAAM,MAAM,MAAM,2BAA2B;CAC7C,MAAM,QAAQ,MAAM,MAAM,UAAU,aAAa;CAGjD,IAAIA;AAGJ,KAAI,MAAM,OAAO,cAAc,OAAO,MAAM,OAAO,WAAW,KAAK,GAAG;AACpE,sBAAoB,OAAO,MAAM,OAAO,WAAW,GAAG;AACtD,MACE,uCAAuC,MAAM,OAAO,WAAW,KAAK,kBAAkB,KACvF;QACI;EAEL,MAAM,cAAc,MAAM,QAAQ,MAC/B,WAAW,OAAO,eAAe,QACnC;AACD,MACE,eACA,YAAY,cACZ,OAAO,YAAY,WAAW,KAAK,GACnC;AACA,uBAAoB,OAAO,YAAY,WAAW,GAAG;AACrD,OACE,6CAA6C,YAAY,WAAW,KAAK,kBAAkB,KAC5F;QAED,KACE,2EACD;;AAIL,KACE,uCAAuC,aAAa,8BACrD;CAID,MAAMC,uBAA2D,EAAE;AAGnE,MAAK,IAAI,cAAc,GAAG,cAAc,MAAM,QAAQ,QAAQ,eAAe;EAC3E,MAAM,SAAS,MAAM,QAAQ;AAG7B,MAAI,OAAO,eAAe,WAAW,OAAO,eAAe,QACzD;EAGF,MAAM,UAAU,cAAc;AAC9B,MAAI,oBAAoB,QAAQ,IAAI,OAAO,WAAW,GAAG;EAMzD,MAAM,qBAAqB,MAAM,sBAHb,MAAM,sBAAsB,YAAY,EAK1D,mBAJqB,EAAE,GAAG,SAAS,CAMpC;AAGD,SAAO,OAAO,sBAAsB,mBAAmB;;AAIzD,KAAI,MAAM,aAAa,SAAS,EAC9B,KAAI;AACF,MAAI,wCAAwC;EAE5C,MAAM,cAAc,MAAM,4BAA4B;EACtD,MAAM,eAAe;EAGrB,MAAM,qBAAqB,MAAM,sBAC/B,aACA,mBAJqB,EAAE,GAAG,cAAc,CAMzC;AAED,MAAI,mBAAmB,eAAe;AACpC,wBAAqB,gBAAgB,mBAAmB;AACxD,OAAI,oDAAoD;;UAEnD,OAAO;AACd,MAAI,kDAAkD,QAAQ;;AAKlE,QAAO;;AAGT,MAAM,iCAAiC,eAAe;CACpD,OAAO;CACP,WAAW,iBAAiB,GAAG,SAAS,aAAa,CAAC;CACtD,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAM,mCAAmC,aAAa;AACpE,SAAO,KAAK,UAAU,OAAO,MAAM,EAAE;;CAExC,CAAC;AAEF,MAAa,6BAA6B,OACxC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,+BAA+B,WAAW,aAAa;UAC7D,OAAO;AACd,UAAQ,MAAM,yCAAyC,MAAM;AAC7D,QAAM"}
1
+ {"version":3,"file":"generateTrackFragmentIndex.js","names":["scrubTask: Promise<Record<number, TrackFragmentIndex> | null>","trackFragmentIndexes: Record<number, TrackFragmentIndex>"],"sources":["../../src/tasks/generateTrackFragmentIndex.ts"],"sourcesContent":["import { idempotentTask } from \"../idempotentTask.js\";\nimport debug from \"debug\";\nimport { basename } from \"node:path\";\nimport { Probe } from \"../Probe.js\";\nimport { generateFragmentIndex } from \"../generateFragmentIndex.js\";\nimport type { TrackFragmentIndex } from \"../Probe.js\";\n\nexport const generateTrackFragmentIndexFromPath = async (\n absolutePath: string,\n) => {\n const log = debug(\"ef:generateTrackFragment\");\n const probe = await Probe.probePath(absolutePath);\n\n const startTimeOffsetMs = probe.startTimeOffsetMs;\n if (startTimeOffsetMs !== undefined) {\n log(`Extracted start_time offset: ${startTimeOffsetMs}ms`);\n } else {\n log(\n \"No format/stream timing offset found - will detect from composition time\",\n );\n }\n\n log(\n `Generating track fragment index for ${absolutePath} using single-track approach`,\n );\n\n // Process all audio/video streams and scrub track in parallel\n const trackTasks = probe.streams\n .map((stream, streamIndex) => {\n if (stream.codec_type !== \"audio\" && stream.codec_type !== \"video\") {\n return null;\n }\n const trackId = streamIndex + 1;\n log(`Processing track ${trackId} (${stream.codec_type})`);\n const trackStream = probe.createTrackReadstream(streamIndex);\n const trackIdMapping = { 0: trackId };\n return generateFragmentIndex(\n trackStream,\n startTimeOffsetMs,\n trackIdMapping,\n );\n })\n .filter(\n (task): task is Promise<Record<number, TrackFragmentIndex>> =>\n task !== null,\n );\n\n const scrubTask: Promise<Record<number, TrackFragmentIndex> | null> =\n probe.videoStreams.length > 0\n ? (async () => {\n try {\n log(\"Generating scrub track fragment index\");\n const scrubStream = probe.createScrubTrackReadstream();\n const scrubTrackId = -1;\n const result = await generateFragmentIndex(\n scrubStream,\n startTimeOffsetMs,\n { 0: scrubTrackId },\n );\n log(\"Scrub track fragment index generated successfully\");\n return result;\n } catch (error) {\n log(`Failed to generate scrub track fragment index: ${error}`);\n return null;\n }\n })()\n : Promise.resolve(null);\n\n const [trackResults, scrubResult] = await Promise.all([\n Promise.all(trackTasks),\n scrubTask,\n ]);\n\n const trackFragmentIndexes: Record<number, TrackFragmentIndex> = {};\n for (const result of trackResults) {\n Object.assign(trackFragmentIndexes, result);\n }\n if (scrubResult) {\n Object.assign(trackFragmentIndexes, scrubResult);\n }\n\n return trackFragmentIndexes;\n};\n\nconst generateTrackFragmentIndexTask = idempotentTask({\n label: \"trackFragmentIndex\",\n filename: (absolutePath) => `${basename(absolutePath)}.tracks.json`,\n runner: async (absolutePath: string) => {\n const index = await generateTrackFragmentIndexFromPath(absolutePath);\n return JSON.stringify(index, null, 2);\n },\n});\n\nexport const generateTrackFragmentIndex = async (\n cacheRoot: string,\n absolutePath: string,\n) => {\n try {\n return await generateTrackFragmentIndexTask(cacheRoot, absolutePath);\n } catch (error) {\n console.trace(\"Error generating track fragment index\", error);\n throw error;\n }\n};\n"],"mappings":";;;;;;;AAOA,MAAa,qCAAqC,OAChD,iBACG;CACH,MAAM,MAAM,MAAM,2BAA2B;CAC7C,MAAM,QAAQ,MAAM,MAAM,UAAU,aAAa;CAEjD,MAAM,oBAAoB,MAAM;AAChC,KAAI,sBAAsB,OACxB,KAAI,gCAAgC,kBAAkB,IAAI;KAE1D,KACE,2EACD;AAGH,KACE,uCAAuC,aAAa,8BACrD;CAGD,MAAM,aAAa,MAAM,QACtB,KAAK,QAAQ,gBAAgB;AAC5B,MAAI,OAAO,eAAe,WAAW,OAAO,eAAe,QACzD,QAAO;EAET,MAAM,UAAU,cAAc;AAC9B,MAAI,oBAAoB,QAAQ,IAAI,OAAO,WAAW,GAAG;AAGzD,SAAO,sBAFa,MAAM,sBAAsB,YAAY,EAI1D,mBAHqB,EAAE,GAAG,SAAS,CAKpC;GACD,CACD,QACE,SACC,SAAS,KACZ;CAEH,MAAMA,YACJ,MAAM,aAAa,SAAS,KACvB,YAAY;AACX,MAAI;AACF,OAAI,wCAAwC;GAC5C,MAAM,cAAc,MAAM,4BAA4B;GACtD,MAAM,eAAe;GACrB,MAAM,SAAS,MAAM,sBACnB,aACA,mBACA,EAAE,GAAG,cAAc,CACpB;AACD,OAAI,oDAAoD;AACxD,UAAO;WACA,OAAO;AACd,OAAI,kDAAkD,QAAQ;AAC9D,UAAO;;KAEP,GACJ,QAAQ,QAAQ,KAAK;CAE3B,MAAM,CAAC,cAAc,eAAe,MAAM,QAAQ,IAAI,CACpD,QAAQ,IAAI,WAAW,EACvB,UACD,CAAC;CAEF,MAAMC,uBAA2D,EAAE;AACnE,MAAK,MAAM,UAAU,aACnB,QAAO,OAAO,sBAAsB,OAAO;AAE7C,KAAI,YACF,QAAO,OAAO,sBAAsB,YAAY;AAGlD,QAAO;;AAGT,MAAM,iCAAiC,eAAe;CACpD,OAAO;CACP,WAAW,iBAAiB,GAAG,SAAS,aAAa,CAAC;CACtD,QAAQ,OAAO,iBAAyB;EACtC,MAAM,QAAQ,MAAM,mCAAmC,aAAa;AACpE,SAAO,KAAK,UAAU,OAAO,MAAM,EAAE;;CAExC,CAAC;AAEF,MAAa,6BAA6B,OACxC,WACA,iBACG;AACH,KAAI;AACF,SAAO,MAAM,+BAA+B,WAAW,aAAa;UAC7D,OAAO;AACd,UAAQ,MAAM,yCAAyC,MAAM;AAC7D,QAAM"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@editframe/assets",
3
- "version": "0.40.0",
3
+ "version": "0.40.2",
4
4
  "description": "",
5
5
  "repository": {
6
6
  "type": "git",