@ztimson/ai-utils 0.7.4 → 0.7.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/asr.js +5 -5
- package/dist/asr.js.map +1 -1
- package/dist/asr.mjs +59 -58
- package/dist/asr.mjs.map +1 -1
- package/dist/index.js +15 -15
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +19 -19
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/asr.js
CHANGED
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports,Symbol.toStringTag,{value:"Module"});const
|
|
1
|
+
"use strict";Object.defineProperty(exports,Symbol.toStringTag,{value:"Module"});const S=require("@xenova/transformers"),l=require("worker_threads"),y=require("node:child_process"),m=require("node:fs"),g=require("node:path"),P=require("node:os"),h=require("wavefile");async function w(){const i=t=>new Promise(n=>{const o=y.spawn(t,["-c","import pyannote.audio"]);o.on("close",a=>n(a===0)),o.on("error",()=>n(!1))});return await i("python3")?"python3":await i("python")?"python":null}async function v(i,t,n,o){const a=`
|
|
2
2
|
import sys
|
|
3
3
|
import json
|
|
4
4
|
import os
|
|
5
5
|
from pyannote.audio import Pipeline
|
|
6
6
|
|
|
7
|
-
os.environ['TORCH_HOME'] = r"${
|
|
8
|
-
pipeline = Pipeline.from_pretrained("pyannote/speaker-diarization-3.1", token="${
|
|
7
|
+
os.environ['TORCH_HOME'] = r"${n}"
|
|
8
|
+
pipeline = Pipeline.from_pretrained("pyannote/speaker-diarization-3.1", token="${o}")
|
|
9
9
|
output = pipeline(sys.argv[1])
|
|
10
10
|
|
|
11
11
|
segments = []
|
|
@@ -13,6 +13,6 @@ for turn, speaker in output.speaker_diarization:
|
|
|
13
13
|
segments.append({"start": turn.start, "end": turn.end, "speaker": speaker})
|
|
14
14
|
|
|
15
15
|
print(json.dumps(segments))
|
|
16
|
-
`;return new Promise((
|
|
17
|
-
`)}function
|
|
16
|
+
`;return new Promise((p,r)=>{let e="";const c=y.spawn(i,["-c",a,t]);c.stdout.on("data",s=>e+=s.toString()),c.stderr.on("data",s=>console.error(s.toString())),c.on("close",s=>{if(s===0)try{p(JSON.parse(e))}catch{r(new Error("Failed to parse diarization output"))}else r(new Error(`Python process exited with code ${s}`))}),c.on("error",r)})}function _(i,t){const n=new Map;let o=0;t.forEach(e=>{n.has(e.speaker)||n.set(e.speaker,++o)});const a=[];let p=-1,r="";return i.forEach(e=>{const c=e.timestamp[0],s=t.find(f=>c>=f.start&&c<=f.end),u=s?n.get(s.speaker):1;u!==p?(r&&a.push(`[Speaker ${p}]: ${r.trim()}`),p=u,r=e.text):r+=e.text}),r&&a.push(`[Speaker ${p}]: ${r.trim()}`),a.join(`
|
|
17
|
+
`)}function $(i){let t,n;try{t=new h.WaveFile(m.readFileSync(i))}catch{n=g.join(m.mkdtempSync(g.join(P.tmpdir(),"audio-")),"converted.wav"),y.execSync(`ffmpeg -i "${i}" -ar 16000 -ac 1 -f wav "${n}"`,{stdio:"ignore"}),t=new h.WaveFile(m.readFileSync(n))}finally{t.toBitDepth("32f"),t.toSampleRate(16e3);const o=t.getSamples();if(Array.isArray(o)){const a=o[0],p=o[1],r=new Float32Array(a.length);for(let e=0;e<a.length;e++)r[e]=(a[e]+p[e])/2;return[n||i,r]}return[n||i,o]}}l.parentPort?.on("message",async({file:i,speaker:t,model:n,modelDir:o,token:a})=>{let p=null;try{const r=await S.pipeline("automatic-speech-recognition",`Xenova/${n}`,{cache_dir:o,quantized:!0}),[e,c]=$(i);p=e!==i?e:null;const s=await w(),[u,f]=await Promise.all([r(c,{return_timestamps:t?"word":!1}),!t||!a||!s?Promise.resolve():v(s,e,o,a)]),d=u.text?.trim()||null;if(!t)return l.parentPort?.postMessage({text:d});if(!a)return l.parentPort?.postMessage({text:d,error:"HuggingFace token required"});if(!s)return l.parentPort?.postMessage({text:d,error:"Speaker diarization unavailable"});const k=_(u.chunks||[],f||[]);l.parentPort?.postMessage({text:k})}catch(r){l.parentPort?.postMessage({error:r.stack||r.message})}finally{p&&m.rmSync(p,{recursive:!0,force:!0})}});exports.canDiarization=w;
|
|
18
18
|
//# sourceMappingURL=asr.js.map
|
package/dist/asr.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"asr.js","sources":["../src/asr.ts"],"sourcesContent":["import { pipeline } from '@xenova/transformers';\nimport { parentPort } from 'worker_threads';\nimport { spawn } from 'node:child_process';\nimport { execSync } from 'node:child_process';\nimport { mkdtempSync, rmSync, readFileSync } from 'node:fs';\nimport { join } from 'node:path';\nimport { tmpdir } from 'node:os';\nimport wavefile from 'wavefile';\n\
|
|
1
|
+
{"version":3,"file":"asr.js","sources":["../src/asr.ts"],"sourcesContent":["import { pipeline } from '@xenova/transformers';\nimport { parentPort } from 'worker_threads';\nimport { spawn } from 'node:child_process';\nimport { execSync } from 'node:child_process';\nimport { mkdtempSync, rmSync, readFileSync } from 'node:fs';\nimport { join } from 'node:path';\nimport { tmpdir } from 'node:os';\nimport wavefile from 'wavefile';\n\nexport async function canDiarization(): Promise<string | null> {\n\tconst checkPython = (cmd: string) => {\n\t\treturn new Promise<boolean>((resolve) => {\n\t\t\tconst proc = spawn(cmd, ['-c', 'import pyannote.audio']);\n\t\t\tproc.on('close', (code: number) => resolve(code === 0));\n\t\t\tproc.on('error', () => resolve(false));\n\t\t});\n\t};\n\tif(await checkPython('python3')) return 'python3';\n\tif(await checkPython('python')) return 'python';\n\treturn null;\n}\n\nasync function runDiarization(binary: string, audioPath: string, dir: string, token: string): Promise<any[]> {\n\tconst script = `\nimport sys\nimport json\nimport os\nfrom pyannote.audio import Pipeline\n\nos.environ['TORCH_HOME'] = r\"${dir}\"\npipeline = Pipeline.from_pretrained(\"pyannote/speaker-diarization-3.1\", token=\"${token}\")\noutput = pipeline(sys.argv[1])\n\nsegments = []\nfor turn, speaker in output.speaker_diarization:\n segments.append({\"start\": turn.start, \"end\": turn.end, \"speaker\": speaker})\n\nprint(json.dumps(segments))\n`;\n\n\treturn new Promise((resolve, reject) => {\n\t\tlet output = '';\n\t\tconst proc = spawn(binary, ['-c', script, audioPath]);\n\t\tproc.stdout.on('data', (data: Buffer) => output += data.toString());\n\t\tproc.stderr.on('data', (data: Buffer) => console.error(data.toString()));\n\t\tproc.on('close', (code: number) => {\n\t\t\tif(code === 0) {\n\t\t\t\ttry {\n\t\t\t\t\tresolve(JSON.parse(output));\n\t\t\t\t} catch (err) {\n\t\t\t\t\treject(new Error('Failed to parse diarization output'));\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\treject(new Error(`Python process exited with code ${code}`));\n\t\t\t}\n\t\t});\n\t\tproc.on('error', reject);\n\t});\n}\n\nfunction combineSpeakerTranscript(chunks: any[], speakers: any[]): string {\n\tconst speakerMap = new Map();\n\tlet speakerCount = 0;\n\tspeakers.forEach((seg: any) => {\n\t\tif(!speakerMap.has(seg.speaker)) speakerMap.set(seg.speaker, ++speakerCount);\n\t});\n\n\tconst lines: string[] = [];\n\tlet currentSpeaker = -1;\n\tlet currentText = '';\n\tchunks.forEach((chunk: any) => {\n\t\tconst time = chunk.timestamp[0];\n\t\tconst speaker = speakers.find((s: any) => time >= s.start && time <= s.end);\n\t\tconst speakerNum = speaker ? speakerMap.get(speaker.speaker) : 1;\n\t\tif (speakerNum !== currentSpeaker) {\n\t\t\tif(currentText) lines.push(`[Speaker ${currentSpeaker}]: ${currentText.trim()}`);\n\t\t\tcurrentSpeaker = speakerNum;\n\t\t\tcurrentText = chunk.text;\n\t\t} else {\n\t\t\tcurrentText += chunk.text;\n\t\t}\n\t});\n\tif(currentText) lines.push(`[Speaker ${currentSpeaker}]: ${currentText.trim()}`);\n\treturn lines.join('\\n');\n}\n\nfunction prepareAudioBuffer(file: string): [string, Float32Array] {\n\tlet wav: any, tmp;\n\ttry {\n\t\twav = new wavefile.WaveFile(readFileSync(file));\n\t} catch(err) {\n\t\ttmp = join(mkdtempSync(join(tmpdir(), 'audio-')), 'converted.wav');\n\t\texecSync(`ffmpeg -i \"${file}\" -ar 16000 -ac 1 -f wav \"${tmp}\"`, { stdio: 'ignore' });\n\t\twav = new wavefile.WaveFile(readFileSync(tmp));\n\t} finally {\n\t\twav.toBitDepth('32f');\n\t\twav.toSampleRate(16000);\n\t\tconst samples = wav.getSamples();\n\t\tif(Array.isArray(samples)) {\n\t\t\tconst left = samples[0];\n\t\t\tconst right = samples[1];\n\t\t\tconst buffer = new Float32Array(left.length);\n\t\t\tfor (let i = 0; i < left.length; i++) buffer[i] = (left[i] + right[i]) / 2;\n\t\t\treturn [tmp || file, buffer];\n\t\t}\n\t\treturn [tmp || file, samples];\n\t}\n}\n\nparentPort?.on('message', async ({ file, speaker, model, modelDir, token }) => {\n\tlet tempFile = null;\n\ttry {\n\t\tconst asr: any = await pipeline('automatic-speech-recognition', `Xenova/${model}`, {cache_dir: modelDir, quantized: true});\n\t\tconst [f, buffer] = prepareAudioBuffer(file);\n\t\ttempFile = f !== file ? f : null;\n\t\tconst hasDiarization = await canDiarization();\n\t\tconst [transcript, speakers] = await Promise.all([\n\t\t\tasr(buffer, {return_timestamps: speaker ? 'word' : false}),\n\t\t\t(!speaker || !token || !hasDiarization) ? Promise.resolve(): runDiarization(hasDiarization, f, modelDir, token),\n\t\t]);\n\n\t\tconst text = transcript.text?.trim() || null;\n\t\tif(!speaker) return parentPort?.postMessage({ text });\n\t\tif(!token) return parentPort?.postMessage({ text, error: 'HuggingFace token required' });\n\t\tif(!hasDiarization) return parentPort?.postMessage({ text, error: 'Speaker diarization unavailable' });\n\n\t\tconst combined = combineSpeakerTranscript(transcript.chunks || [], speakers || []);\n\t\tparentPort?.postMessage({ text: combined });\n\t} catch (err: any) {\n\t\tparentPort?.postMessage({ error: err.stack || err.message });\n\t} finally {\n\t\tif(tempFile) rmSync(tempFile, { recursive: true, force: true });\n\t}\n});\n"],"names":["canDiarization","checkPython","cmd","resolve","proc","spawn","code","runDiarization","binary","audioPath","dir","token","script","reject","output","data","combineSpeakerTranscript","chunks","speakers","speakerMap","speakerCount","seg","lines","currentSpeaker","currentText","chunk","time","speaker","s","speakerNum","prepareAudioBuffer","file","wav","tmp","wavefile","readFileSync","join","mkdtempSync","tmpdir","execSync","samples","left","right","buffer","i","parentPort","model","modelDir","tempFile","asr","pipeline","f","hasDiarization","transcript","text","combined","err"],"mappings":"2QASA,eAAsBA,GAAyC,CAC9D,MAAMC,EAAeC,GACb,IAAI,QAAkBC,GAAY,CACxC,MAAMC,EAAOC,EAAAA,MAAMH,EAAK,CAAC,KAAM,uBAAuB,CAAC,EACvDE,EAAK,GAAG,QAAUE,GAAiBH,EAAQG,IAAS,CAAC,CAAC,EACtDF,EAAK,GAAG,QAAS,IAAMD,EAAQ,EAAK,CAAC,CACtC,CAAC,EAEF,OAAG,MAAMF,EAAY,SAAS,EAAU,UACrC,MAAMA,EAAY,QAAQ,EAAU,SAChC,IACR,CAEA,eAAeM,EAAeC,EAAgBC,EAAmBC,EAAaC,EAA+B,CAC5G,MAAMC,EAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,+BAMeF,CAAG;AAAA,iFAC+CC,CAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAUrF,OAAO,IAAI,QAAQ,CAACR,EAASU,IAAW,CACvC,IAAIC,EAAS,GACb,MAAMV,EAAOC,EAAAA,MAAMG,EAAQ,CAAC,KAAMI,EAAQH,CAAS,CAAC,EACpDL,EAAK,OAAO,GAAG,OAASW,GAAiBD,GAAUC,EAAK,UAAU,EAClEX,EAAK,OAAO,GAAG,OAASW,GAAiB,QAAQ,MAAMA,EAAK,SAAA,CAAU,CAAC,EACvEX,EAAK,GAAG,QAAUE,GAAiB,CAClC,GAAGA,IAAS,EACX,GAAI,CACHH,EAAQ,KAAK,MAAMW,CAAM,CAAC,CAC3B,MAAc,CACbD,EAAO,IAAI,MAAM,oCAAoC,CAAC,CACvD,MAEAA,EAAO,IAAI,MAAM,mCAAmCP,CAAI,EAAE,CAAC,CAE7D,CAAC,EACDF,EAAK,GAAG,QAASS,CAAM,CACxB,CAAC,CACF,CAEA,SAASG,EAAyBC,EAAeC,EAAyB,CACzE,MAAMC,MAAiB,IACvB,IAAIC,EAAe,EACnBF,EAAS,QAASG,GAAa,CAC1BF,EAAW,IAAIE,EAAI,OAAO,GAAGF,EAAW,IAAIE,EAAI,QAAS,EAAED,CAAY,CAC5E,CAAC,EAED,MAAME,EAAkB,CAAA,EACxB,IAAIC,EAAiB,GACjBC,EAAc,GAClB,OAAAP,EAAO,QAASQ,GAAe,CAC9B,MAAMC,EAAOD,EAAM,UAAU,CAAC,EACxBE,EAAUT,EAAS,KAAMU,GAAWF,GAAQE,EAAE,OAASF,GAAQE,EAAE,GAAG,EACpEC,EAAaF,EAAUR,EAAW,IAAIQ,EAAQ,OAAO,EAAI,EAC3DE,IAAeN,GACfC,KAAmB,KAAK,YAAYD,CAAc,MAAMC,EAAY,KAAA,CAAM,EAAE,EAC/ED,EAAiBM,EACjBL,EAAcC,EAAM,MAEpBD,GAAeC,EAAM,IAEvB,CAAC,EACED,KAAmB,KAAK,YAAYD,CAAc,MAAMC,EAAY,KAAA,CAAM,EAAE,EACxEF,EAAM,KAAK;AAAA,CAAI,CACvB,CAEA,SAASQ,EAAmBC,EAAsC,CACjE,IAAIC,EAAUC,EACd,GAAI,CACHD,EAAM,IAAIE,EAAS,SAASC,EAAAA,aAAaJ,CAAI,CAAC,CAC/C,MAAa,CACZE,EAAMG,EAAAA,KAAKC,EAAAA,YAAYD,EAAAA,KAAKE,EAAAA,OAAA,EAAU,QAAQ,CAAC,EAAG,eAAe,EACjEC,WAAS,cAAcR,CAAI,6BAA6BE,CAAG,IAAK,CAAE,MAAO,SAAU,EACnFD,EAAM,IAAIE,EAAS,SAASC,EAAAA,aAAaF,CAAG,CAAC,CAC9C,QAAA,CACCD,EAAI,WAAW,KAAK,EACpBA,EAAI,aAAa,IAAK,EACtB,MAAMQ,EAAUR,EAAI,WAAA,EACpB,GAAG,MAAM,QAAQQ,CAAO,EAAG,CAC1B,MAAMC,EAAOD,EAAQ,CAAC,EAChBE,EAAQF,EAAQ,CAAC,EACjBG,EAAS,IAAI,aAAaF,EAAK,MAAM,EAC3C,QAASG,EAAI,EAAGA,EAAIH,EAAK,OAAQG,IAAKD,EAAOC,CAAC,GAAKH,EAAKG,CAAC,EAAIF,EAAME,CAAC,GAAK,EACzE,MAAO,CAACX,GAAOF,EAAMY,CAAM,CAC5B,CACA,MAAO,CAACV,GAAOF,EAAMS,CAAO,CAC7B,CACD,CAEAK,EAAAA,YAAY,GAAG,UAAW,MAAO,CAAE,KAAAd,EAAM,QAAAJ,EAAS,MAAAmB,EAAO,SAAAC,EAAU,MAAApC,KAAY,CAC9E,IAAIqC,EAAW,KACf,GAAI,CACH,MAAMC,EAAW,MAAMC,EAAAA,SAAS,+BAAgC,UAAUJ,CAAK,GAAI,CAAC,UAAWC,EAAU,UAAW,EAAA,CAAK,EACnH,CAACI,EAAGR,CAAM,EAAIb,EAAmBC,CAAI,EAC3CiB,EAAWG,IAAMpB,EAAOoB,EAAI,KAC5B,MAAMC,EAAiB,MAAMpD,EAAA,EACvB,CAACqD,EAAYnC,CAAQ,EAAI,MAAM,QAAQ,IAAI,CAChD+B,EAAIN,EAAQ,CAAC,kBAAmBhB,EAAU,OAAS,GAAM,EACxD,CAACA,GAAW,CAAChB,GAAS,CAACyC,EAAkB,QAAQ,QAAA,EAAW7C,EAAe6C,EAAgBD,EAAGJ,EAAUpC,CAAK,CAAA,CAC9G,EAEK2C,EAAOD,EAAW,MAAM,KAAA,GAAU,KACxC,GAAG,CAAC1B,EAAS,OAAOkB,EAAAA,YAAY,YAAY,CAAE,KAAAS,EAAM,EACpD,GAAG,CAAC3C,EAAO,OAAOkC,cAAY,YAAY,CAAE,KAAAS,EAAM,MAAO,6BAA8B,EACvF,GAAG,CAACF,EAAgB,OAAOP,cAAY,YAAY,CAAE,KAAAS,EAAM,MAAO,kCAAmC,EAErG,MAAMC,EAAWvC,EAAyBqC,EAAW,QAAU,CAAA,EAAInC,GAAY,EAAE,EACjF2B,EAAAA,YAAY,YAAY,CAAE,KAAMU,CAAA,CAAU,CAC3C,OAASC,EAAU,CAClBX,EAAAA,YAAY,YAAY,CAAE,MAAOW,EAAI,OAASA,EAAI,QAAS,CAC5D,QAAA,CACIR,YAAiBA,EAAU,CAAE,UAAW,GAAM,MAAO,GAAM,CAC/D,CACD,CAAC"}
|
package/dist/asr.mjs
CHANGED
|
@@ -1,27 +1,26 @@
|
|
|
1
|
-
import { pipeline as
|
|
1
|
+
import { pipeline as k } from "@xenova/transformers";
|
|
2
2
|
import { parentPort as m } from "worker_threads";
|
|
3
|
-
import { execSync as
|
|
4
|
-
import { rmSync as
|
|
3
|
+
import { execSync as S, spawn as h } from "node:child_process";
|
|
4
|
+
import { rmSync as v, readFileSync as d, mkdtempSync as $ } from "node:fs";
|
|
5
5
|
import { join as y } from "node:path";
|
|
6
|
-
import { tmpdir as
|
|
6
|
+
import { tmpdir as x } from "node:os";
|
|
7
7
|
import g from "wavefile";
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
n.on("close", (o) => t(o === 0)), n.on("error", () => t(!1));
|
|
8
|
+
async function P() {
|
|
9
|
+
const i = (r) => new Promise((n) => {
|
|
10
|
+
const o = h(r, ["-c", "import pyannote.audio"]);
|
|
11
|
+
o.on("close", (a) => n(a === 0)), o.on("error", () => n(!1));
|
|
13
12
|
});
|
|
14
|
-
return await
|
|
13
|
+
return await i("python3") ? "python3" : await i("python") ? "python" : null;
|
|
15
14
|
}
|
|
16
|
-
async function
|
|
17
|
-
const
|
|
15
|
+
async function z(i, r, n, o) {
|
|
16
|
+
const a = `
|
|
18
17
|
import sys
|
|
19
18
|
import json
|
|
20
19
|
import os
|
|
21
20
|
from pyannote.audio import Pipeline
|
|
22
21
|
|
|
23
|
-
os.environ['TORCH_HOME'] = r"${
|
|
24
|
-
pipeline = Pipeline.from_pretrained("pyannote/speaker-diarization-3.1", token="${
|
|
22
|
+
os.environ['TORCH_HOME'] = r"${n}"
|
|
23
|
+
pipeline = Pipeline.from_pretrained("pyannote/speaker-diarization-3.1", token="${o}")
|
|
25
24
|
output = pipeline(sys.argv[1])
|
|
26
25
|
|
|
27
26
|
segments = []
|
|
@@ -30,71 +29,73 @@ for turn, speaker in output.speaker_diarization:
|
|
|
30
29
|
|
|
31
30
|
print(json.dumps(segments))
|
|
32
31
|
`;
|
|
33
|
-
return new Promise((
|
|
32
|
+
return new Promise((p, t) => {
|
|
34
33
|
let e = "";
|
|
35
|
-
const c = h(
|
|
36
|
-
c.stdout.on("data", (
|
|
37
|
-
if (
|
|
34
|
+
const c = h(i, ["-c", a, r]);
|
|
35
|
+
c.stdout.on("data", (s) => e += s.toString()), c.stderr.on("data", (s) => console.error(s.toString())), c.on("close", (s) => {
|
|
36
|
+
if (s === 0)
|
|
38
37
|
try {
|
|
39
|
-
|
|
38
|
+
p(JSON.parse(e));
|
|
40
39
|
} catch {
|
|
41
|
-
|
|
40
|
+
t(new Error("Failed to parse diarization output"));
|
|
42
41
|
}
|
|
43
42
|
else
|
|
44
|
-
|
|
45
|
-
}), c.on("error",
|
|
43
|
+
t(new Error(`Python process exited with code ${s}`));
|
|
44
|
+
}), c.on("error", t);
|
|
46
45
|
});
|
|
47
46
|
}
|
|
48
|
-
function
|
|
49
|
-
const
|
|
50
|
-
let
|
|
47
|
+
function M(i, r) {
|
|
48
|
+
const n = /* @__PURE__ */ new Map();
|
|
49
|
+
let o = 0;
|
|
51
50
|
r.forEach((e) => {
|
|
52
|
-
|
|
51
|
+
n.has(e.speaker) || n.set(e.speaker, ++o);
|
|
53
52
|
});
|
|
54
|
-
const
|
|
55
|
-
let
|
|
56
|
-
return
|
|
57
|
-
const c = e.timestamp[0],
|
|
58
|
-
u !==
|
|
59
|
-
}),
|
|
53
|
+
const a = [];
|
|
54
|
+
let p = -1, t = "";
|
|
55
|
+
return i.forEach((e) => {
|
|
56
|
+
const c = e.timestamp[0], s = r.find((f) => c >= f.start && c <= f.end), u = s ? n.get(s.speaker) : 1;
|
|
57
|
+
u !== p ? (t && a.push(`[Speaker ${p}]: ${t.trim()}`), p = u, t = e.text) : t += e.text;
|
|
58
|
+
}), t && a.push(`[Speaker ${p}]: ${t.trim()}`), a.join(`
|
|
60
59
|
`);
|
|
61
60
|
}
|
|
62
|
-
function
|
|
63
|
-
let r,
|
|
61
|
+
function F(i) {
|
|
62
|
+
let r, n;
|
|
64
63
|
try {
|
|
65
|
-
r = new g.WaveFile(d(
|
|
64
|
+
r = new g.WaveFile(d(i));
|
|
66
65
|
} catch {
|
|
67
|
-
|
|
66
|
+
n = y($(y(x(), "audio-")), "converted.wav"), S(`ffmpeg -i "${i}" -ar 16000 -ac 1 -f wav "${n}"`, { stdio: "ignore" }), r = new g.WaveFile(d(n));
|
|
68
67
|
} finally {
|
|
69
68
|
r.toBitDepth("32f"), r.toSampleRate(16e3);
|
|
70
|
-
const
|
|
71
|
-
if (Array.isArray(
|
|
72
|
-
const
|
|
73
|
-
for (let e = 0; e <
|
|
74
|
-
return [
|
|
69
|
+
const o = r.getSamples();
|
|
70
|
+
if (Array.isArray(o)) {
|
|
71
|
+
const a = o[0], p = o[1], t = new Float32Array(a.length);
|
|
72
|
+
for (let e = 0; e < a.length; e++) t[e] = (a[e] + p[e]) / 2;
|
|
73
|
+
return [n || i, t];
|
|
75
74
|
}
|
|
76
|
-
return [
|
|
75
|
+
return [n || i, o];
|
|
77
76
|
}
|
|
78
77
|
}
|
|
79
|
-
m?.on("message", async ({ file:
|
|
78
|
+
m?.on("message", async ({ file: i, speaker: r, model: n, modelDir: o, token: a }) => {
|
|
79
|
+
let p = null;
|
|
80
80
|
try {
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
if (!
|
|
89
|
-
if (!
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
|
|
81
|
+
const t = await k("automatic-speech-recognition", `Xenova/${n}`, { cache_dir: o, quantized: !0 }), [e, c] = F(i);
|
|
82
|
+
p = e !== i ? e : null;
|
|
83
|
+
const s = await P(), [u, f] = await Promise.all([
|
|
84
|
+
t(c, { return_timestamps: r ? "word" : !1 }),
|
|
85
|
+
!r || !a || !s ? Promise.resolve() : z(s, e, o, a)
|
|
86
|
+
]), l = u.text?.trim() || null;
|
|
87
|
+
if (!r) return m?.postMessage({ text: l });
|
|
88
|
+
if (!a) return m?.postMessage({ text: l, error: "HuggingFace token required" });
|
|
89
|
+
if (!s) return m?.postMessage({ text: l, error: "Speaker diarization unavailable" });
|
|
90
|
+
const w = M(u.chunks || [], f || []);
|
|
91
|
+
m?.postMessage({ text: w });
|
|
92
|
+
} catch (t) {
|
|
93
|
+
m?.postMessage({ error: t.stack || t.message });
|
|
94
|
+
} finally {
|
|
95
|
+
p && v(p, { recursive: !0, force: !0 });
|
|
95
96
|
}
|
|
96
97
|
});
|
|
97
98
|
export {
|
|
98
|
-
|
|
99
|
+
P as canDiarization
|
|
99
100
|
};
|
|
100
101
|
//# sourceMappingURL=asr.mjs.map
|
package/dist/asr.mjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"asr.mjs","sources":["../src/asr.ts"],"sourcesContent":["import { pipeline } from '@xenova/transformers';\nimport { parentPort } from 'worker_threads';\nimport { spawn } from 'node:child_process';\nimport { execSync } from 'node:child_process';\nimport { mkdtempSync, rmSync, readFileSync } from 'node:fs';\nimport { join } from 'node:path';\nimport { tmpdir } from 'node:os';\nimport wavefile from 'wavefile';\n\
|
|
1
|
+
{"version":3,"file":"asr.mjs","sources":["../src/asr.ts"],"sourcesContent":["import { pipeline } from '@xenova/transformers';\nimport { parentPort } from 'worker_threads';\nimport { spawn } from 'node:child_process';\nimport { execSync } from 'node:child_process';\nimport { mkdtempSync, rmSync, readFileSync } from 'node:fs';\nimport { join } from 'node:path';\nimport { tmpdir } from 'node:os';\nimport wavefile from 'wavefile';\n\nexport async function canDiarization(): Promise<string | null> {\n\tconst checkPython = (cmd: string) => {\n\t\treturn new Promise<boolean>((resolve) => {\n\t\t\tconst proc = spawn(cmd, ['-c', 'import pyannote.audio']);\n\t\t\tproc.on('close', (code: number) => resolve(code === 0));\n\t\t\tproc.on('error', () => resolve(false));\n\t\t});\n\t};\n\tif(await checkPython('python3')) return 'python3';\n\tif(await checkPython('python')) return 'python';\n\treturn null;\n}\n\nasync function runDiarization(binary: string, audioPath: string, dir: string, token: string): Promise<any[]> {\n\tconst script = `\nimport sys\nimport json\nimport os\nfrom pyannote.audio import Pipeline\n\nos.environ['TORCH_HOME'] = r\"${dir}\"\npipeline = Pipeline.from_pretrained(\"pyannote/speaker-diarization-3.1\", token=\"${token}\")\noutput = pipeline(sys.argv[1])\n\nsegments = []\nfor turn, speaker in output.speaker_diarization:\n segments.append({\"start\": turn.start, \"end\": turn.end, \"speaker\": speaker})\n\nprint(json.dumps(segments))\n`;\n\n\treturn new Promise((resolve, reject) => {\n\t\tlet output = '';\n\t\tconst proc = spawn(binary, ['-c', script, audioPath]);\n\t\tproc.stdout.on('data', (data: Buffer) => output += data.toString());\n\t\tproc.stderr.on('data', (data: Buffer) => console.error(data.toString()));\n\t\tproc.on('close', (code: number) => {\n\t\t\tif(code === 0) {\n\t\t\t\ttry {\n\t\t\t\t\tresolve(JSON.parse(output));\n\t\t\t\t} catch (err) {\n\t\t\t\t\treject(new Error('Failed to parse diarization output'));\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\treject(new Error(`Python process exited with code ${code}`));\n\t\t\t}\n\t\t});\n\t\tproc.on('error', reject);\n\t});\n}\n\nfunction combineSpeakerTranscript(chunks: any[], speakers: any[]): string {\n\tconst speakerMap = new Map();\n\tlet speakerCount = 0;\n\tspeakers.forEach((seg: any) => {\n\t\tif(!speakerMap.has(seg.speaker)) speakerMap.set(seg.speaker, ++speakerCount);\n\t});\n\n\tconst lines: string[] = [];\n\tlet currentSpeaker = -1;\n\tlet currentText = '';\n\tchunks.forEach((chunk: any) => {\n\t\tconst time = chunk.timestamp[0];\n\t\tconst speaker = speakers.find((s: any) => time >= s.start && time <= s.end);\n\t\tconst speakerNum = speaker ? speakerMap.get(speaker.speaker) : 1;\n\t\tif (speakerNum !== currentSpeaker) {\n\t\t\tif(currentText) lines.push(`[Speaker ${currentSpeaker}]: ${currentText.trim()}`);\n\t\t\tcurrentSpeaker = speakerNum;\n\t\t\tcurrentText = chunk.text;\n\t\t} else {\n\t\t\tcurrentText += chunk.text;\n\t\t}\n\t});\n\tif(currentText) lines.push(`[Speaker ${currentSpeaker}]: ${currentText.trim()}`);\n\treturn lines.join('\\n');\n}\n\nfunction prepareAudioBuffer(file: string): [string, Float32Array] {\n\tlet wav: any, tmp;\n\ttry {\n\t\twav = new wavefile.WaveFile(readFileSync(file));\n\t} catch(err) {\n\t\ttmp = join(mkdtempSync(join(tmpdir(), 'audio-')), 'converted.wav');\n\t\texecSync(`ffmpeg -i \"${file}\" -ar 16000 -ac 1 -f wav \"${tmp}\"`, { stdio: 'ignore' });\n\t\twav = new wavefile.WaveFile(readFileSync(tmp));\n\t} finally {\n\t\twav.toBitDepth('32f');\n\t\twav.toSampleRate(16000);\n\t\tconst samples = wav.getSamples();\n\t\tif(Array.isArray(samples)) {\n\t\t\tconst left = samples[0];\n\t\t\tconst right = samples[1];\n\t\t\tconst buffer = new Float32Array(left.length);\n\t\t\tfor (let i = 0; i < left.length; i++) buffer[i] = (left[i] + right[i]) / 2;\n\t\t\treturn [tmp || file, buffer];\n\t\t}\n\t\treturn [tmp || file, samples];\n\t}\n}\n\nparentPort?.on('message', async ({ file, speaker, model, modelDir, token }) => {\n\tlet tempFile = null;\n\ttry {\n\t\tconst asr: any = await pipeline('automatic-speech-recognition', `Xenova/${model}`, {cache_dir: modelDir, quantized: true});\n\t\tconst [f, buffer] = prepareAudioBuffer(file);\n\t\ttempFile = f !== file ? f : null;\n\t\tconst hasDiarization = await canDiarization();\n\t\tconst [transcript, speakers] = await Promise.all([\n\t\t\tasr(buffer, {return_timestamps: speaker ? 'word' : false}),\n\t\t\t(!speaker || !token || !hasDiarization) ? Promise.resolve(): runDiarization(hasDiarization, f, modelDir, token),\n\t\t]);\n\n\t\tconst text = transcript.text?.trim() || null;\n\t\tif(!speaker) return parentPort?.postMessage({ text });\n\t\tif(!token) return parentPort?.postMessage({ text, error: 'HuggingFace token required' });\n\t\tif(!hasDiarization) return parentPort?.postMessage({ text, error: 'Speaker diarization unavailable' });\n\n\t\tconst combined = combineSpeakerTranscript(transcript.chunks || [], speakers || []);\n\t\tparentPort?.postMessage({ text: combined });\n\t} catch (err: any) {\n\t\tparentPort?.postMessage({ error: err.stack || err.message });\n\t} finally {\n\t\tif(tempFile) rmSync(tempFile, { recursive: true, force: true });\n\t}\n});\n"],"names":["canDiarization","checkPython","cmd","resolve","proc","spawn","code","runDiarization","binary","audioPath","dir","token","script","reject","output","data","combineSpeakerTranscript","chunks","speakers","speakerMap","speakerCount","seg","lines","currentSpeaker","currentText","chunk","time","speaker","s","speakerNum","prepareAudioBuffer","file","wav","tmp","wavefile","readFileSync","join","mkdtempSync","tmpdir","execSync","samples","left","right","buffer","i","parentPort","model","modelDir","tempFile","asr","pipeline","f","hasDiarization","transcript","text","combined","err"],"mappings":";;;;;;;AASA,eAAsBA,IAAyC;AAC9D,QAAMC,IAAc,CAACC,MACb,IAAI,QAAiB,CAACC,MAAY;AACxC,UAAMC,IAAOC,EAAMH,GAAK,CAAC,MAAM,uBAAuB,CAAC;AACvD,IAAAE,EAAK,GAAG,SAAS,CAACE,MAAiBH,EAAQG,MAAS,CAAC,CAAC,GACtDF,EAAK,GAAG,SAAS,MAAMD,EAAQ,EAAK,CAAC;AAAA,EACtC,CAAC;AAEF,SAAG,MAAMF,EAAY,SAAS,IAAU,YACrC,MAAMA,EAAY,QAAQ,IAAU,WAChC;AACR;AAEA,eAAeM,EAAeC,GAAgBC,GAAmBC,GAAaC,GAA+B;AAC5G,QAAMC,IAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,+BAMeF,CAAG;AAAA,iFAC+CC,CAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAUrF,SAAO,IAAI,QAAQ,CAACR,GAASU,MAAW;AACvC,QAAIC,IAAS;AACb,UAAMV,IAAOC,EAAMG,GAAQ,CAAC,MAAMI,GAAQH,CAAS,CAAC;AACpD,IAAAL,EAAK,OAAO,GAAG,QAAQ,CAACW,MAAiBD,KAAUC,EAAK,UAAU,GAClEX,EAAK,OAAO,GAAG,QAAQ,CAACW,MAAiB,QAAQ,MAAMA,EAAK,SAAA,CAAU,CAAC,GACvEX,EAAK,GAAG,SAAS,CAACE,MAAiB;AAClC,UAAGA,MAAS;AACX,YAAI;AACH,UAAAH,EAAQ,KAAK,MAAMW,CAAM,CAAC;AAAA,QAC3B,QAAc;AACb,UAAAD,EAAO,IAAI,MAAM,oCAAoC,CAAC;AAAA,QACvD;AAAA;AAEA,QAAAA,EAAO,IAAI,MAAM,mCAAmCP,CAAI,EAAE,CAAC;AAAA,IAE7D,CAAC,GACDF,EAAK,GAAG,SAASS,CAAM;AAAA,EACxB,CAAC;AACF;AAEA,SAASG,EAAyBC,GAAeC,GAAyB;AACzE,QAAMC,wBAAiB,IAAA;AACvB,MAAIC,IAAe;AACnB,EAAAF,EAAS,QAAQ,CAACG,MAAa;AAC9B,IAAIF,EAAW,IAAIE,EAAI,OAAO,KAAGF,EAAW,IAAIE,EAAI,SAAS,EAAED,CAAY;AAAA,EAC5E,CAAC;AAED,QAAME,IAAkB,CAAA;AACxB,MAAIC,IAAiB,IACjBC,IAAc;AAClB,SAAAP,EAAO,QAAQ,CAACQ,MAAe;AAC9B,UAAMC,IAAOD,EAAM,UAAU,CAAC,GACxBE,IAAUT,EAAS,KAAK,CAACU,MAAWF,KAAQE,EAAE,SAASF,KAAQE,EAAE,GAAG,GACpEC,IAAaF,IAAUR,EAAW,IAAIQ,EAAQ,OAAO,IAAI;AAC/D,IAAIE,MAAeN,KACfC,OAAmB,KAAK,YAAYD,CAAc,MAAMC,EAAY,KAAA,CAAM,EAAE,GAC/ED,IAAiBM,GACjBL,IAAcC,EAAM,QAEpBD,KAAeC,EAAM;AAAA,EAEvB,CAAC,GACED,OAAmB,KAAK,YAAYD,CAAc,MAAMC,EAAY,KAAA,CAAM,EAAE,GACxEF,EAAM,KAAK;AAAA,CAAI;AACvB;AAEA,SAASQ,EAAmBC,GAAsC;AACjE,MAAIC,GAAUC;AACd,MAAI;AACH,IAAAD,IAAM,IAAIE,EAAS,SAASC,EAAaJ,CAAI,CAAC;AAAA,EAC/C,QAAa;AACZ,IAAAE,IAAMG,EAAKC,EAAYD,EAAKE,EAAA,GAAU,QAAQ,CAAC,GAAG,eAAe,GACjEC,EAAS,cAAcR,CAAI,6BAA6BE,CAAG,KAAK,EAAE,OAAO,UAAU,GACnFD,IAAM,IAAIE,EAAS,SAASC,EAAaF,CAAG,CAAC;AAAA,EAC9C,UAAA;AACC,IAAAD,EAAI,WAAW,KAAK,GACpBA,EAAI,aAAa,IAAK;AACtB,UAAMQ,IAAUR,EAAI,WAAA;AACpB,QAAG,MAAM,QAAQQ,CAAO,GAAG;AAC1B,YAAMC,IAAOD,EAAQ,CAAC,GAChBE,IAAQF,EAAQ,CAAC,GACjBG,IAAS,IAAI,aAAaF,EAAK,MAAM;AAC3C,eAASG,IAAI,GAAGA,IAAIH,EAAK,QAAQG,IAAK,CAAAD,EAAOC,CAAC,KAAKH,EAAKG,CAAC,IAAIF,EAAME,CAAC,KAAK;AACzE,aAAO,CAACX,KAAOF,GAAMY,CAAM;AAAA,IAC5B;AACA,WAAO,CAACV,KAAOF,GAAMS,CAAO;AAAA,EAC7B;AACD;AAEAK,GAAY,GAAG,WAAW,OAAO,EAAE,MAAAd,GAAM,SAAAJ,GAAS,OAAAmB,GAAO,UAAAC,GAAU,OAAApC,QAAY;AAC9E,MAAIqC,IAAW;AACf,MAAI;AACH,UAAMC,IAAW,MAAMC,EAAS,gCAAgC,UAAUJ,CAAK,IAAI,EAAC,WAAWC,GAAU,WAAW,GAAA,CAAK,GACnH,CAACI,GAAGR,CAAM,IAAIb,EAAmBC,CAAI;AAC3C,IAAAiB,IAAWG,MAAMpB,IAAOoB,IAAI;AAC5B,UAAMC,IAAiB,MAAMpD,EAAA,GACvB,CAACqD,GAAYnC,CAAQ,IAAI,MAAM,QAAQ,IAAI;AAAA,MAChD+B,EAAIN,GAAQ,EAAC,mBAAmBhB,IAAU,SAAS,IAAM;AAAA,MACxD,CAACA,KAAW,CAAChB,KAAS,CAACyC,IAAkB,QAAQ,QAAA,IAAW7C,EAAe6C,GAAgBD,GAAGJ,GAAUpC,CAAK;AAAA,IAAA,CAC9G,GAEK2C,IAAOD,EAAW,MAAM,KAAA,KAAU;AACxC,QAAG,CAAC1B,EAAS,QAAOkB,GAAY,YAAY,EAAE,MAAAS,GAAM;AACpD,QAAG,CAAC3C,EAAO,QAAOkC,GAAY,YAAY,EAAE,MAAAS,GAAM,OAAO,8BAA8B;AACvF,QAAG,CAACF,EAAgB,QAAOP,GAAY,YAAY,EAAE,MAAAS,GAAM,OAAO,mCAAmC;AAErG,UAAMC,IAAWvC,EAAyBqC,EAAW,UAAU,CAAA,GAAInC,KAAY,EAAE;AACjF,IAAA2B,GAAY,YAAY,EAAE,MAAMU,EAAA,CAAU;AAAA,EAC3C,SAASC,GAAU;AAClB,IAAAX,GAAY,YAAY,EAAE,OAAOW,EAAI,SAASA,EAAI,SAAS;AAAA,EAC5D,UAAA;AACC,IAAGR,OAAiBA,GAAU,EAAE,WAAW,IAAM,OAAO,IAAM;AAAA,EAC/D;AACD,CAAC;"}
|
package/dist/index.js
CHANGED
|
@@ -1,31 +1,31 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports,Symbol.toStringTag,{value:"Module"});const R=require("node:os"),f=require("@ztimson/utils"),$=require("@anthropic-ai/sdk"),v=require("openai"),x=require("worker_threads"),S=require("url"),w=require("path"),
|
|
1
|
+
"use strict";Object.defineProperty(exports,Symbol.toStringTag,{value:"Module"});const R=require("node:os"),f=require("@ztimson/utils"),$=require("@anthropic-ai/sdk"),v=require("openai"),x=require("worker_threads"),S=require("url"),w=require("path"),T=require("./asr.js"),L=require("tesseract.js");require("./embedder.js");const N=require("cheerio"),j=require("@ztimson/node-utils");var g=typeof document<"u"?document.currentScript:null;function q(h){const r=Object.create(null,{[Symbol.toStringTag]:{value:"Module"}});if(h){for(const e in h)if(e!=="default"){const t=Object.getOwnPropertyDescriptor(h,e);Object.defineProperty(r,e,t.get?t:{enumerable:!0,get:()=>h[e]})}}return r.default=h,Object.freeze(r)}const C=q(R),D=q(N);class k{}class P extends k{constructor(r,e,t){super(),this.ai=r,this.apiToken=e,this.model=t,this.client=new $.Anthropic({apiKey:e})}client;toStandard(r){const e=Date.now(),t=[];for(let c of r)if(typeof c.content=="string")t.push({timestamp:e,...c});else{const n=c.content?.filter(s=>s.type=="text").map(s=>s.text).join(`
|
|
2
2
|
|
|
3
|
-
`);
|
|
3
|
+
`);n&&t.push({timestamp:e,role:c.role,content:n}),c.content.forEach(s=>{if(s.type=="tool_use")t.push({timestamp:e,role:"tool",id:s.id,name:s.name,args:s.input,content:void 0});else if(s.type=="tool_result"){const l=t.findLast(o=>o.id==s.tool_use_id);l&&(l[s.is_error?"error":"content"]=s.content)}})}return t}fromStandard(r){for(let e=0;e<r.length;e++)if(r[e].role=="tool"){const t=r[e];r.splice(e,1,{role:"assistant",content:[{type:"tool_use",id:t.id,name:t.name,input:t.args}]},{role:"user",content:[{type:"tool_result",tool_use_id:t.id,is_error:!!t.error,content:t.error||t.content}]}),e++}return r.map(({timestamp:e,...t})=>t)}ask(r,e={}){const t=new AbortController;return Object.assign(new Promise(async c=>{let n=this.fromStandard([...e.history||[],{role:"user",content:r,timestamp:Date.now()}]);const s=e.tools||this.ai.options.llm?.tools||[],l={model:e.model||this.model,max_tokens:e.max_tokens||this.ai.options.llm?.max_tokens||4096,system:e.system||this.ai.options.llm?.system||"",temperature:e.temperature||this.ai.options.llm?.temperature||.7,tools:s.map(m=>({name:m.name,description:m.description,input_schema:{type:"object",properties:m.args?f.objectMap(m.args,(i,u)=>({...u,required:void 0})):{},required:m.args?Object.entries(m.args).filter(i=>i[1].required).map(i=>i[0]):[]},fn:void 0})),messages:n,stream:!!e.stream};let o,a=!0;do{if(o=await this.client.messages.create(l).catch(i=>{throw i.message+=`
|
|
4
4
|
|
|
5
5
|
Messages:
|
|
6
|
-
${JSON.stringify(
|
|
6
|
+
${JSON.stringify(n,null,2)}`,i}),e.stream){a?a=!1:e.stream({text:`
|
|
7
7
|
|
|
8
|
-
`}),o.content=[];for await(const i of o){if(t.signal.aborted)break;if(i.type==="content_block_start")i.content_block.type==="text"?o.content.push({type:"text",text:""}):i.content_block.type==="tool_use"&&o.content.push({type:"tool_use",id:i.content_block.id,name:i.content_block.name,input:""});else if(i.type==="content_block_delta")if(i.delta.type==="text_delta"){const
|
|
8
|
+
`}),o.content=[];for await(const i of o){if(t.signal.aborted)break;if(i.type==="content_block_start")i.content_block.type==="text"?o.content.push({type:"text",text:""}):i.content_block.type==="tool_use"&&o.content.push({type:"tool_use",id:i.content_block.id,name:i.content_block.name,input:""});else if(i.type==="content_block_delta")if(i.delta.type==="text_delta"){const u=i.delta.text;o.content.at(-1).text+=u,e.stream({text:u})}else i.delta.type==="input_json_delta"&&(o.content.at(-1).input+=i.delta.partial_json);else if(i.type==="content_block_stop"){const u=o.content.at(-1);u.input!=null&&(u.input=u.input?f.JSONAttemptParse(u.input,{}):{})}else if(i.type==="message_stop")break}}const m=o.content.filter(i=>i.type==="tool_use");if(m.length&&!t.signal.aborted){n.push({role:"assistant",content:o.content});const i=await Promise.all(m.map(async u=>{const p=s.find(f.findByProp("name",u.name));if(e.stream&&e.stream({tool:u.name}),!p)return{tool_use_id:u.id,is_error:!0,content:"Tool not found"};try{const d=await p.fn(u.input,e?.stream,this.ai);return{type:"tool_result",tool_use_id:u.id,content:f.JSONSanitize(d)}}catch(d){return{type:"tool_result",tool_use_id:u.id,is_error:!0,content:d?.message||d?.toString()||"Unknown"}}}));n.push({role:"user",content:i}),l.messages=n}}while(!t.signal.aborted&&o.content.some(m=>m.type==="tool_use"));n.push({role:"assistant",content:o.content.filter(m=>m.type=="text").map(m=>m.text).join(`
|
|
9
9
|
|
|
10
|
-
`)}),
|
|
10
|
+
`)}),n=this.toStandard(n),e.stream&&e.stream({done:!0}),e.history&&e.history.splice(0,e.history.length,...n),c(n.at(-1)?.content)}),{abort:()=>t.abort()})}}class _ extends k{constructor(r,e,t,c){super(),this.ai=r,this.host=e,this.token=t,this.model=c,this.client=new v.OpenAI(f.clean({baseURL:e,apiKey:t}))}client;toStandard(r){for(let e=0;e<r.length;e++){const t=r[e];if(t.role==="assistant"&&t.tool_calls){const c=t.tool_calls.map(n=>({role:"tool",id:n.id,name:n.function.name,args:f.JSONAttemptParse(n.function.arguments,{}),timestamp:t.timestamp}));r.splice(e,1,...c),e+=c.length-1}else if(t.role==="tool"&&t.content){const c=r.find(n=>t.tool_call_id==n.id);c&&(t.content.includes('"error":')?c.error=t.content:c.content=t.content),r.splice(e,1),e--}r[e]?.timestamp||(r[e].timestamp=Date.now())}return r}fromStandard(r){return r.reduce((e,t)=>{if(t.role==="tool")e.push({role:"assistant",content:null,tool_calls:[{id:t.id,type:"function",function:{name:t.name,arguments:JSON.stringify(t.args)}}],refusal:null,annotations:[]},{role:"tool",tool_call_id:t.id,content:t.error||t.content});else{const{timestamp:c,...n}=t;e.push(n)}return e},[])}ask(r,e={}){const t=new AbortController;return Object.assign(new Promise(async(c,n)=>{e.system&&e.history?.[0]?.role!="system"&&e.history?.splice(0,0,{role:"system",content:e.system,timestamp:Date.now()});let s=this.fromStandard([...e.history||[],{role:"user",content:r,timestamp:Date.now()}]);const l=e.tools||this.ai.options.llm?.tools||[],o={model:e.model||this.model,messages:s,stream:!!e.stream,max_tokens:e.max_tokens||this.ai.options.llm?.max_tokens||4096,temperature:e.temperature||this.ai.options.llm?.temperature||.7,tools:l.map(i=>({type:"function",function:{name:i.name,description:i.description,parameters:{type:"object",properties:i.args?f.objectMap(i.args,(u,p)=>({...p,required:void 0})):{},required:i.args?Object.entries(i.args).filter(u=>u[1].required).map(u=>u[0]):[]}}}))};let a,m=!0;do{if(a=await this.client.chat.completions.create(o).catch(u=>{throw u.message+=`
|
|
11
11
|
|
|
12
12
|
Messages:
|
|
13
|
-
${JSON.stringify(
|
|
13
|
+
${JSON.stringify(s,null,2)}`,u}),e.stream){m?m=!1:e.stream({text:`
|
|
14
14
|
|
|
15
|
-
`}),a.choices=[{message:{content:"",tool_calls:[]}}];for await(const
|
|
15
|
+
`}),a.choices=[{message:{content:"",tool_calls:[]}}];for await(const u of a){if(t.signal.aborted)break;u.choices[0].delta.content&&(a.choices[0].message.content+=u.choices[0].delta.content,e.stream({text:u.choices[0].delta.content})),u.choices[0].delta.tool_calls&&(a.choices[0].message.tool_calls=u.choices[0].delta.tool_calls)}}const i=a.choices[0].message.tool_calls||[];if(i.length&&!t.signal.aborted){s.push(a.choices[0].message);const u=await Promise.all(i.map(async p=>{const d=l?.find(f.findByProp("name",p.function.name));if(e.stream&&e.stream({tool:p.function.name}),!d)return{role:"tool",tool_call_id:p.id,content:'{"error": "Tool not found"}'};try{const y=f.JSONAttemptParse(p.function.arguments,{}),b=await d.fn(y,e.stream,this.ai);return{role:"tool",tool_call_id:p.id,content:f.JSONSanitize(b)}}catch(y){return{role:"tool",tool_call_id:p.id,content:f.JSONSanitize({error:y?.message||y?.toString()||"Unknown"})}}}));s.push(...u),o.messages=s}}while(!t.signal.aborted&&a.choices?.[0]?.message?.tool_calls?.length);s.push({role:"assistant",content:a.choices[0].message.content||""}),s=this.toStandard(s),e.stream&&e.stream({done:!0}),e.history&&e.history.splice(0,e.history.length,...s),c(s.at(-1)?.content)}),{abort:()=>t.abort()})}}class J{constructor(r){this.ai=r,r.options.llm?.models&&Object.entries(r.options.llm.models).forEach(([e,t])=>{this.defaultModel||(this.defaultModel=e),t.proto=="anthropic"?this.models[e]=new P(this.ai,t.token,e):t.proto=="ollama"?this.models[e]=new _(this.ai,t.host,"not-needed",e):t.proto=="openai"&&(this.models[e]=new _(this.ai,t.host||null,t.token,e))})}defaultModel;models={};ask(r,e={}){const t=e.model||this.defaultModel;if(!this.models[t])throw new Error(`Model does not exist: ${t}`);let c=()=>{};return Object.assign(new Promise(async n=>{if(e.history||(e.history=[]),e.memory){e.system=(e.system||"")+`
|
|
16
16
|
You have passive persistent memory never make any mention of your memory capabilities and what you can/cannot remember
|
|
17
|
-
`;const l=async(a,m,i=50)=>{const[
|
|
17
|
+
`;const l=async(a,m,i=50)=>{const[u,p]=await Promise.all([m?this.embedding(m):Promise.resolve(null),a?this.embedding(a):Promise.resolve(null)]);return(e.memory||[]).map(d=>({...d,score:u?this.cosineSimilarity(d.embeddings[0],u[0].embedding):1})).filter(d=>d.score>=.8).map(d=>({...d,score:p?this.cosineSimilarity(d.embeddings[1],p[0].embedding):d.score})).filter(d=>d.score>=.2).toSorted((d,y)=>d.score-y.score).slice(0,i)},o=await l(r);o.length&&e.history.push({role:"assistant",content:`Things I remembered:
|
|
18
18
|
`+o.map(a=>`${a.owner}: ${a.fact}`).join(`
|
|
19
|
-
`)}),e.tools=[...e.tools||[],{name:"read_memory",description:"Check your long-term memory for more information",args:{subject:{type:"string",description:"Find information by a subject topic, can be used with or without query argument"},query:{type:"string",description:"Search memory based on a query, can be used with or without subject argument"},limit:{type:"number",description:"Result limit, default 5"}},fn:a=>{if(!a.subject&&!a.query)throw new Error("Either a subject or query argument is required");return l(a.query,a.subject,a.limit||5)}}]}const
|
|
19
|
+
`)}),e.tools=[...e.tools||[],{name:"read_memory",description:"Check your long-term memory for more information",args:{subject:{type:"string",description:"Find information by a subject topic, can be used with or without query argument"},query:{type:"string",description:"Search memory based on a query, can be used with or without subject argument"},limit:{type:"number",description:"Result limit, default 5"}},fn:a=>{if(!a.subject&&!a.query)throw new Error("Either a subject or query argument is required");return l(a.query,a.subject,a.limit||5)}}]}const s=await this.models[t].ask(r,e);if(e.memory){const l=e.history?.findIndex(o=>o.role=="assistant"&&o.content.startsWith("Things I remembered:"));l!=null&&l>=0&&e.history?.splice(l,1)}if(e.compress||e.memory){let l=null;if(e.compress)l=await this.ai.language.compressHistory(e.history,e.compress.max,e.compress.min,e),e.history.splice(0,e.history.length,...l.history);else{const o=e.history?.findLastIndex(a=>a.role=="user")??-1;l=await this.ai.language.compressHistory(o!=-1?e.history.slice(o):e.history,0,0,e)}if(e.memory){const o=e.memory.filter(a=>!l.memory.some(m=>this.cosineSimilarity(a.embeddings[1],m.embeddings[1])>.8)).concat(l.memory);e.memory.splice(0,e.memory.length,...o)}}return n(s)}),{abort:c})}async compressHistory(r,e,t,c){if(this.estimateTokens(r)<e)return{history:r,memory:[]};let n=0,s=0;for(let d of r.toReversed())if(s+=this.estimateTokens(d.content),s<t)n++;else break;if(r.length<=n)return{history:r,memory:[]};const l=r[0].role=="system"?r[0]:null,o=n==0?[]:r.slice(-n),a=(n==0?r:r.slice(0,-n)).filter(d=>d.role==="assistant"||d.role==="user"),m=await this.json(a.map(d=>`${d.role}: ${d.content}`).join(`
|
|
20
20
|
|
|
21
|
-
`),"{summary: string, facts: [[subject, fact]]}",{system:"Create the smallest summary possible, no more than 500 tokens. Create a list of NEW facts (split by subject [pro]noun and fact) about what you learned from this conversation that you didn't already know or get from a tool call or system prompt. Focus only on new information about people, topics, or facts. Avoid generating facts about the AI.",model:c?.model,temperature:c?.temperature||.3}),i=new Date,
|
|
21
|
+
`),"{summary: string, facts: [[subject, fact]]}",{system:"Create the smallest summary possible, no more than 500 tokens. Create a list of NEW facts (split by subject [pro]noun and fact) about what you learned from this conversation that you didn't already know or get from a tool call or system prompt. Focus only on new information about people, topics, or facts. Avoid generating facts about the AI.",model:c?.model,temperature:c?.temperature||.3}),i=new Date,u=await Promise.all((m?.facts||[])?.map(async([d,y])=>{const b=await Promise.all([this.embedding(d),this.embedding(`${d}: ${y}`)]);return{owner:d,fact:y,embeddings:[b[0][0].embedding,b[1][0].embedding],timestamp:i}})),p=[{role:"assistant",content:`Conversation Summary: ${m?.summary}`,timestamp:Date.now()},...o];return l&&p.splice(0,0,l),{history:p,memory:u}}cosineSimilarity(r,e){if(r.length!==e.length)throw new Error("Vectors must be same length");let t=0,c=0,n=0;for(let l=0;l<r.length;l++)t+=r[l]*e[l],c+=r[l]*r[l],n+=e[l]*e[l];const s=Math.sqrt(c)*Math.sqrt(n);return s===0?0:t/s}chunk(r,e=500,t=50){const c=(o,a="")=>o?Object.entries(o).flatMap(([m,i])=>{const u=a?`${a}${isNaN(+m)?`.${m}`:`[${m}]`}`:m;return typeof i=="object"&&!Array.isArray(i)?c(i,u):`${u}: ${Array.isArray(i)?i.join(", "):i}`}):[],s=(typeof r=="object"?c(r):r.split(`
|
|
22
22
|
`)).flatMap(o=>[...o.split(/\s+/).filter(Boolean),`
|
|
23
|
-
`]),l=[];for(let o=0;o<
|
|
24
|
-
`))>e&&a)break;a=
|
|
25
|
-
`).trim();i&&l.push(i),o=Math.max(m-t,m===o?o+1:m)}return l}async embedding(r,e={}){let{maxTokens:t=500,overlapTokens:c=50}=e;const
|
|
23
|
+
`]),l=[];for(let o=0;o<s.length;){let a="",m=o;for(;m<s.length;){const u=a+(a?" ":"")+s[m];if(this.estimateTokens(u.replace(/\s*\n\s*/g,`
|
|
24
|
+
`))>e&&a)break;a=u,m++}const i=a.replace(/\s*\n\s*/g,`
|
|
25
|
+
`).trim();i&&l.push(i),o=Math.max(m-t,m===o?o+1:m)}return l}async embedding(r,e={}){let{maxTokens:t=500,overlapTokens:c=50}=e;const n=o=>new Promise((a,m)=>{const i=new x.Worker(w.join(w.dirname(S.fileURLToPath(typeof document>"u"?require("url").pathToFileURL(__filename).href:g&&g.tagName.toUpperCase()==="SCRIPT"&&g.src||new URL("index.js",document.baseURI).href)),"embedder.js")),u=({embedding:d})=>{i.terminate(),a(d)},p=d=>{i.terminate(),m(d)};i.on("message",u),i.on("error",p),i.on("exit",d=>{d!==0&&m(new Error(`Worker exited with code ${d}`))}),i.postMessage({text:o,model:this.ai.options?.embedder||"bge-small-en-v1.5",modelDir:this.ai.options.path})}),s=this.chunk(r,t,c),l=[];for(let o=0;o<s.length;o++){const a=s[o],m=await n(a);l.push({index:o,embedding:m,text:a,tokens:this.estimateTokens(a)})}return l}estimateTokens(r){const e=JSON.stringify(r);return Math.ceil(e.length/4*1.2)}fuzzyMatch(r,...e){if(e.length<2)throw new Error("Requires at least 2 strings to compare");const t=(s,l=10)=>s.toLowerCase().split("").map((o,a)=>o.charCodeAt(0)*(a+1)%l/l).slice(0,l),c=t(r),n=e.map(s=>t(s)).map(s=>this.cosineSimilarity(c,s));return{avg:n.reduce((s,l)=>s+l,0)/n.length,max:Math.max(...n),similarities:n}}async json(r,e,t){let c=await this.ask(r,{...t,system:(t?.system?`${t.system}
|
|
26
26
|
`:"")+`Only respond using a JSON code block matching this schema:
|
|
27
27
|
\`\`\`json
|
|
28
28
|
${e}
|
|
29
|
-
\`\`\``});if(!c)return{};const
|
|
30
|
-
`),'{1: "Detected Name"}',{system:"Use
|
|
29
|
+
\`\`\``});if(!c)return{};const n=/```(?:.+)?\s*([\s\S]*?)```/.exec(c),s=n?n[1].trim():c;return f.JSONAttemptParse(s,{})}summarize(r,e,t){return this.ask(r,{system:`Generate a brief summary <= ${e} tokens. Output nothing else`,temperature:.3,...t})}}class O{constructor(r){this.ai=r}asr(r,e={}){const{model:t=this.ai.options.asr||"whisper-base",speaker:c=!1}=e;let n=!1;const s=()=>{n=!0};let l=new Promise((o,a)=>{const m=new x.Worker(w.join(w.dirname(S.fileURLToPath(typeof document>"u"?require("url").pathToFileURL(__filename).href:g&&g.tagName.toUpperCase()==="SCRIPT"&&g.src||new URL("index.js",document.baseURI).href)),"asr.js")),i=({text:p,warning:d,error:y})=>{setTimeout(()=>m.terminate(),1e3),!n&&(y?a(new Error(y)):(d&&console.warn(d),o(p)))},u=p=>{setTimeout(()=>m.terminate(),1e3),n||a(p)};m.on("message",i),m.on("error",u),m.on("exit",p=>{p!==0&&!n&&a(new Error(`Worker exited with code ${p}`))}),m.postMessage({file:r,model:t,speaker:c,modelDir:this.ai.options.path,token:this.ai.options.hfToken})});if(e.speaker=="id"){if(!this.ai.language.defaultModel)throw new Error("Configure an LLM for advanced ASR speaker detection");l=l.then(async o=>{if(!o)return o;let a=this.ai.language.chunk(o,500,0);a.length>4&&(a=[...a.slice(0,3),a.at(-1)]);const m=await this.ai.language.json(a.join(`
|
|
30
|
+
`),'{1: "Detected Name", 2: "Second Name"}',{system:"Use the following transcript to identify speakers. Only identify speakers you are positive about, dont mention speakers you are unsure about in your response",temperature:.1});return Object.entries(m).forEach(([i,u])=>{o=o.replaceAll(`[Speaker ${i}]`,`[${u}]`)}),o})}return Object.assign(l,{abort:s})}canDiarization=()=>T.canDiarization().then(r=>!!r)}class M{constructor(r){this.ai=r}ocr(r){let e;const t=new Promise(async c=>{e=await L.createWorker(this.ai.options.ocr||"eng",2,{cachePath:this.ai.options.path});const{data:n}=await e.recognize(r);await e.terminate(),c(n.text.trim()||null)});return Object.assign(t,{abort:()=>e?.terminate()})}}class W{constructor(r){this.options=r,r.path||(r.path=C.tmpdir()),process.env.TRANSFORMERS_CACHE=r.path,this.audio=new O(this),this.language=new J(this),this.vision=new M(this)}audio;language;vision}const A={name:"cli",description:"Use the command line interface, returns any output",args:{command:{type:"string",description:"Command to run",required:!0}},fn:h=>j.$`${h.command}`},z={name:"get_datetime",description:"Get current UTC date / time",args:{},fn:async()=>new Date().toUTCString()},I={name:"exec",description:"Run code/scripts",args:{language:{type:"string",description:"Execution language",enum:["cli","node","python"],required:!0},code:{type:"string",description:"Code to execute",required:!0}},fn:async(h,r,e)=>{try{switch(h.type){case"bash":return await A.fn({command:h.code},r,e);case"node":return await U.fn({code:h.code},r,e);case"python":return await E.fn({code:h.code},r,e)}}catch(t){return{error:t?.message||t.toString()}}}},F={name:"fetch",description:"Make HTTP request to URL",args:{url:{type:"string",description:"URL to fetch",required:!0},method:{type:"string",description:"HTTP method to use",enum:["GET","POST","PUT","DELETE"],default:"GET"},headers:{type:"object",description:"HTTP headers to send",default:{}},body:{type:"object",description:"HTTP body to send"}},fn:h=>new f.Http({url:h.url,headers:h.headers}).request({method:h.method||"GET",body:h.body})},U={name:"exec_javascript",description:"Execute commonjs javascript",args:{code:{type:"string",description:"CommonJS javascript",required:!0}},fn:async h=>{const r=f.consoleInterceptor(null),e=await f.fn({console:r},h.code,!0).catch(t=>r.output.error.push(t));return{...r.output,return:e,stdout:void 0,stderr:void 0}}},E={name:"exec_javascript",description:"Execute commonjs javascript",args:{code:{type:"string",description:"CommonJS javascript",required:!0}},fn:async h=>({result:j.$Sync`python -c "${h.code}"`})},H={name:"read_webpage",description:"Extract clean, structured content from a webpage. Use after web_search to read specific URLs",args:{url:{type:"string",description:"URL to extract content from",required:!0},focus:{type:"string",description:'Optional: What aspect to focus on (e.g., "pricing", "features", "contact info")'}},fn:async h=>{const r=await fetch(h.url,{headers:{"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64)"}}).then(s=>s.text()).catch(s=>{throw new Error(`Failed to fetch: ${s.message}`)}),e=D.load(r);e('script, style, nav, footer, header, aside, iframe, noscript, [role="navigation"], [role="banner"], .ad, .ads, .cookie, .popup').remove();const t={title:e('meta[property="og:title"]').attr("content")||e("title").text()||"",description:e('meta[name="description"]').attr("content")||e('meta[property="og:description"]').attr("content")||""};let c="";const n=["article","main",'[role="main"]',".content",".post",".entry","body"];for(const s of n){const l=e(s).first();if(l.length&&l.text().trim().length>200){c=l.text();break}}return c||(c=e("body").text()),c=c.replace(/\s+/g," ").trim().slice(0,8e3),{url:h.url,title:t.title.trim(),description:t.description.trim(),content:c,focus:h.focus}}},B={name:"web_search",description:"Use duckduckgo (anonymous) to find find relevant online resources. Returns a list of URLs that works great with the `read_webpage` tool",args:{query:{type:"string",description:"Search string",required:!0},length:{type:"string",description:"Number of results to return",default:5}},fn:async h=>{const r=await fetch(`https://html.duckduckgo.com/html/?q=${encodeURIComponent(h.query)}`,{headers:{"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64)","Accept-Language":"en-US,en;q=0.9"}}).then(n=>n.text());let e,t=/<a .*?href="(.+?)".+?<\/a>/g;const c=new f.ASet;for(;(e=t.exec(r))!==null;){let n=/uddg=(.+)&?/.exec(decodeURIComponent(e[1]))?.[1];if(n&&(n=decodeURIComponent(n)),n&&c.add(n),c.size>=(h.length||5))break}return c}};exports.canDiarization=T.canDiarization;exports.Ai=W;exports.Anthropic=P;exports.Audio=O;exports.CliTool=A;exports.DateTimeTool=z;exports.ExecTool=I;exports.FetchTool=F;exports.JSTool=U;exports.LLMProvider=k;exports.OpenAi=_;exports.PythonTool=E;exports.ReadWebpageTool=H;exports.Vision=M;exports.WebSearchTool=B;
|
|
31
31
|
//# sourceMappingURL=index.js.map
|