sophontalk-services 0.0.3 → 0.0.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist-lib/sophontalk-services.cjs +1 -1
- package/dist-lib/sophontalk-services.js +43 -46
- package/dist-lib/workers/feature.worker.js +2060 -0
- package/dist-lib/workers/inference.worker.js +2500 -0
- package/package.json +2 -3
- package/dist-lib/assets/feature.worker-Dx6moind.js +0 -145
- package/dist-lib/assets/inference.worker-BPSCUMiL.js +0 -178
|
@@ -1 +1 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports,Symbol.toStringTag,{value:"Module"});const
|
|
1
|
+
"use strict";Object.defineProperty(exports,Symbol.toStringTag,{value:"Module"});const u=require("onnxruntime-web");var o=typeof document<"u"?document.currentScript:null;const _=16e3,f=25,m=10,w=80,R=4,p=.97,M=1e-10,k=20,F=!0,g=!0,l=25,L=1e3/l,S=s=>{if(s===128)return 4;if(s===96)return 3;throw new Error(`Unsupported crop size: ${s}. Only nano (96) and tiny (128) are supported.`)},O=(s,e,t)=>{if(e<=1)return{nextIndex:0,nextDirection:1};let r=s+t,n=t;return r>=e?(r=e-2,n=-1):r<0&&(r=1,n=1),{nextIndex:r,nextDirection:n}};async function T(s={}){const{jsonUrl:e="/complete_dataset.json",zipUrl:t="/processed_images.zip"}=s,[r,n]=await Promise.all([fetch(e),fetch(t)]);if(!r.ok)throw new Error(`无法加载 ${e}`);if(!n.ok)throw new Error(`无法加载 ${t}`);const i=await r.json(),E=await n.arrayBuffer();return{dataset:i,zipBuffer:E}}class P{worker=new Worker(new URL("./workers/feature.worker.js",typeof document>"u"?require("url").pathToFileURL(__filename).href:o&&o.tagName.toUpperCase()==="SCRIPT"&&o.src||new URL("sophontalk-services.cjs",document.baseURI).href),{type:"module"});async process(e){const t=e.getChannelData(0),r=e.numberOfChannels>1?e.getChannelData(1):void 0,n={leftChannel:t,rightChannel:r,sampleRate:e.sampleRate},i=[t.buffer];return r&&i.push(r.buffer),new Promise((E,h)=>{const c=a=>{a.data?.status==="success"?(this.worker.removeEventListener("message",c),this.worker.removeEventListener("error",d),E(a.data.payload)):a.data?.status==="error"&&(this.worker.removeEventListener("message",c),this.worker.removeEventListener("error",d),h(new Error(a.data.error||"特征提取失败")))},d=a=>{this.worker.removeEventListener("message",c),this.worker.removeEventListener("error",d),h(a)};this.worker.addEventListener("message",c),this.worker.addEventListener("error",d),this.worker.postMessage(n,i)})}dispose(){this.worker.terminate()}}class U{worker;handlers={};constructor(e={}){this.handlers=e,this.worker=new Worker(new URL("./workers/inference.worker.js",typeof document>"u"?require("url").pathToFileURL(__filename).href:o&&o.tagName.toUpperCase()==="SCRIPT"&&o.src||new URL("sophontalk-services.cjs",document.baseURI).href),{type:"module"}),this.worker.onmessage=this.handleMessage.bind(this)}handleMessage(e){const{type:t,payload:r}=e.data;switch(t){case"ready":this.handlers.onReady?.();break;case"start":this.handlers.onStart?.(r.totalFrames);break;case"frame":this.handlers.onFrame?.(r.frame,r.index);break;case"done":this.handlers.onDone?.();break;case"error":this.handlers.onError?.(r);break}}init(e){this.worker.postMessage({type:"init",modelPath:e.modelPath,dataset:e.dataset,zipBuffer:e.zipBuffer,blendingMask:e.blendingMask,wasmPaths:e.wasmPaths||u.env.wasm.wasmPaths},[e.zipBuffer,e.blendingMask])}run(e){this.worker.postMessage({type:"run",audioFeatures:e.audioFeatures,audioDimensions:e.audioDimensions},[e.audioFeatures.buffer])}stop(){this.worker.postMessage({type:"stop"})}terminate(){this.worker.terminate()}}exports.DEFAULT_LOW_FREQ=k;exports.ENERGY_FLOOR=M;exports.FRAME_INTERVAL_MS=L;exports.FRAME_LENGTH_MS=f;exports.FRAME_SHIFT_MS=m;exports.FeatureExtractor=P;exports.InferenceEngine=U;exports.NUM_MEL_BINS=w;exports.NUM_SEQUENCE_FRAMES=R;exports.PREEMPH_COEFF=p;exports.REMOVE_DC_OFFSET=F;exports.ROUND_TO_POWER_OF_TWO=g;exports.SAMPLE_RATE=_;exports.TARGET_FPS=l;exports.calculatePingPongState=O;exports.getBorder=S;exports.loadDataset=T;
|
|
@@ -1,69 +1,66 @@
|
|
|
1
|
-
|
|
1
|
+
import { env as l } from "onnxruntime-web";
|
|
2
|
+
const m = 16e3, k = 25, u = 10, f = 80, p = 4, g = 0.97, _ = 1e-10, M = 20, R = !0, F = !0, w = 25, L = 1e3 / w, y = (s) => {
|
|
2
3
|
if (s === 128) return 4;
|
|
3
4
|
if (s === 96) return 3;
|
|
4
5
|
throw new Error(`Unsupported crop size: ${s}. Only nano (96) and tiny (128) are supported.`);
|
|
5
6
|
}, P = (s, e, t) => {
|
|
6
7
|
if (e <= 1) return { nextIndex: 0, nextDirection: 1 };
|
|
7
|
-
let r = s + t,
|
|
8
|
-
return r >= e ? (r = e - 2,
|
|
8
|
+
let r = s + t, a = t;
|
|
9
|
+
return r >= e ? (r = e - 2, a = -1) : r < 0 && (r = 1, a = 1), { nextIndex: r, nextDirection: a };
|
|
9
10
|
};
|
|
10
11
|
async function b(s = {}) {
|
|
11
12
|
const {
|
|
12
13
|
jsonUrl: e = "/complete_dataset.json",
|
|
13
14
|
zipUrl: t = "/processed_images.zip"
|
|
14
|
-
} = s, [r,
|
|
15
|
+
} = s, [r, a] = await Promise.all([
|
|
15
16
|
fetch(e),
|
|
16
17
|
fetch(t)
|
|
17
18
|
]);
|
|
18
19
|
if (!r.ok) throw new Error(`无法加载 ${e}`);
|
|
19
|
-
if (!
|
|
20
|
-
const o = await r.json(), c = await
|
|
20
|
+
if (!a.ok) throw new Error(`无法加载 ${t}`);
|
|
21
|
+
const o = await r.json(), c = await a.arrayBuffer();
|
|
21
22
|
return { dataset: o, zipBuffer: c };
|
|
22
23
|
}
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
24
|
+
class v {
|
|
25
|
+
worker = new Worker(
|
|
26
|
+
new URL(
|
|
27
|
+
/* @vite-ignore */
|
|
28
|
+
"./workers/feature.worker.js",
|
|
29
|
+
import.meta.url
|
|
30
|
+
),
|
|
31
|
+
{ type: "module" }
|
|
30
32
|
);
|
|
31
|
-
}
|
|
32
|
-
class O {
|
|
33
|
-
worker = new l();
|
|
34
33
|
async process(e) {
|
|
35
|
-
const t = e.getChannelData(0), r = e.numberOfChannels > 1 ? e.getChannelData(1) : void 0,
|
|
34
|
+
const t = e.getChannelData(0), r = e.numberOfChannels > 1 ? e.getChannelData(1) : void 0, a = {
|
|
36
35
|
leftChannel: t,
|
|
37
36
|
rightChannel: r,
|
|
38
37
|
sampleRate: e.sampleRate
|
|
39
38
|
}, o = [t.buffer];
|
|
40
39
|
return r && o.push(r.buffer), new Promise((c, d) => {
|
|
41
|
-
const i = (
|
|
42
|
-
|
|
43
|
-
}, h = (
|
|
44
|
-
this.worker.removeEventListener("message", i), this.worker.removeEventListener("error", h), d(
|
|
40
|
+
const i = (n) => {
|
|
41
|
+
n.data?.status === "success" ? (this.worker.removeEventListener("message", i), this.worker.removeEventListener("error", h), c(n.data.payload)) : n.data?.status === "error" && (this.worker.removeEventListener("message", i), this.worker.removeEventListener("error", h), d(new Error(n.data.error || "特征提取失败")));
|
|
42
|
+
}, h = (n) => {
|
|
43
|
+
this.worker.removeEventListener("message", i), this.worker.removeEventListener("error", h), d(n);
|
|
45
44
|
};
|
|
46
|
-
this.worker.addEventListener("message", i), this.worker.addEventListener("error", h), this.worker.postMessage(
|
|
45
|
+
this.worker.addEventListener("message", i), this.worker.addEventListener("error", h), this.worker.postMessage(a, o);
|
|
47
46
|
});
|
|
48
47
|
}
|
|
49
48
|
dispose() {
|
|
50
49
|
this.worker.terminate();
|
|
51
50
|
}
|
|
52
51
|
}
|
|
53
|
-
|
|
54
|
-
return new Worker(
|
|
55
|
-
"./assets/inference.worker-BPSCUMiL.js",
|
|
56
|
-
{
|
|
57
|
-
type: "module",
|
|
58
|
-
name: s?.name
|
|
59
|
-
}
|
|
60
|
-
);
|
|
61
|
-
}
|
|
62
|
-
class v {
|
|
52
|
+
class O {
|
|
63
53
|
worker;
|
|
64
54
|
handlers = {};
|
|
65
55
|
constructor(e = {}) {
|
|
66
|
-
this.handlers = e, this.worker = new
|
|
56
|
+
this.handlers = e, this.worker = new Worker(
|
|
57
|
+
new URL(
|
|
58
|
+
/* @vite-ignore */
|
|
59
|
+
"./workers/inference.worker.js",
|
|
60
|
+
import.meta.url
|
|
61
|
+
),
|
|
62
|
+
{ type: "module" }
|
|
63
|
+
), this.worker.onmessage = this.handleMessage.bind(this);
|
|
67
64
|
}
|
|
68
65
|
handleMessage(e) {
|
|
69
66
|
const { type: t, payload: r } = e.data;
|
|
@@ -93,7 +90,7 @@ class v {
|
|
|
93
90
|
dataset: e.dataset,
|
|
94
91
|
zipBuffer: e.zipBuffer,
|
|
95
92
|
blendingMask: e.blendingMask,
|
|
96
|
-
wasmPaths: e.wasmPaths
|
|
93
|
+
wasmPaths: e.wasmPaths || l.wasm.wasmPaths
|
|
97
94
|
},
|
|
98
95
|
[e.zipBuffer, e.blendingMask]
|
|
99
96
|
);
|
|
@@ -116,20 +113,20 @@ class v {
|
|
|
116
113
|
}
|
|
117
114
|
}
|
|
118
115
|
export {
|
|
119
|
-
|
|
116
|
+
M as DEFAULT_LOW_FREQ,
|
|
120
117
|
_ as ENERGY_FLOOR,
|
|
121
118
|
L as FRAME_INTERVAL_MS,
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
126
|
-
|
|
127
|
-
|
|
128
|
-
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
119
|
+
k as FRAME_LENGTH_MS,
|
|
120
|
+
u as FRAME_SHIFT_MS,
|
|
121
|
+
v as FeatureExtractor,
|
|
122
|
+
O as InferenceEngine,
|
|
123
|
+
f as NUM_MEL_BINS,
|
|
124
|
+
p as NUM_SEQUENCE_FRAMES,
|
|
125
|
+
g as PREEMPH_COEFF,
|
|
126
|
+
R as REMOVE_DC_OFFSET,
|
|
127
|
+
F as ROUND_TO_POWER_OF_TWO,
|
|
128
|
+
m as SAMPLE_RATE,
|
|
129
|
+
w as TARGET_FPS,
|
|
133
130
|
P as calculatePingPongState,
|
|
134
131
|
y as getBorder,
|
|
135
132
|
b as loadDataset
|