@tensamin/audio 0.2.5 → 0.2.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -9
- package/dist/{chunk-LDVWEWAA.mjs → chunk-47YTMONE.mjs} +2 -2
- package/dist/{chunk-BVDZ4WDU.mjs → chunk-7DOV5ZIH.mjs} +2 -2
- package/dist/{chunk-NYIMTVTV.mjs → chunk-FYX5A3OM.mjs} +2 -2
- package/dist/{chunk-YH7I47TT.mjs → chunk-HCEWI3JP.mjs} +29 -9
- package/dist/{chunk-Y3UA67NM.mjs → chunk-XHU4WPFD.mjs} +1 -1
- package/dist/defaults.js +2 -2
- package/dist/defaults.mjs +1 -1
- package/dist/index.js +29 -9
- package/dist/index.mjs +5 -5
- package/dist/livekit/integration.js +29 -9
- package/dist/livekit/integration.mjs +5 -5
- package/dist/pipeline/audio-pipeline.js +2 -2
- package/dist/pipeline/audio-pipeline.mjs +3 -3
- package/dist/pipeline/remote-audio-monitor.js +1 -1
- package/dist/pipeline/remote-audio-monitor.mjs +3 -3
- package/dist/vad/vad-state.js +1 -1
- package/dist/vad/vad-state.mjs +2 -2
- package/package.json +2 -2
package/README.md
CHANGED
|
@@ -87,12 +87,4 @@ output: {
|
|
|
87
87
|
maxGainDb?: number; // default: 6.0
|
|
88
88
|
smoothTransitions?: boolean;// default: true
|
|
89
89
|
}
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
### LiveKit mute handling
|
|
93
|
-
|
|
94
|
-
```ts
|
|
95
|
-
muteWhenSilent?: boolean; // default: false
|
|
96
|
-
````
|
|
97
|
-
|
|
98
|
-
When `muteWhenSilent` is `true`, the library automatically calls `track.mute()` when silence is detected and `track.unmute()` when speech resumes (only if it muted the track itself).
|
|
90
|
+
```
|
|
@@ -3,11 +3,11 @@ import {
|
|
|
3
3
|
} from "./chunk-QNQK6QFB.mjs";
|
|
4
4
|
import {
|
|
5
5
|
LevelBasedVAD
|
|
6
|
-
} from "./chunk-
|
|
6
|
+
} from "./chunk-XHU4WPFD.mjs";
|
|
7
7
|
import {
|
|
8
8
|
DEFAULT_REMOTE_SPEAKING_OPTIONS,
|
|
9
9
|
DEFAULT_SPEAKING_DETECTION_CONFIG
|
|
10
|
-
} from "./chunk-
|
|
10
|
+
} from "./chunk-FYX5A3OM.mjs";
|
|
11
11
|
import {
|
|
12
12
|
getAudioContext,
|
|
13
13
|
registerPipeline,
|
|
@@ -6,13 +6,13 @@ import {
|
|
|
6
6
|
} from "./chunk-QNQK6QFB.mjs";
|
|
7
7
|
import {
|
|
8
8
|
LevelBasedVAD
|
|
9
|
-
} from "./chunk-
|
|
9
|
+
} from "./chunk-XHU4WPFD.mjs";
|
|
10
10
|
import {
|
|
11
11
|
DEFAULT_LIVEKIT_SPEAKING_OPTIONS,
|
|
12
12
|
DEFAULT_NOISE_SUPPRESSION_CONFIG,
|
|
13
13
|
DEFAULT_OUTPUT_GAIN_CONFIG,
|
|
14
14
|
DEFAULT_SPEAKING_DETECTION_CONFIG
|
|
15
|
-
} from "./chunk-
|
|
15
|
+
} from "./chunk-FYX5A3OM.mjs";
|
|
16
16
|
import {
|
|
17
17
|
getAudioContext,
|
|
18
18
|
registerPipeline,
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
// src/defaults.ts
|
|
2
2
|
var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
3
|
-
minDb: -
|
|
3
|
+
minDb: -60,
|
|
4
4
|
maxDb: -20,
|
|
5
5
|
speakOnRatio: 0.5,
|
|
6
6
|
speakOffRatio: 0.3,
|
|
@@ -10,7 +10,7 @@ var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
|
10
10
|
};
|
|
11
11
|
var DEFAULT_NOISE_SUPPRESSION_CONFIG = {
|
|
12
12
|
enabled: true,
|
|
13
|
-
noiseReductionLevel:
|
|
13
|
+
noiseReductionLevel: 60
|
|
14
14
|
};
|
|
15
15
|
var DEFAULT_OUTPUT_GAIN_CONFIG = {
|
|
16
16
|
speechGain: 1,
|
|
@@ -1,9 +1,9 @@
|
|
|
1
1
|
import {
|
|
2
2
|
createAudioPipeline
|
|
3
|
-
} from "./chunk-
|
|
3
|
+
} from "./chunk-7DOV5ZIH.mjs";
|
|
4
4
|
import {
|
|
5
5
|
createRemoteAudioMonitor
|
|
6
|
-
} from "./chunk-
|
|
6
|
+
} from "./chunk-47YTMONE.mjs";
|
|
7
7
|
|
|
8
8
|
// src/livekit/integration.ts
|
|
9
9
|
async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
@@ -21,16 +21,30 @@ async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
|
21
21
|
const listeners = /* @__PURE__ */ new Set();
|
|
22
22
|
let mutedByController = false;
|
|
23
23
|
let currentState = pipeline.state;
|
|
24
|
+
let gainNode = null;
|
|
25
|
+
if (options.muteWhenSilent) {
|
|
26
|
+
try {
|
|
27
|
+
const audioContext = new AudioContext();
|
|
28
|
+
const source = audioContext.createMediaStreamSource(
|
|
29
|
+
new MediaStream([pipeline.processedTrack])
|
|
30
|
+
);
|
|
31
|
+
gainNode = audioContext.createGain();
|
|
32
|
+
source.connect(gainNode);
|
|
33
|
+
gainNode.connect(audioContext.destination);
|
|
34
|
+
} catch (error) {
|
|
35
|
+
console.error("Failed to create gain node for volume control", error);
|
|
36
|
+
}
|
|
37
|
+
}
|
|
24
38
|
const speakingHandler = (state) => {
|
|
25
39
|
currentState = state;
|
|
26
40
|
listeners.forEach((listener) => listener(state));
|
|
27
|
-
if (options.muteWhenSilent) {
|
|
28
|
-
if (!state.speaking &&
|
|
29
|
-
|
|
41
|
+
if (options.muteWhenSilent && gainNode) {
|
|
42
|
+
if (!state.speaking && gainNode.gain.value > 0) {
|
|
43
|
+
gainNode.gain.value = 0;
|
|
30
44
|
mutedByController = true;
|
|
31
45
|
}
|
|
32
46
|
if (state.speaking && mutedByController) {
|
|
33
|
-
|
|
47
|
+
gainNode.gain.value = 1;
|
|
34
48
|
mutedByController = false;
|
|
35
49
|
}
|
|
36
50
|
}
|
|
@@ -62,9 +76,15 @@ async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
|
62
76
|
pipeline.events.off("speakingChange", speakingHandler);
|
|
63
77
|
pipeline.events.off("error", errorHandler);
|
|
64
78
|
listeners.clear();
|
|
65
|
-
if (
|
|
66
|
-
|
|
67
|
-
|
|
79
|
+
if (gainNode) {
|
|
80
|
+
if (mutedByController && gainNode.gain.value === 0) {
|
|
81
|
+
gainNode.gain.value = 1;
|
|
82
|
+
mutedByController = false;
|
|
83
|
+
}
|
|
84
|
+
try {
|
|
85
|
+
gainNode.disconnect();
|
|
86
|
+
} catch (error) {
|
|
87
|
+
}
|
|
68
88
|
}
|
|
69
89
|
pipeline.dispose();
|
|
70
90
|
if (originalTrack.readyState === "live") {
|
package/dist/defaults.js
CHANGED
|
@@ -28,7 +28,7 @@ __export(defaults_exports, {
|
|
|
28
28
|
});
|
|
29
29
|
module.exports = __toCommonJS(defaults_exports);
|
|
30
30
|
var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
31
|
-
minDb: -
|
|
31
|
+
minDb: -60,
|
|
32
32
|
maxDb: -20,
|
|
33
33
|
speakOnRatio: 0.5,
|
|
34
34
|
speakOffRatio: 0.3,
|
|
@@ -38,7 +38,7 @@ var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
|
38
38
|
};
|
|
39
39
|
var DEFAULT_NOISE_SUPPRESSION_CONFIG = {
|
|
40
40
|
enabled: true,
|
|
41
|
-
noiseReductionLevel:
|
|
41
|
+
noiseReductionLevel: 60
|
|
42
42
|
};
|
|
43
43
|
var DEFAULT_OUTPUT_GAIN_CONFIG = {
|
|
44
44
|
speechGain: 1,
|
package/dist/defaults.mjs
CHANGED
package/dist/index.js
CHANGED
|
@@ -42,7 +42,7 @@ module.exports = __toCommonJS(index_exports);
|
|
|
42
42
|
|
|
43
43
|
// src/defaults.ts
|
|
44
44
|
var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
45
|
-
minDb: -
|
|
45
|
+
minDb: -60,
|
|
46
46
|
maxDb: -20,
|
|
47
47
|
speakOnRatio: 0.5,
|
|
48
48
|
speakOffRatio: 0.3,
|
|
@@ -52,7 +52,7 @@ var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
|
52
52
|
};
|
|
53
53
|
var DEFAULT_NOISE_SUPPRESSION_CONFIG = {
|
|
54
54
|
enabled: true,
|
|
55
|
-
noiseReductionLevel:
|
|
55
|
+
noiseReductionLevel: 60
|
|
56
56
|
};
|
|
57
57
|
var DEFAULT_OUTPUT_GAIN_CONFIG = {
|
|
58
58
|
speechGain: 1,
|
|
@@ -503,16 +503,30 @@ async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
|
503
503
|
const listeners = /* @__PURE__ */ new Set();
|
|
504
504
|
let mutedByController = false;
|
|
505
505
|
let currentState = pipeline.state;
|
|
506
|
+
let gainNode = null;
|
|
507
|
+
if (options.muteWhenSilent) {
|
|
508
|
+
try {
|
|
509
|
+
const audioContext = new AudioContext();
|
|
510
|
+
const source = audioContext.createMediaStreamSource(
|
|
511
|
+
new MediaStream([pipeline.processedTrack])
|
|
512
|
+
);
|
|
513
|
+
gainNode = audioContext.createGain();
|
|
514
|
+
source.connect(gainNode);
|
|
515
|
+
gainNode.connect(audioContext.destination);
|
|
516
|
+
} catch (error) {
|
|
517
|
+
console.error("Failed to create gain node for volume control", error);
|
|
518
|
+
}
|
|
519
|
+
}
|
|
506
520
|
const speakingHandler = (state) => {
|
|
507
521
|
currentState = state;
|
|
508
522
|
listeners.forEach((listener) => listener(state));
|
|
509
|
-
if (options.muteWhenSilent) {
|
|
510
|
-
if (!state.speaking &&
|
|
511
|
-
|
|
523
|
+
if (options.muteWhenSilent && gainNode) {
|
|
524
|
+
if (!state.speaking && gainNode.gain.value > 0) {
|
|
525
|
+
gainNode.gain.value = 0;
|
|
512
526
|
mutedByController = true;
|
|
513
527
|
}
|
|
514
528
|
if (state.speaking && mutedByController) {
|
|
515
|
-
|
|
529
|
+
gainNode.gain.value = 1;
|
|
516
530
|
mutedByController = false;
|
|
517
531
|
}
|
|
518
532
|
}
|
|
@@ -544,9 +558,15 @@ async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
|
544
558
|
pipeline.events.off("speakingChange", speakingHandler);
|
|
545
559
|
pipeline.events.off("error", errorHandler);
|
|
546
560
|
listeners.clear();
|
|
547
|
-
if (
|
|
548
|
-
|
|
549
|
-
|
|
561
|
+
if (gainNode) {
|
|
562
|
+
if (mutedByController && gainNode.gain.value === 0) {
|
|
563
|
+
gainNode.gain.value = 1;
|
|
564
|
+
mutedByController = false;
|
|
565
|
+
}
|
|
566
|
+
try {
|
|
567
|
+
gainNode.disconnect();
|
|
568
|
+
} catch (error) {
|
|
569
|
+
}
|
|
550
570
|
}
|
|
551
571
|
pipeline.dispose();
|
|
552
572
|
if (originalTrack.readyState === "live") {
|
package/dist/index.mjs
CHANGED
|
@@ -2,19 +2,19 @@ import "./chunk-WBQAMGXK.mjs";
|
|
|
2
2
|
import {
|
|
3
3
|
attachSpeakingDetectionToRemoteTrack,
|
|
4
4
|
attachSpeakingDetectionToTrack
|
|
5
|
-
} from "./chunk-
|
|
6
|
-
import "./chunk-
|
|
5
|
+
} from "./chunk-HCEWI3JP.mjs";
|
|
6
|
+
import "./chunk-7DOV5ZIH.mjs";
|
|
7
7
|
import "./chunk-IS37FHDN.mjs";
|
|
8
|
-
import "./chunk-
|
|
8
|
+
import "./chunk-47YTMONE.mjs";
|
|
9
9
|
import "./chunk-QNQK6QFB.mjs";
|
|
10
|
-
import "./chunk-
|
|
10
|
+
import "./chunk-XHU4WPFD.mjs";
|
|
11
11
|
import {
|
|
12
12
|
DEFAULT_LIVEKIT_SPEAKING_OPTIONS,
|
|
13
13
|
DEFAULT_NOISE_SUPPRESSION_CONFIG,
|
|
14
14
|
DEFAULT_OUTPUT_GAIN_CONFIG,
|
|
15
15
|
DEFAULT_REMOTE_SPEAKING_OPTIONS,
|
|
16
16
|
DEFAULT_SPEAKING_DETECTION_CONFIG
|
|
17
|
-
} from "./chunk-
|
|
17
|
+
} from "./chunk-FYX5A3OM.mjs";
|
|
18
18
|
import "./chunk-OZ7KMC4S.mjs";
|
|
19
19
|
export {
|
|
20
20
|
DEFAULT_LIVEKIT_SPEAKING_OPTIONS,
|
|
@@ -159,7 +159,7 @@ async function createLevelDetectorNode(context, onLevel, options) {
|
|
|
159
159
|
|
|
160
160
|
// src/defaults.ts
|
|
161
161
|
var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
162
|
-
minDb: -
|
|
162
|
+
minDb: -60,
|
|
163
163
|
maxDb: -20,
|
|
164
164
|
speakOnRatio: 0.5,
|
|
165
165
|
speakOffRatio: 0.3,
|
|
@@ -169,7 +169,7 @@ var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
|
169
169
|
};
|
|
170
170
|
var DEFAULT_NOISE_SUPPRESSION_CONFIG = {
|
|
171
171
|
enabled: true,
|
|
172
|
-
noiseReductionLevel:
|
|
172
|
+
noiseReductionLevel: 60
|
|
173
173
|
};
|
|
174
174
|
var DEFAULT_OUTPUT_GAIN_CONFIG = {
|
|
175
175
|
speechGain: 1,
|
|
@@ -498,16 +498,30 @@ async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
|
498
498
|
const listeners = /* @__PURE__ */ new Set();
|
|
499
499
|
let mutedByController = false;
|
|
500
500
|
let currentState = pipeline.state;
|
|
501
|
+
let gainNode = null;
|
|
502
|
+
if (options.muteWhenSilent) {
|
|
503
|
+
try {
|
|
504
|
+
const audioContext = new AudioContext();
|
|
505
|
+
const source = audioContext.createMediaStreamSource(
|
|
506
|
+
new MediaStream([pipeline.processedTrack])
|
|
507
|
+
);
|
|
508
|
+
gainNode = audioContext.createGain();
|
|
509
|
+
source.connect(gainNode);
|
|
510
|
+
gainNode.connect(audioContext.destination);
|
|
511
|
+
} catch (error) {
|
|
512
|
+
console.error("Failed to create gain node for volume control", error);
|
|
513
|
+
}
|
|
514
|
+
}
|
|
501
515
|
const speakingHandler = (state) => {
|
|
502
516
|
currentState = state;
|
|
503
517
|
listeners.forEach((listener) => listener(state));
|
|
504
|
-
if (options.muteWhenSilent) {
|
|
505
|
-
if (!state.speaking &&
|
|
506
|
-
|
|
518
|
+
if (options.muteWhenSilent && gainNode) {
|
|
519
|
+
if (!state.speaking && gainNode.gain.value > 0) {
|
|
520
|
+
gainNode.gain.value = 0;
|
|
507
521
|
mutedByController = true;
|
|
508
522
|
}
|
|
509
523
|
if (state.speaking && mutedByController) {
|
|
510
|
-
|
|
524
|
+
gainNode.gain.value = 1;
|
|
511
525
|
mutedByController = false;
|
|
512
526
|
}
|
|
513
527
|
}
|
|
@@ -539,9 +553,15 @@ async function attachSpeakingDetectionToTrack(track, options = {}) {
|
|
|
539
553
|
pipeline.events.off("speakingChange", speakingHandler);
|
|
540
554
|
pipeline.events.off("error", errorHandler);
|
|
541
555
|
listeners.clear();
|
|
542
|
-
if (
|
|
543
|
-
|
|
544
|
-
|
|
556
|
+
if (gainNode) {
|
|
557
|
+
if (mutedByController && gainNode.gain.value === 0) {
|
|
558
|
+
gainNode.gain.value = 1;
|
|
559
|
+
mutedByController = false;
|
|
560
|
+
}
|
|
561
|
+
try {
|
|
562
|
+
gainNode.disconnect();
|
|
563
|
+
} catch (error) {
|
|
564
|
+
}
|
|
545
565
|
}
|
|
546
566
|
pipeline.dispose();
|
|
547
567
|
if (originalTrack.readyState === "live") {
|
|
@@ -1,13 +1,13 @@
|
|
|
1
1
|
import {
|
|
2
2
|
attachSpeakingDetectionToRemoteTrack,
|
|
3
3
|
attachSpeakingDetectionToTrack
|
|
4
|
-
} from "../chunk-
|
|
5
|
-
import "../chunk-
|
|
4
|
+
} from "../chunk-HCEWI3JP.mjs";
|
|
5
|
+
import "../chunk-7DOV5ZIH.mjs";
|
|
6
6
|
import "../chunk-IS37FHDN.mjs";
|
|
7
|
-
import "../chunk-
|
|
7
|
+
import "../chunk-47YTMONE.mjs";
|
|
8
8
|
import "../chunk-QNQK6QFB.mjs";
|
|
9
|
-
import "../chunk-
|
|
10
|
-
import "../chunk-
|
|
9
|
+
import "../chunk-XHU4WPFD.mjs";
|
|
10
|
+
import "../chunk-FYX5A3OM.mjs";
|
|
11
11
|
import "../chunk-OZ7KMC4S.mjs";
|
|
12
12
|
export {
|
|
13
13
|
attachSpeakingDetectionToRemoteTrack,
|
|
@@ -156,7 +156,7 @@ async function createLevelDetectorNode(context, onLevel, options) {
|
|
|
156
156
|
|
|
157
157
|
// src/defaults.ts
|
|
158
158
|
var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
159
|
-
minDb: -
|
|
159
|
+
minDb: -60,
|
|
160
160
|
maxDb: -20,
|
|
161
161
|
speakOnRatio: 0.5,
|
|
162
162
|
speakOffRatio: 0.3,
|
|
@@ -166,7 +166,7 @@ var DEFAULT_SPEAKING_DETECTION_CONFIG = {
|
|
|
166
166
|
};
|
|
167
167
|
var DEFAULT_NOISE_SUPPRESSION_CONFIG = {
|
|
168
168
|
enabled: true,
|
|
169
|
-
noiseReductionLevel:
|
|
169
|
+
noiseReductionLevel: 60
|
|
170
170
|
};
|
|
171
171
|
var DEFAULT_OUTPUT_GAIN_CONFIG = {
|
|
172
172
|
speechGain: 1,
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
import {
|
|
2
2
|
createAudioPipeline
|
|
3
|
-
} from "../chunk-
|
|
3
|
+
} from "../chunk-7DOV5ZIH.mjs";
|
|
4
4
|
import "../chunk-IS37FHDN.mjs";
|
|
5
5
|
import "../chunk-QNQK6QFB.mjs";
|
|
6
|
-
import "../chunk-
|
|
7
|
-
import "../chunk-
|
|
6
|
+
import "../chunk-XHU4WPFD.mjs";
|
|
7
|
+
import "../chunk-FYX5A3OM.mjs";
|
|
8
8
|
import "../chunk-OZ7KMC4S.mjs";
|
|
9
9
|
export {
|
|
10
10
|
createAudioPipeline
|
|
@@ -1,9 +1,9 @@
|
|
|
1
1
|
import {
|
|
2
2
|
createRemoteAudioMonitor
|
|
3
|
-
} from "../chunk-
|
|
3
|
+
} from "../chunk-47YTMONE.mjs";
|
|
4
4
|
import "../chunk-QNQK6QFB.mjs";
|
|
5
|
-
import "../chunk-
|
|
6
|
-
import "../chunk-
|
|
5
|
+
import "../chunk-XHU4WPFD.mjs";
|
|
6
|
+
import "../chunk-FYX5A3OM.mjs";
|
|
7
7
|
import "../chunk-OZ7KMC4S.mjs";
|
|
8
8
|
export {
|
|
9
9
|
createRemoteAudioMonitor
|
package/dist/vad/vad-state.js
CHANGED
package/dist/vad/vad-state.mjs
CHANGED
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@tensamin/audio",
|
|
3
|
-
"version": "0.2.
|
|
3
|
+
"version": "0.2.7",
|
|
4
4
|
"author": {
|
|
5
5
|
"email": "aloisianer@proton.me",
|
|
6
6
|
"name": "Alois"
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
"typescript": "^5.9.3"
|
|
20
20
|
},
|
|
21
21
|
"dependencies": {
|
|
22
|
-
"deepfilternet3-noise-filter": "^1.1.
|
|
22
|
+
"deepfilternet3-noise-filter": "^1.1.4",
|
|
23
23
|
"mitt": "^3.0.1"
|
|
24
24
|
},
|
|
25
25
|
"peerDependencies": {
|