@elizaos/app-core 2.0.0-alpha.47 → 2.0.0-alpha.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +2 -0
- package/dist/hooks/useVoiceChat.d.ts +1 -36
- package/dist/hooks/useVoiceChat.d.ts.map +1 -1
- package/dist/hooks/useVoiceChat.js +9 -4
- package/dist/i18n/locales/en.json +1193 -1193
- package/dist/i18n/locales/es.json +1193 -1193
- package/dist/i18n/locales/ko.json +1193 -1193
- package/dist/i18n/locales/pt.json +1193 -1193
- package/dist/i18n/locales/zh-CN.json +1193 -1193
- package/package.json +4 -4
- package/src/hooks/useVoiceChat.ts +9 -12
- package/dist/package.json +0 -182
- package/dist/styles/anime.css +0 -6324
- package/dist/styles/base.css +0 -196
- package/dist/styles/onboarding-game.css +0 -738
- package/dist/styles/styles.css +0 -2087
- package/dist/styles/xterm.css +0 -241
package/.turbo/turbo-build.log
CHANGED
|
@@ -1,2 +1,4 @@
|
|
|
1
1
|
$ bun run build:dist
|
|
2
2
|
$ test -f ../ui/dist/index.d.ts || (cd ../ui && bun run build) && rm -rf dist && tsc -p tsconfig.build.json && node ../../scripts/copy-package-assets.mjs packages/app-core src/styles src/i18n/locales && node ../../scripts/prepare-package-dist.mjs packages/app-core
|
|
3
|
+
src/hooks/useVoiceChat.ts(980,25): error TS2551: Property '__ELIZA_API_TOKEN__' does not exist on type 'Window & typeof globalThis'. Did you mean '__MILADY_API_TOKEN__'?
|
|
4
|
+
src/hooks/useVoiceChat.ts(981,22): error TS2551: Property '__ELIZA_API_TOKEN__' does not exist on type 'Window & typeof globalThis'. Did you mean '__MILADY_API_TOKEN__'?
|
|
@@ -8,41 +8,6 @@
|
|
|
8
8
|
* STT: Web Speech API (SpeechRecognition) for user voice input.
|
|
9
9
|
*/
|
|
10
10
|
import type { VoiceConfig } from "../api/client";
|
|
11
|
-
interface SpeechRecognitionInstance extends EventTarget {
|
|
12
|
-
continuous: boolean;
|
|
13
|
-
interimResults: boolean;
|
|
14
|
-
lang: string;
|
|
15
|
-
onstart: (() => void) | null;
|
|
16
|
-
onend: (() => void) | null;
|
|
17
|
-
onerror: ((event: {
|
|
18
|
-
error: string;
|
|
19
|
-
}) => void) | null;
|
|
20
|
-
onresult: ((event: SpeechRecognitionResultEvent) => void) | null;
|
|
21
|
-
start(): void;
|
|
22
|
-
stop(): void;
|
|
23
|
-
abort(): void;
|
|
24
|
-
}
|
|
25
|
-
interface SpeechRecognitionResultEvent {
|
|
26
|
-
results: SpeechRecognitionResultList;
|
|
27
|
-
resultIndex: number;
|
|
28
|
-
}
|
|
29
|
-
interface SpeechRecognitionResultList {
|
|
30
|
-
length: number;
|
|
31
|
-
[index: number]: {
|
|
32
|
-
isFinal: boolean;
|
|
33
|
-
0: {
|
|
34
|
-
transcript: string;
|
|
35
|
-
confidence: number;
|
|
36
|
-
};
|
|
37
|
-
};
|
|
38
|
-
}
|
|
39
|
-
type SpeechRecognitionCtor = new () => SpeechRecognitionInstance;
|
|
40
|
-
declare global {
|
|
41
|
-
interface Window {
|
|
42
|
-
SpeechRecognition?: SpeechRecognitionCtor;
|
|
43
|
-
webkitSpeechRecognition?: SpeechRecognitionCtor;
|
|
44
|
-
}
|
|
45
|
-
}
|
|
46
11
|
type SpeechSegmentKind = "full" | "first-sentence" | "remainder";
|
|
47
12
|
type SpeechProviderKind = "elevenlabs" | "browser";
|
|
48
13
|
export type VoiceCaptureMode = "idle" | "compose" | "push-to-talk";
|
|
@@ -64,7 +29,7 @@ export interface VoiceChatOptions {
|
|
|
64
29
|
onTranscriptPreview?: (text: string, event: VoiceTranscriptPreviewEvent) => void;
|
|
65
30
|
/** Called when playback of a speech segment starts */
|
|
66
31
|
onPlaybackStart?: (event: VoicePlaybackStartEvent) => void;
|
|
67
|
-
/** True when the user is authenticated to
|
|
32
|
+
/** True when the user is authenticated to Eliza Cloud */
|
|
68
33
|
cloudConnected?: boolean;
|
|
69
34
|
/** Whether user speech should immediately interrupt assistant playback */
|
|
70
35
|
interruptOnSpeech?: boolean;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"useVoiceChat.d.ts","sourceRoot":"","sources":["../../src/hooks/useVoiceChat.ts"],"names":[],"mappings":"AAAA;;;;;;;;GAQG;AAYH,OAAO,KAAK,EAAE,WAAW,EAAE,MAAM,eAAe,CAAC;
|
|
1
|
+
{"version":3,"file":"useVoiceChat.d.ts","sourceRoot":"","sources":["../../src/hooks/useVoiceChat.ts"],"names":[],"mappings":"AAAA;;;;;;;;GAQG;AAYH,OAAO,KAAK,EAAE,WAAW,EAAE,MAAM,eAAe,CAAC;AAkDjD,KAAK,iBAAiB,GAAG,MAAM,GAAG,gBAAgB,GAAG,WAAW,CAAC;AACjE,KAAK,kBAAkB,GAAG,YAAY,GAAG,SAAS,CAAC;AACnD,MAAM,MAAM,gBAAgB,GAAG,MAAM,GAAG,SAAS,GAAG,cAAc,CAAC;AAEnE,MAAM,WAAW,uBAAuB;IACtC,IAAI,EAAE,MAAM,CAAC;IACb,OAAO,EAAE,iBAAiB,CAAC;IAC3B,QAAQ,EAAE,kBAAkB,CAAC;IAC7B,MAAM,EAAE,OAAO,CAAC;IAChB,WAAW,EAAE,MAAM,CAAC;CACrB;AAED,MAAM,WAAW,2BAA2B;IAC1C,IAAI,EAAE,OAAO,CAAC,gBAAgB,EAAE,MAAM,CAAC,CAAC;IACxC,OAAO,EAAE,OAAO,CAAC;CAClB;AAED,MAAM,WAAW,gBAAgB;IAC/B,sDAAsD;IACtD,YAAY,EAAE,CAAC,IAAI,EAAE,MAAM,KAAK,IAAI,CAAC;IACrC,yDAAyD;IACzD,mBAAmB,CAAC,EAAE,CACpB,IAAI,EAAE,MAAM,EACZ,KAAK,EAAE,2BAA2B,KAC/B,IAAI,CAAC;IACV,sDAAsD;IACtD,eAAe,CAAC,EAAE,CAAC,KAAK,EAAE,uBAAuB,KAAK,IAAI,CAAC;IAC3D,yDAAyD;IACzD,cAAc,CAAC,EAAE,OAAO,CAAC;IACzB,0EAA0E;IAC1E,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,yDAAyD;IACzD,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,iEAAiE;IACjE,WAAW,CAAC,EAAE,WAAW,GAAG,IAAI,CAAC;CAClC;AAED,MAAM,WAAW,cAAc;IAC7B,8CAA8C;IAC9C,WAAW,EAAE,OAAO,CAAC;IACrB,+BAA+B;IAC/B,WAAW,EAAE,gBAAgB,CAAC;IAC9B,8CAA8C;IAC9C,UAAU,EAAE,OAAO,CAAC;IACpB,gDAAgD;IAChD,SAAS,EAAE,MAAM,CAAC;IAClB,kDAAkD;IAClD,iBAAiB,EAAE,MAAM,CAAC;IAC1B,0CAA0C;IAC1C,SAAS,EAAE,OAAO,CAAC;IACnB,iEAAiE;IACjE,kBAAkB,EAAE,OAAO,CAAC;IAC5B,oCAAoC;IACpC,eAAe,EAAE,MAAM,IAAI,CAAC;IAC5B,0DAA0D;IAC1D,cAAc,EAAE,CAAC,IAAI,CAAC,EAAE,OAAO,CAAC,gBAAgB,EAAE,MAAM,CAAC,KAAK,OAAO,CAAC,IAAI,CAAC,CAAC;IAC5E,6DAA6D;IAC7D,aAAa,EAAE,CAAC,OAAO,CAAC,EAAE;QAAE,MAAM,CAAC,EAAE,OAAO,CAAA;KAAE,KAAK,OAAO,CAAC,IAAI,CAAC,CAAC;IACjE,4CAA4C;IAC5C,KAAK,EAAE,CAAC,IAAI,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE;QAAE,MAAM,CAAC,EAAE,OAAO,CAAA;KAAE,KAAK,IAAI,CAAC;IAC9D,gEAAgE;IAChE,oBAAoB,EAAE,CACpB,SAAS,EAAE,MAAM,EACjB,IAAI,EAAE,MAAM,EACZ,OAAO,EAAE,OAAO,KACb,IAAI,CAAC;IACV,8BAA8B;IAC9B,YAAY,EAAE,MAAM,IAAI,CAAC;CAC1B;AA6CD,wBAAgB,iBAAiB,CAAC,YAAY,EAAE,MAAM,GAAG,MAAM,CAM9D;AAqBD,iBAAS,eAAe,CAAC,KAAK,EAAE,MAAM,GAAG,MAAM,CAI9C;AAED,iBAAS,kBAAkB,CAAC,IAAI,EAAE,MAAM,GAAG;IACzC,QAAQ,EAAE,OAAO,CAAC;IAClB,aAAa,EAAE,MAAM,CAAC;IACtB,SAAS,EAAE,MAAM,CAAC;CACnB,CA6BA;AAED,iBAAS,cAAc,CAAC,QAAQ,EAAE,MAAM,EAAE,aAAa,EAAE,MAAM,GAAG,MAAM,CAkBvE;AAED,iBAAS,qBAAqB,CAAC,IAAI,EAAE,MAAM,EAAE,OAAO,EAAE,OAAO,GAAG,MAAM,CA0BrE;AAaD,iBAAS,2BAA2B,CAClC,MAAM,EAAE,WAAW,GAAG,IAAI,GAAG,SAAS,EACtC,OAAO,CAAC,EAAE;IAAE,cAAc,CAAC,EAAE,OAAO,CAAA;CAAE,GACrC,WAAW,GAAG,IAAI,CAiDpB;AAED,eAAO,MAAM,oBAAoB;;;;;;CAMhC,CAAC;AAgBF,wBAAgB,YAAY,CAAC,OAAO,EAAE,gBAAgB,GAAG,cAAc,CA6hCtE"}
|
|
@@ -14,6 +14,11 @@ import { getTalkModePlugin, } from "../bridge/native-plugins";
|
|
|
14
14
|
import { resolveApiUrl } from "../utils";
|
|
15
15
|
import { sanitizeSpeechText } from "../utils/spoken-text";
|
|
16
16
|
import { mergeStreamingText } from "../utils/streaming-text";
|
|
17
|
+
/** Access browser SpeechRecognition APIs which may live under a vendor prefix. */
|
|
18
|
+
function getSpeechRecognitionCtor() {
|
|
19
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
20
|
+
return window.SpeechRecognition ?? window.webkitSpeechRecognition;
|
|
21
|
+
}
|
|
17
22
|
const DEFAULT_ELEVEN_MODEL = "eleven_flash_v2_5";
|
|
18
23
|
const DEFAULT_ELEVEN_VOICE = "EXAVITQu4vr4xnSDxMaL";
|
|
19
24
|
const MAX_SPOKEN_CHARS = 360;
|
|
@@ -314,7 +319,7 @@ export function useVoiceChat(options) {
|
|
|
314
319
|
}, []);
|
|
315
320
|
// ── Init ──────────────────────────────────────────────────────────
|
|
316
321
|
useEffect(() => {
|
|
317
|
-
const SpeechRecognitionAPI =
|
|
322
|
+
const SpeechRecognitionAPI = getSpeechRecognitionCtor();
|
|
318
323
|
const canUseMicrophone = typeof navigator !== "undefined" &&
|
|
319
324
|
typeof navigator.mediaDevices?.getUserMedia === "function";
|
|
320
325
|
setSupported(shouldPreferNativeTalkMode()
|
|
@@ -448,7 +453,7 @@ export function useVoiceChat(options) {
|
|
|
448
453
|
talkModeHandlesRef.current = [transcriptHandle, errorHandle, stateHandle];
|
|
449
454
|
}, [applyTranscriptUpdate]);
|
|
450
455
|
const startBrowserRecognition = useCallback((mode) => {
|
|
451
|
-
const SpeechRecognitionAPI =
|
|
456
|
+
const SpeechRecognitionAPI = getSpeechRecognitionCtor();
|
|
452
457
|
if (!SpeechRecognitionAPI)
|
|
453
458
|
return false;
|
|
454
459
|
const recognition = new SpeechRecognitionAPI();
|
|
@@ -684,8 +689,8 @@ export function useVoiceChat(options) {
|
|
|
684
689
|
},
|
|
685
690
|
};
|
|
686
691
|
const apiToken = typeof window !== "undefined" &&
|
|
687
|
-
typeof window.
|
|
688
|
-
? window.
|
|
692
|
+
typeof window.__ELIZA_API_TOKEN__ === "string"
|
|
693
|
+
? window.__ELIZA_API_TOKEN__.trim()
|
|
689
694
|
: "";
|
|
690
695
|
const fetchViaProxy = async () => {
|
|
691
696
|
return fetch(resolveElevenProxyEndpoint(), {
|