@layercode/js-sdk 2.8.1 → 2.8.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/layercode-js-sdk.esm.js +116 -14
- package/dist/layercode-js-sdk.esm.js.map +1 -1
- package/dist/layercode-js-sdk.min.js +116 -14
- package/dist/layercode-js-sdk.min.js.map +1 -1
- package/dist/types/index.d.ts +14 -0
- package/dist/types/interfaces.d.ts +6 -2
- package/dist/types/wavtools/lib/analysis/audio_analysis.d.ts +1 -1
- package/package.json +1 -1
package/dist/types/index.d.ts
CHANGED
|
@@ -84,6 +84,13 @@ interface LayercodeClientOptions {
|
|
|
84
84
|
audioInput?: boolean;
|
|
85
85
|
/** Whether audio output is enabled. I.e. do we play the sound in the browser client */
|
|
86
86
|
audioOutput?: boolean;
|
|
87
|
+
/**
|
|
88
|
+
* When true, defers actual audio hardware initialization (AudioContext, mic permissions)
|
|
89
|
+
* until setAudioInput(true) or setAudioOutput(true) is called.
|
|
90
|
+
* The server will still be told this is a voice session, but no browser audio APIs are touched until needed.
|
|
91
|
+
* This avoids Chrome's autoplay policy blocking AudioContext before user gesture.
|
|
92
|
+
*/
|
|
93
|
+
deferAudioInit?: boolean;
|
|
87
94
|
/** Fired when audio input flag changes */
|
|
88
95
|
audioInputChanged?: (audioInput: boolean) => void;
|
|
89
96
|
/** Fired when audio output flag changes */
|
|
@@ -143,9 +150,11 @@ declare class LayercodeClient implements ILayercodeClient {
|
|
|
143
150
|
private canInterrupt;
|
|
144
151
|
private userIsSpeaking;
|
|
145
152
|
private agentIsSpeaking;
|
|
153
|
+
private agentIsPlayingAudio;
|
|
146
154
|
private recorderStarted;
|
|
147
155
|
private readySent;
|
|
148
156
|
private currentTurnId;
|
|
157
|
+
private sentReplayFinishedForDisabledOutput;
|
|
149
158
|
private audioBuffer;
|
|
150
159
|
private vadConfig;
|
|
151
160
|
private deviceId;
|
|
@@ -220,6 +229,11 @@ declare class LayercodeClient implements ILayercodeClient {
|
|
|
220
229
|
audioInputDisconnect(): Promise<void>;
|
|
221
230
|
setAudioInput(state: boolean): Promise<void>;
|
|
222
231
|
setAudioOutput(state: boolean): Promise<void>;
|
|
232
|
+
/**
|
|
233
|
+
* Syncs the reported agentSpeaking state with the actual audio playback state.
|
|
234
|
+
* Called when audioOutput is enabled to ensure proper state synchronization.
|
|
235
|
+
*/
|
|
236
|
+
private _syncAgentSpeakingState;
|
|
223
237
|
/** Emitters for audio flags */
|
|
224
238
|
private _emitAudioInput;
|
|
225
239
|
private _emitAudioOutput;
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
export type LayercodeMessageType = 'client.audio' | 'trigger.turn.start' | 'trigger.turn.end' | 'trigger.response.audio.replay_finished' | 'vad_events' | 'client.ready' | 'client.response.text' | 'client.response.data' | 'turn.start' | 'response.audio' | 'response.text' | 'response.data' | 'user.transcript.interim_delta' | 'user.transcript.delta' | 'user.transcript';
|
|
1
|
+
export type LayercodeMessageType = 'client.audio' | 'trigger.turn.start' | 'trigger.turn.end' | 'trigger.response.audio.replay_finished' | 'vad_events' | 'client.ready' | 'client.response.text' | 'client.response.data' | 'turn.start' | 'response.audio' | 'response.text' | 'response.data' | 'response.end' | 'user.transcript.interim_delta' | 'user.transcript.delta' | 'user.transcript';
|
|
2
2
|
export interface BaseLayercodeMessage {
|
|
3
3
|
type: LayercodeMessageType;
|
|
4
4
|
event_id?: string;
|
|
@@ -68,6 +68,10 @@ export interface ServerResponseUserTranscript extends BaseLayercodeMessage {
|
|
|
68
68
|
content: string;
|
|
69
69
|
turn_id: string;
|
|
70
70
|
}
|
|
71
|
-
export
|
|
71
|
+
export interface ServerResponseEndMessage extends BaseLayercodeMessage {
|
|
72
|
+
type: 'response.end';
|
|
73
|
+
turn_id: string;
|
|
74
|
+
}
|
|
75
|
+
export type ServerMessage = ServerTurnMessage | ServerResponseAudioMessage | ServerResponseTextMessage | ServerResponseDataMessage | ServerResponseEndMessage | ServerResponseUserTranscriptInterimDelta | ServerResponseUserTranscriptDelta | ServerResponseUserTranscript;
|
|
72
76
|
export type ClientMessage = ClientAudioMessage | ClientTriggerTurnMessage | ClientTriggerResponseAudioReplayFinishedMessage | ClientVadEventsMessage | ClientReadyMessage | ClientResponseTextMessage | ClientResponseDataMessage;
|
|
73
77
|
export type LayercodeMessage = ClientMessage | ServerMessage;
|
|
@@ -31,7 +31,7 @@ export class AudioAnalysis {
|
|
|
31
31
|
constructor(audioElement: HTMLAudioElement, audioBuffer?: AudioBuffer | null);
|
|
32
32
|
fftResults: any[];
|
|
33
33
|
audio: HTMLAudioElement;
|
|
34
|
-
context:
|
|
34
|
+
context: AudioContext | OfflineAudioContext;
|
|
35
35
|
analyser: AnalyserNode;
|
|
36
36
|
sampleRate: number;
|
|
37
37
|
audioBuffer: AudioBuffer | null;
|
package/package.json
CHANGED