@volley/recognition-client-sdk 0.1.297 → 0.1.381
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/browser.bundled.d.ts +52 -14
- package/dist/index.bundled.d.ts +64 -24
- package/dist/index.d.ts +1 -1
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +146 -90
- package/dist/index.js.map +4 -4
- package/dist/recog-client-sdk.browser.js +85 -81
- package/dist/recog-client-sdk.browser.js.map +4 -4
- package/dist/recognition-client.d.ts +4 -0
- package/dist/recognition-client.d.ts.map +1 -1
- package/dist/simplified-vgf-recognition-client.d.ts +2 -0
- package/dist/simplified-vgf-recognition-client.d.ts.map +1 -1
- package/dist/vgf-recognition-state.d.ts +9 -0
- package/dist/vgf-recognition-state.d.ts.map +1 -1
- package/package.json +23 -23
- package/src/index.ts +2 -0
- package/src/recognition-client.ts +54 -2
- package/src/simplified-vgf-recognition-client.spec.ts +294 -12
- package/src/simplified-vgf-recognition-client.ts +68 -12
- package/src/vgf-recognition-state.ts +12 -0
|
@@ -10,9 +10,12 @@ import { z } from 'zod';
|
|
|
10
10
|
declare enum RecognitionProvider {
|
|
11
11
|
ASSEMBLYAI = "assemblyai",
|
|
12
12
|
DEEPGRAM = "deepgram",
|
|
13
|
+
ELEVENLABS = "elevenlabs",
|
|
14
|
+
FIREWORKS = "fireworks",
|
|
13
15
|
GOOGLE = "google",
|
|
14
16
|
GEMINI_BATCH = "gemini-batch",
|
|
15
|
-
OPENAI_BATCH = "openai-batch"
|
|
17
|
+
OPENAI_BATCH = "openai-batch",
|
|
18
|
+
OPENAI_REALTIME = "openai-realtime"
|
|
16
19
|
}
|
|
17
20
|
/**
|
|
18
21
|
* ASR API type - distinguishes between streaming and file-based transcription APIs
|
|
@@ -34,23 +37,54 @@ declare enum DeepgramModel {
|
|
|
34
37
|
/**
|
|
35
38
|
* Google Cloud Speech models
|
|
36
39
|
* @see https://cloud.google.com/speech-to-text/docs/transcription-model
|
|
40
|
+
* @see https://cloud.google.com/speech-to-text/v2/docs/chirp_3-model
|
|
37
41
|
*/
|
|
38
42
|
declare enum GoogleModel {
|
|
43
|
+
CHIRP_3 = "chirp_3",
|
|
44
|
+
CHIRP_2 = "chirp_2",
|
|
45
|
+
CHIRP = "chirp",
|
|
39
46
|
LATEST_LONG = "latest_long",
|
|
40
47
|
LATEST_SHORT = "latest_short",
|
|
41
48
|
TELEPHONY = "telephony",
|
|
42
49
|
TELEPHONY_SHORT = "telephony_short",
|
|
43
|
-
MEDICAL_DICTATION = "medical_dictation",
|
|
44
|
-
MEDICAL_CONVERSATION = "medical_conversation",
|
|
45
50
|
DEFAULT = "default",
|
|
46
51
|
COMMAND_AND_SEARCH = "command_and_search",
|
|
47
52
|
PHONE_CALL = "phone_call",
|
|
48
53
|
VIDEO = "video"
|
|
49
54
|
}
|
|
55
|
+
/**
|
|
56
|
+
* Fireworks AI models for ASR
|
|
57
|
+
* @see https://docs.fireworks.ai/guides/querying-asr-models
|
|
58
|
+
* @see https://fireworks.ai/models/fireworks/fireworks-asr-large
|
|
59
|
+
*/
|
|
60
|
+
declare enum FireworksModel {
|
|
61
|
+
ASR_V1 = "fireworks-asr-large",
|
|
62
|
+
ASR_V2 = "fireworks-asr-v2",
|
|
63
|
+
WHISPER_V3 = "whisper-v3",
|
|
64
|
+
WHISPER_V3_TURBO = "whisper-v3-turbo"
|
|
65
|
+
}
|
|
66
|
+
/**
|
|
67
|
+
* ElevenLabs Scribe models for speech-to-text
|
|
68
|
+
* @see https://elevenlabs.io/blog/introducing-scribe-v2-realtime
|
|
69
|
+
* @see https://elevenlabs.io/docs/cookbooks/speech-to-text/streaming
|
|
70
|
+
* @see https://elevenlabs.io/docs/api-reference/speech-to-text/convert
|
|
71
|
+
*/
|
|
72
|
+
declare enum ElevenLabsModel {
|
|
73
|
+
SCRIBE_V2_REALTIME = "scribe_v2_realtime",
|
|
74
|
+
SCRIBE_V1 = "scribe_v1"
|
|
75
|
+
}
|
|
76
|
+
/**
|
|
77
|
+
* OpenAI Realtime API transcription models
|
|
78
|
+
* These are the verified `input_audio_transcription.model` values.
|
|
79
|
+
* @see https://platform.openai.com/docs/guides/realtime
|
|
80
|
+
*/
|
|
81
|
+
declare enum OpenAIRealtimeModel {
|
|
82
|
+
GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe"
|
|
83
|
+
}
|
|
50
84
|
/**
|
|
51
85
|
* Type alias for any model from any provider
|
|
52
86
|
*/
|
|
53
|
-
type RecognitionModel = DeepgramModel | GoogleModel | string;
|
|
87
|
+
type RecognitionModel = DeepgramModel | GoogleModel | FireworksModel | ElevenLabsModel | OpenAIRealtimeModel | string;
|
|
54
88
|
|
|
55
89
|
/**
|
|
56
90
|
* Audio encoding types
|
|
@@ -180,10 +214,10 @@ declare enum Language {
|
|
|
180
214
|
* Message type discriminator for recognition results V1
|
|
181
215
|
*/
|
|
182
216
|
declare enum RecognitionResultTypeV1 {
|
|
183
|
-
TRANSCRIPTION = "Transcription"
|
|
184
|
-
FUNCTION_CALL = "FunctionCall"
|
|
185
|
-
METADATA = "Metadata"
|
|
186
|
-
ERROR = "Error"
|
|
217
|
+
TRANSCRIPTION = "Transcription",
|
|
218
|
+
FUNCTION_CALL = "FunctionCall",
|
|
219
|
+
METADATA = "Metadata",
|
|
220
|
+
ERROR = "Error",
|
|
187
221
|
CLIENT_CONTROL_MESSAGE = "ClientControlMessage"
|
|
188
222
|
}
|
|
189
223
|
/**
|
|
@@ -312,12 +346,12 @@ type MetadataResultV1 = z.infer<typeof MetadataResultSchemaV1>;
|
|
|
312
346
|
* Error type enum V1 - categorizes different types of errors
|
|
313
347
|
*/
|
|
314
348
|
declare enum ErrorTypeV1 {
|
|
315
|
-
AUTHENTICATION_ERROR = "authentication_error"
|
|
316
|
-
VALIDATION_ERROR = "validation_error"
|
|
317
|
-
PROVIDER_ERROR = "provider_error"
|
|
318
|
-
TIMEOUT_ERROR = "timeout_error"
|
|
319
|
-
QUOTA_EXCEEDED = "quota_exceeded"
|
|
320
|
-
CONNECTION_ERROR = "connection_error"
|
|
349
|
+
AUTHENTICATION_ERROR = "authentication_error",
|
|
350
|
+
VALIDATION_ERROR = "validation_error",
|
|
351
|
+
PROVIDER_ERROR = "provider_error",
|
|
352
|
+
TIMEOUT_ERROR = "timeout_error",
|
|
353
|
+
QUOTA_EXCEEDED = "quota_exceeded",
|
|
354
|
+
CONNECTION_ERROR = "connection_error",
|
|
321
355
|
UNKNOWN_ERROR = "unknown_error"
|
|
322
356
|
}
|
|
323
357
|
/**
|
|
@@ -1103,6 +1137,10 @@ declare class RealTimeTwoWayWebSocketRecognitionClient extends WebSocketAudioCli
|
|
|
1103
1137
|
getStats(): IRecognitionClientStats;
|
|
1104
1138
|
protected onConnected(): void;
|
|
1105
1139
|
protected onDisconnected(code: number, reason: string): void;
|
|
1140
|
+
/**
|
|
1141
|
+
* Get human-readable description for WebSocket close code
|
|
1142
|
+
*/
|
|
1143
|
+
private getCloseCodeDescription;
|
|
1106
1144
|
protected onError(error: Event): void;
|
|
1107
1145
|
protected onMessage(msg: {
|
|
1108
1146
|
v: number;
|
package/dist/index.bundled.d.ts
CHANGED
|
@@ -10,9 +10,12 @@ import { z } from 'zod';
|
|
|
10
10
|
declare enum RecognitionProvider {
|
|
11
11
|
ASSEMBLYAI = "assemblyai",
|
|
12
12
|
DEEPGRAM = "deepgram",
|
|
13
|
+
ELEVENLABS = "elevenlabs",
|
|
14
|
+
FIREWORKS = "fireworks",
|
|
13
15
|
GOOGLE = "google",
|
|
14
16
|
GEMINI_BATCH = "gemini-batch",
|
|
15
|
-
OPENAI_BATCH = "openai-batch"
|
|
17
|
+
OPENAI_BATCH = "openai-batch",
|
|
18
|
+
OPENAI_REALTIME = "openai-realtime"
|
|
16
19
|
}
|
|
17
20
|
/**
|
|
18
21
|
* ASR API type - distinguishes between streaming and file-based transcription APIs
|
|
@@ -34,23 +37,54 @@ declare enum DeepgramModel {
|
|
|
34
37
|
/**
|
|
35
38
|
* Google Cloud Speech models
|
|
36
39
|
* @see https://cloud.google.com/speech-to-text/docs/transcription-model
|
|
40
|
+
* @see https://cloud.google.com/speech-to-text/v2/docs/chirp_3-model
|
|
37
41
|
*/
|
|
38
42
|
declare enum GoogleModel {
|
|
43
|
+
CHIRP_3 = "chirp_3",
|
|
44
|
+
CHIRP_2 = "chirp_2",
|
|
45
|
+
CHIRP = "chirp",
|
|
39
46
|
LATEST_LONG = "latest_long",
|
|
40
47
|
LATEST_SHORT = "latest_short",
|
|
41
48
|
TELEPHONY = "telephony",
|
|
42
49
|
TELEPHONY_SHORT = "telephony_short",
|
|
43
|
-
MEDICAL_DICTATION = "medical_dictation",
|
|
44
|
-
MEDICAL_CONVERSATION = "medical_conversation",
|
|
45
50
|
DEFAULT = "default",
|
|
46
51
|
COMMAND_AND_SEARCH = "command_and_search",
|
|
47
52
|
PHONE_CALL = "phone_call",
|
|
48
53
|
VIDEO = "video"
|
|
49
54
|
}
|
|
55
|
+
/**
|
|
56
|
+
* Fireworks AI models for ASR
|
|
57
|
+
* @see https://docs.fireworks.ai/guides/querying-asr-models
|
|
58
|
+
* @see https://fireworks.ai/models/fireworks/fireworks-asr-large
|
|
59
|
+
*/
|
|
60
|
+
declare enum FireworksModel {
|
|
61
|
+
ASR_V1 = "fireworks-asr-large",
|
|
62
|
+
ASR_V2 = "fireworks-asr-v2",
|
|
63
|
+
WHISPER_V3 = "whisper-v3",
|
|
64
|
+
WHISPER_V3_TURBO = "whisper-v3-turbo"
|
|
65
|
+
}
|
|
66
|
+
/**
|
|
67
|
+
* ElevenLabs Scribe models for speech-to-text
|
|
68
|
+
* @see https://elevenlabs.io/blog/introducing-scribe-v2-realtime
|
|
69
|
+
* @see https://elevenlabs.io/docs/cookbooks/speech-to-text/streaming
|
|
70
|
+
* @see https://elevenlabs.io/docs/api-reference/speech-to-text/convert
|
|
71
|
+
*/
|
|
72
|
+
declare enum ElevenLabsModel {
|
|
73
|
+
SCRIBE_V2_REALTIME = "scribe_v2_realtime",
|
|
74
|
+
SCRIBE_V1 = "scribe_v1"
|
|
75
|
+
}
|
|
76
|
+
/**
|
|
77
|
+
* OpenAI Realtime API transcription models
|
|
78
|
+
* These are the verified `input_audio_transcription.model` values.
|
|
79
|
+
* @see https://platform.openai.com/docs/guides/realtime
|
|
80
|
+
*/
|
|
81
|
+
declare enum OpenAIRealtimeModel {
|
|
82
|
+
GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe"
|
|
83
|
+
}
|
|
50
84
|
/**
|
|
51
85
|
* Type alias for any model from any provider
|
|
52
86
|
*/
|
|
53
|
-
type RecognitionModel = DeepgramModel | GoogleModel | string;
|
|
87
|
+
type RecognitionModel = DeepgramModel | GoogleModel | FireworksModel | ElevenLabsModel | OpenAIRealtimeModel | string;
|
|
54
88
|
|
|
55
89
|
/**
|
|
56
90
|
* Audio encoding types
|
|
@@ -180,10 +214,10 @@ declare enum Language {
|
|
|
180
214
|
* Message type discriminator for recognition results V1
|
|
181
215
|
*/
|
|
182
216
|
declare enum RecognitionResultTypeV1 {
|
|
183
|
-
TRANSCRIPTION = "Transcription"
|
|
184
|
-
FUNCTION_CALL = "FunctionCall"
|
|
185
|
-
METADATA = "Metadata"
|
|
186
|
-
ERROR = "Error"
|
|
217
|
+
TRANSCRIPTION = "Transcription",
|
|
218
|
+
FUNCTION_CALL = "FunctionCall",
|
|
219
|
+
METADATA = "Metadata",
|
|
220
|
+
ERROR = "Error",
|
|
187
221
|
CLIENT_CONTROL_MESSAGE = "ClientControlMessage"
|
|
188
222
|
}
|
|
189
223
|
/**
|
|
@@ -312,12 +346,12 @@ type MetadataResultV1 = z.infer<typeof MetadataResultSchemaV1>;
|
|
|
312
346
|
* Error type enum V1 - categorizes different types of errors
|
|
313
347
|
*/
|
|
314
348
|
declare enum ErrorTypeV1 {
|
|
315
|
-
AUTHENTICATION_ERROR = "authentication_error"
|
|
316
|
-
VALIDATION_ERROR = "validation_error"
|
|
317
|
-
PROVIDER_ERROR = "provider_error"
|
|
318
|
-
TIMEOUT_ERROR = "timeout_error"
|
|
319
|
-
QUOTA_EXCEEDED = "quota_exceeded"
|
|
320
|
-
CONNECTION_ERROR = "connection_error"
|
|
349
|
+
AUTHENTICATION_ERROR = "authentication_error",
|
|
350
|
+
VALIDATION_ERROR = "validation_error",
|
|
351
|
+
PROVIDER_ERROR = "provider_error",
|
|
352
|
+
TIMEOUT_ERROR = "timeout_error",
|
|
353
|
+
QUOTA_EXCEEDED = "quota_exceeded",
|
|
354
|
+
CONNECTION_ERROR = "connection_error",
|
|
321
355
|
UNKNOWN_ERROR = "unknown_error"
|
|
322
356
|
}
|
|
323
357
|
/**
|
|
@@ -353,7 +387,7 @@ type ErrorResultV1 = z.infer<typeof ErrorResultSchemaV1>;
|
|
|
353
387
|
* In the long run audio client(mic) should know it, rather than servers.
|
|
354
388
|
*/
|
|
355
389
|
declare enum ClientControlActionV1 {
|
|
356
|
-
READY_FOR_UPLOADING_RECORDING = "ready_for_uploading_recording"
|
|
390
|
+
READY_FOR_UPLOADING_RECORDING = "ready_for_uploading_recording",
|
|
357
391
|
STOP_RECORDING = "stop_recording"
|
|
358
392
|
}
|
|
359
393
|
|
|
@@ -1239,14 +1273,11 @@ declare function createDefaultASRConfig(overrides?: PartialASRRequestConfig): AS
|
|
|
1239
1273
|
* @see https://ai.google.dev/gemini-api/docs/api-versions
|
|
1240
1274
|
*/
|
|
1241
1275
|
declare enum GeminiModel {
|
|
1242
|
-
GEMINI_2_5_PRO = "gemini-2.5-pro"
|
|
1243
|
-
GEMINI_2_5_FLASH = "gemini-2.5-flash"
|
|
1244
|
-
GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite"
|
|
1245
|
-
GEMINI_2_0_FLASH_LATEST = "gemini-2.0-flash-latest"
|
|
1246
|
-
|
|
1247
|
-
GEMINI_2_0_FLASH_EXP = "gemini-2.0-flash-exp",// Experimental version
|
|
1248
|
-
GEMINI_1_5_FLASH = "gemini-1.5-flash",
|
|
1249
|
-
GEMINI_1_5_PRO = "gemini-1.5-pro"
|
|
1276
|
+
GEMINI_2_5_PRO = "gemini-2.5-pro",
|
|
1277
|
+
GEMINI_2_5_FLASH = "gemini-2.5-flash",
|
|
1278
|
+
GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",
|
|
1279
|
+
GEMINI_2_0_FLASH_LATEST = "gemini-2.0-flash-latest",
|
|
1280
|
+
GEMINI_2_0_FLASH_EXP = "gemini-2.0-flash-exp"
|
|
1250
1281
|
}
|
|
1251
1282
|
|
|
1252
1283
|
/**
|
|
@@ -1782,6 +1813,10 @@ declare class RealTimeTwoWayWebSocketRecognitionClient extends WebSocketAudioCli
|
|
|
1782
1813
|
getStats(): IRecognitionClientStats;
|
|
1783
1814
|
protected onConnected(): void;
|
|
1784
1815
|
protected onDisconnected(code: number, reason: string): void;
|
|
1816
|
+
/**
|
|
1817
|
+
* Get human-readable description for WebSocket close code
|
|
1818
|
+
*/
|
|
1819
|
+
private getCloseCodeDescription;
|
|
1785
1820
|
protected onError(error: Event): void;
|
|
1786
1821
|
protected onMessage(msg: {
|
|
1787
1822
|
v: number;
|
|
@@ -2033,6 +2068,7 @@ declare const RecognitionVGFStateSchema: z.ZodObject<{
|
|
|
2033
2068
|
functionCallConfidence: z.ZodOptional<z.ZodNumber>;
|
|
2034
2069
|
finalFunctionCallTimestamp: z.ZodOptional<z.ZodString>;
|
|
2035
2070
|
promptSlotMap: z.ZodOptional<z.ZodRecord<z.ZodString, z.ZodArray<z.ZodString, "many">>>;
|
|
2071
|
+
recognitionActionProcessingState: z.ZodOptional<z.ZodString>;
|
|
2036
2072
|
}, "strip", z.ZodTypeAny, {
|
|
2037
2073
|
audioUtteranceId: string;
|
|
2038
2074
|
pendingTranscript: string;
|
|
@@ -2049,6 +2085,7 @@ declare const RecognitionVGFStateSchema: z.ZodObject<{
|
|
|
2049
2085
|
functionCallConfidence?: number | undefined;
|
|
2050
2086
|
finalFunctionCallTimestamp?: string | undefined;
|
|
2051
2087
|
promptSlotMap?: Record<string, string[]> | undefined;
|
|
2088
|
+
recognitionActionProcessingState?: string | undefined;
|
|
2052
2089
|
}, {
|
|
2053
2090
|
audioUtteranceId: string;
|
|
2054
2091
|
startRecordingStatus?: string | undefined;
|
|
@@ -2065,6 +2102,7 @@ declare const RecognitionVGFStateSchema: z.ZodObject<{
|
|
|
2065
2102
|
functionCallConfidence?: number | undefined;
|
|
2066
2103
|
finalFunctionCallTimestamp?: string | undefined;
|
|
2067
2104
|
promptSlotMap?: Record<string, string[]> | undefined;
|
|
2105
|
+
recognitionActionProcessingState?: string | undefined;
|
|
2068
2106
|
}>;
|
|
2069
2107
|
type RecognitionState = z.infer<typeof RecognitionVGFStateSchema>;
|
|
2070
2108
|
declare const RecordingStatus: {
|
|
@@ -2197,6 +2235,8 @@ declare class SimplifiedVGFRecognitionClient implements ISimplifiedVGFRecognitio
|
|
|
2197
2235
|
private state;
|
|
2198
2236
|
private isRecordingAudio;
|
|
2199
2237
|
private stateChangeCallback;
|
|
2238
|
+
private expectedUuid;
|
|
2239
|
+
private logger;
|
|
2200
2240
|
constructor(config: SimplifiedVGFClientConfig);
|
|
2201
2241
|
connect(): Promise<void>;
|
|
2202
2242
|
sendAudio(audioData: ArrayBuffer | ArrayBufferView | Blob): void;
|
|
@@ -2298,5 +2338,5 @@ declare function getRecognitionConductorHttpBase(stage?: Stage | string | null |
|
|
|
2298
2338
|
declare function getRecognitionConductorWsBase(stage?: Stage | string | null | undefined): string;
|
|
2299
2339
|
declare function getRecognitionConductorHost(stage?: Stage | string | null | undefined): string;
|
|
2300
2340
|
|
|
2301
|
-
export { AudioEncoding, ClientControlActionV1, ClientState, ConfigBuilder, ConnectionError, ControlSignalTypeV1 as ControlSignal, ControlSignalTypeV1, DeepgramModel, ErrorTypeV1, FinalTranscriptStability, GeminiModel, GoogleModel, Language, OpenAIModel, RECOGNITION_CONDUCTOR_BASES, RECOGNITION_SERVICE_BASES, RealTimeTwoWayWebSocketRecognitionClient, RecognitionContextTypeV1, RecognitionError, RecognitionProvider, RecognitionResultTypeV1, RecognitionVGFStateSchema, RecordingStatus, STAGES, SampleRate, SimplifiedVGFRecognitionClient, TimeoutError, TranscriptionStatus, ValidationError, createClient, createClientWithBuilder, createDefaultASRConfig, createInitialRecognitionState, createSimplifiedVGFClient, getRecognitionConductorBase, getRecognitionConductorHost, getRecognitionConductorHttpBase, getRecognitionConductorWsBase, getRecognitionServiceBase, getRecognitionServiceHost, getRecognitionServiceHttpBase, getRecognitionServiceWsBase, getUserFriendlyMessage, isExceptionImmediatelyAvailable, isNormalDisconnection, isValidRecordingStatusTransition, normalizeStage };
|
|
2341
|
+
export { AudioEncoding, ClientControlActionV1, ClientState, ConfigBuilder, ConnectionError, ControlSignalTypeV1 as ControlSignal, ControlSignalTypeV1, DeepgramModel, ElevenLabsModel, ErrorTypeV1, FinalTranscriptStability, FireworksModel, GeminiModel, GoogleModel, Language, OpenAIModel, RECOGNITION_CONDUCTOR_BASES, RECOGNITION_SERVICE_BASES, RealTimeTwoWayWebSocketRecognitionClient, RecognitionContextTypeV1, RecognitionError, RecognitionProvider, RecognitionResultTypeV1, RecognitionVGFStateSchema, RecordingStatus, STAGES, SampleRate, SimplifiedVGFRecognitionClient, TimeoutError, TranscriptionStatus, ValidationError, createClient, createClientWithBuilder, createDefaultASRConfig, createInitialRecognitionState, createSimplifiedVGFClient, getRecognitionConductorBase, getRecognitionConductorHost, getRecognitionConductorHttpBase, getRecognitionConductorWsBase, getRecognitionServiceBase, getRecognitionServiceHost, getRecognitionServiceHttpBase, getRecognitionServiceWsBase, getUserFriendlyMessage, isExceptionImmediatelyAvailable, isNormalDisconnection, isValidRecordingStatusTransition, normalizeStage };
|
|
2302
2342
|
export type { ASRRequestConfig, ASRRequestV1, AuthenticationException, ConnectionException, ErrorResultV1, FunctionCallResultV1, GameContextV1, IRecognitionClient, IRecognitionClientConfig, IRecognitionClientStats, ISimplifiedVGFRecognitionClient, MetadataResultV1, ProviderException, QuotaExceededException, RealTimeTwoWayWebSocketRecognitionClientConfig, RecognitionCallbackUrl, RecognitionException, RecognitionState, RecordingStatusType, SimplifiedVGFClientConfig, SlotMap, Stage, TimeoutException, TranscriptionResult, TranscriptionResultV1, TranscriptionStatusType, UnknownException, ValidationException };
|
package/dist/index.d.ts
CHANGED
|
@@ -10,6 +10,6 @@ export { SimplifiedVGFRecognitionClient, createSimplifiedVGFClient, type ISimpli
|
|
|
10
10
|
export { type RecognitionState, RecognitionVGFStateSchema, RecordingStatus, TranscriptionStatus, type RecordingStatusType, type TranscriptionStatusType, createInitialRecognitionState, isValidRecordingStatusTransition } from './vgf-recognition-state.js';
|
|
11
11
|
export { AudioEncoding } from '@recog/websocket';
|
|
12
12
|
export { type GameContextV1, type SlotMap, RecognitionContextTypeV1, ControlSignalTypeV1, ControlSignalTypeV1 as ControlSignal, // Alias for backward compatibility
|
|
13
|
-
type TranscriptionResultV1, type FunctionCallResultV1, type MetadataResultV1, type ErrorResultV1, RecognitionResultTypeV1, ClientControlActionV1, type ASRRequestConfig, type ASRRequestV1, FinalTranscriptStability, createDefaultASRConfig, RecognitionProvider, DeepgramModel, GoogleModel, GeminiModel, OpenAIModel, Language, SampleRate, STAGES, type Stage } from '@recog/shared-types';
|
|
13
|
+
type TranscriptionResultV1, type FunctionCallResultV1, type MetadataResultV1, type ErrorResultV1, RecognitionResultTypeV1, ClientControlActionV1, type ASRRequestConfig, type ASRRequestV1, FinalTranscriptStability, createDefaultASRConfig, RecognitionProvider, DeepgramModel, ElevenLabsModel, FireworksModel, GoogleModel, GeminiModel, OpenAIModel, Language, SampleRate, STAGES, type Stage } from '@recog/shared-types';
|
|
14
14
|
export { getRecognitionServiceBase, getRecognitionServiceHttpBase, getRecognitionServiceWsBase, getRecognitionServiceHost, getRecognitionConductorBase, getRecognitionConductorHttpBase, getRecognitionConductorWsBase, getRecognitionConductorHost, normalizeStage, RECOGNITION_SERVICE_BASES, RECOGNITION_CONDUCTOR_BASES } from '@recog/shared-config';
|
|
15
15
|
//# sourceMappingURL=index.d.ts.map
|
package/dist/index.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AACA,OAAO,EACL,wCAAwC,EACxC,KAAK,8CAA8C,EACnD,KAAK,mBAAmB,EACxB,qBAAqB,EACtB,MAAM,yBAAyB,CAAC;AAGjC,OAAO,EACL,KAAK,kBAAkB,EACvB,KAAK,wBAAwB,EAC7B,KAAK,uBAAuB,EAC5B,KAAK,sBAAsB,EAC3B,WAAW,EACZ,MAAM,+BAA+B,CAAC;AAGvC,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AAGpD,OAAO,EAAE,YAAY,EAAE,uBAAuB,EAAE,MAAM,cAAc,CAAC;AAGrE,OAAO,EACL,gBAAgB,EAChB,eAAe,EACf,YAAY,EACZ,eAAe,EAChB,MAAM,aAAa,CAAC;AAGrB,OAAO,EAAE,WAAW,EAAE,MAAM,qBAAqB,CAAC;AAGlD,YAAY,EACV,oBAAoB,EACpB,mBAAmB,EACnB,gBAAgB,EAChB,mBAAmB,EACnB,uBAAuB,EACvB,iBAAiB,EACjB,sBAAsB,EACtB,gBAAgB,EACjB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,+BAA+B,EAC/B,sBAAsB,EACvB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,8BAA8B,EAC9B,yBAAyB,EACzB,KAAK,+BAA+B,EACpC,KAAK,yBAAyB,EAC/B,MAAM,wCAAwC,CAAC;AAEhD,OAAO,EACL,KAAK,gBAAgB,EACrB,yBAAyB,EACzB,eAAe,EACf,mBAAmB,EACnB,KAAK,mBAAmB,EACxB,KAAK,uBAAuB,EAC5B,6BAA6B,EAC7B,gCAAgC,EACjC,MAAM,4BAA4B,CAAC;AAGpC,OAAO,EAAE,aAAa,EAAE,MAAM,kBAAkB,CAAC;AAGjD,OAAO,EAEL,KAAK,aAAa,EAClB,KAAK,OAAO,EACZ,wBAAwB,EACxB,mBAAmB,EACnB,mBAAmB,IAAI,aAAa,EAAG,mCAAmC;AAG1E,KAAK,qBAAqB,EAC1B,KAAK,oBAAoB,EACzB,KAAK,gBAAgB,EACrB,KAAK,aAAa,EAClB,uBAAuB,EACvB,qBAAqB,EAGrB,KAAK,gBAAgB,EACrB,KAAK,YAAY,EACjB,wBAAwB,EACxB,sBAAsB,EACtB,mBAAmB,EACnB,aAAa,EACb,WAAW,EACX,WAAW,EACX,WAAW,EACX,QAAQ,EACR,UAAU,EAGV,MAAM,EACN,KAAK,KAAK,EACX,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,yBAAyB,EACzB,6BAA6B,EAC7B,2BAA2B,EAC3B,yBAAyB,EACzB,2BAA2B,EAC3B,+BAA+B,EAC/B,6BAA6B,EAC7B,2BAA2B,EAC3B,cAAc,EACd,yBAAyB,EACzB,2BAA2B,EAC5B,MAAM,sBAAsB,CAAC"}
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AACA,OAAO,EACL,wCAAwC,EACxC,KAAK,8CAA8C,EACnD,KAAK,mBAAmB,EACxB,qBAAqB,EACtB,MAAM,yBAAyB,CAAC;AAGjC,OAAO,EACL,KAAK,kBAAkB,EACvB,KAAK,wBAAwB,EAC7B,KAAK,uBAAuB,EAC5B,KAAK,sBAAsB,EAC3B,WAAW,EACZ,MAAM,+BAA+B,CAAC;AAGvC,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AAGpD,OAAO,EAAE,YAAY,EAAE,uBAAuB,EAAE,MAAM,cAAc,CAAC;AAGrE,OAAO,EACL,gBAAgB,EAChB,eAAe,EACf,YAAY,EACZ,eAAe,EAChB,MAAM,aAAa,CAAC;AAGrB,OAAO,EAAE,WAAW,EAAE,MAAM,qBAAqB,CAAC;AAGlD,YAAY,EACV,oBAAoB,EACpB,mBAAmB,EACnB,gBAAgB,EAChB,mBAAmB,EACnB,uBAAuB,EACvB,iBAAiB,EACjB,sBAAsB,EACtB,gBAAgB,EACjB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,+BAA+B,EAC/B,sBAAsB,EACvB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,8BAA8B,EAC9B,yBAAyB,EACzB,KAAK,+BAA+B,EACpC,KAAK,yBAAyB,EAC/B,MAAM,wCAAwC,CAAC;AAEhD,OAAO,EACL,KAAK,gBAAgB,EACrB,yBAAyB,EACzB,eAAe,EACf,mBAAmB,EACnB,KAAK,mBAAmB,EACxB,KAAK,uBAAuB,EAC5B,6BAA6B,EAC7B,gCAAgC,EACjC,MAAM,4BAA4B,CAAC;AAGpC,OAAO,EAAE,aAAa,EAAE,MAAM,kBAAkB,CAAC;AAGjD,OAAO,EAEL,KAAK,aAAa,EAClB,KAAK,OAAO,EACZ,wBAAwB,EACxB,mBAAmB,EACnB,mBAAmB,IAAI,aAAa,EAAG,mCAAmC;AAG1E,KAAK,qBAAqB,EAC1B,KAAK,oBAAoB,EACzB,KAAK,gBAAgB,EACrB,KAAK,aAAa,EAClB,uBAAuB,EACvB,qBAAqB,EAGrB,KAAK,gBAAgB,EACrB,KAAK,YAAY,EACjB,wBAAwB,EACxB,sBAAsB,EACtB,mBAAmB,EACnB,aAAa,EACb,eAAe,EACf,cAAc,EACd,WAAW,EACX,WAAW,EACX,WAAW,EACX,QAAQ,EACR,UAAU,EAGV,MAAM,EACN,KAAK,KAAK,EACX,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,yBAAyB,EACzB,6BAA6B,EAC7B,2BAA2B,EAC3B,yBAAyB,EACzB,2BAA2B,EAC3B,+BAA+B,EAC/B,6BAA6B,EAC7B,2BAA2B,EAC3B,cAAc,EACd,yBAAyB,EACzB,2BAA2B,EAC5B,MAAM,sBAAsB,CAAC"}
|