@volley/recognition-client-sdk 0.1.297 → 0.1.382

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -10,9 +10,12 @@ import { z } from 'zod';
10
10
  declare enum RecognitionProvider {
11
11
  ASSEMBLYAI = "assemblyai",
12
12
  DEEPGRAM = "deepgram",
13
+ ELEVENLABS = "elevenlabs",
14
+ FIREWORKS = "fireworks",
13
15
  GOOGLE = "google",
14
16
  GEMINI_BATCH = "gemini-batch",
15
- OPENAI_BATCH = "openai-batch"
17
+ OPENAI_BATCH = "openai-batch",
18
+ OPENAI_REALTIME = "openai-realtime"
16
19
  }
17
20
  /**
18
21
  * ASR API type - distinguishes between streaming and file-based transcription APIs
@@ -34,23 +37,54 @@ declare enum DeepgramModel {
34
37
  /**
35
38
  * Google Cloud Speech models
36
39
  * @see https://cloud.google.com/speech-to-text/docs/transcription-model
40
+ * @see https://cloud.google.com/speech-to-text/v2/docs/chirp_3-model
37
41
  */
38
42
  declare enum GoogleModel {
43
+ CHIRP_3 = "chirp_3",
44
+ CHIRP_2 = "chirp_2",
45
+ CHIRP = "chirp",
39
46
  LATEST_LONG = "latest_long",
40
47
  LATEST_SHORT = "latest_short",
41
48
  TELEPHONY = "telephony",
42
49
  TELEPHONY_SHORT = "telephony_short",
43
- MEDICAL_DICTATION = "medical_dictation",
44
- MEDICAL_CONVERSATION = "medical_conversation",
45
50
  DEFAULT = "default",
46
51
  COMMAND_AND_SEARCH = "command_and_search",
47
52
  PHONE_CALL = "phone_call",
48
53
  VIDEO = "video"
49
54
  }
55
+ /**
56
+ * Fireworks AI models for ASR
57
+ * @see https://docs.fireworks.ai/guides/querying-asr-models
58
+ * @see https://fireworks.ai/models/fireworks/fireworks-asr-large
59
+ */
60
+ declare enum FireworksModel {
61
+ ASR_V1 = "fireworks-asr-large",
62
+ ASR_V2 = "fireworks-asr-v2",
63
+ WHISPER_V3 = "whisper-v3",
64
+ WHISPER_V3_TURBO = "whisper-v3-turbo"
65
+ }
66
+ /**
67
+ * ElevenLabs Scribe models for speech-to-text
68
+ * @see https://elevenlabs.io/blog/introducing-scribe-v2-realtime
69
+ * @see https://elevenlabs.io/docs/cookbooks/speech-to-text/streaming
70
+ * @see https://elevenlabs.io/docs/api-reference/speech-to-text/convert
71
+ */
72
+ declare enum ElevenLabsModel {
73
+ SCRIBE_V2_REALTIME = "scribe_v2_realtime",
74
+ SCRIBE_V1 = "scribe_v1"
75
+ }
76
+ /**
77
+ * OpenAI Realtime API transcription models
78
+ * These are the verified `input_audio_transcription.model` values.
79
+ * @see https://platform.openai.com/docs/guides/realtime
80
+ */
81
+ declare enum OpenAIRealtimeModel {
82
+ GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe"
83
+ }
50
84
  /**
51
85
  * Type alias for any model from any provider
52
86
  */
53
- type RecognitionModel = DeepgramModel | GoogleModel | string;
87
+ type RecognitionModel = DeepgramModel | GoogleModel | FireworksModel | ElevenLabsModel | OpenAIRealtimeModel | string;
54
88
 
55
89
  /**
56
90
  * Audio encoding types
@@ -180,10 +214,10 @@ declare enum Language {
180
214
  * Message type discriminator for recognition results V1
181
215
  */
182
216
  declare enum RecognitionResultTypeV1 {
183
- TRANSCRIPTION = "Transcription",// Transcript message contains all in the history. result of STT(Speech to text)
184
- FUNCTION_CALL = "FunctionCall",// Not supported in P1.result of STF(Speedch to function call) Function call schema
185
- METADATA = "Metadata",// Metadata message contains all the timestamps, provider info, and ASR config
186
- ERROR = "Error",// Error message contains the error details
217
+ TRANSCRIPTION = "Transcription",
218
+ FUNCTION_CALL = "FunctionCall",
219
+ METADATA = "Metadata",
220
+ ERROR = "Error",
187
221
  CLIENT_CONTROL_MESSAGE = "ClientControlMessage"
188
222
  }
189
223
  /**
@@ -312,12 +346,12 @@ type MetadataResultV1 = z.infer<typeof MetadataResultSchemaV1>;
312
346
  * Error type enum V1 - categorizes different types of errors
313
347
  */
314
348
  declare enum ErrorTypeV1 {
315
- AUTHENTICATION_ERROR = "authentication_error",// Authentication/authorization failures
316
- VALIDATION_ERROR = "validation_error",// Invalid input or configuration
317
- PROVIDER_ERROR = "provider_error",// Error from ASR provider (Deepgram, Google, etc.) Unlikely to happen with fall
318
- TIMEOUT_ERROR = "timeout_error",// Request or operation timeout. Likely business logic did not handle timeout.
319
- QUOTA_EXCEEDED = "quota_exceeded",// Quota or rate limit exceeded. Unlikely to happen with fallbakcs
320
- CONNECTION_ERROR = "connection_error",// Connection establishment or network error
349
+ AUTHENTICATION_ERROR = "authentication_error",
350
+ VALIDATION_ERROR = "validation_error",
351
+ PROVIDER_ERROR = "provider_error",
352
+ TIMEOUT_ERROR = "timeout_error",
353
+ QUOTA_EXCEEDED = "quota_exceeded",
354
+ CONNECTION_ERROR = "connection_error",
321
355
  UNKNOWN_ERROR = "unknown_error"
322
356
  }
323
357
  /**
@@ -1103,6 +1137,10 @@ declare class RealTimeTwoWayWebSocketRecognitionClient extends WebSocketAudioCli
1103
1137
  getStats(): IRecognitionClientStats;
1104
1138
  protected onConnected(): void;
1105
1139
  protected onDisconnected(code: number, reason: string): void;
1140
+ /**
1141
+ * Get human-readable description for WebSocket close code
1142
+ */
1143
+ private getCloseCodeDescription;
1106
1144
  protected onError(error: Event): void;
1107
1145
  protected onMessage(msg: {
1108
1146
  v: number;
@@ -10,9 +10,12 @@ import { z } from 'zod';
10
10
  declare enum RecognitionProvider {
11
11
  ASSEMBLYAI = "assemblyai",
12
12
  DEEPGRAM = "deepgram",
13
+ ELEVENLABS = "elevenlabs",
14
+ FIREWORKS = "fireworks",
13
15
  GOOGLE = "google",
14
16
  GEMINI_BATCH = "gemini-batch",
15
- OPENAI_BATCH = "openai-batch"
17
+ OPENAI_BATCH = "openai-batch",
18
+ OPENAI_REALTIME = "openai-realtime"
16
19
  }
17
20
  /**
18
21
  * ASR API type - distinguishes between streaming and file-based transcription APIs
@@ -34,23 +37,54 @@ declare enum DeepgramModel {
34
37
  /**
35
38
  * Google Cloud Speech models
36
39
  * @see https://cloud.google.com/speech-to-text/docs/transcription-model
40
+ * @see https://cloud.google.com/speech-to-text/v2/docs/chirp_3-model
37
41
  */
38
42
  declare enum GoogleModel {
43
+ CHIRP_3 = "chirp_3",
44
+ CHIRP_2 = "chirp_2",
45
+ CHIRP = "chirp",
39
46
  LATEST_LONG = "latest_long",
40
47
  LATEST_SHORT = "latest_short",
41
48
  TELEPHONY = "telephony",
42
49
  TELEPHONY_SHORT = "telephony_short",
43
- MEDICAL_DICTATION = "medical_dictation",
44
- MEDICAL_CONVERSATION = "medical_conversation",
45
50
  DEFAULT = "default",
46
51
  COMMAND_AND_SEARCH = "command_and_search",
47
52
  PHONE_CALL = "phone_call",
48
53
  VIDEO = "video"
49
54
  }
55
+ /**
56
+ * Fireworks AI models for ASR
57
+ * @see https://docs.fireworks.ai/guides/querying-asr-models
58
+ * @see https://fireworks.ai/models/fireworks/fireworks-asr-large
59
+ */
60
+ declare enum FireworksModel {
61
+ ASR_V1 = "fireworks-asr-large",
62
+ ASR_V2 = "fireworks-asr-v2",
63
+ WHISPER_V3 = "whisper-v3",
64
+ WHISPER_V3_TURBO = "whisper-v3-turbo"
65
+ }
66
+ /**
67
+ * ElevenLabs Scribe models for speech-to-text
68
+ * @see https://elevenlabs.io/blog/introducing-scribe-v2-realtime
69
+ * @see https://elevenlabs.io/docs/cookbooks/speech-to-text/streaming
70
+ * @see https://elevenlabs.io/docs/api-reference/speech-to-text/convert
71
+ */
72
+ declare enum ElevenLabsModel {
73
+ SCRIBE_V2_REALTIME = "scribe_v2_realtime",
74
+ SCRIBE_V1 = "scribe_v1"
75
+ }
76
+ /**
77
+ * OpenAI Realtime API transcription models
78
+ * These are the verified `input_audio_transcription.model` values.
79
+ * @see https://platform.openai.com/docs/guides/realtime
80
+ */
81
+ declare enum OpenAIRealtimeModel {
82
+ GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe"
83
+ }
50
84
  /**
51
85
  * Type alias for any model from any provider
52
86
  */
53
- type RecognitionModel = DeepgramModel | GoogleModel | string;
87
+ type RecognitionModel = DeepgramModel | GoogleModel | FireworksModel | ElevenLabsModel | OpenAIRealtimeModel | string;
54
88
 
55
89
  /**
56
90
  * Audio encoding types
@@ -180,10 +214,10 @@ declare enum Language {
180
214
  * Message type discriminator for recognition results V1
181
215
  */
182
216
  declare enum RecognitionResultTypeV1 {
183
- TRANSCRIPTION = "Transcription",// Transcript message contains all in the history. result of STT(Speech to text)
184
- FUNCTION_CALL = "FunctionCall",// Not supported in P1.result of STF(Speedch to function call) Function call schema
185
- METADATA = "Metadata",// Metadata message contains all the timestamps, provider info, and ASR config
186
- ERROR = "Error",// Error message contains the error details
217
+ TRANSCRIPTION = "Transcription",
218
+ FUNCTION_CALL = "FunctionCall",
219
+ METADATA = "Metadata",
220
+ ERROR = "Error",
187
221
  CLIENT_CONTROL_MESSAGE = "ClientControlMessage"
188
222
  }
189
223
  /**
@@ -312,12 +346,12 @@ type MetadataResultV1 = z.infer<typeof MetadataResultSchemaV1>;
312
346
  * Error type enum V1 - categorizes different types of errors
313
347
  */
314
348
  declare enum ErrorTypeV1 {
315
- AUTHENTICATION_ERROR = "authentication_error",// Authentication/authorization failures
316
- VALIDATION_ERROR = "validation_error",// Invalid input or configuration
317
- PROVIDER_ERROR = "provider_error",// Error from ASR provider (Deepgram, Google, etc.) Unlikely to happen with fall
318
- TIMEOUT_ERROR = "timeout_error",// Request or operation timeout. Likely business logic did not handle timeout.
319
- QUOTA_EXCEEDED = "quota_exceeded",// Quota or rate limit exceeded. Unlikely to happen with fallbakcs
320
- CONNECTION_ERROR = "connection_error",// Connection establishment or network error
349
+ AUTHENTICATION_ERROR = "authentication_error",
350
+ VALIDATION_ERROR = "validation_error",
351
+ PROVIDER_ERROR = "provider_error",
352
+ TIMEOUT_ERROR = "timeout_error",
353
+ QUOTA_EXCEEDED = "quota_exceeded",
354
+ CONNECTION_ERROR = "connection_error",
321
355
  UNKNOWN_ERROR = "unknown_error"
322
356
  }
323
357
  /**
@@ -353,7 +387,7 @@ type ErrorResultV1 = z.infer<typeof ErrorResultSchemaV1>;
353
387
  * In the long run audio client(mic) should know it, rather than servers.
354
388
  */
355
389
  declare enum ClientControlActionV1 {
356
- READY_FOR_UPLOADING_RECORDING = "ready_for_uploading_recording",// Server is ready to receive audio, client should start uploading recording (includes audioUtteranceId)
390
+ READY_FOR_UPLOADING_RECORDING = "ready_for_uploading_recording",
357
391
  STOP_RECORDING = "stop_recording"
358
392
  }
359
393
 
@@ -1239,14 +1273,11 @@ declare function createDefaultASRConfig(overrides?: PartialASRRequestConfig): AS
1239
1273
  * @see https://ai.google.dev/gemini-api/docs/api-versions
1240
1274
  */
1241
1275
  declare enum GeminiModel {
1242
- GEMINI_2_5_PRO = "gemini-2.5-pro",// State-of-the-art thinking model
1243
- GEMINI_2_5_FLASH = "gemini-2.5-flash",// Best price-performance balance
1244
- GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",// Fastest, most cost-efficient
1245
- GEMINI_2_0_FLASH_LATEST = "gemini-2.0-flash-latest",// Auto-updated to latest 2.0 flash
1246
- GEMINI_2_0_FLASH = "gemini-2.0-flash-002",// Specific stable version
1247
- GEMINI_2_0_FLASH_EXP = "gemini-2.0-flash-exp",// Experimental version
1248
- GEMINI_1_5_FLASH = "gemini-1.5-flash",
1249
- GEMINI_1_5_PRO = "gemini-1.5-pro"
1276
+ GEMINI_2_5_PRO = "gemini-2.5-pro",
1277
+ GEMINI_2_5_FLASH = "gemini-2.5-flash",
1278
+ GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",
1279
+ GEMINI_2_0_FLASH_LATEST = "gemini-2.0-flash-latest",
1280
+ GEMINI_2_0_FLASH_EXP = "gemini-2.0-flash-exp"
1250
1281
  }
1251
1282
 
1252
1283
  /**
@@ -1782,6 +1813,10 @@ declare class RealTimeTwoWayWebSocketRecognitionClient extends WebSocketAudioCli
1782
1813
  getStats(): IRecognitionClientStats;
1783
1814
  protected onConnected(): void;
1784
1815
  protected onDisconnected(code: number, reason: string): void;
1816
+ /**
1817
+ * Get human-readable description for WebSocket close code
1818
+ */
1819
+ private getCloseCodeDescription;
1785
1820
  protected onError(error: Event): void;
1786
1821
  protected onMessage(msg: {
1787
1822
  v: number;
@@ -2033,6 +2068,7 @@ declare const RecognitionVGFStateSchema: z.ZodObject<{
2033
2068
  functionCallConfidence: z.ZodOptional<z.ZodNumber>;
2034
2069
  finalFunctionCallTimestamp: z.ZodOptional<z.ZodString>;
2035
2070
  promptSlotMap: z.ZodOptional<z.ZodRecord<z.ZodString, z.ZodArray<z.ZodString, "many">>>;
2071
+ recognitionActionProcessingState: z.ZodOptional<z.ZodString>;
2036
2072
  }, "strip", z.ZodTypeAny, {
2037
2073
  audioUtteranceId: string;
2038
2074
  pendingTranscript: string;
@@ -2049,6 +2085,7 @@ declare const RecognitionVGFStateSchema: z.ZodObject<{
2049
2085
  functionCallConfidence?: number | undefined;
2050
2086
  finalFunctionCallTimestamp?: string | undefined;
2051
2087
  promptSlotMap?: Record<string, string[]> | undefined;
2088
+ recognitionActionProcessingState?: string | undefined;
2052
2089
  }, {
2053
2090
  audioUtteranceId: string;
2054
2091
  startRecordingStatus?: string | undefined;
@@ -2065,6 +2102,7 @@ declare const RecognitionVGFStateSchema: z.ZodObject<{
2065
2102
  functionCallConfidence?: number | undefined;
2066
2103
  finalFunctionCallTimestamp?: string | undefined;
2067
2104
  promptSlotMap?: Record<string, string[]> | undefined;
2105
+ recognitionActionProcessingState?: string | undefined;
2068
2106
  }>;
2069
2107
  type RecognitionState = z.infer<typeof RecognitionVGFStateSchema>;
2070
2108
  declare const RecordingStatus: {
@@ -2197,6 +2235,8 @@ declare class SimplifiedVGFRecognitionClient implements ISimplifiedVGFRecognitio
2197
2235
  private state;
2198
2236
  private isRecordingAudio;
2199
2237
  private stateChangeCallback;
2238
+ private expectedUuid;
2239
+ private logger;
2200
2240
  constructor(config: SimplifiedVGFClientConfig);
2201
2241
  connect(): Promise<void>;
2202
2242
  sendAudio(audioData: ArrayBuffer | ArrayBufferView | Blob): void;
@@ -2298,5 +2338,5 @@ declare function getRecognitionConductorHttpBase(stage?: Stage | string | null |
2298
2338
  declare function getRecognitionConductorWsBase(stage?: Stage | string | null | undefined): string;
2299
2339
  declare function getRecognitionConductorHost(stage?: Stage | string | null | undefined): string;
2300
2340
 
2301
- export { AudioEncoding, ClientControlActionV1, ClientState, ConfigBuilder, ConnectionError, ControlSignalTypeV1 as ControlSignal, ControlSignalTypeV1, DeepgramModel, ErrorTypeV1, FinalTranscriptStability, GeminiModel, GoogleModel, Language, OpenAIModel, RECOGNITION_CONDUCTOR_BASES, RECOGNITION_SERVICE_BASES, RealTimeTwoWayWebSocketRecognitionClient, RecognitionContextTypeV1, RecognitionError, RecognitionProvider, RecognitionResultTypeV1, RecognitionVGFStateSchema, RecordingStatus, STAGES, SampleRate, SimplifiedVGFRecognitionClient, TimeoutError, TranscriptionStatus, ValidationError, createClient, createClientWithBuilder, createDefaultASRConfig, createInitialRecognitionState, createSimplifiedVGFClient, getRecognitionConductorBase, getRecognitionConductorHost, getRecognitionConductorHttpBase, getRecognitionConductorWsBase, getRecognitionServiceBase, getRecognitionServiceHost, getRecognitionServiceHttpBase, getRecognitionServiceWsBase, getUserFriendlyMessage, isExceptionImmediatelyAvailable, isNormalDisconnection, isValidRecordingStatusTransition, normalizeStage };
2341
+ export { AudioEncoding, ClientControlActionV1, ClientState, ConfigBuilder, ConnectionError, ControlSignalTypeV1 as ControlSignal, ControlSignalTypeV1, DeepgramModel, ElevenLabsModel, ErrorTypeV1, FinalTranscriptStability, FireworksModel, GeminiModel, GoogleModel, Language, OpenAIModel, RECOGNITION_CONDUCTOR_BASES, RECOGNITION_SERVICE_BASES, RealTimeTwoWayWebSocketRecognitionClient, RecognitionContextTypeV1, RecognitionError, RecognitionProvider, RecognitionResultTypeV1, RecognitionVGFStateSchema, RecordingStatus, STAGES, SampleRate, SimplifiedVGFRecognitionClient, TimeoutError, TranscriptionStatus, ValidationError, createClient, createClientWithBuilder, createDefaultASRConfig, createInitialRecognitionState, createSimplifiedVGFClient, getRecognitionConductorBase, getRecognitionConductorHost, getRecognitionConductorHttpBase, getRecognitionConductorWsBase, getRecognitionServiceBase, getRecognitionServiceHost, getRecognitionServiceHttpBase, getRecognitionServiceWsBase, getUserFriendlyMessage, isExceptionImmediatelyAvailable, isNormalDisconnection, isValidRecordingStatusTransition, normalizeStage };
2302
2342
  export type { ASRRequestConfig, ASRRequestV1, AuthenticationException, ConnectionException, ErrorResultV1, FunctionCallResultV1, GameContextV1, IRecognitionClient, IRecognitionClientConfig, IRecognitionClientStats, ISimplifiedVGFRecognitionClient, MetadataResultV1, ProviderException, QuotaExceededException, RealTimeTwoWayWebSocketRecognitionClientConfig, RecognitionCallbackUrl, RecognitionException, RecognitionState, RecordingStatusType, SimplifiedVGFClientConfig, SlotMap, Stage, TimeoutException, TranscriptionResult, TranscriptionResultV1, TranscriptionStatusType, UnknownException, ValidationException };
package/dist/index.d.ts CHANGED
@@ -10,6 +10,6 @@ export { SimplifiedVGFRecognitionClient, createSimplifiedVGFClient, type ISimpli
10
10
  export { type RecognitionState, RecognitionVGFStateSchema, RecordingStatus, TranscriptionStatus, type RecordingStatusType, type TranscriptionStatusType, createInitialRecognitionState, isValidRecordingStatusTransition } from './vgf-recognition-state.js';
11
11
  export { AudioEncoding } from '@recog/websocket';
12
12
  export { type GameContextV1, type SlotMap, RecognitionContextTypeV1, ControlSignalTypeV1, ControlSignalTypeV1 as ControlSignal, // Alias for backward compatibility
13
- type TranscriptionResultV1, type FunctionCallResultV1, type MetadataResultV1, type ErrorResultV1, RecognitionResultTypeV1, ClientControlActionV1, type ASRRequestConfig, type ASRRequestV1, FinalTranscriptStability, createDefaultASRConfig, RecognitionProvider, DeepgramModel, GoogleModel, GeminiModel, OpenAIModel, Language, SampleRate, STAGES, type Stage } from '@recog/shared-types';
13
+ type TranscriptionResultV1, type FunctionCallResultV1, type MetadataResultV1, type ErrorResultV1, RecognitionResultTypeV1, ClientControlActionV1, type ASRRequestConfig, type ASRRequestV1, FinalTranscriptStability, createDefaultASRConfig, RecognitionProvider, DeepgramModel, ElevenLabsModel, FireworksModel, GoogleModel, GeminiModel, OpenAIModel, Language, SampleRate, STAGES, type Stage } from '@recog/shared-types';
14
14
  export { getRecognitionServiceBase, getRecognitionServiceHttpBase, getRecognitionServiceWsBase, getRecognitionServiceHost, getRecognitionConductorBase, getRecognitionConductorHttpBase, getRecognitionConductorWsBase, getRecognitionConductorHost, normalizeStage, RECOGNITION_SERVICE_BASES, RECOGNITION_CONDUCTOR_BASES } from '@recog/shared-config';
15
15
  //# sourceMappingURL=index.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AACA,OAAO,EACL,wCAAwC,EACxC,KAAK,8CAA8C,EACnD,KAAK,mBAAmB,EACxB,qBAAqB,EACtB,MAAM,yBAAyB,CAAC;AAGjC,OAAO,EACL,KAAK,kBAAkB,EACvB,KAAK,wBAAwB,EAC7B,KAAK,uBAAuB,EAC5B,KAAK,sBAAsB,EAC3B,WAAW,EACZ,MAAM,+BAA+B,CAAC;AAGvC,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AAGpD,OAAO,EAAE,YAAY,EAAE,uBAAuB,EAAE,MAAM,cAAc,CAAC;AAGrE,OAAO,EACL,gBAAgB,EAChB,eAAe,EACf,YAAY,EACZ,eAAe,EAChB,MAAM,aAAa,CAAC;AAGrB,OAAO,EAAE,WAAW,EAAE,MAAM,qBAAqB,CAAC;AAGlD,YAAY,EACV,oBAAoB,EACpB,mBAAmB,EACnB,gBAAgB,EAChB,mBAAmB,EACnB,uBAAuB,EACvB,iBAAiB,EACjB,sBAAsB,EACtB,gBAAgB,EACjB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,+BAA+B,EAC/B,sBAAsB,EACvB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,8BAA8B,EAC9B,yBAAyB,EACzB,KAAK,+BAA+B,EACpC,KAAK,yBAAyB,EAC/B,MAAM,wCAAwC,CAAC;AAEhD,OAAO,EACL,KAAK,gBAAgB,EACrB,yBAAyB,EACzB,eAAe,EACf,mBAAmB,EACnB,KAAK,mBAAmB,EACxB,KAAK,uBAAuB,EAC5B,6BAA6B,EAC7B,gCAAgC,EACjC,MAAM,4BAA4B,CAAC;AAGpC,OAAO,EAAE,aAAa,EAAE,MAAM,kBAAkB,CAAC;AAGjD,OAAO,EAEL,KAAK,aAAa,EAClB,KAAK,OAAO,EACZ,wBAAwB,EACxB,mBAAmB,EACnB,mBAAmB,IAAI,aAAa,EAAG,mCAAmC;AAG1E,KAAK,qBAAqB,EAC1B,KAAK,oBAAoB,EACzB,KAAK,gBAAgB,EACrB,KAAK,aAAa,EAClB,uBAAuB,EACvB,qBAAqB,EAGrB,KAAK,gBAAgB,EACrB,KAAK,YAAY,EACjB,wBAAwB,EACxB,sBAAsB,EACtB,mBAAmB,EACnB,aAAa,EACb,WAAW,EACX,WAAW,EACX,WAAW,EACX,QAAQ,EACR,UAAU,EAGV,MAAM,EACN,KAAK,KAAK,EACX,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,yBAAyB,EACzB,6BAA6B,EAC7B,2BAA2B,EAC3B,yBAAyB,EACzB,2BAA2B,EAC3B,+BAA+B,EAC/B,6BAA6B,EAC7B,2BAA2B,EAC3B,cAAc,EACd,yBAAyB,EACzB,2BAA2B,EAC5B,MAAM,sBAAsB,CAAC"}
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AACA,OAAO,EACL,wCAAwC,EACxC,KAAK,8CAA8C,EACnD,KAAK,mBAAmB,EACxB,qBAAqB,EACtB,MAAM,yBAAyB,CAAC;AAGjC,OAAO,EACL,KAAK,kBAAkB,EACvB,KAAK,wBAAwB,EAC7B,KAAK,uBAAuB,EAC5B,KAAK,sBAAsB,EAC3B,WAAW,EACZ,MAAM,+BAA+B,CAAC;AAGvC,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AAGpD,OAAO,EAAE,YAAY,EAAE,uBAAuB,EAAE,MAAM,cAAc,CAAC;AAGrE,OAAO,EACL,gBAAgB,EAChB,eAAe,EACf,YAAY,EACZ,eAAe,EAChB,MAAM,aAAa,CAAC;AAGrB,OAAO,EAAE,WAAW,EAAE,MAAM,qBAAqB,CAAC;AAGlD,YAAY,EACV,oBAAoB,EACpB,mBAAmB,EACnB,gBAAgB,EAChB,mBAAmB,EACnB,uBAAuB,EACvB,iBAAiB,EACjB,sBAAsB,EACtB,gBAAgB,EACjB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,+BAA+B,EAC/B,sBAAsB,EACvB,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,8BAA8B,EAC9B,yBAAyB,EACzB,KAAK,+BAA+B,EACpC,KAAK,yBAAyB,EAC/B,MAAM,wCAAwC,CAAC;AAEhD,OAAO,EACL,KAAK,gBAAgB,EACrB,yBAAyB,EACzB,eAAe,EACf,mBAAmB,EACnB,KAAK,mBAAmB,EACxB,KAAK,uBAAuB,EAC5B,6BAA6B,EAC7B,gCAAgC,EACjC,MAAM,4BAA4B,CAAC;AAGpC,OAAO,EAAE,aAAa,EAAE,MAAM,kBAAkB,CAAC;AAGjD,OAAO,EAEL,KAAK,aAAa,EAClB,KAAK,OAAO,EACZ,wBAAwB,EACxB,mBAAmB,EACnB,mBAAmB,IAAI,aAAa,EAAG,mCAAmC;AAG1E,KAAK,qBAAqB,EAC1B,KAAK,oBAAoB,EACzB,KAAK,gBAAgB,EACrB,KAAK,aAAa,EAClB,uBAAuB,EACvB,qBAAqB,EAGrB,KAAK,gBAAgB,EACrB,KAAK,YAAY,EACjB,wBAAwB,EACxB,sBAAsB,EACtB,mBAAmB,EACnB,aAAa,EACb,eAAe,EACf,cAAc,EACd,WAAW,EACX,WAAW,EACX,WAAW,EACX,QAAQ,EACR,UAAU,EAGV,MAAM,EACN,KAAK,KAAK,EACX,MAAM,qBAAqB,CAAC;AAG7B,OAAO,EACL,yBAAyB,EACzB,6BAA6B,EAC7B,2BAA2B,EAC3B,yBAAyB,EACzB,2BAA2B,EAC3B,+BAA+B,EAC/B,6BAA6B,EAC7B,2BAA2B,EAC3B,cAAc,EACd,yBAAyB,EACzB,2BAA2B,EAC5B,MAAM,sBAAsB,CAAC"}