voice-router-dev 0.3.2 → 0.3.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.mts CHANGED
@@ -1,5 +1,5 @@
1
- import { GladiaLanguage as TranscriptionLanguageCodeEnum, GladiaTranslationLanguage as TranslationLanguageCodeEnum, DeepgramRedact as ListenV1RedactParameterOneOfItem, DeepgramTopicMode as SharedCustomTopicModeParameter, DeepgramEncoding as ListenV1EncodingParameter, GladiaEncoding as StreamingSupportedEncodingEnum, GladiaBitDepth as StreamingSupportedBitDepthEnum, GladiaSampleRate as StreamingSupportedSampleRateEnum, GladiaModel as StreamingSupportedModels, DeepgramModelType, AssemblyAISampleRateType, AssemblyAIEncodingType, AssemblyAISpeechModelType } from './constants.mjs';
2
- export { AssemblyAIEncoding, AssemblyAISampleRate, AssemblyAISpeechModel, DeepgramModel } from './constants.mjs';
1
+ import { DeepgramModelType, DeepgramTopicModeType, DeepgramRedactType, AssemblyAISampleRateType, AssemblyAIEncodingType, AssemblyAISpeechModelType } from './constants.mjs';
2
+ export { AssemblyAIEncoding, AssemblyAISampleRate, AssemblyAISpeechModel, DeepgramEncoding, DeepgramModel, DeepgramRedact, DeepgramTopicMode, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaSampleRate, GladiaTranslationLanguage } from './constants.mjs';
3
3
 
4
4
  /**
5
5
  * Unified audio encoding types for Voice Router SDK
@@ -700,6 +700,118 @@ interface CallbackConfigDto {
700
700
  method?: CallbackMethodEnum;
701
701
  }
702
702
 
703
+ /**
704
+ * Generated by orval v7.9.0 🍺
705
+ * Do not edit manually.
706
+ * Gladia Control API
707
+ * OpenAPI spec version: 1.0
708
+ */
709
+ /**
710
+ * Specify the language in which it will be pronounced when sound comparison occurs. Default to transcription language.
711
+ */
712
+ type TranscriptionLanguageCodeEnum = (typeof TranscriptionLanguageCodeEnum)[keyof typeof TranscriptionLanguageCodeEnum];
713
+ declare const TranscriptionLanguageCodeEnum: {
714
+ readonly af: "af";
715
+ readonly am: "am";
716
+ readonly ar: "ar";
717
+ readonly as: "as";
718
+ readonly az: "az";
719
+ readonly ba: "ba";
720
+ readonly be: "be";
721
+ readonly bg: "bg";
722
+ readonly bn: "bn";
723
+ readonly bo: "bo";
724
+ readonly br: "br";
725
+ readonly bs: "bs";
726
+ readonly ca: "ca";
727
+ readonly cs: "cs";
728
+ readonly cy: "cy";
729
+ readonly da: "da";
730
+ readonly de: "de";
731
+ readonly el: "el";
732
+ readonly en: "en";
733
+ readonly es: "es";
734
+ readonly et: "et";
735
+ readonly eu: "eu";
736
+ readonly fa: "fa";
737
+ readonly fi: "fi";
738
+ readonly fo: "fo";
739
+ readonly fr: "fr";
740
+ readonly gl: "gl";
741
+ readonly gu: "gu";
742
+ readonly ha: "ha";
743
+ readonly haw: "haw";
744
+ readonly he: "he";
745
+ readonly hi: "hi";
746
+ readonly hr: "hr";
747
+ readonly ht: "ht";
748
+ readonly hu: "hu";
749
+ readonly hy: "hy";
750
+ readonly id: "id";
751
+ readonly is: "is";
752
+ readonly it: "it";
753
+ readonly ja: "ja";
754
+ readonly jw: "jw";
755
+ readonly ka: "ka";
756
+ readonly kk: "kk";
757
+ readonly km: "km";
758
+ readonly kn: "kn";
759
+ readonly ko: "ko";
760
+ readonly la: "la";
761
+ readonly lb: "lb";
762
+ readonly ln: "ln";
763
+ readonly lo: "lo";
764
+ readonly lt: "lt";
765
+ readonly lv: "lv";
766
+ readonly mg: "mg";
767
+ readonly mi: "mi";
768
+ readonly mk: "mk";
769
+ readonly ml: "ml";
770
+ readonly mn: "mn";
771
+ readonly mr: "mr";
772
+ readonly ms: "ms";
773
+ readonly mt: "mt";
774
+ readonly my: "my";
775
+ readonly ne: "ne";
776
+ readonly nl: "nl";
777
+ readonly nn: "nn";
778
+ readonly no: "no";
779
+ readonly oc: "oc";
780
+ readonly pa: "pa";
781
+ readonly pl: "pl";
782
+ readonly ps: "ps";
783
+ readonly pt: "pt";
784
+ readonly ro: "ro";
785
+ readonly ru: "ru";
786
+ readonly sa: "sa";
787
+ readonly sd: "sd";
788
+ readonly si: "si";
789
+ readonly sk: "sk";
790
+ readonly sl: "sl";
791
+ readonly sn: "sn";
792
+ readonly so: "so";
793
+ readonly sq: "sq";
794
+ readonly sr: "sr";
795
+ readonly su: "su";
796
+ readonly sv: "sv";
797
+ readonly sw: "sw";
798
+ readonly ta: "ta";
799
+ readonly te: "te";
800
+ readonly tg: "tg";
801
+ readonly th: "th";
802
+ readonly tk: "tk";
803
+ readonly tl: "tl";
804
+ readonly tr: "tr";
805
+ readonly tt: "tt";
806
+ readonly uk: "uk";
807
+ readonly ur: "ur";
808
+ readonly uz: "uz";
809
+ readonly vi: "vi";
810
+ readonly yi: "yi";
811
+ readonly yo: "yo";
812
+ readonly zh: "zh";
813
+ };
814
+
703
815
  /**
704
816
  * Generated by orval v7.9.0 🍺
705
817
  * Do not edit manually.
@@ -935,6 +1047,119 @@ interface SummarizationConfigDTO {
935
1047
  type?: SummaryTypesEnum;
936
1048
  }
937
1049
 
1050
+ /**
1051
+ * Generated by orval v7.9.0 🍺
1052
+ * Do not edit manually.
1053
+ * Gladia Control API
1054
+ * OpenAPI spec version: 1.0
1055
+ */
1056
+ /**
1057
+ * Target language in `iso639-1` format you want the transcription translated to
1058
+ */
1059
+ type TranslationLanguageCodeEnum = (typeof TranslationLanguageCodeEnum)[keyof typeof TranslationLanguageCodeEnum];
1060
+ declare const TranslationLanguageCodeEnum: {
1061
+ readonly af: "af";
1062
+ readonly am: "am";
1063
+ readonly ar: "ar";
1064
+ readonly as: "as";
1065
+ readonly az: "az";
1066
+ readonly ba: "ba";
1067
+ readonly be: "be";
1068
+ readonly bg: "bg";
1069
+ readonly bn: "bn";
1070
+ readonly bo: "bo";
1071
+ readonly br: "br";
1072
+ readonly bs: "bs";
1073
+ readonly ca: "ca";
1074
+ readonly cs: "cs";
1075
+ readonly cy: "cy";
1076
+ readonly da: "da";
1077
+ readonly de: "de";
1078
+ readonly el: "el";
1079
+ readonly en: "en";
1080
+ readonly es: "es";
1081
+ readonly et: "et";
1082
+ readonly eu: "eu";
1083
+ readonly fa: "fa";
1084
+ readonly fi: "fi";
1085
+ readonly fo: "fo";
1086
+ readonly fr: "fr";
1087
+ readonly gl: "gl";
1088
+ readonly gu: "gu";
1089
+ readonly ha: "ha";
1090
+ readonly haw: "haw";
1091
+ readonly he: "he";
1092
+ readonly hi: "hi";
1093
+ readonly hr: "hr";
1094
+ readonly ht: "ht";
1095
+ readonly hu: "hu";
1096
+ readonly hy: "hy";
1097
+ readonly id: "id";
1098
+ readonly is: "is";
1099
+ readonly it: "it";
1100
+ readonly ja: "ja";
1101
+ readonly jw: "jw";
1102
+ readonly ka: "ka";
1103
+ readonly kk: "kk";
1104
+ readonly km: "km";
1105
+ readonly kn: "kn";
1106
+ readonly ko: "ko";
1107
+ readonly la: "la";
1108
+ readonly lb: "lb";
1109
+ readonly ln: "ln";
1110
+ readonly lo: "lo";
1111
+ readonly lt: "lt";
1112
+ readonly lv: "lv";
1113
+ readonly mg: "mg";
1114
+ readonly mi: "mi";
1115
+ readonly mk: "mk";
1116
+ readonly ml: "ml";
1117
+ readonly mn: "mn";
1118
+ readonly mr: "mr";
1119
+ readonly ms: "ms";
1120
+ readonly mt: "mt";
1121
+ readonly my: "my";
1122
+ readonly ne: "ne";
1123
+ readonly nl: "nl";
1124
+ readonly nn: "nn";
1125
+ readonly no: "no";
1126
+ readonly oc: "oc";
1127
+ readonly pa: "pa";
1128
+ readonly pl: "pl";
1129
+ readonly ps: "ps";
1130
+ readonly pt: "pt";
1131
+ readonly ro: "ro";
1132
+ readonly ru: "ru";
1133
+ readonly sa: "sa";
1134
+ readonly sd: "sd";
1135
+ readonly si: "si";
1136
+ readonly sk: "sk";
1137
+ readonly sl: "sl";
1138
+ readonly sn: "sn";
1139
+ readonly so: "so";
1140
+ readonly sq: "sq";
1141
+ readonly sr: "sr";
1142
+ readonly su: "su";
1143
+ readonly sv: "sv";
1144
+ readonly sw: "sw";
1145
+ readonly ta: "ta";
1146
+ readonly te: "te";
1147
+ readonly tg: "tg";
1148
+ readonly th: "th";
1149
+ readonly tk: "tk";
1150
+ readonly tl: "tl";
1151
+ readonly tr: "tr";
1152
+ readonly tt: "tt";
1153
+ readonly uk: "uk";
1154
+ readonly ur: "ur";
1155
+ readonly uz: "uz";
1156
+ readonly vi: "vi";
1157
+ readonly wo: "wo";
1158
+ readonly yi: "yi";
1159
+ readonly yo: "yo";
1160
+ readonly zh: "zh";
1161
+ };
1162
+
938
1163
  /**
939
1164
  * Generated by orval v7.9.0 🍺
940
1165
  * Do not edit manually.
@@ -4130,12 +4355,16 @@ interface Transcription {
4130
4355
  /**
4131
4356
  * Generated by orval v7.9.0 🍺
4132
4357
  * Do not edit manually.
4133
- * Deepgram API Specification
4134
- * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4135
-
4136
- * OpenAPI spec version: 1.0.0
4358
+ * Gladia Control API
4359
+ * OpenAPI spec version: 1.0
4137
4360
  */
4138
- type ListenV1ModelParameter = "nova-3" | "nova-3-general" | "nova-3-medical" | "nova-2" | "nova-2-general" | "nova-2-meeting" | "nova-2-finance" | "nova-2-conversationalai" | "nova-2-voicemail" | "nova-2-video" | "nova-2-medical" | "nova-2-drivethru" | "nova-2-automotive" | "nova" | "nova-general" | "nova-phonecall" | "nova-medical" | "enhanced" | "enhanced-general" | "enhanced-meeting" | "enhanced-phonecall" | "enhanced-finance" | "base" | "meeting" | "phonecall" | "finance" | "conversationalai" | "voicemail" | "video" | string;
4361
+ /**
4362
+ * The model used to process the audio. "solaria-1" is used by default.
4363
+ */
4364
+ type StreamingSupportedModels = (typeof StreamingSupportedModels)[keyof typeof StreamingSupportedModels];
4365
+ declare const StreamingSupportedModels: {
4366
+ readonly "solaria-1": "solaria-1";
4367
+ };
4139
4368
 
4140
4369
  /**
4141
4370
  * Generated by orval v7.9.0 🍺
@@ -4254,6 +4483,44 @@ type SharedTopicsParameter = boolean;
4254
4483
  */
4255
4484
  type SharedCustomTopicParameter = string | string[];
4256
4485
 
4486
+ /**
4487
+ * Generated by orval v7.9.0 🍺
4488
+ * Do not edit manually.
4489
+ * Deepgram API Specification
4490
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4491
+
4492
+ * OpenAPI spec version: 1.0.0
4493
+ */
4494
+ /**
4495
+ * SharedCustomTopicModeParameter type definition
4496
+ */
4497
+ /**
4498
+ * SharedCustomTopicModeParameter type definition
4499
+ */
4500
+ /**
4501
+ * SharedCustomTopicModeParameter type definition
4502
+ */
4503
+ /**
4504
+ * SharedCustomTopicModeParameter type definition
4505
+ */
4506
+ /**
4507
+ * SharedCustomTopicModeParameter type definition
4508
+ */
4509
+ /**
4510
+ * SharedCustomTopicModeParameter type definition
4511
+ */
4512
+ /**
4513
+ * SharedCustomTopicModeParameter type definition
4514
+ */
4515
+ /**
4516
+ * SharedCustomTopicModeParameter type definition
4517
+ */
4518
+ type SharedCustomTopicModeParameter = typeof SharedCustomTopicModeParameter[keyof typeof SharedCustomTopicModeParameter];
4519
+ declare const SharedCustomTopicModeParameter: {
4520
+ readonly extended: "extended";
4521
+ readonly strict: "strict";
4522
+ };
4523
+
4257
4524
  /**
4258
4525
  * Generated by orval v7.9.0 🍺
4259
4526
  * Do not edit manually.
@@ -4364,6 +4631,48 @@ type ListenV1DiarizeParameter = boolean;
4364
4631
  */
4365
4632
  type ListenV1DictationParameter = boolean;
4366
4633
 
4634
+ /**
4635
+ * Generated by orval v7.9.0 🍺
4636
+ * Do not edit manually.
4637
+ * Deepgram API Specification
4638
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4639
+
4640
+ * OpenAPI spec version: 1.0.0
4641
+ */
4642
+ /**
4643
+ * ListenV1EncodingParameter type definition
4644
+ */
4645
+ /**
4646
+ * ListenV1EncodingParameter type definition
4647
+ */
4648
+ /**
4649
+ * ListenV1EncodingParameter type definition
4650
+ */
4651
+ /**
4652
+ * ListenV1EncodingParameter type definition
4653
+ */
4654
+ /**
4655
+ * ListenV1EncodingParameter type definition
4656
+ */
4657
+ /**
4658
+ * ListenV1EncodingParameter type definition
4659
+ */
4660
+ /**
4661
+ * ListenV1EncodingParameter type definition
4662
+ */
4663
+ /**
4664
+ * ListenV1EncodingParameter type definition
4665
+ */
4666
+ type ListenV1EncodingParameter = typeof ListenV1EncodingParameter[keyof typeof ListenV1EncodingParameter];
4667
+ declare const ListenV1EncodingParameter: {
4668
+ readonly linear16: "linear16";
4669
+ readonly flac: "flac";
4670
+ readonly mulaw: "mulaw";
4671
+ readonly opus: "opus";
4672
+ readonly speex: "speex";
4673
+ readonly g729: "g729";
4674
+ };
4675
+
4367
4676
  /**
4368
4677
  * Generated by orval v7.9.0 🍺
4369
4678
  * Do not edit manually.
@@ -4426,6 +4735,16 @@ type ListenV1LanguageParameter = string;
4426
4735
  */
4427
4736
  type ListenV1MeasurementsParameter = boolean;
4428
4737
 
4738
+ /**
4739
+ * Generated by orval v7.9.0 🍺
4740
+ * Do not edit manually.
4741
+ * Deepgram API Specification
4742
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4743
+
4744
+ * OpenAPI spec version: 1.0.0
4745
+ */
4746
+ type ListenV1ModelParameter = "nova-3" | "nova-3-general" | "nova-3-medical" | "nova-2" | "nova-2-general" | "nova-2-meeting" | "nova-2-finance" | "nova-2-conversationalai" | "nova-2-voicemail" | "nova-2-video" | "nova-2-medical" | "nova-2-drivethru" | "nova-2-automotive" | "nova" | "nova-general" | "nova-phonecall" | "nova-medical" | "enhanced" | "enhanced-general" | "enhanced-meeting" | "enhanced-phonecall" | "enhanced-finance" | "base" | "meeting" | "phonecall" | "finance" | "conversationalai" | "voicemail" | "video" | string;
4747
+
4429
4748
  /**
4430
4749
  * Generated by orval v7.9.0 🍺
4431
4750
  * Do not edit manually.
@@ -4491,6 +4810,21 @@ type ListenV1ProfanityFilterParameter = boolean;
4491
4810
  */
4492
4811
  type ListenV1PunctuateParameter = boolean;
4493
4812
 
4813
+ /**
4814
+ * Generated by orval v7.9.0 🍺
4815
+ * Do not edit manually.
4816
+ * Deepgram API Specification
4817
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4818
+
4819
+ * OpenAPI spec version: 1.0.0
4820
+ */
4821
+ type ListenV1RedactParameterOneOfItem = (typeof ListenV1RedactParameterOneOfItem)[keyof typeof ListenV1RedactParameterOneOfItem];
4822
+ declare const ListenV1RedactParameterOneOfItem: {
4823
+ readonly pci: "pci";
4824
+ readonly pii: "pii";
4825
+ readonly numbers: "numbers";
4826
+ };
4827
+
4494
4828
  /**
4495
4829
  * Generated by orval v7.9.0 🍺
4496
4830
  * Do not edit manually.
@@ -5445,6 +5779,62 @@ type StreamingRequestCustomMetadata = {
5445
5779
  [key: string]: unknown;
5446
5780
  };
5447
5781
 
5782
+ /**
5783
+ * Generated by orval v7.9.0 🍺
5784
+ * Do not edit manually.
5785
+ * Gladia Control API
5786
+ * OpenAPI spec version: 1.0
5787
+ */
5788
+ /**
5789
+ * The bit depth of the audio stream
5790
+ */
5791
+ type StreamingSupportedBitDepthEnum = (typeof StreamingSupportedBitDepthEnum)[keyof typeof StreamingSupportedBitDepthEnum];
5792
+ declare const StreamingSupportedBitDepthEnum: {
5793
+ readonly NUMBER_8: 8;
5794
+ readonly NUMBER_16: 16;
5795
+ readonly NUMBER_24: 24;
5796
+ readonly NUMBER_32: 32;
5797
+ };
5798
+
5799
+ /**
5800
+ * Generated by orval v7.9.0 🍺
5801
+ * Do not edit manually.
5802
+ * Gladia Control API
5803
+ * OpenAPI spec version: 1.0
5804
+ */
5805
+ /**
5806
+ * The encoding format of the audio stream. Supported formats:
5807
+ - PCM: 8, 16, 24, and 32 bits
5808
+ - A-law: 8 bits
5809
+ - μ-law: 8 bits
5810
+
5811
+ Note: No need to add WAV headers to raw audio as the API supports both formats.
5812
+ */
5813
+ type StreamingSupportedEncodingEnum = (typeof StreamingSupportedEncodingEnum)[keyof typeof StreamingSupportedEncodingEnum];
5814
+ declare const StreamingSupportedEncodingEnum: {
5815
+ readonly "wav/pcm": "wav/pcm";
5816
+ readonly "wav/alaw": "wav/alaw";
5817
+ readonly "wav/ulaw": "wav/ulaw";
5818
+ };
5819
+
5820
+ /**
5821
+ * Generated by orval v7.9.0 🍺
5822
+ * Do not edit manually.
5823
+ * Gladia Control API
5824
+ * OpenAPI spec version: 1.0
5825
+ */
5826
+ /**
5827
+ * The sample rate of the audio stream
5828
+ */
5829
+ type StreamingSupportedSampleRateEnum = (typeof StreamingSupportedSampleRateEnum)[keyof typeof StreamingSupportedSampleRateEnum];
5830
+ declare const StreamingSupportedSampleRateEnum: {
5831
+ readonly NUMBER_8000: 8000;
5832
+ readonly NUMBER_16000: 16000;
5833
+ readonly NUMBER_32000: 32000;
5834
+ readonly NUMBER_44100: 44100;
5835
+ readonly NUMBER_48000: 48000;
5836
+ };
5837
+
5448
5838
  /**
5449
5839
  * Generated by orval v7.9.0 🍺
5450
5840
  * Do not edit manually.
@@ -5624,7 +6014,7 @@ interface DeepgramStreamingOptions {
5624
6014
  * Custom topic detection mode
5625
6015
  * Use `DeepgramTopicMode` const for autocomplete
5626
6016
  */
5627
- customTopicMode?: SharedCustomTopicModeParameter;
6017
+ customTopicMode?: DeepgramTopicModeType;
5628
6018
  /** Enable intent recognition */
5629
6019
  intents?: boolean;
5630
6020
  /** Custom intent definitions */
@@ -5633,7 +6023,7 @@ interface DeepgramStreamingOptions {
5633
6023
  * Custom intent detection mode
5634
6024
  * Use `DeepgramTopicMode` const for autocomplete
5635
6025
  */
5636
- customIntentMode?: SharedCustomTopicModeParameter;
6026
+ customIntentMode?: DeepgramTopicModeType;
5637
6027
  /** Enable summarization */
5638
6028
  summarize?: boolean;
5639
6029
  /** Custom vocabulary/keywords for boosting */
@@ -5652,7 +6042,7 @@ interface DeepgramStreamingOptions {
5652
6042
  * { redact: [DeepgramRedact.pii, DeepgramRedact.pci] }
5653
6043
  * ```
5654
6044
  */
5655
- redact?: boolean | ListenV1RedactParameterOneOfItem[];
6045
+ redact?: boolean | DeepgramRedactType[];
5656
6046
  /** Callback URL for webhooks */
5657
6047
  callback?: string;
5658
6048
  /** Extra metadata to include in response */
@@ -5669,139 +6059,592 @@ interface DeepgramStreamingOptions {
5669
6059
  vadThreshold?: number;
5670
6060
  }
5671
6061
 
5672
- /**
5673
- * AssemblyAI streaming options
5674
- *
5675
- * Based on the v3 Universal Streaming API parameters from the AssemblyAI SDK.
5676
- * Supports advanced features like VAD tuning, end-of-turn detection, and profanity filtering.
5677
- *
5678
- * @see https://www.assemblyai.com/docs/speech-to-text/streaming
6062
+ /**
6063
+ * AssemblyAI streaming options
6064
+ *
6065
+ * Based on the v3 Universal Streaming API parameters from the AssemblyAI SDK.
6066
+ * Supports advanced features like VAD tuning, end-of-turn detection, and profanity filtering.
6067
+ *
6068
+ * @see https://www.assemblyai.com/docs/speech-to-text/streaming
6069
+ */
6070
+ interface AssemblyAIStreamingOptions {
6071
+ /**
6072
+ * Sample rate in Hz
6073
+ * Use `AssemblyAISampleRate` const for autocomplete:
6074
+ * @example
6075
+ * ```typescript
6076
+ * import { AssemblyAISampleRate } from '@meeting-baas/sdk'
6077
+ * { sampleRate: AssemblyAISampleRate.rate16000 }
6078
+ * ```
6079
+ */
6080
+ sampleRate?: AssemblyAISampleRateType;
6081
+ /**
6082
+ * Audio encoding format
6083
+ * Use `AssemblyAIEncoding` const for autocomplete:
6084
+ * @example
6085
+ * ```typescript
6086
+ * import { AssemblyAIEncoding } from '@meeting-baas/sdk'
6087
+ * { encoding: AssemblyAIEncoding.pcmS16le }
6088
+ * ```
6089
+ */
6090
+ encoding?: AssemblyAIEncodingType;
6091
+ /**
6092
+ * Speech model to use
6093
+ *
6094
+ * Strict union type - only accepts valid AssemblyAI streaming models.
6095
+ * Use `AssemblyAISpeechModel` const for autocomplete:
6096
+ * @example
6097
+ * ```typescript
6098
+ * import { AssemblyAISpeechModel } from 'voice-router-dev'
6099
+ * { speechModel: AssemblyAISpeechModel.english }
6100
+ * { speechModel: AssemblyAISpeechModel.multilingual }
6101
+ * // Or use string literals directly:
6102
+ * { speechModel: "universal-streaming-english" }
6103
+ * ```
6104
+ */
6105
+ speechModel?: AssemblyAISpeechModelType;
6106
+ /** Enable automatic language detection */
6107
+ languageDetection?: boolean;
6108
+ /**
6109
+ * Confidence threshold for end-of-turn detection (0-1)
6110
+ * Higher values require more confidence before ending a turn
6111
+ * @default 0.5
6112
+ */
6113
+ endOfTurnConfidenceThreshold?: number;
6114
+ /**
6115
+ * Minimum silence duration (ms) to trigger end-of-turn when confident
6116
+ * Only applies when confidence is above threshold
6117
+ * @default 1000
6118
+ */
6119
+ minEndOfTurnSilenceWhenConfident?: number;
6120
+ /**
6121
+ * Maximum silence duration (ms) before forcing end-of-turn
6122
+ * Regardless of confidence level
6123
+ * @default 20000
6124
+ */
6125
+ maxTurnSilence?: number;
6126
+ /**
6127
+ * VAD sensitivity threshold (0-1)
6128
+ * Lower values are more sensitive to quiet speech
6129
+ */
6130
+ vadThreshold?: number;
6131
+ /**
6132
+ * Enable real-time text formatting of turns
6133
+ * Applies punctuation, capitalization, and formatting
6134
+ */
6135
+ formatTurns?: boolean;
6136
+ /** Filter profanity in real-time transcription */
6137
+ filterProfanity?: boolean;
6138
+ /**
6139
+ * Key terms to boost in recognition
6140
+ * Increases recognition accuracy for specific words/phrases
6141
+ */
6142
+ keyterms?: string[];
6143
+ /**
6144
+ * Key term prompting for context
6145
+ * Provides additional context about the terms to improve recognition
6146
+ */
6147
+ keytermsPrompt?: string[];
6148
+ /**
6149
+ * Inactivity timeout in milliseconds
6150
+ * Session will close if no audio is received for this duration
6151
+ */
6152
+ inactivityTimeout?: number;
6153
+ /**
6154
+ * Use token-based authentication
6155
+ * If true, will create a temporary token before connecting
6156
+ */
6157
+ useToken?: boolean;
6158
+ /**
6159
+ * Token expiration time in seconds (minimum 60)
6160
+ * Only used if useToken is true
6161
+ * @default 3600
6162
+ */
6163
+ tokenExpiresIn?: number;
6164
+ }
6165
+ /**
6166
+ * AssemblyAI dynamic configuration update
6167
+ * Can be sent mid-stream to adjust parameters
6168
+ */
6169
+ type AssemblyAIUpdateConfiguration = Omit<StreamingUpdateConfiguration, "type">;
6170
+ /**
6171
+ * Union of all provider-specific streaming options
6172
+ */
6173
+ type ProviderStreamingOptions = ({
6174
+ provider: "gladia";
6175
+ } & GladiaStreamingOptions) | ({
6176
+ provider: "deepgram";
6177
+ } & DeepgramStreamingOptions) | ({
6178
+ provider: "assemblyai";
6179
+ } & AssemblyAIStreamingOptions);
6180
+ /**
6181
+ * Type-safe streaming options for a specific provider
6182
+ */
6183
+ type StreamingOptionsForProvider<P extends StreamingProvider> = P extends "gladia" ? GladiaStreamingOptions : P extends "deepgram" ? DeepgramStreamingOptions : P extends "assemblyai" ? AssemblyAIStreamingOptions : never;
6184
+ /**
6185
+ * Type-safe transcribeStream parameters for a specific provider
6186
+ */
6187
+ interface TranscribeStreamParams<P extends StreamingProvider> {
6188
+ /** Streaming options specific to this provider */
6189
+ options?: StreamingOptionsForProvider<P> & {
6190
+ provider: P;
6191
+ };
6192
+ /** Event callbacks */
6193
+ callbacks?: StreamingCallbacks;
6194
+ }
6195
+
6196
+ /**
6197
+ * Generated by orval v7.9.0 🍺
6198
+ * Do not edit manually.
6199
+ * AssemblyAI API
6200
+ * AssemblyAI API
6201
+ * OpenAPI spec version: 1.3.4
6202
+ */
6203
+ /**
6204
+ * Get transcripts that were created after this transcript ID
6205
+ */
6206
+ type AfterId = string;
6207
+
6208
+ /**
6209
+ * Generated by orval v7.9.0 🍺
6210
+ * Do not edit manually.
6211
+ * AssemblyAI API
6212
+ * AssemblyAI API
6213
+ * OpenAPI spec version: 1.3.4
6214
+ */
6215
+ /**
6216
+ * Get transcripts that were created before this transcript ID
6217
+ */
6218
+ type BeforeId = string;
6219
+
6220
+ /**
6221
+ * Generated by orval v7.9.0 🍺
6222
+ * Do not edit manually.
6223
+ * AssemblyAI API
6224
+ * AssemblyAI API
6225
+ * OpenAPI spec version: 1.3.4
6226
+ */
6227
+ /**
6228
+ * Only get transcripts created on this date
6229
+ */
6230
+ type CreatedOn = string;
6231
+
6232
+ /**
6233
+ * Generated by orval v7.9.0 🍺
6234
+ * Do not edit manually.
6235
+ * AssemblyAI API
6236
+ * AssemblyAI API
6237
+ * OpenAPI spec version: 1.3.4
6238
+ */
6239
+ /**
6240
+ * Maximum amount of transcripts to retrieve
6241
+ */
6242
+ type Limit = number;
6243
+
6244
+ /**
6245
+ * Generated by orval v7.9.0 🍺
6246
+ * Do not edit manually.
6247
+ * AssemblyAI API
6248
+ * AssemblyAI API
6249
+ * OpenAPI spec version: 1.3.4
6250
+ */
6251
+ /**
6252
+ * Only get throttled transcripts, overrides the status filter
6253
+ */
6254
+ type ThrottledOnly = boolean;
6255
+
6256
+ /**
6257
+ * Generated by orval v7.9.0 🍺
6258
+ * Do not edit manually.
6259
+ * AssemblyAI API
6260
+ * AssemblyAI API
6261
+ * OpenAPI spec version: 1.3.4
6262
+ */
6263
+
6264
+ type ListTranscriptsParams = {
6265
+ /**
6266
+ * Maximum amount of transcripts to retrieve
6267
+ */
6268
+ limit?: Limit;
6269
+ /**
6270
+ * Filter by transcript status
6271
+ */
6272
+ status?: TranscriptStatus;
6273
+ /**
6274
+ * Only get transcripts created on this date
6275
+ */
6276
+ created_on?: CreatedOn;
6277
+ /**
6278
+ * Get transcripts that were created before this transcript ID
6279
+ */
6280
+ before_id?: BeforeId;
6281
+ /**
6282
+ * Get transcripts that were created after this transcript ID
6283
+ */
6284
+ after_id?: AfterId;
6285
+ /**
6286
+ * Only get throttled transcripts, overrides the status filter
6287
+ */
6288
+ throttled_only?: ThrottledOnly;
6289
+ };
6290
+
6291
+ /**
6292
+ * Generated by orval v7.9.0 🍺
6293
+ * Do not edit manually.
6294
+ * Gladia Control API
6295
+ * OpenAPI spec version: 1.0
6296
+ */
6297
+ type TranscriptionControllerListV2KindItem = (typeof TranscriptionControllerListV2KindItem)[keyof typeof TranscriptionControllerListV2KindItem];
6298
+ declare const TranscriptionControllerListV2KindItem: {
6299
+ readonly "pre-recorded": "pre-recorded";
6300
+ readonly live: "live";
6301
+ };
6302
+
6303
+ /**
6304
+ * Generated by orval v7.9.0 🍺
6305
+ * Do not edit manually.
6306
+ * Gladia Control API
6307
+ * OpenAPI spec version: 1.0
6308
+ */
6309
+ type TranscriptionControllerListV2StatusItem = (typeof TranscriptionControllerListV2StatusItem)[keyof typeof TranscriptionControllerListV2StatusItem];
6310
+ declare const TranscriptionControllerListV2StatusItem: {
6311
+ readonly queued: "queued";
6312
+ readonly processing: "processing";
6313
+ readonly done: "done";
6314
+ readonly error: "error";
6315
+ };
6316
+
6317
+ /**
6318
+ * Generated by orval v7.9.0 🍺
6319
+ * Do not edit manually.
6320
+ * Gladia Control API
6321
+ * OpenAPI spec version: 1.0
6322
+ */
6323
+
6324
+ type TranscriptionControllerListV2Params = {
6325
+ /**
6326
+ * The starting point for pagination. A value of 0 starts from the first item.
6327
+ */
6328
+ offset?: number;
6329
+ /**
6330
+ * The maximum number of items to return. Useful for pagination and controlling data payload size.
6331
+ */
6332
+ limit?: number;
6333
+ /**
6334
+ * Filter items relevant to a specific date in ISO format (YYYY-MM-DD).
6335
+ */
6336
+ date?: string;
6337
+ /**
6338
+ * Include items that occurred before the specified date in ISO format.
6339
+ */
6340
+ before_date?: string;
6341
+ /**
6342
+ * Filter for items after the specified date. Use with `before_date` for a range. Date in ISO format.
6343
+ */
6344
+ after_date?: string;
6345
+ /**
6346
+ * Filter the list based on item status. Accepts multiple values from the predefined list.
6347
+ */
6348
+ status?: TranscriptionControllerListV2StatusItem[];
6349
+ custom_metadata?: {
6350
+ [key: string]: unknown;
6351
+ };
6352
+ /**
6353
+ * Filter the list based on the item type. Supports multiple values from the predefined list.
6354
+ */
6355
+ kind?: TranscriptionControllerListV2KindItem[];
6356
+ };
6357
+
6358
+ /**
6359
+ * Generated by orval v7.9.0 🍺
6360
+ * Do not edit manually.
6361
+ * Deepgram API Specification
6362
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6363
+
6364
+ * OpenAPI spec version: 1.0.0
6365
+ */
6366
+ /**
6367
+ * Start date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
6368
+ */
6369
+ type ManageV1StartDateTimeParameter = string;
6370
+
6371
+ /**
6372
+ * Generated by orval v7.9.0 🍺
6373
+ * Do not edit manually.
6374
+ * Deepgram API Specification
6375
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6376
+
6377
+ * OpenAPI spec version: 1.0.0
6378
+ */
6379
+ /**
6380
+ * End date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
6381
+ */
6382
+ type ManageV1EndDateTimeParameter = string;
6383
+
6384
+ /**
6385
+ * Generated by orval v7.9.0 🍺
6386
+ * Do not edit manually.
6387
+ * Deepgram API Specification
6388
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6389
+
6390
+ * OpenAPI spec version: 1.0.0
6391
+ */
6392
+ /**
6393
+ * Number of results to return per page. Default 10. Range [1,1000]
6394
+ */
6395
+ type ManageV1LimitParameter = number;
6396
+
6397
+ /**
6398
+ * Generated by orval v7.9.0 🍺
6399
+ * Do not edit manually.
6400
+ * Deepgram API Specification
6401
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6402
+
6403
+ * OpenAPI spec version: 1.0.0
6404
+ */
6405
+ /**
6406
+ * Navigate and return the results to retrieve specific portions of information of the response
6407
+ */
6408
+ type ManageV1PageParameter = number;
6409
+
6410
+ /**
6411
+ * Generated by orval v7.9.0 🍺
6412
+ * Do not edit manually.
6413
+ * Deepgram API Specification
6414
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6415
+
6416
+ * OpenAPI spec version: 1.0.0
6417
+ */
6418
+ /**
6419
+ * Filter for requests where a specific accessor was used
6420
+ */
6421
+ type ManageV1FilterAccessorParameter = string;
6422
+
6423
+ /**
6424
+ * Generated by orval v7.9.0 🍺
6425
+ * Do not edit manually.
6426
+ * Deepgram API Specification
6427
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6428
+
6429
+ * OpenAPI spec version: 1.0.0
6430
+ */
6431
+ /**
6432
+ * Filter for a specific request id
6433
+ */
6434
+ type ManageV1FilterRequestIdParameter = string;
6435
+
6436
+ /**
6437
+ * Generated by orval v7.9.0 🍺
6438
+ * Do not edit manually.
6439
+ * Deepgram API Specification
6440
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6441
+
6442
+ * OpenAPI spec version: 1.0.0
6443
+ */
6444
+ /**
6445
+ * Deployment type for the requests
6446
+ */
6447
+ /**
6448
+ * ManageV1FilterDeploymentParameter type definition
6449
+ */
6450
+ /**
6451
+ * ManageV1FilterDeploymentParameter type definition
6452
+ */
6453
+ /**
6454
+ * ManageV1FilterDeploymentParameter type definition
6455
+ */
6456
+ /**
6457
+ * ManageV1FilterDeploymentParameter type definition
6458
+ */
6459
+ /**
6460
+ * ManageV1FilterDeploymentParameter type definition
6461
+ */
6462
+ /**
6463
+ * ManageV1FilterDeploymentParameter type definition
6464
+ */
6465
+ /**
6466
+ * ManageV1FilterDeploymentParameter type definition
6467
+ */
6468
+ /**
6469
+ * ManageV1FilterDeploymentParameter type definition
6470
+ */
6471
+ type ManageV1FilterDeploymentParameter = typeof ManageV1FilterDeploymentParameter[keyof typeof ManageV1FilterDeploymentParameter];
6472
+ declare const ManageV1FilterDeploymentParameter: {
6473
+ readonly hosted: "hosted";
6474
+ readonly beta: "beta";
6475
+ };
6476
+
6477
+ /**
6478
+ * Generated by orval v7.9.0 🍺
6479
+ * Do not edit manually.
6480
+ * Deepgram API Specification
6481
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6482
+
6483
+ * OpenAPI spec version: 1.0.0
6484
+ */
6485
+ /**
6486
+ * ManageV1FilterEndpointParameter type definition
6487
+ */
6488
+ /**
6489
+ * ManageV1FilterEndpointParameter type definition
6490
+ */
6491
+ /**
6492
+ * ManageV1FilterEndpointParameter type definition
6493
+ */
6494
+ /**
6495
+ * ManageV1FilterEndpointParameter type definition
6496
+ */
6497
+ /**
6498
+ * ManageV1FilterEndpointParameter type definition
6499
+ */
6500
+ /**
6501
+ * ManageV1FilterEndpointParameter type definition
6502
+ */
6503
+ /**
6504
+ * ManageV1FilterEndpointParameter type definition
6505
+ */
6506
+ /**
6507
+ * ManageV1FilterEndpointParameter type definition
6508
+ */
6509
+ type ManageV1FilterEndpointParameter = typeof ManageV1FilterEndpointParameter[keyof typeof ManageV1FilterEndpointParameter];
6510
+ declare const ManageV1FilterEndpointParameter: {
6511
+ readonly listen: "listen";
6512
+ readonly read: "read";
6513
+ readonly speak: "speak";
6514
+ readonly agent: "agent";
6515
+ };
6516
+
6517
+ /**
6518
+ * Generated by orval v7.9.0 🍺
6519
+ * Do not edit manually.
6520
+ * Deepgram API Specification
6521
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6522
+
6523
+ * OpenAPI spec version: 1.0.0
6524
+ */
6525
+ /**
6526
+ * Method type for the request
6527
+ */
6528
+ /**
6529
+ * ManageV1FilterMethodParameter type definition
6530
+ */
6531
+ /**
6532
+ * ManageV1FilterMethodParameter type definition
6533
+ */
6534
+ /**
6535
+ * ManageV1FilterMethodParameter type definition
6536
+ */
6537
+ /**
6538
+ * ManageV1FilterMethodParameter type definition
6539
+ */
6540
+ /**
6541
+ * ManageV1FilterMethodParameter type definition
6542
+ */
6543
+ /**
6544
+ * ManageV1FilterMethodParameter type definition
6545
+ */
6546
+ /**
6547
+ * ManageV1FilterMethodParameter type definition
6548
+ */
6549
+ /**
6550
+ * ManageV1FilterMethodParameter type definition
6551
+ */
6552
+ type ManageV1FilterMethodParameter = typeof ManageV1FilterMethodParameter[keyof typeof ManageV1FilterMethodParameter];
6553
+ declare const ManageV1FilterMethodParameter: {
6554
+ readonly sync: "sync";
6555
+ readonly async: "async";
6556
+ readonly streaming: "streaming";
6557
+ };
6558
+
6559
+ /**
6560
+ * Generated by orval v7.9.0 🍺
6561
+ * Do not edit manually.
6562
+ * Deepgram API Specification
6563
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6564
+
6565
+ * OpenAPI spec version: 1.0.0
6566
+ */
6567
+ /**
6568
+ * ManageV1FilterStatusParameter type definition
6569
+ */
6570
+ /**
6571
+ * ManageV1FilterStatusParameter type definition
6572
+ */
6573
+ /**
6574
+ * ManageV1FilterStatusParameter type definition
6575
+ */
6576
+ /**
6577
+ * ManageV1FilterStatusParameter type definition
6578
+ */
6579
+ /**
6580
+ * ManageV1FilterStatusParameter type definition
6581
+ */
6582
+ /**
6583
+ * ManageV1FilterStatusParameter type definition
6584
+ */
6585
+ /**
6586
+ * ManageV1FilterStatusParameter type definition
6587
+ */
6588
+ /**
6589
+ * ManageV1FilterStatusParameter type definition
6590
+ */
6591
+ type ManageV1FilterStatusParameter = typeof ManageV1FilterStatusParameter[keyof typeof ManageV1FilterStatusParameter];
6592
+ declare const ManageV1FilterStatusParameter: {
6593
+ readonly succeeded: "succeeded";
6594
+ readonly failed: "failed";
6595
+ };
6596
+
6597
+ /**
6598
+ * Generated by orval v7.9.0 🍺
6599
+ * Do not edit manually.
6600
+ * Deepgram API Specification
6601
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
6602
+
6603
+ * OpenAPI spec version: 1.0.0
5679
6604
  */
5680
- interface AssemblyAIStreamingOptions {
5681
- /**
5682
- * Sample rate in Hz
5683
- * Use `AssemblyAISampleRate` const for autocomplete:
5684
- * @example
5685
- * ```typescript
5686
- * import { AssemblyAISampleRate } from '@meeting-baas/sdk'
5687
- * { sampleRate: AssemblyAISampleRate.rate16000 }
5688
- * ```
5689
- */
5690
- sampleRate?: AssemblyAISampleRateType;
5691
- /**
5692
- * Audio encoding format
5693
- * Use `AssemblyAIEncoding` const for autocomplete:
5694
- * @example
5695
- * ```typescript
5696
- * import { AssemblyAIEncoding } from '@meeting-baas/sdk'
5697
- * { encoding: AssemblyAIEncoding.pcmS16le }
5698
- * ```
5699
- */
5700
- encoding?: AssemblyAIEncodingType;
5701
- /**
5702
- * Speech model to use
5703
- *
5704
- * Strict union type - only accepts valid AssemblyAI streaming models.
5705
- * Use `AssemblyAISpeechModel` const for autocomplete:
5706
- * @example
5707
- * ```typescript
5708
- * import { AssemblyAISpeechModel } from 'voice-router-dev'
5709
- * { speechModel: AssemblyAISpeechModel.english }
5710
- * { speechModel: AssemblyAISpeechModel.multilingual }
5711
- * // Or use string literals directly:
5712
- * { speechModel: "universal-streaming-english" }
5713
- * ```
5714
- */
5715
- speechModel?: AssemblyAISpeechModelType;
5716
- /** Enable automatic language detection */
5717
- languageDetection?: boolean;
6605
+
6606
+ type ManageV1ProjectsRequestsListParams = {
5718
6607
  /**
5719
- * Confidence threshold for end-of-turn detection (0-1)
5720
- * Higher values require more confidence before ending a turn
5721
- * @default 0.5
6608
+ * Start date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
5722
6609
  */
5723
- endOfTurnConfidenceThreshold?: number;
6610
+ start?: ManageV1StartDateTimeParameter;
5724
6611
  /**
5725
- * Minimum silence duration (ms) to trigger end-of-turn when confident
5726
- * Only applies when confidence is above threshold
5727
- * @default 1000
6612
+ * End date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
5728
6613
  */
5729
- minEndOfTurnSilenceWhenConfident?: number;
6614
+ end?: ManageV1EndDateTimeParameter;
5730
6615
  /**
5731
- * Maximum silence duration (ms) before forcing end-of-turn
5732
- * Regardless of confidence level
5733
- * @default 20000
6616
+ * Number of results to return per page. Default 10. Range [1,1000]
5734
6617
  */
5735
- maxTurnSilence?: number;
6618
+ limit?: ManageV1LimitParameter;
5736
6619
  /**
5737
- * VAD sensitivity threshold (0-1)
5738
- * Lower values are more sensitive to quiet speech
6620
+ * Navigate and return the results to retrieve specific portions of information of the response
5739
6621
  */
5740
- vadThreshold?: number;
6622
+ page?: ManageV1PageParameter;
5741
6623
  /**
5742
- * Enable real-time text formatting of turns
5743
- * Applies punctuation, capitalization, and formatting
6624
+ * Filter for requests where a specific accessor was used
5744
6625
  */
5745
- formatTurns?: boolean;
5746
- /** Filter profanity in real-time transcription */
5747
- filterProfanity?: boolean;
6626
+ accessor?: ManageV1FilterAccessorParameter;
5748
6627
  /**
5749
- * Key terms to boost in recognition
5750
- * Increases recognition accuracy for specific words/phrases
6628
+ * Filter for a specific request id
5751
6629
  */
5752
- keyterms?: string[];
6630
+ request_id?: ManageV1FilterRequestIdParameter;
5753
6631
  /**
5754
- * Key term prompting for context
5755
- * Provides additional context about the terms to improve recognition
6632
+ * Filter for requests where a specific deployment was used
5756
6633
  */
5757
- keytermsPrompt?: string[];
6634
+ deployment?: ManageV1FilterDeploymentParameter;
5758
6635
  /**
5759
- * Inactivity timeout in milliseconds
5760
- * Session will close if no audio is received for this duration
6636
+ * Filter for requests where a specific endpoint was used
5761
6637
  */
5762
- inactivityTimeout?: number;
6638
+ endpoint?: ManageV1FilterEndpointParameter;
5763
6639
  /**
5764
- * Use token-based authentication
5765
- * If true, will create a temporary token before connecting
6640
+ * Filter for requests where a specific method was used
5766
6641
  */
5767
- useToken?: boolean;
6642
+ method?: ManageV1FilterMethodParameter;
5768
6643
  /**
5769
- * Token expiration time in seconds (minimum 60)
5770
- * Only used if useToken is true
5771
- * @default 3600
6644
+ * Filter for requests that succeeded (status code < 300) or failed (status code >=400)
5772
6645
  */
5773
- tokenExpiresIn?: number;
5774
- }
5775
- /**
5776
- * AssemblyAI dynamic configuration update
5777
- * Can be sent mid-stream to adjust parameters
5778
- */
5779
- type AssemblyAIUpdateConfiguration = Omit<StreamingUpdateConfiguration, "type">;
5780
- /**
5781
- * Union of all provider-specific streaming options
5782
- */
5783
- type ProviderStreamingOptions = ({
5784
- provider: "gladia";
5785
- } & GladiaStreamingOptions) | ({
5786
- provider: "deepgram";
5787
- } & DeepgramStreamingOptions) | ({
5788
- provider: "assemblyai";
5789
- } & AssemblyAIStreamingOptions);
5790
- /**
5791
- * Type-safe streaming options for a specific provider
5792
- */
5793
- type StreamingOptionsForProvider<P extends StreamingProvider> = P extends "gladia" ? GladiaStreamingOptions : P extends "deepgram" ? DeepgramStreamingOptions : P extends "assemblyai" ? AssemblyAIStreamingOptions : never;
5794
- /**
5795
- * Type-safe transcribeStream parameters for a specific provider
5796
- */
5797
- interface TranscribeStreamParams<P extends StreamingProvider> {
5798
- /** Streaming options specific to this provider */
5799
- options?: StreamingOptionsForProvider<P> & {
5800
- provider: P;
5801
- };
5802
- /** Event callbacks */
5803
- callbacks?: StreamingCallbacks;
5804
- }
6646
+ status?: ManageV1FilterStatusParameter;
6647
+ };
5805
6648
 
5806
6649
  /**
5807
6650
  * Unified types for the Voice Router SDK
@@ -5816,15 +6659,20 @@ type SpeechmaticsOperatingPoint = "standard" | "enhanced";
5816
6659
  /**
5817
6660
  * Unified transcription model type with autocomplete for all providers
5818
6661
  *
5819
- * Includes all known models from:
5820
- * - Deepgram: nova-3, nova-2, enhanced, base, whisper, etc.
6662
+ * Strict union type - only accepts valid models from each provider:
6663
+ * - Deepgram: nova-3, nova-2, enhanced, base, etc.
5821
6664
  * - AssemblyAI: best, slam-1, universal
5822
6665
  * - Gladia: solaria-1
5823
6666
  * - Speechmatics: standard, enhanced
5824
6667
  *
5825
- * Also accepts any string for future/custom models.
6668
+ * Use provider const objects for autocomplete:
6669
+ * @example
6670
+ * ```typescript
6671
+ * import { DeepgramModel } from 'voice-router-dev'
6672
+ * { model: DeepgramModel["nova-3"] }
6673
+ * ```
5826
6674
  */
5827
- type TranscriptionModel = ListenV1ModelParameter | StreamingSupportedModels | SpeechModel | SpeechmaticsOperatingPoint;
6675
+ type TranscriptionModel = DeepgramModelType | StreamingSupportedModels | SpeechModel | SpeechmaticsOperatingPoint;
5828
6676
  /**
5829
6677
  * Unified transcription language type with autocomplete for all providers
5830
6678
  *
@@ -5944,6 +6792,60 @@ interface ProviderCapabilities {
5944
6792
  entityDetection: boolean;
5945
6793
  /** PII redaction */
5946
6794
  piiRedaction: boolean;
6795
+ /** List/fetch previous transcriptions */
6796
+ listTranscripts: boolean;
6797
+ /** Delete existing transcriptions */
6798
+ deleteTranscript: boolean;
6799
+ /** Download original audio file from transcription (only Gladia supports this) */
6800
+ getAudioFile?: boolean;
6801
+ }
6802
+
6803
+ /**
6804
+ * Options for listing transcripts with date/time filtering
6805
+ *
6806
+ * Providers support different filtering capabilities:
6807
+ * - AssemblyAI: status, created_on, before_id, after_id, throttled_only
6808
+ * - Gladia: status, date, before_date, after_date, custom_metadata
6809
+ * - Azure: status, skip, top, filter (OData)
6810
+ * - Deepgram: start, end, status, page, request_id, endpoint (requires projectId)
6811
+ *
6812
+ * @example Filter by date
6813
+ * ```typescript
6814
+ * await adapter.listTranscripts({
6815
+ * date: '2026-01-07', // Exact date (ISO format)
6816
+ * status: 'completed',
6817
+ * limit: 50
6818
+ * })
6819
+ * ```
6820
+ *
6821
+ * @example Filter by date range
6822
+ * ```typescript
6823
+ * await adapter.listTranscripts({
6824
+ * afterDate: '2026-01-01',
6825
+ * beforeDate: '2026-01-31',
6826
+ * limit: 100
6827
+ * })
6828
+ * ```
6829
+ */
6830
+ interface ListTranscriptsOptions {
6831
+ /** Maximum number of transcripts to retrieve */
6832
+ limit?: number;
6833
+ /** Pagination offset (skip N results) */
6834
+ offset?: number;
6835
+ /** Filter by transcript status */
6836
+ status?: "queued" | "processing" | "completed" | "error" | string;
6837
+ /** Filter by exact date (ISO format: YYYY-MM-DD) */
6838
+ date?: string;
6839
+ /** Filter for transcripts created before this date (ISO format) */
6840
+ beforeDate?: string;
6841
+ /** Filter for transcripts created after this date (ISO format) */
6842
+ afterDate?: string;
6843
+ /** AssemblyAI-specific list options */
6844
+ assemblyai?: Partial<ListTranscriptsParams>;
6845
+ /** Gladia-specific list options */
6846
+ gladia?: Partial<TranscriptionControllerListV2Params>;
6847
+ /** Deepgram-specific list options (request history) */
6848
+ deepgram?: Partial<ManageV1ProjectsRequestsListParams>;
5947
6849
  }
5948
6850
  /**
5949
6851
  * Audio input for transcription
@@ -6091,6 +6993,135 @@ interface Utterance {
6091
6993
  * Transcription status
6092
6994
  */
6093
6995
  type TranscriptionStatus = "queued" | "processing" | "completed" | "error";
6996
+ /**
6997
+ * Transcript metadata with typed common fields
6998
+ *
6999
+ * Contains provider-agnostic metadata fields that are commonly available.
7000
+ * Provider-specific fields can be accessed via the index signature.
7001
+ *
7002
+ * @example
7003
+ * ```typescript
7004
+ * const { transcripts } = await router.listTranscripts('assemblyai', { limit: 20 });
7005
+ * transcripts.forEach(item => {
7006
+ * console.log(item.data?.metadata?.audioUrl); // string | undefined
7007
+ * console.log(item.data?.metadata?.createdAt); // string | undefined
7008
+ * console.log(item.data?.metadata?.audioDuration); // number | undefined
7009
+ * });
7010
+ * ```
7011
+ */
7012
+ interface TranscriptMetadata {
7013
+ /**
7014
+ * Original audio URL/source you provided to the API (echoed back).
7015
+ * This is NOT a provider-hosted URL - it's what you sent when creating the transcription.
7016
+ */
7017
+ sourceAudioUrl?: string;
7018
+ /**
7019
+ * True if the provider stored the audio and it can be downloaded via adapter.getAudioFile().
7020
+ * Currently only Gladia supports this - other providers discard audio after processing.
7021
+ *
7022
+ * @example
7023
+ * ```typescript
7024
+ * if (item.data?.metadata?.audioFileAvailable) {
7025
+ * const audio = await gladiaAdapter.getAudioFile(item.data.id)
7026
+ * // audio.data is a Blob
7027
+ * }
7028
+ * ```
7029
+ */
7030
+ audioFileAvailable?: boolean;
7031
+ /** Resource URL for the transcript */
7032
+ resourceUrl?: string;
7033
+ /** Creation timestamp (ISO 8601) */
7034
+ createdAt?: string;
7035
+ /** Completion timestamp (ISO 8601) */
7036
+ completedAt?: string;
7037
+ /** Last action timestamp (Azure) */
7038
+ lastActionAt?: string;
7039
+ /** Audio duration in seconds */
7040
+ audioDuration?: number;
7041
+ /** Transcript type */
7042
+ kind?: "batch" | "pre-recorded" | "live" | "streaming";
7043
+ /** Display name (Azure) */
7044
+ displayName?: string;
7045
+ /** Files URL (Azure) */
7046
+ filesUrl?: string;
7047
+ /** Custom metadata (Gladia) */
7048
+ customMetadata?: Record<string, unknown>;
7049
+ /** Provider-specific fields */
7050
+ [key: string]: unknown;
7051
+ }
7052
+ /**
7053
+ * Transcript data structure
7054
+ *
7055
+ * Contains the core transcript information returned by getTranscript and listTranscripts.
7056
+ *
7057
+ * @example
7058
+ * ```typescript
7059
+ * const result = await router.getTranscript('abc123', 'assemblyai');
7060
+ * if (result.success && result.data) {
7061
+ * console.log(result.data.id); // string
7062
+ * console.log(result.data.text); // string
7063
+ * console.log(result.data.status); // TranscriptionStatus
7064
+ * console.log(result.data.metadata); // TranscriptMetadata
7065
+ * }
7066
+ * ```
7067
+ */
7068
+ interface TranscriptData {
7069
+ /** Unique transcript ID */
7070
+ id: string;
7071
+ /** Full transcribed text (empty for list items) */
7072
+ text: string;
7073
+ /** Overall confidence score (0-1) */
7074
+ confidence?: number;
7075
+ /** Transcription status */
7076
+ status: TranscriptionStatus;
7077
+ /** Detected or specified language code */
7078
+ language?: string;
7079
+ /** Audio duration in seconds */
7080
+ duration?: number;
7081
+ /** Speaker diarization results */
7082
+ speakers?: Speaker[];
7083
+ /** Word-level transcription with timestamps */
7084
+ words?: Word[];
7085
+ /** Utterances (speaker turns) */
7086
+ utterances?: Utterance[];
7087
+ /** Summary of the content (if summarization enabled) */
7088
+ summary?: string;
7089
+ /** Transcript metadata */
7090
+ metadata?: TranscriptMetadata;
7091
+ /** Creation timestamp (shorthand for metadata.createdAt) */
7092
+ createdAt?: string;
7093
+ /** Completion timestamp (shorthand for metadata.completedAt) */
7094
+ completedAt?: string;
7095
+ }
7096
+ /**
7097
+ * Response from listTranscripts
7098
+ *
7099
+ * @example
7100
+ * ```typescript
7101
+ * import type { ListTranscriptsResponse } from 'voice-router-dev';
7102
+ *
7103
+ * const response: ListTranscriptsResponse = await router.listTranscripts('assemblyai', {
7104
+ * status: 'completed',
7105
+ * limit: 50
7106
+ * });
7107
+ *
7108
+ * response.transcripts.forEach(item => {
7109
+ * console.log(item.data?.id, item.data?.status);
7110
+ * });
7111
+ *
7112
+ * if (response.hasMore) {
7113
+ * // Fetch next page
7114
+ * }
7115
+ * ```
7116
+ */
7117
+ interface ListTranscriptsResponse {
7118
+ /** List of transcripts */
7119
+ transcripts: UnifiedTranscriptResponse[];
7120
+ /** Total count (if available from provider) */
7121
+ total?: number;
7122
+ /** Whether more results are available */
7123
+ hasMore?: boolean;
7124
+ }
6094
7125
  /**
6095
7126
  * Map of provider names to their raw response types
6096
7127
  * Enables type-safe access to provider-specific raw responses
@@ -6140,34 +7171,7 @@ interface UnifiedTranscriptResponse<P extends TranscriptionProvider = Transcript
6140
7171
  /** Provider that performed the transcription */
6141
7172
  provider: P;
6142
7173
  /** Transcription data (only present on success) */
6143
- data?: {
6144
- /** Unique transcription ID */
6145
- id: string;
6146
- /** Full transcribed text */
6147
- text: string;
6148
- /** Overall confidence score (0-1) */
6149
- confidence?: number;
6150
- /** Transcription status */
6151
- status: TranscriptionStatus;
6152
- /** Detected or specified language code */
6153
- language?: string;
6154
- /** Audio duration in seconds */
6155
- duration?: number;
6156
- /** Speaker diarization results */
6157
- speakers?: Speaker[];
6158
- /** Word-level transcription with timestamps */
6159
- words?: Word[];
6160
- /** Utterances (speaker turns) */
6161
- utterances?: Utterance[];
6162
- /** Summary of the content (if summarization enabled) */
6163
- summary?: string;
6164
- /** Additional provider-specific metadata */
6165
- metadata?: Record<string, unknown>;
6166
- /** Creation timestamp */
6167
- createdAt?: string;
6168
- /** Completion timestamp */
6169
- completedAt?: string;
6170
- };
7174
+ data?: TranscriptData;
6171
7175
  /**
6172
7176
  * Extended provider-specific data (fully typed from OpenAPI specs)
6173
7177
  *
@@ -6678,14 +7682,22 @@ interface TranscriptionAdapter {
6678
7682
  success: boolean;
6679
7683
  }>;
6680
7684
  /**
6681
- * List recent transcriptions
6682
- * Not all providers support listing
7685
+ * List recent transcriptions with filtering
7686
+ *
7687
+ * Supports date/time filtering, status filtering, and pagination.
7688
+ * Not all providers support all options - use provider-specific passthrough for full control.
7689
+ *
7690
+ * @example Filter by date range
7691
+ * ```typescript
7692
+ * const { transcripts } = await adapter.listTranscripts({
7693
+ * afterDate: '2026-01-01',
7694
+ * beforeDate: '2026-01-31',
7695
+ * status: 'completed',
7696
+ * limit: 50
7697
+ * })
7698
+ * ```
6683
7699
  */
6684
- listTranscripts?(options?: {
6685
- limit?: number;
6686
- offset?: number;
6687
- status?: string;
6688
- }): Promise<{
7700
+ listTranscripts?(options?: ListTranscriptsOptions): Promise<{
6689
7701
  transcripts: UnifiedTranscriptResponse[];
6690
7702
  total?: number;
6691
7703
  hasMore?: boolean;
@@ -6981,14 +7993,22 @@ declare class VoiceRouter {
6981
7993
  success: boolean;
6982
7994
  }>;
6983
7995
  /**
6984
- * List recent transcriptions
6985
- * Not all providers support this operation
7996
+ * List recent transcriptions with filtering
7997
+ *
7998
+ * Supports date/time filtering, status filtering, and pagination.
7999
+ * Not all providers support this operation.
8000
+ *
8001
+ * @example Filter by date range
8002
+ * ```typescript
8003
+ * const { transcripts } = await router.listTranscripts('assemblyai', {
8004
+ * afterDate: '2026-01-01',
8005
+ * beforeDate: '2026-01-31',
8006
+ * status: 'completed',
8007
+ * limit: 50
8008
+ * })
8009
+ * ```
6986
8010
  */
6987
- listTranscripts(provider: TranscriptionProvider, options?: {
6988
- limit?: number;
6989
- offset?: number;
6990
- status?: string;
6991
- }): Promise<{
8011
+ listTranscripts(provider: TranscriptionProvider, options?: ListTranscriptsOptions): Promise<{
6992
8012
  transcripts: UnifiedTranscriptResponse[];
6993
8013
  total?: number;
6994
8014
  hasMore?: boolean;
@@ -7301,33 +8321,127 @@ declare class GladiaAdapter extends BaseAdapter {
7301
8321
  */
7302
8322
  private extractUtterances;
7303
8323
  /**
7304
- * Delete a transcription job and its associated data
8324
+ * Delete a transcription job and its associated data
8325
+ *
8326
+ * Removes the transcription data from Gladia's servers. This action is
8327
+ * irreversible. Supports both pre-recorded and streaming job IDs.
8328
+ *
8329
+ * @param transcriptId - The ID of the transcript/job to delete
8330
+ * @param jobType - Type of job: 'pre-recorded' or 'streaming' (defaults to 'pre-recorded')
8331
+ * @returns Promise with success status
8332
+ *
8333
+ * @example Delete a pre-recorded transcript
8334
+ * ```typescript
8335
+ * const result = await adapter.deleteTranscript('abc123');
8336
+ * if (result.success) {
8337
+ * console.log('Transcript deleted successfully');
8338
+ * }
8339
+ * ```
8340
+ *
8341
+ * @example Delete a streaming job
8342
+ * ```typescript
8343
+ * const result = await adapter.deleteTranscript('stream-456', 'streaming');
8344
+ * ```
8345
+ *
8346
+ * @see https://docs.gladia.io/
8347
+ */
8348
+ deleteTranscript(transcriptId: string, jobType?: "pre-recorded" | "streaming"): Promise<{
8349
+ success: boolean;
8350
+ }>;
8351
+ /**
8352
+ * Download the original audio file from a transcription
8353
+ *
8354
+ * Gladia stores the audio files used for transcription and allows downloading them.
8355
+ * This works for both pre-recorded and streaming (live) transcriptions.
8356
+ *
8357
+ * @param transcriptId - The ID of the transcript/job
8358
+ * @param jobType - Type of job: 'pre-recorded' or 'streaming' (defaults to 'pre-recorded')
8359
+ * @returns Promise with the audio file as a Blob, or error
8360
+ *
8361
+ * @example Download audio from a pre-recorded transcription
8362
+ * ```typescript
8363
+ * const result = await adapter.getAudioFile('abc123');
8364
+ * if (result.success && result.data) {
8365
+ * // Save to file (Node.js)
8366
+ * const buffer = Buffer.from(await result.data.arrayBuffer());
8367
+ * fs.writeFileSync('audio.mp3', buffer);
8368
+ *
8369
+ * // Or create download URL (browser)
8370
+ * const url = URL.createObjectURL(result.data);
8371
+ * }
8372
+ * ```
8373
+ *
8374
+ * @example Download audio from a live/streaming session
8375
+ * ```typescript
8376
+ * const result = await adapter.getAudioFile('stream-456', 'streaming');
8377
+ * if (result.success) {
8378
+ * console.log('Audio file size:', result.data?.size);
8379
+ * }
8380
+ * ```
8381
+ *
8382
+ * @see https://docs.gladia.io/
8383
+ */
8384
+ getAudioFile(transcriptId: string, jobType?: "pre-recorded" | "streaming"): Promise<{
8385
+ success: boolean;
8386
+ data?: Blob;
8387
+ error?: {
8388
+ code: string;
8389
+ message: string;
8390
+ };
8391
+ }>;
8392
+ /**
8393
+ * List recent transcriptions with filtering
7305
8394
  *
7306
- * Removes the transcription data from Gladia's servers. This action is
7307
- * irreversible. Supports both pre-recorded and streaming job IDs.
8395
+ * Retrieves a list of transcription jobs (both pre-recorded and streaming)
8396
+ * with optional filtering by status, date, and custom metadata.
7308
8397
  *
7309
- * @param transcriptId - The ID of the transcript/job to delete
7310
- * @param jobType - Type of job: 'pre-recorded' or 'streaming' (defaults to 'pre-recorded')
7311
- * @returns Promise with success status
8398
+ * @param options - Filtering and pagination options
8399
+ * @param options.limit - Maximum number of transcripts to return
8400
+ * @param options.offset - Pagination offset (skip N results)
8401
+ * @param options.status - Filter by status (queued, processing, done, error)
8402
+ * @param options.date - Filter by exact date (ISO format YYYY-MM-DD)
8403
+ * @param options.beforeDate - Filter for transcripts before this date
8404
+ * @param options.afterDate - Filter for transcripts after this date
8405
+ * @param options.gladia - Full Gladia-specific options (custom_metadata, kind, etc.)
8406
+ * @returns List of transcripts with pagination info
7312
8407
  *
7313
- * @example Delete a pre-recorded transcript
8408
+ * @example List recent transcripts
7314
8409
  * ```typescript
7315
- * const result = await adapter.deleteTranscript('abc123');
7316
- * if (result.success) {
7317
- * console.log('Transcript deleted successfully');
7318
- * }
8410
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
8411
+ * limit: 50,
8412
+ * status: 'done'
8413
+ * })
7319
8414
  * ```
7320
8415
  *
7321
- * @example Delete a streaming job
8416
+ * @example Filter by date range
7322
8417
  * ```typescript
7323
- * const result = await adapter.deleteTranscript('stream-456', 'streaming');
8418
+ * const { transcripts } = await adapter.listTranscripts({
8419
+ * afterDate: '2026-01-01',
8420
+ * beforeDate: '2026-01-31',
8421
+ * limit: 100
8422
+ * })
8423
+ * ```
8424
+ *
8425
+ * @example Filter by custom metadata
8426
+ * ```typescript
8427
+ * const { transcripts } = await adapter.listTranscripts({
8428
+ * gladia: {
8429
+ * custom_metadata: { project: 'my-project' }
8430
+ * }
8431
+ * })
7324
8432
  * ```
7325
8433
  *
7326
8434
  * @see https://docs.gladia.io/
7327
8435
  */
7328
- deleteTranscript(transcriptId: string, jobType?: "pre-recorded" | "streaming"): Promise<{
7329
- success: boolean;
8436
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
8437
+ transcripts: UnifiedTranscriptResponse[];
8438
+ total?: number;
8439
+ hasMore?: boolean;
7330
8440
  }>;
8441
+ /**
8442
+ * Normalize a transcript list item to unified format
8443
+ */
8444
+ private normalizeListItem;
7331
8445
  /**
7332
8446
  * Stream audio for real-time transcription
7333
8447
  *
@@ -7631,6 +8745,57 @@ declare class AssemblyAIAdapter extends BaseAdapter {
7631
8745
  deleteTranscript(transcriptId: string): Promise<{
7632
8746
  success: boolean;
7633
8747
  }>;
8748
+ /**
8749
+ * List recent transcriptions with filtering
8750
+ *
8751
+ * Retrieves a list of transcripts with optional filtering by status and date.
8752
+ * Transcripts are sorted from newest to oldest and can be retrieved for the
8753
+ * last 90 days of usage.
8754
+ *
8755
+ * @param options - Filtering and pagination options
8756
+ * @param options.limit - Maximum number of transcripts (max 200)
8757
+ * @param options.status - Filter by status (queued, processing, completed, error)
8758
+ * @param options.date - Filter by exact date (ISO format YYYY-MM-DD)
8759
+ * @param options.assemblyai - Full AssemblyAI-specific options (before_id, after_id, etc.)
8760
+ * @returns List of transcripts with pagination info
8761
+ *
8762
+ * @example List recent transcripts
8763
+ * ```typescript
8764
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
8765
+ * limit: 50,
8766
+ * status: 'completed'
8767
+ * })
8768
+ * ```
8769
+ *
8770
+ * @example Filter by date
8771
+ * ```typescript
8772
+ * const { transcripts } = await adapter.listTranscripts({
8773
+ * date: '2026-01-07',
8774
+ * limit: 100
8775
+ * })
8776
+ * ```
8777
+ *
8778
+ * @example Use cursor pagination
8779
+ * ```typescript
8780
+ * const { transcripts } = await adapter.listTranscripts({
8781
+ * assemblyai: {
8782
+ * after_id: 'abc123', // Get transcripts after this ID
8783
+ * limit: 50
8784
+ * }
8785
+ * })
8786
+ * ```
8787
+ *
8788
+ * @see https://www.assemblyai.com/docs/api-reference/transcripts/list
8789
+ */
8790
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
8791
+ transcripts: UnifiedTranscriptResponse[];
8792
+ total?: number;
8793
+ hasMore?: boolean;
8794
+ }>;
8795
+ /**
8796
+ * Normalize a transcript list item to unified format
8797
+ */
8798
+ private normalizeListItem;
7634
8799
  /**
7635
8800
  * Build AssemblyAI transcription request from unified options
7636
8801
  */
@@ -7819,7 +8984,10 @@ declare class DeepgramAdapter extends BaseAdapter {
7819
8984
  private client?;
7820
8985
  protected baseUrl: string;
7821
8986
  private wsBaseUrl;
7822
- initialize(config: ProviderConfig): void;
8987
+ private projectId?;
8988
+ initialize(config: ProviderConfig & {
8989
+ projectId?: string;
8990
+ }): void;
7823
8991
  /**
7824
8992
  * Submit audio for transcription
7825
8993
  *
@@ -7867,12 +9035,30 @@ declare class DeepgramAdapter extends BaseAdapter {
7867
9035
  /**
7868
9036
  * Get transcription result by ID
7869
9037
  *
7870
- * Note: Deepgram processes synchronously, so this method is primarily
7871
- * for retrieving cached results if you've stored the request ID.
7872
- * The initial transcribe() call already returns complete results.
9038
+ * Retrieves a previous transcription from Deepgram's request history.
7873
9039
  *
7874
- * @param transcriptId - Request ID from Deepgram
7875
- * @returns Normalized transcription response
9040
+ * Unlike the list endpoint, getting a single request DOES include the full
9041
+ * transcript response. Requires `projectId` to be set during initialization.
9042
+ *
9043
+ * @param transcriptId - Request ID from a previous transcription
9044
+ * @returns Full transcript response including text, words, and metadata
9045
+ *
9046
+ * @example Get a transcript by request ID
9047
+ * ```typescript
9048
+ * const adapter = new DeepgramAdapter()
9049
+ * adapter.initialize({
9050
+ * apiKey: process.env.DEEPGRAM_API_KEY,
9051
+ * projectId: process.env.DEEPGRAM_PROJECT_ID
9052
+ * })
9053
+ *
9054
+ * const result = await adapter.getTranscript('abc123-request-id')
9055
+ * if (result.success) {
9056
+ * console.log(result.data?.text)
9057
+ * console.log(result.data?.words)
9058
+ * }
9059
+ * ```
9060
+ *
9061
+ * @see https://developers.deepgram.com/reference/get-request
7876
9062
  */
7877
9063
  getTranscript(transcriptId: string): Promise<UnifiedTranscriptResponse>;
7878
9064
  /**
@@ -8012,6 +9198,52 @@ declare class DeepgramAdapter extends BaseAdapter {
8012
9198
  * Handle all WebSocket message types from Deepgram streaming
8013
9199
  */
8014
9200
  private handleWebSocketMessage;
9201
+ /**
9202
+ * List recent transcription requests from Deepgram's request history
9203
+ *
9204
+ * **Important:** Deepgram processes synchronously and doesn't store transcript content.
9205
+ * This method returns request metadata (IDs, timestamps, status) but NOT the actual
9206
+ * transcript text. Use this for auditing, billing analysis, or request tracking.
9207
+ *
9208
+ * Requires `projectId` to be set during initialization.
9209
+ *
9210
+ * @param options - Filtering and pagination options
9211
+ * @param options.limit - Maximum number of requests (default 10, max 1000)
9212
+ * @param options.status - Filter by status: 'succeeded' or 'failed'
9213
+ * @param options.afterDate - Start date (YYYY-MM-DD or ISO 8601)
9214
+ * @param options.beforeDate - End date (YYYY-MM-DD or ISO 8601)
9215
+ * @param options.deepgram - Provider-specific params (page, request_id, etc.)
9216
+ * @returns List of transcription request metadata
9217
+ *
9218
+ * @example List recent transcription requests
9219
+ * ```typescript
9220
+ * const adapter = new DeepgramAdapter()
9221
+ * adapter.initialize({
9222
+ * apiKey: process.env.DEEPGRAM_API_KEY,
9223
+ * projectId: process.env.DEEPGRAM_PROJECT_ID
9224
+ * })
9225
+ *
9226
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
9227
+ * limit: 50,
9228
+ * status: 'succeeded'
9229
+ * })
9230
+ *
9231
+ * transcripts.forEach(t => {
9232
+ * console.log(t.data?.id, t.data?.metadata?.createdAt)
9233
+ * })
9234
+ * ```
9235
+ *
9236
+ * @see https://developers.deepgram.com/reference/get-all-requests
9237
+ */
9238
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
9239
+ transcripts: UnifiedTranscriptResponse[];
9240
+ total?: number;
9241
+ hasMore?: boolean;
9242
+ }>;
9243
+ /**
9244
+ * Normalize a Deepgram request history item to unified format
9245
+ */
9246
+ private normalizeRequestItem;
8015
9247
  }
8016
9248
  /**
8017
9249
  * Factory function to create a Deepgram adapter
@@ -8154,6 +9386,48 @@ declare class AzureSTTAdapter extends BaseAdapter {
8154
9386
  deleteTranscript(transcriptId: string): Promise<{
8155
9387
  success: boolean;
8156
9388
  }>;
9389
+ /**
9390
+ * List recent transcriptions with filtering
9391
+ *
9392
+ * Retrieves a list of transcription jobs for the authenticated subscription.
9393
+ * Azure uses OData filtering for advanced queries.
9394
+ *
9395
+ * @param options - Filtering and pagination options
9396
+ * @param options.limit - Maximum number of transcripts (maps to 'top')
9397
+ * @param options.offset - Pagination offset (maps to 'skip')
9398
+ * @param options.status - Filter by status (uses OData filter)
9399
+ * @returns List of transcripts with pagination info
9400
+ *
9401
+ * @example List recent transcripts
9402
+ * ```typescript
9403
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
9404
+ * limit: 50
9405
+ * })
9406
+ * ```
9407
+ *
9408
+ * @example Filter by status using OData
9409
+ * ```typescript
9410
+ * const { transcripts } = await adapter.listTranscripts({
9411
+ * status: 'Succeeded',
9412
+ * limit: 100
9413
+ * })
9414
+ * ```
9415
+ *
9416
+ * @see https://learn.microsoft.com/azure/cognitive-services/speech-service/batch-transcription
9417
+ */
9418
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
9419
+ transcripts: UnifiedTranscriptResponse[];
9420
+ total?: number;
9421
+ hasMore?: boolean;
9422
+ }>;
9423
+ /**
9424
+ * Map unified status to Azure status format using generated enum
9425
+ */
9426
+ private mapStatusToAzure;
9427
+ /**
9428
+ * Normalize a transcript list item to unified format
9429
+ */
9430
+ private normalizeListItem;
8157
9431
  /**
8158
9432
  * Build Azure-specific transcription properties
8159
9433
  */
@@ -11927,73 +13201,6 @@ type StreamingControllerInitStreamingSessionV2Params = {
11927
13201
 
11928
13202
  type TranscriptionControllerGetTranscriptV2200 = PreRecordedResponse | StreamingResponse;
11929
13203
 
11930
- /**
11931
- * Generated by orval v7.9.0 🍺
11932
- * Do not edit manually.
11933
- * Gladia Control API
11934
- * OpenAPI spec version: 1.0
11935
- */
11936
- type TranscriptionControllerListV2KindItem = (typeof TranscriptionControllerListV2KindItem)[keyof typeof TranscriptionControllerListV2KindItem];
11937
- declare const TranscriptionControllerListV2KindItem: {
11938
- readonly "pre-recorded": "pre-recorded";
11939
- readonly live: "live";
11940
- };
11941
-
11942
- /**
11943
- * Generated by orval v7.9.0 🍺
11944
- * Do not edit manually.
11945
- * Gladia Control API
11946
- * OpenAPI spec version: 1.0
11947
- */
11948
- type TranscriptionControllerListV2StatusItem = (typeof TranscriptionControllerListV2StatusItem)[keyof typeof TranscriptionControllerListV2StatusItem];
11949
- declare const TranscriptionControllerListV2StatusItem: {
11950
- readonly queued: "queued";
11951
- readonly processing: "processing";
11952
- readonly done: "done";
11953
- readonly error: "error";
11954
- };
11955
-
11956
- /**
11957
- * Generated by orval v7.9.0 🍺
11958
- * Do not edit manually.
11959
- * Gladia Control API
11960
- * OpenAPI spec version: 1.0
11961
- */
11962
-
11963
- type TranscriptionControllerListV2Params = {
11964
- /**
11965
- * The starting point for pagination. A value of 0 starts from the first item.
11966
- */
11967
- offset?: number;
11968
- /**
11969
- * The maximum number of items to return. Useful for pagination and controlling data payload size.
11970
- */
11971
- limit?: number;
11972
- /**
11973
- * Filter items relevant to a specific date in ISO format (YYYY-MM-DD).
11974
- */
11975
- date?: string;
11976
- /**
11977
- * Include items that occurred before the specified date in ISO format.
11978
- */
11979
- before_date?: string;
11980
- /**
11981
- * Filter for items after the specified date. Use with `before_date` for a range. Date in ISO format.
11982
- */
11983
- after_date?: string;
11984
- /**
11985
- * Filter the list based on item status. Accepts multiple values from the predefined list.
11986
- */
11987
- status?: TranscriptionControllerListV2StatusItem[];
11988
- custom_metadata?: {
11989
- [key: string]: unknown;
11990
- };
11991
- /**
11992
- * Filter the list based on the item type. Supports multiple values from the predefined list.
11993
- */
11994
- kind?: TranscriptionControllerListV2KindItem[];
11995
- };
11996
-
11997
13204
  /**
11998
13205
  * Generated by orval v7.9.0 🍺
11999
13206
  * Do not edit manually.
@@ -13233,101 +14440,6 @@ interface ListTranscriptParams {
13233
14440
  throttled_only?: boolean;
13234
14441
  }
13235
14442
 
13236
- /**
13237
- * Generated by orval v7.9.0 🍺
13238
- * Do not edit manually.
13239
- * AssemblyAI API
13240
- * AssemblyAI API
13241
- * OpenAPI spec version: 1.3.4
13242
- */
13243
- /**
13244
- * Get transcripts that were created after this transcript ID
13245
- */
13246
- type AfterId = string;
13247
-
13248
- /**
13249
- * Generated by orval v7.9.0 🍺
13250
- * Do not edit manually.
13251
- * AssemblyAI API
13252
- * AssemblyAI API
13253
- * OpenAPI spec version: 1.3.4
13254
- */
13255
- /**
13256
- * Get transcripts that were created before this transcript ID
13257
- */
13258
- type BeforeId = string;
13259
-
13260
- /**
13261
- * Generated by orval v7.9.0 🍺
13262
- * Do not edit manually.
13263
- * AssemblyAI API
13264
- * AssemblyAI API
13265
- * OpenAPI spec version: 1.3.4
13266
- */
13267
- /**
13268
- * Only get transcripts created on this date
13269
- */
13270
- type CreatedOn = string;
13271
-
13272
- /**
13273
- * Generated by orval v7.9.0 🍺
13274
- * Do not edit manually.
13275
- * AssemblyAI API
13276
- * AssemblyAI API
13277
- * OpenAPI spec version: 1.3.4
13278
- */
13279
- /**
13280
- * Maximum amount of transcripts to retrieve
13281
- */
13282
- type Limit = number;
13283
-
13284
- /**
13285
- * Generated by orval v7.9.0 🍺
13286
- * Do not edit manually.
13287
- * AssemblyAI API
13288
- * AssemblyAI API
13289
- * OpenAPI spec version: 1.3.4
13290
- */
13291
- /**
13292
- * Only get throttled transcripts, overrides the status filter
13293
- */
13294
- type ThrottledOnly = boolean;
13295
-
13296
- /**
13297
- * Generated by orval v7.9.0 🍺
13298
- * Do not edit manually.
13299
- * AssemblyAI API
13300
- * AssemblyAI API
13301
- * OpenAPI spec version: 1.3.4
13302
- */
13303
-
13304
- type ListTranscriptsParams = {
13305
- /**
13306
- * Maximum amount of transcripts to retrieve
13307
- */
13308
- limit?: Limit;
13309
- /**
13310
- * Filter by transcript status
13311
- */
13312
- status?: TranscriptStatus;
13313
- /**
13314
- * Only get transcripts created on this date
13315
- */
13316
- created_on?: CreatedOn;
13317
- /**
13318
- * Get transcripts that were created before this transcript ID
13319
- */
13320
- before_id?: BeforeId;
13321
- /**
13322
- * Get transcripts that were created after this transcript ID
13323
- */
13324
- after_id?: AfterId;
13325
- /**
13326
- * Only get throttled transcripts, overrides the status filter
13327
- */
13328
- throttled_only?: ThrottledOnly;
13329
- };
13330
-
13331
14443
  /**
13332
14444
  * Generated by orval v7.9.0 🍺
13333
14445
  * Do not edit manually.
@@ -13939,4 +15051,4 @@ declare namespace index {
13939
15051
  export { index_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, type index_AutoHighlightResult as AutoHighlightResult, type index_AutoHighlightsResult as AutoHighlightsResult, type index_BadRequestResponse as BadRequestResponse, type index_CannotAccessUploadedFileResponse as CannotAccessUploadedFileResponse, type index_Chapter as Chapter, type index_ContentSafetyLabel as ContentSafetyLabel, type index_ContentSafetyLabelResult as ContentSafetyLabelResult, type index_ContentSafetyLabelsResult as ContentSafetyLabelsResult, type index_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, type index_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, type index_CreateRealtimeTemporaryTokenParams as CreateRealtimeTemporaryTokenParams, type index_Entity as Entity, index_EntityType as EntityType, type Error$1 as Error, type index_GatewayTimeoutResponse as GatewayTimeoutResponse, type index_GetSubtitlesParams as GetSubtitlesParams, type index_InternalServerErrorResponse as InternalServerErrorResponse, type index_LemurActionItemsParams as LemurActionItemsParams, type index_LemurActionItemsParamsAllOf as LemurActionItemsParamsAllOf, type index_LemurActionItemsResponse as LemurActionItemsResponse, type index_LemurBaseParams as LemurBaseParams, type index_LemurBaseParamsContext as LemurBaseParamsContext, type index_LemurBaseParamsContextOneOf as LemurBaseParamsContextOneOf, type index_LemurBaseParamsFinalModel as LemurBaseParamsFinalModel, type index_LemurBaseResponse as LemurBaseResponse, index_LemurModel as LemurModel, type index_LemurQuestion as LemurQuestion, type index_LemurQuestionAnswer as LemurQuestionAnswer, type index_LemurQuestionAnswerParams as LemurQuestionAnswerParams, type index_LemurQuestionAnswerParamsAllOf as LemurQuestionAnswerParamsAllOf, type index_LemurQuestionAnswerResponse as LemurQuestionAnswerResponse, type index_LemurQuestionAnswerResponseAllOf as LemurQuestionAnswerResponseAllOf, type index_LemurQuestionContext as LemurQuestionContext, type index_LemurQuestionContextOneOf as LemurQuestionContextOneOf, type index_LemurResponse as LemurResponse, type index_LemurStringResponse as LemurStringResponse, type index_LemurStringResponseAllOf as LemurStringResponseAllOf, type index_LemurSummaryParams as LemurSummaryParams, type index_LemurSummaryParamsAllOf as LemurSummaryParamsAllOf, type index_LemurSummaryResponse as LemurSummaryResponse, type index_LemurTaskParams as LemurTaskParams, type index_LemurTaskParamsAllOf as LemurTaskParamsAllOf, type index_LemurTaskResponse as LemurTaskResponse, type index_LemurUsage as LemurUsage, type index_ListTranscriptParams as ListTranscriptParams, type index_ListTranscriptsParams as ListTranscriptsParams, type index_NotFoundResponse as NotFoundResponse, type index_PageDetails as PageDetails, type index_PageDetailsNextUrl as PageDetailsNextUrl, type index_PageDetailsPrevUrl as PageDetailsPrevUrl, type index_ParagraphsResponse as ParagraphsResponse, index_PiiPolicy as PiiPolicy, type index_PurgeLemurRequestDataResponse as PurgeLemurRequestDataResponse, type index_RealtimeTemporaryTokenResponse as RealtimeTemporaryTokenResponse, index_RedactPiiAudioQuality as RedactPiiAudioQuality, type index_RedactedAudioNotification as RedactedAudioNotification, type index_RedactedAudioResponse as RedactedAudioResponse, index_RedactedAudioStatus as RedactedAudioStatus, type index_SentencesResponse as SentencesResponse, index_Sentiment as Sentiment, type SentimentAnalysisResult$1 as SentimentAnalysisResult, type index_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, type index_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, type index_ServiceUnavailableResponse as ServiceUnavailableResponse, type index_SeverityScoreSummary as SeverityScoreSummary, index_SpeechModel as SpeechModel, index_SubstitutionPolicy as SubstitutionPolicy, index_SubtitleFormat as SubtitleFormat, index_SummaryModel as SummaryModel, index_SummaryType as SummaryType, type index_Timestamp as Timestamp, type index_TooManyRequestsResponse as TooManyRequestsResponse, type index_TopicDetectionModelResult as TopicDetectionModelResult, type index_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, type index_TopicDetectionResult as TopicDetectionResult, type index_TopicDetectionResultLabelsItem as TopicDetectionResultLabelsItem, type index_Transcript as Transcript, type index_TranscriptAudioDuration as TranscriptAudioDuration, type index_TranscriptAudioEndAt as TranscriptAudioEndAt, type index_TranscriptAudioStartFrom as TranscriptAudioStartFrom, type index_TranscriptAutoChapters as TranscriptAutoChapters, type index_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index_TranscriptBoostParam as TranscriptBoostParam, type index_TranscriptBoostParamProperty as TranscriptBoostParamProperty, type index_TranscriptChapters as TranscriptChapters, type index_TranscriptConfidence as TranscriptConfidence, type index_TranscriptContentSafety as TranscriptContentSafety, type index_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, type index_TranscriptCustomSpelling as TranscriptCustomSpelling, type index_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, type index_TranscriptCustomTopics as TranscriptCustomTopics, type index_TranscriptDisfluencies as TranscriptDisfluencies, type index_TranscriptEntities as TranscriptEntities, type index_TranscriptEntityDetection as TranscriptEntityDetection, type index_TranscriptFilterProfanity as TranscriptFilterProfanity, type index_TranscriptFormatText as TranscriptFormatText, type index_TranscriptIabCategories as TranscriptIabCategories, type index_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index_TranscriptLanguageCode as TranscriptLanguageCode, type index_TranscriptLanguageCodeProperty as TranscriptLanguageCodeProperty, type index_TranscriptLanguageConfidence as TranscriptLanguageConfidence, type index_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, type index_TranscriptLanguageDetection as TranscriptLanguageDetection, type index_TranscriptList as TranscriptList, type index_TranscriptListItem as TranscriptListItem, type index_TranscriptListItemCompleted as TranscriptListItemCompleted, type index_TranscriptListItemError as TranscriptListItemError, type index_TranscriptMultichannel as TranscriptMultichannel, type index_TranscriptOptionalParams as TranscriptOptionalParams, type index_TranscriptOptionalParamsLanguageCode as TranscriptOptionalParamsLanguageCode, type index_TranscriptOptionalParamsLanguageCodeOneOf as TranscriptOptionalParamsLanguageCodeOneOf, type index_TranscriptOptionalParamsRedactPiiSub as TranscriptOptionalParamsRedactPiiSub, type index_TranscriptOptionalParamsSpeakersExpected as TranscriptOptionalParamsSpeakersExpected, type index_TranscriptOptionalParamsSpeechModel as TranscriptOptionalParamsSpeechModel, type index_TranscriptOptionalParamsSpeechThreshold as TranscriptOptionalParamsSpeechThreshold, type index_TranscriptOptionalParamsWebhookAuthHeaderName as TranscriptOptionalParamsWebhookAuthHeaderName, type index_TranscriptOptionalParamsWebhookAuthHeaderValue as TranscriptOptionalParamsWebhookAuthHeaderValue, type index_TranscriptParagraph as TranscriptParagraph, type index_TranscriptParams as TranscriptParams, type index_TranscriptParamsAllOf as TranscriptParamsAllOf, type index_TranscriptPunctuate as TranscriptPunctuate, type index_TranscriptReadyNotification as TranscriptReadyNotification, index_TranscriptReadyStatus as TranscriptReadyStatus, type index_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, type index_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, type index_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, type index_TranscriptSentence as TranscriptSentence, type index_TranscriptSentenceChannel as TranscriptSentenceChannel, type index_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, type index_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, type index_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, type index_TranscriptSpeakerLabels as TranscriptSpeakerLabels, type index_TranscriptSpeakersExpected as TranscriptSpeakersExpected, type index_TranscriptSpeechModel as TranscriptSpeechModel, type index_TranscriptSpeechThreshold as TranscriptSpeechThreshold, type index_TranscriptSpeedBoost as TranscriptSpeedBoost, index_TranscriptStatus as TranscriptStatus, type index_TranscriptSummary as TranscriptSummary, type index_TranscriptSummaryModel as TranscriptSummaryModel, type index_TranscriptSummaryType as TranscriptSummaryType, type index_TranscriptText as TranscriptText, type index_TranscriptThrottled as TranscriptThrottled, type index_TranscriptUtterance as TranscriptUtterance, type index_TranscriptUtteranceChannel as TranscriptUtteranceChannel, type index_TranscriptUtterances as TranscriptUtterances, type index_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, type index_TranscriptWebhookNotification as TranscriptWebhookNotification, type index_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, type index_TranscriptWebhookUrl as TranscriptWebhookUrl, type index_TranscriptWord as TranscriptWord, type index_TranscriptWordChannel as TranscriptWordChannel, type index_TranscriptWordSpeaker as TranscriptWordSpeaker, type index_TranscriptWords as TranscriptWords, type index_UnauthorizedResponse as UnauthorizedResponse, type index_UploadedFile as UploadedFile, type index_WordSearchMatch as WordSearchMatch, type index_WordSearchParams as WordSearchParams, type index_WordSearchResponse as WordSearchResponse, type index_WordSearchTimestamp as WordSearchTimestamp };
13940
15052
  }
13941
15053
 
13942
- export { AssemblyAIAdapter, type Chapter as AssemblyAIChapter, type ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, AssemblyAIEncodingType, type Entity as AssemblyAIEntity, type AssemblyAIExtendedData, type AutoHighlightsResult as AssemblyAIHighlightsResult, type TranscriptOptionalParams as AssemblyAIOptions, AssemblyAISampleRateType, type SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAISpeechModelType, type AssemblyAIStreamingOptions, type TopicDetectionModelResult as AssemblyAITopicsResult, index as AssemblyAITypes, type AssemblyAIUpdateConfiguration, AssemblyAIWebhookHandler, type TranscriptWebhookNotification as AssemblyAIWebhookPayload, type AudioAckEvent, type AudioChunk, type AudioInput, AudioResponseFormat, AudioTranscriptionModel, AzureSTTAdapter, AzureWebhookHandler, BaseAdapter, BaseWebhookHandler, type BatchOnlyProvider, type ChapterizationEvent, DeepgramAdapter, ListenV1EncodingParameter as DeepgramEncoding, type DeepgramExtendedData, type ListenV1ResponseMetadata as DeepgramMetadata, DeepgramModelType, type ListenV1MediaTranscribeParams as DeepgramOptions, ListenV1RedactParameterOneOfItem as DeepgramRedact, ListenV1RedactParameterOneOfItem as DeepgramRedactType, type DeepgramStreamingOptions, SharedCustomTopicModeParameter as DeepgramTopicMode, SharedCustomTopicModeParameter as DeepgramTopicModeType, DeepgramWebhookHandler, type ListenV1Response as DeepgramWebhookPayload, type EntityEvent, GladiaAdapter, type AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, type AudioToLlmListDTO as GladiaAudioToLlmResult, StreamingSupportedBitDepthEnum as GladiaBitDepth, type ChapterizationDTO as GladiaChapters, type CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, StreamingSupportedEncodingEnum as GladiaEncoding, type NamedEntityRecognitionDTO as GladiaEntities, type GladiaExtendedData, TranscriptionLanguageCodeEnum as GladiaLanguage, StreamingSupportedModels as GladiaModel, type ModerationDTO as GladiaModeration, type InitTranscriptionRequest as GladiaOptions, StreamingSupportedSampleRateEnum as GladiaSampleRate, type SentimentAnalysisDTO as GladiaSentiment, type SpeakerReidentificationDTO as GladiaSpeakerReidentification, type GladiaStreamingOptions, type StreamingRequest as GladiaStreamingRequest, type StructuredDataExtractionDTO as GladiaStructuredData, type TranslationDTO as GladiaTranslation, TranslationLanguageCodeEnum as GladiaTranslationLanguage, index$1 as GladiaTypes, type CallbackTranscriptionErrorPayload as GladiaWebhookErrorPayload, GladiaWebhookHandler, type GladiaWebhookPayload, type CallbackTranscriptionSuccessPayload as GladiaWebhookSuccessPayload, type LifecycleEvent, ListenV1EncodingParameter, type ListenV1LanguageParameter, type ListenV1ModelParameter, type ListenV1VersionParameter, OpenAIWhisperAdapter, type CreateTranscriptionRequest as OpenAIWhisperOptions, type ProviderCapabilities, type ProviderConfig, type ProviderExtendedDataMap, type ProviderRawResponseMap, type ProviderStreamingOptions, type ProviderWebhookPayloadMap, type SentimentEvent, type SessionStatus, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type Speaker, type SpeechEvent, SpeechmaticsAdapter, type SpeechmaticsOperatingPoint, SpeechmaticsWebhookHandler, type StreamEvent, type StreamEventType, type StreamingCallbacks, type StreamingOptions, type StreamingOptionsForProvider, type StreamingProvider, type StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedSampleRateEnum, type SummarizationEvent, type TranscribeOptions, type TranscribeStreamParams, type TranscriptionAdapter, type TranscriptionLanguage, type TranscriptionModel, type TranscriptionProvider, type TranscriptionStatus, type TranslationEvent, type UnifiedTranscriptResponse, type UnifiedWebhookEvent, type Utterance, VoiceRouter, type VoiceRouterConfig, type WebhookEventType, WebhookRouter, type WebhookRouterOptions, type WebhookRouterResult, type WebhookValidation, type WebhookVerificationOptions, type Word, createAssemblyAIAdapter, createAssemblyAIWebhookHandler, createAzureSTTAdapter, createAzureWebhookHandler, createDeepgramAdapter, createDeepgramWebhookHandler, createGladiaAdapter, createGladiaWebhookHandler, createOpenAIWhisperAdapter, createSpeechmaticsAdapter, createVoiceRouter, createWebhookRouter };
15054
+ export { AssemblyAIAdapter, type Chapter as AssemblyAIChapter, type ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, AssemblyAIEncodingType, type Entity as AssemblyAIEntity, type AssemblyAIExtendedData, type AutoHighlightsResult as AssemblyAIHighlightsResult, type TranscriptOptionalParams as AssemblyAIOptions, AssemblyAISampleRateType, type SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAISpeechModelType, type AssemblyAIStreamingOptions, type TopicDetectionModelResult as AssemblyAITopicsResult, index as AssemblyAITypes, type AssemblyAIUpdateConfiguration, AssemblyAIWebhookHandler, type TranscriptWebhookNotification as AssemblyAIWebhookPayload, type AudioAckEvent, type AudioChunk, type AudioInput, AudioResponseFormat, AudioTranscriptionModel, AzureSTTAdapter, AzureWebhookHandler, BaseAdapter, BaseWebhookHandler, type BatchOnlyProvider, type ChapterizationEvent, DeepgramAdapter, type DeepgramExtendedData, type ListenV1ResponseMetadata as DeepgramMetadata, DeepgramModelType, type ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramRedactType, type DeepgramStreamingOptions, DeepgramTopicModeType, DeepgramWebhookHandler, type ListenV1Response as DeepgramWebhookPayload, type EntityEvent, GladiaAdapter, type AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, type AudioToLlmListDTO as GladiaAudioToLlmResult, type ChapterizationDTO as GladiaChapters, type CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, type NamedEntityRecognitionDTO as GladiaEntities, type GladiaExtendedData, type ModerationDTO as GladiaModeration, type InitTranscriptionRequest as GladiaOptions, type SentimentAnalysisDTO as GladiaSentiment, type SpeakerReidentificationDTO as GladiaSpeakerReidentification, type GladiaStreamingOptions, type StreamingRequest as GladiaStreamingRequest, type StructuredDataExtractionDTO as GladiaStructuredData, type TranslationDTO as GladiaTranslation, index$1 as GladiaTypes, type CallbackTranscriptionErrorPayload as GladiaWebhookErrorPayload, GladiaWebhookHandler, type GladiaWebhookPayload, type CallbackTranscriptionSuccessPayload as GladiaWebhookSuccessPayload, type LifecycleEvent, type ListTranscriptsOptions, type ListTranscriptsResponse, ListenV1EncodingParameter, type ListenV1LanguageParameter, type ListenV1ModelParameter, type ListenV1VersionParameter, OpenAIWhisperAdapter, type CreateTranscriptionRequest as OpenAIWhisperOptions, type ProviderCapabilities, type ProviderConfig, type ProviderExtendedDataMap, type ProviderRawResponseMap, type ProviderStreamingOptions, type ProviderWebhookPayloadMap, type SentimentEvent, type SessionStatus, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type Speaker, type SpeechEvent, SpeechmaticsAdapter, type SpeechmaticsOperatingPoint, SpeechmaticsWebhookHandler, type StreamEvent, type StreamEventType, type StreamingCallbacks, type StreamingOptions, type StreamingOptionsForProvider, type StreamingProvider, type StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedSampleRateEnum, type SummarizationEvent, type TranscribeOptions, type TranscribeStreamParams, type TranscriptData, type TranscriptMetadata, type TranscriptionAdapter, type TranscriptionLanguage, type TranscriptionModel, type TranscriptionProvider, type TranscriptionStatus, type TranslationEvent, type UnifiedTranscriptResponse, type UnifiedWebhookEvent, type Utterance, VoiceRouter, type VoiceRouterConfig, type WebhookEventType, WebhookRouter, type WebhookRouterOptions, type WebhookRouterResult, type WebhookValidation, type WebhookVerificationOptions, type Word, createAssemblyAIAdapter, createAssemblyAIWebhookHandler, createAzureSTTAdapter, createAzureWebhookHandler, createDeepgramAdapter, createDeepgramWebhookHandler, createGladiaAdapter, createGladiaWebhookHandler, createOpenAIWhisperAdapter, createSpeechmaticsAdapter, createVoiceRouter, createWebhookRouter };