voice-router-dev 0.3.1 → 0.3.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.mts CHANGED
@@ -1,5 +1,5 @@
1
- import { GladiaLanguage as TranscriptionLanguageCodeEnum, GladiaTranslationLanguage as TranslationLanguageCodeEnum, DeepgramRedact as ListenV1RedactParameterOneOfItem, DeepgramTopicMode as SharedCustomTopicModeParameter, DeepgramEncoding as ListenV1EncodingParameter, GladiaEncoding as StreamingSupportedEncodingEnum, GladiaBitDepth as StreamingSupportedBitDepthEnum, GladiaSampleRate as StreamingSupportedSampleRateEnum, GladiaModel as StreamingSupportedModels, DeepgramModelType, AssemblyAISampleRateType, AssemblyAIEncodingType, AssemblyAISpeechModelType } from './constants.mjs';
2
- export { AssemblyAIEncoding, AssemblyAISampleRate, AssemblyAISpeechModel, DeepgramModel } from './constants.mjs';
1
+ import { DeepgramModelType, DeepgramTopicModeType, DeepgramRedactType, AssemblyAISampleRateType, AssemblyAIEncodingType, AssemblyAISpeechModelType } from './constants.mjs';
2
+ export { AssemblyAIEncoding, AssemblyAISampleRate, AssemblyAISpeechModel, DeepgramEncoding, DeepgramModel, DeepgramRedact, DeepgramTopicMode, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaSampleRate, GladiaTranslationLanguage } from './constants.mjs';
3
3
 
4
4
  /**
5
5
  * Unified audio encoding types for Voice Router SDK
@@ -700,6 +700,118 @@ interface CallbackConfigDto {
700
700
  method?: CallbackMethodEnum;
701
701
  }
702
702
 
703
+ /**
704
+ * Generated by orval v7.9.0 🍺
705
+ * Do not edit manually.
706
+ * Gladia Control API
707
+ * OpenAPI spec version: 1.0
708
+ */
709
+ /**
710
+ * Specify the language in which it will be pronounced when sound comparison occurs. Default to transcription language.
711
+ */
712
+ type TranscriptionLanguageCodeEnum = (typeof TranscriptionLanguageCodeEnum)[keyof typeof TranscriptionLanguageCodeEnum];
713
+ declare const TranscriptionLanguageCodeEnum: {
714
+ readonly af: "af";
715
+ readonly am: "am";
716
+ readonly ar: "ar";
717
+ readonly as: "as";
718
+ readonly az: "az";
719
+ readonly ba: "ba";
720
+ readonly be: "be";
721
+ readonly bg: "bg";
722
+ readonly bn: "bn";
723
+ readonly bo: "bo";
724
+ readonly br: "br";
725
+ readonly bs: "bs";
726
+ readonly ca: "ca";
727
+ readonly cs: "cs";
728
+ readonly cy: "cy";
729
+ readonly da: "da";
730
+ readonly de: "de";
731
+ readonly el: "el";
732
+ readonly en: "en";
733
+ readonly es: "es";
734
+ readonly et: "et";
735
+ readonly eu: "eu";
736
+ readonly fa: "fa";
737
+ readonly fi: "fi";
738
+ readonly fo: "fo";
739
+ readonly fr: "fr";
740
+ readonly gl: "gl";
741
+ readonly gu: "gu";
742
+ readonly ha: "ha";
743
+ readonly haw: "haw";
744
+ readonly he: "he";
745
+ readonly hi: "hi";
746
+ readonly hr: "hr";
747
+ readonly ht: "ht";
748
+ readonly hu: "hu";
749
+ readonly hy: "hy";
750
+ readonly id: "id";
751
+ readonly is: "is";
752
+ readonly it: "it";
753
+ readonly ja: "ja";
754
+ readonly jw: "jw";
755
+ readonly ka: "ka";
756
+ readonly kk: "kk";
757
+ readonly km: "km";
758
+ readonly kn: "kn";
759
+ readonly ko: "ko";
760
+ readonly la: "la";
761
+ readonly lb: "lb";
762
+ readonly ln: "ln";
763
+ readonly lo: "lo";
764
+ readonly lt: "lt";
765
+ readonly lv: "lv";
766
+ readonly mg: "mg";
767
+ readonly mi: "mi";
768
+ readonly mk: "mk";
769
+ readonly ml: "ml";
770
+ readonly mn: "mn";
771
+ readonly mr: "mr";
772
+ readonly ms: "ms";
773
+ readonly mt: "mt";
774
+ readonly my: "my";
775
+ readonly ne: "ne";
776
+ readonly nl: "nl";
777
+ readonly nn: "nn";
778
+ readonly no: "no";
779
+ readonly oc: "oc";
780
+ readonly pa: "pa";
781
+ readonly pl: "pl";
782
+ readonly ps: "ps";
783
+ readonly pt: "pt";
784
+ readonly ro: "ro";
785
+ readonly ru: "ru";
786
+ readonly sa: "sa";
787
+ readonly sd: "sd";
788
+ readonly si: "si";
789
+ readonly sk: "sk";
790
+ readonly sl: "sl";
791
+ readonly sn: "sn";
792
+ readonly so: "so";
793
+ readonly sq: "sq";
794
+ readonly sr: "sr";
795
+ readonly su: "su";
796
+ readonly sv: "sv";
797
+ readonly sw: "sw";
798
+ readonly ta: "ta";
799
+ readonly te: "te";
800
+ readonly tg: "tg";
801
+ readonly th: "th";
802
+ readonly tk: "tk";
803
+ readonly tl: "tl";
804
+ readonly tr: "tr";
805
+ readonly tt: "tt";
806
+ readonly uk: "uk";
807
+ readonly ur: "ur";
808
+ readonly uz: "uz";
809
+ readonly vi: "vi";
810
+ readonly yi: "yi";
811
+ readonly yo: "yo";
812
+ readonly zh: "zh";
813
+ };
814
+
703
815
  /**
704
816
  * Generated by orval v7.9.0 🍺
705
817
  * Do not edit manually.
@@ -935,6 +1047,119 @@ interface SummarizationConfigDTO {
935
1047
  type?: SummaryTypesEnum;
936
1048
  }
937
1049
 
1050
+ /**
1051
+ * Generated by orval v7.9.0 🍺
1052
+ * Do not edit manually.
1053
+ * Gladia Control API
1054
+ * OpenAPI spec version: 1.0
1055
+ */
1056
+ /**
1057
+ * Target language in `iso639-1` format you want the transcription translated to
1058
+ */
1059
+ type TranslationLanguageCodeEnum = (typeof TranslationLanguageCodeEnum)[keyof typeof TranslationLanguageCodeEnum];
1060
+ declare const TranslationLanguageCodeEnum: {
1061
+ readonly af: "af";
1062
+ readonly am: "am";
1063
+ readonly ar: "ar";
1064
+ readonly as: "as";
1065
+ readonly az: "az";
1066
+ readonly ba: "ba";
1067
+ readonly be: "be";
1068
+ readonly bg: "bg";
1069
+ readonly bn: "bn";
1070
+ readonly bo: "bo";
1071
+ readonly br: "br";
1072
+ readonly bs: "bs";
1073
+ readonly ca: "ca";
1074
+ readonly cs: "cs";
1075
+ readonly cy: "cy";
1076
+ readonly da: "da";
1077
+ readonly de: "de";
1078
+ readonly el: "el";
1079
+ readonly en: "en";
1080
+ readonly es: "es";
1081
+ readonly et: "et";
1082
+ readonly eu: "eu";
1083
+ readonly fa: "fa";
1084
+ readonly fi: "fi";
1085
+ readonly fo: "fo";
1086
+ readonly fr: "fr";
1087
+ readonly gl: "gl";
1088
+ readonly gu: "gu";
1089
+ readonly ha: "ha";
1090
+ readonly haw: "haw";
1091
+ readonly he: "he";
1092
+ readonly hi: "hi";
1093
+ readonly hr: "hr";
1094
+ readonly ht: "ht";
1095
+ readonly hu: "hu";
1096
+ readonly hy: "hy";
1097
+ readonly id: "id";
1098
+ readonly is: "is";
1099
+ readonly it: "it";
1100
+ readonly ja: "ja";
1101
+ readonly jw: "jw";
1102
+ readonly ka: "ka";
1103
+ readonly kk: "kk";
1104
+ readonly km: "km";
1105
+ readonly kn: "kn";
1106
+ readonly ko: "ko";
1107
+ readonly la: "la";
1108
+ readonly lb: "lb";
1109
+ readonly ln: "ln";
1110
+ readonly lo: "lo";
1111
+ readonly lt: "lt";
1112
+ readonly lv: "lv";
1113
+ readonly mg: "mg";
1114
+ readonly mi: "mi";
1115
+ readonly mk: "mk";
1116
+ readonly ml: "ml";
1117
+ readonly mn: "mn";
1118
+ readonly mr: "mr";
1119
+ readonly ms: "ms";
1120
+ readonly mt: "mt";
1121
+ readonly my: "my";
1122
+ readonly ne: "ne";
1123
+ readonly nl: "nl";
1124
+ readonly nn: "nn";
1125
+ readonly no: "no";
1126
+ readonly oc: "oc";
1127
+ readonly pa: "pa";
1128
+ readonly pl: "pl";
1129
+ readonly ps: "ps";
1130
+ readonly pt: "pt";
1131
+ readonly ro: "ro";
1132
+ readonly ru: "ru";
1133
+ readonly sa: "sa";
1134
+ readonly sd: "sd";
1135
+ readonly si: "si";
1136
+ readonly sk: "sk";
1137
+ readonly sl: "sl";
1138
+ readonly sn: "sn";
1139
+ readonly so: "so";
1140
+ readonly sq: "sq";
1141
+ readonly sr: "sr";
1142
+ readonly su: "su";
1143
+ readonly sv: "sv";
1144
+ readonly sw: "sw";
1145
+ readonly ta: "ta";
1146
+ readonly te: "te";
1147
+ readonly tg: "tg";
1148
+ readonly th: "th";
1149
+ readonly tk: "tk";
1150
+ readonly tl: "tl";
1151
+ readonly tr: "tr";
1152
+ readonly tt: "tt";
1153
+ readonly uk: "uk";
1154
+ readonly ur: "ur";
1155
+ readonly uz: "uz";
1156
+ readonly vi: "vi";
1157
+ readonly wo: "wo";
1158
+ readonly yi: "yi";
1159
+ readonly yo: "yo";
1160
+ readonly zh: "zh";
1161
+ };
1162
+
938
1163
  /**
939
1164
  * Generated by orval v7.9.0 🍺
940
1165
  * Do not edit manually.
@@ -4130,12 +4355,16 @@ interface Transcription {
4130
4355
  /**
4131
4356
  * Generated by orval v7.9.0 🍺
4132
4357
  * Do not edit manually.
4133
- * Deepgram API Specification
4134
- * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4135
-
4136
- * OpenAPI spec version: 1.0.0
4358
+ * Gladia Control API
4359
+ * OpenAPI spec version: 1.0
4137
4360
  */
4138
- type ListenV1ModelParameter = "nova-3" | "nova-3-general" | "nova-3-medical" | "nova-2" | "nova-2-general" | "nova-2-meeting" | "nova-2-finance" | "nova-2-conversationalai" | "nova-2-voicemail" | "nova-2-video" | "nova-2-medical" | "nova-2-drivethru" | "nova-2-automotive" | "nova" | "nova-general" | "nova-phonecall" | "nova-medical" | "enhanced" | "enhanced-general" | "enhanced-meeting" | "enhanced-phonecall" | "enhanced-finance" | "base" | "meeting" | "phonecall" | "finance" | "conversationalai" | "voicemail" | "video" | string;
4361
+ /**
4362
+ * The model used to process the audio. "solaria-1" is used by default.
4363
+ */
4364
+ type StreamingSupportedModels = (typeof StreamingSupportedModels)[keyof typeof StreamingSupportedModels];
4365
+ declare const StreamingSupportedModels: {
4366
+ readonly "solaria-1": "solaria-1";
4367
+ };
4139
4368
 
4140
4369
  /**
4141
4370
  * Generated by orval v7.9.0 🍺
@@ -4254,6 +4483,44 @@ type SharedTopicsParameter = boolean;
4254
4483
  */
4255
4484
  type SharedCustomTopicParameter = string | string[];
4256
4485
 
4486
+ /**
4487
+ * Generated by orval v7.9.0 🍺
4488
+ * Do not edit manually.
4489
+ * Deepgram API Specification
4490
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4491
+
4492
+ * OpenAPI spec version: 1.0.0
4493
+ */
4494
+ /**
4495
+ * SharedCustomTopicModeParameter type definition
4496
+ */
4497
+ /**
4498
+ * SharedCustomTopicModeParameter type definition
4499
+ */
4500
+ /**
4501
+ * SharedCustomTopicModeParameter type definition
4502
+ */
4503
+ /**
4504
+ * SharedCustomTopicModeParameter type definition
4505
+ */
4506
+ /**
4507
+ * SharedCustomTopicModeParameter type definition
4508
+ */
4509
+ /**
4510
+ * SharedCustomTopicModeParameter type definition
4511
+ */
4512
+ /**
4513
+ * SharedCustomTopicModeParameter type definition
4514
+ */
4515
+ /**
4516
+ * SharedCustomTopicModeParameter type definition
4517
+ */
4518
+ type SharedCustomTopicModeParameter = typeof SharedCustomTopicModeParameter[keyof typeof SharedCustomTopicModeParameter];
4519
+ declare const SharedCustomTopicModeParameter: {
4520
+ readonly extended: "extended";
4521
+ readonly strict: "strict";
4522
+ };
4523
+
4257
4524
  /**
4258
4525
  * Generated by orval v7.9.0 🍺
4259
4526
  * Do not edit manually.
@@ -4364,6 +4631,48 @@ type ListenV1DiarizeParameter = boolean;
4364
4631
  */
4365
4632
  type ListenV1DictationParameter = boolean;
4366
4633
 
4634
+ /**
4635
+ * Generated by orval v7.9.0 🍺
4636
+ * Do not edit manually.
4637
+ * Deepgram API Specification
4638
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4639
+
4640
+ * OpenAPI spec version: 1.0.0
4641
+ */
4642
+ /**
4643
+ * ListenV1EncodingParameter type definition
4644
+ */
4645
+ /**
4646
+ * ListenV1EncodingParameter type definition
4647
+ */
4648
+ /**
4649
+ * ListenV1EncodingParameter type definition
4650
+ */
4651
+ /**
4652
+ * ListenV1EncodingParameter type definition
4653
+ */
4654
+ /**
4655
+ * ListenV1EncodingParameter type definition
4656
+ */
4657
+ /**
4658
+ * ListenV1EncodingParameter type definition
4659
+ */
4660
+ /**
4661
+ * ListenV1EncodingParameter type definition
4662
+ */
4663
+ /**
4664
+ * ListenV1EncodingParameter type definition
4665
+ */
4666
+ type ListenV1EncodingParameter = typeof ListenV1EncodingParameter[keyof typeof ListenV1EncodingParameter];
4667
+ declare const ListenV1EncodingParameter: {
4668
+ readonly linear16: "linear16";
4669
+ readonly flac: "flac";
4670
+ readonly mulaw: "mulaw";
4671
+ readonly opus: "opus";
4672
+ readonly speex: "speex";
4673
+ readonly g729: "g729";
4674
+ };
4675
+
4367
4676
  /**
4368
4677
  * Generated by orval v7.9.0 🍺
4369
4678
  * Do not edit manually.
@@ -4426,6 +4735,16 @@ type ListenV1LanguageParameter = string;
4426
4735
  */
4427
4736
  type ListenV1MeasurementsParameter = boolean;
4428
4737
 
4738
+ /**
4739
+ * Generated by orval v7.9.0 🍺
4740
+ * Do not edit manually.
4741
+ * Deepgram API Specification
4742
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4743
+
4744
+ * OpenAPI spec version: 1.0.0
4745
+ */
4746
+ type ListenV1ModelParameter = "nova-3" | "nova-3-general" | "nova-3-medical" | "nova-2" | "nova-2-general" | "nova-2-meeting" | "nova-2-finance" | "nova-2-conversationalai" | "nova-2-voicemail" | "nova-2-video" | "nova-2-medical" | "nova-2-drivethru" | "nova-2-automotive" | "nova" | "nova-general" | "nova-phonecall" | "nova-medical" | "enhanced" | "enhanced-general" | "enhanced-meeting" | "enhanced-phonecall" | "enhanced-finance" | "base" | "meeting" | "phonecall" | "finance" | "conversationalai" | "voicemail" | "video" | string;
4747
+
4429
4748
  /**
4430
4749
  * Generated by orval v7.9.0 🍺
4431
4750
  * Do not edit manually.
@@ -4491,6 +4810,21 @@ type ListenV1ProfanityFilterParameter = boolean;
4491
4810
  */
4492
4811
  type ListenV1PunctuateParameter = boolean;
4493
4812
 
4813
+ /**
4814
+ * Generated by orval v7.9.0 🍺
4815
+ * Do not edit manually.
4816
+ * Deepgram API Specification
4817
+ * APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
4818
+
4819
+ * OpenAPI spec version: 1.0.0
4820
+ */
4821
+ type ListenV1RedactParameterOneOfItem = (typeof ListenV1RedactParameterOneOfItem)[keyof typeof ListenV1RedactParameterOneOfItem];
4822
+ declare const ListenV1RedactParameterOneOfItem: {
4823
+ readonly pci: "pci";
4824
+ readonly pii: "pii";
4825
+ readonly numbers: "numbers";
4826
+ };
4827
+
4494
4828
  /**
4495
4829
  * Generated by orval v7.9.0 🍺
4496
4830
  * Do not edit manually.
@@ -5445,6 +5779,62 @@ type StreamingRequestCustomMetadata = {
5445
5779
  [key: string]: unknown;
5446
5780
  };
5447
5781
 
5782
+ /**
5783
+ * Generated by orval v7.9.0 🍺
5784
+ * Do not edit manually.
5785
+ * Gladia Control API
5786
+ * OpenAPI spec version: 1.0
5787
+ */
5788
+ /**
5789
+ * The bit depth of the audio stream
5790
+ */
5791
+ type StreamingSupportedBitDepthEnum = (typeof StreamingSupportedBitDepthEnum)[keyof typeof StreamingSupportedBitDepthEnum];
5792
+ declare const StreamingSupportedBitDepthEnum: {
5793
+ readonly NUMBER_8: 8;
5794
+ readonly NUMBER_16: 16;
5795
+ readonly NUMBER_24: 24;
5796
+ readonly NUMBER_32: 32;
5797
+ };
5798
+
5799
+ /**
5800
+ * Generated by orval v7.9.0 🍺
5801
+ * Do not edit manually.
5802
+ * Gladia Control API
5803
+ * OpenAPI spec version: 1.0
5804
+ */
5805
+ /**
5806
+ * The encoding format of the audio stream. Supported formats:
5807
+ - PCM: 8, 16, 24, and 32 bits
5808
+ - A-law: 8 bits
5809
+ - μ-law: 8 bits
5810
+
5811
+ Note: No need to add WAV headers to raw audio as the API supports both formats.
5812
+ */
5813
+ type StreamingSupportedEncodingEnum = (typeof StreamingSupportedEncodingEnum)[keyof typeof StreamingSupportedEncodingEnum];
5814
+ declare const StreamingSupportedEncodingEnum: {
5815
+ readonly "wav/pcm": "wav/pcm";
5816
+ readonly "wav/alaw": "wav/alaw";
5817
+ readonly "wav/ulaw": "wav/ulaw";
5818
+ };
5819
+
5820
+ /**
5821
+ * Generated by orval v7.9.0 🍺
5822
+ * Do not edit manually.
5823
+ * Gladia Control API
5824
+ * OpenAPI spec version: 1.0
5825
+ */
5826
+ /**
5827
+ * The sample rate of the audio stream
5828
+ */
5829
+ type StreamingSupportedSampleRateEnum = (typeof StreamingSupportedSampleRateEnum)[keyof typeof StreamingSupportedSampleRateEnum];
5830
+ declare const StreamingSupportedSampleRateEnum: {
5831
+ readonly NUMBER_8000: 8000;
5832
+ readonly NUMBER_16000: 16000;
5833
+ readonly NUMBER_32000: 32000;
5834
+ readonly NUMBER_44100: 44100;
5835
+ readonly NUMBER_48000: 48000;
5836
+ };
5837
+
5448
5838
  /**
5449
5839
  * Generated by orval v7.9.0 🍺
5450
5840
  * Do not edit manually.
@@ -5502,7 +5892,6 @@ interface StreamingRequest {
5502
5892
  callback_config?: CallbackConfig;
5503
5893
  }
5504
5894
 
5505
- type StreamingSpeechModel = "universal-streaming-english" | "universal-streaming-multilingual";
5506
5895
  type StreamingUpdateConfiguration = {
5507
5896
  type: "UpdateConfiguration";
5508
5897
  end_of_turn_confidence_threshold?: number;
@@ -5574,15 +5963,19 @@ interface DeepgramStreamingOptions {
5574
5963
  language?: ListenV1LanguageParameter;
5575
5964
  /**
5576
5965
  * Model to use for transcription
5966
+ *
5967
+ * Strict union type - only accepts valid Deepgram models.
5577
5968
  * Use `DeepgramModel` const for autocomplete:
5578
5969
  * @example
5579
5970
  * ```typescript
5580
- * import { DeepgramModel } from '@meeting-baas/sdk'
5581
- * { model: DeepgramModel.nova3 }
5582
- * { model: DeepgramModel.nova2Medical }
5971
+ * import { DeepgramModel } from 'voice-router-dev'
5972
+ * { model: DeepgramModel["nova-3"] }
5973
+ * { model: DeepgramModel["nova-2-medical"] }
5974
+ * // Or use string literals directly:
5975
+ * { model: "nova-3" }
5583
5976
  * ```
5584
5977
  */
5585
- model?: ListenV1ModelParameter | DeepgramModelType;
5978
+ model?: DeepgramModelType;
5586
5979
  /** Model version (e.g., 'latest') */
5587
5980
  version?: ListenV1VersionParameter;
5588
5981
  /** Enable language detection */
@@ -5621,7 +6014,7 @@ interface DeepgramStreamingOptions {
5621
6014
  * Custom topic detection mode
5622
6015
  * Use `DeepgramTopicMode` const for autocomplete
5623
6016
  */
5624
- customTopicMode?: SharedCustomTopicModeParameter;
6017
+ customTopicMode?: DeepgramTopicModeType;
5625
6018
  /** Enable intent recognition */
5626
6019
  intents?: boolean;
5627
6020
  /** Custom intent definitions */
@@ -5630,7 +6023,7 @@ interface DeepgramStreamingOptions {
5630
6023
  * Custom intent detection mode
5631
6024
  * Use `DeepgramTopicMode` const for autocomplete
5632
6025
  */
5633
- customIntentMode?: SharedCustomTopicModeParameter;
6026
+ customIntentMode?: DeepgramTopicModeType;
5634
6027
  /** Enable summarization */
5635
6028
  summarize?: boolean;
5636
6029
  /** Custom vocabulary/keywords for boosting */
@@ -5649,7 +6042,7 @@ interface DeepgramStreamingOptions {
5649
6042
  * { redact: [DeepgramRedact.pii, DeepgramRedact.pci] }
5650
6043
  * ```
5651
6044
  */
5652
- redact?: boolean | ListenV1RedactParameterOneOfItem[];
6045
+ redact?: boolean | DeepgramRedactType[];
5653
6046
  /** Callback URL for webhooks */
5654
6047
  callback?: string;
5655
6048
  /** Extra metadata to include in response */
@@ -5697,15 +6090,19 @@ interface AssemblyAIStreamingOptions {
5697
6090
  encoding?: AssemblyAIEncodingType;
5698
6091
  /**
5699
6092
  * Speech model to use
6093
+ *
6094
+ * Strict union type - only accepts valid AssemblyAI streaming models.
5700
6095
  * Use `AssemblyAISpeechModel` const for autocomplete:
5701
6096
  * @example
5702
6097
  * ```typescript
5703
- * import { AssemblyAISpeechModel } from '@meeting-baas/sdk'
6098
+ * import { AssemblyAISpeechModel } from 'voice-router-dev'
5704
6099
  * { speechModel: AssemblyAISpeechModel.english }
5705
6100
  * { speechModel: AssemblyAISpeechModel.multilingual }
6101
+ * // Or use string literals directly:
6102
+ * { speechModel: "universal-streaming-english" }
5706
6103
  * ```
5707
6104
  */
5708
- speechModel?: AssemblyAISpeechModelType | StreamingSpeechModel;
6105
+ speechModel?: AssemblyAISpeechModelType;
5709
6106
  /** Enable automatic language detection */
5710
6107
  languageDetection?: boolean;
5711
6108
  /**
@@ -5796,6 +6193,168 @@ interface TranscribeStreamParams<P extends StreamingProvider> {
5796
6193
  callbacks?: StreamingCallbacks;
5797
6194
  }
5798
6195
 
6196
+ /**
6197
+ * Generated by orval v7.9.0 🍺
6198
+ * Do not edit manually.
6199
+ * AssemblyAI API
6200
+ * AssemblyAI API
6201
+ * OpenAPI spec version: 1.3.4
6202
+ */
6203
+ /**
6204
+ * Get transcripts that were created after this transcript ID
6205
+ */
6206
+ type AfterId = string;
6207
+
6208
+ /**
6209
+ * Generated by orval v7.9.0 🍺
6210
+ * Do not edit manually.
6211
+ * AssemblyAI API
6212
+ * AssemblyAI API
6213
+ * OpenAPI spec version: 1.3.4
6214
+ */
6215
+ /**
6216
+ * Get transcripts that were created before this transcript ID
6217
+ */
6218
+ type BeforeId = string;
6219
+
6220
+ /**
6221
+ * Generated by orval v7.9.0 🍺
6222
+ * Do not edit manually.
6223
+ * AssemblyAI API
6224
+ * AssemblyAI API
6225
+ * OpenAPI spec version: 1.3.4
6226
+ */
6227
+ /**
6228
+ * Only get transcripts created on this date
6229
+ */
6230
+ type CreatedOn = string;
6231
+
6232
+ /**
6233
+ * Generated by orval v7.9.0 🍺
6234
+ * Do not edit manually.
6235
+ * AssemblyAI API
6236
+ * AssemblyAI API
6237
+ * OpenAPI spec version: 1.3.4
6238
+ */
6239
+ /**
6240
+ * Maximum amount of transcripts to retrieve
6241
+ */
6242
+ type Limit = number;
6243
+
6244
+ /**
6245
+ * Generated by orval v7.9.0 🍺
6246
+ * Do not edit manually.
6247
+ * AssemblyAI API
6248
+ * AssemblyAI API
6249
+ * OpenAPI spec version: 1.3.4
6250
+ */
6251
+ /**
6252
+ * Only get throttled transcripts, overrides the status filter
6253
+ */
6254
+ type ThrottledOnly = boolean;
6255
+
6256
+ /**
6257
+ * Generated by orval v7.9.0 🍺
6258
+ * Do not edit manually.
6259
+ * AssemblyAI API
6260
+ * AssemblyAI API
6261
+ * OpenAPI spec version: 1.3.4
6262
+ */
6263
+
6264
+ type ListTranscriptsParams = {
6265
+ /**
6266
+ * Maximum amount of transcripts to retrieve
6267
+ */
6268
+ limit?: Limit;
6269
+ /**
6270
+ * Filter by transcript status
6271
+ */
6272
+ status?: TranscriptStatus;
6273
+ /**
6274
+ * Only get transcripts created on this date
6275
+ */
6276
+ created_on?: CreatedOn;
6277
+ /**
6278
+ * Get transcripts that were created before this transcript ID
6279
+ */
6280
+ before_id?: BeforeId;
6281
+ /**
6282
+ * Get transcripts that were created after this transcript ID
6283
+ */
6284
+ after_id?: AfterId;
6285
+ /**
6286
+ * Only get throttled transcripts, overrides the status filter
6287
+ */
6288
+ throttled_only?: ThrottledOnly;
6289
+ };
6290
+
6291
+ /**
6292
+ * Generated by orval v7.9.0 🍺
6293
+ * Do not edit manually.
6294
+ * Gladia Control API
6295
+ * OpenAPI spec version: 1.0
6296
+ */
6297
+ type TranscriptionControllerListV2KindItem = (typeof TranscriptionControllerListV2KindItem)[keyof typeof TranscriptionControllerListV2KindItem];
6298
+ declare const TranscriptionControllerListV2KindItem: {
6299
+ readonly "pre-recorded": "pre-recorded";
6300
+ readonly live: "live";
6301
+ };
6302
+
6303
+ /**
6304
+ * Generated by orval v7.9.0 🍺
6305
+ * Do not edit manually.
6306
+ * Gladia Control API
6307
+ * OpenAPI spec version: 1.0
6308
+ */
6309
+ type TranscriptionControllerListV2StatusItem = (typeof TranscriptionControllerListV2StatusItem)[keyof typeof TranscriptionControllerListV2StatusItem];
6310
+ declare const TranscriptionControllerListV2StatusItem: {
6311
+ readonly queued: "queued";
6312
+ readonly processing: "processing";
6313
+ readonly done: "done";
6314
+ readonly error: "error";
6315
+ };
6316
+
6317
+ /**
6318
+ * Generated by orval v7.9.0 🍺
6319
+ * Do not edit manually.
6320
+ * Gladia Control API
6321
+ * OpenAPI spec version: 1.0
6322
+ */
6323
+
6324
+ type TranscriptionControllerListV2Params = {
6325
+ /**
6326
+ * The starting point for pagination. A value of 0 starts from the first item.
6327
+ */
6328
+ offset?: number;
6329
+ /**
6330
+ * The maximum number of items to return. Useful for pagination and controlling data payload size.
6331
+ */
6332
+ limit?: number;
6333
+ /**
6334
+ * Filter items relevant to a specific date in ISO format (YYYY-MM-DD).
6335
+ */
6336
+ date?: string;
6337
+ /**
6338
+ * Include items that occurred before the specified date in ISO format.
6339
+ */
6340
+ before_date?: string;
6341
+ /**
6342
+ * Filter for items after the specified date. Use with `before_date` for a range. Date in ISO format.
6343
+ */
6344
+ after_date?: string;
6345
+ /**
6346
+ * Filter the list based on item status. Accepts multiple values from the predefined list.
6347
+ */
6348
+ status?: TranscriptionControllerListV2StatusItem[];
6349
+ custom_metadata?: {
6350
+ [key: string]: unknown;
6351
+ };
6352
+ /**
6353
+ * Filter the list based on the item type. Supports multiple values from the predefined list.
6354
+ */
6355
+ kind?: TranscriptionControllerListV2KindItem[];
6356
+ };
6357
+
5799
6358
  /**
5800
6359
  * Unified types for the Voice Router SDK
5801
6360
  * These types provide a provider-agnostic interface for transcription services
@@ -5809,15 +6368,20 @@ type SpeechmaticsOperatingPoint = "standard" | "enhanced";
5809
6368
  /**
5810
6369
  * Unified transcription model type with autocomplete for all providers
5811
6370
  *
5812
- * Includes all known models from:
5813
- * - Deepgram: nova-3, nova-2, enhanced, base, whisper, etc.
6371
+ * Strict union type - only accepts valid models from each provider:
6372
+ * - Deepgram: nova-3, nova-2, enhanced, base, etc.
5814
6373
  * - AssemblyAI: best, slam-1, universal
5815
6374
  * - Gladia: solaria-1
5816
6375
  * - Speechmatics: standard, enhanced
5817
6376
  *
5818
- * Also accepts any string for future/custom models.
6377
+ * Use provider const objects for autocomplete:
6378
+ * @example
6379
+ * ```typescript
6380
+ * import { DeepgramModel } from 'voice-router-dev'
6381
+ * { model: DeepgramModel["nova-3"] }
6382
+ * ```
5819
6383
  */
5820
- type TranscriptionModel = ListenV1ModelParameter | StreamingSupportedModels | SpeechModel | SpeechmaticsOperatingPoint;
6384
+ type TranscriptionModel = DeepgramModelType | StreamingSupportedModels | SpeechModel | SpeechmaticsOperatingPoint;
5821
6385
  /**
5822
6386
  * Unified transcription language type with autocomplete for all providers
5823
6387
  *
@@ -5937,6 +6501,55 @@ interface ProviderCapabilities {
5937
6501
  entityDetection: boolean;
5938
6502
  /** PII redaction */
5939
6503
  piiRedaction: boolean;
6504
+ /** List/fetch previous transcriptions */
6505
+ listTranscripts: boolean;
6506
+ /** Delete existing transcriptions */
6507
+ deleteTranscript: boolean;
6508
+ }
6509
+
6510
+ /**
6511
+ * Options for listing transcripts with date/time filtering
6512
+ *
6513
+ * Providers support different filtering capabilities:
6514
+ * - AssemblyAI: status, created_on, before_id, after_id, throttled_only
6515
+ * - Gladia: status, date, before_date, after_date, custom_metadata
6516
+ * - Azure: status, skip, top, filter (OData)
6517
+ *
6518
+ * @example Filter by date
6519
+ * ```typescript
6520
+ * await adapter.listTranscripts({
6521
+ * date: '2026-01-07', // Exact date (ISO format)
6522
+ * status: 'completed',
6523
+ * limit: 50
6524
+ * })
6525
+ * ```
6526
+ *
6527
+ * @example Filter by date range
6528
+ * ```typescript
6529
+ * await adapter.listTranscripts({
6530
+ * afterDate: '2026-01-01',
6531
+ * beforeDate: '2026-01-31',
6532
+ * limit: 100
6533
+ * })
6534
+ * ```
6535
+ */
6536
+ interface ListTranscriptsOptions {
6537
+ /** Maximum number of transcripts to retrieve */
6538
+ limit?: number;
6539
+ /** Pagination offset (skip N results) */
6540
+ offset?: number;
6541
+ /** Filter by transcript status */
6542
+ status?: "queued" | "processing" | "completed" | "error" | string;
6543
+ /** Filter by exact date (ISO format: YYYY-MM-DD) */
6544
+ date?: string;
6545
+ /** Filter for transcripts created before this date (ISO format) */
6546
+ beforeDate?: string;
6547
+ /** Filter for transcripts created after this date (ISO format) */
6548
+ afterDate?: string;
6549
+ /** AssemblyAI-specific list options */
6550
+ assemblyai?: Partial<ListTranscriptsParams>;
6551
+ /** Gladia-specific list options */
6552
+ gladia?: Partial<TranscriptionControllerListV2Params>;
5940
6553
  }
5941
6554
  /**
5942
6555
  * Audio input for transcription
@@ -6671,14 +7284,22 @@ interface TranscriptionAdapter {
6671
7284
  success: boolean;
6672
7285
  }>;
6673
7286
  /**
6674
- * List recent transcriptions
6675
- * Not all providers support listing
7287
+ * List recent transcriptions with filtering
7288
+ *
7289
+ * Supports date/time filtering, status filtering, and pagination.
7290
+ * Not all providers support all options - use provider-specific passthrough for full control.
7291
+ *
7292
+ * @example Filter by date range
7293
+ * ```typescript
7294
+ * const { transcripts } = await adapter.listTranscripts({
7295
+ * afterDate: '2026-01-01',
7296
+ * beforeDate: '2026-01-31',
7297
+ * status: 'completed',
7298
+ * limit: 50
7299
+ * })
7300
+ * ```
6676
7301
  */
6677
- listTranscripts?(options?: {
6678
- limit?: number;
6679
- offset?: number;
6680
- status?: string;
6681
- }): Promise<{
7302
+ listTranscripts?(options?: ListTranscriptsOptions): Promise<{
6682
7303
  transcripts: UnifiedTranscriptResponse[];
6683
7304
  total?: number;
6684
7305
  hasMore?: boolean;
@@ -6974,14 +7595,22 @@ declare class VoiceRouter {
6974
7595
  success: boolean;
6975
7596
  }>;
6976
7597
  /**
6977
- * List recent transcriptions
6978
- * Not all providers support this operation
7598
+ * List recent transcriptions with filtering
7599
+ *
7600
+ * Supports date/time filtering, status filtering, and pagination.
7601
+ * Not all providers support this operation.
7602
+ *
7603
+ * @example Filter by date range
7604
+ * ```typescript
7605
+ * const { transcripts } = await router.listTranscripts('assemblyai', {
7606
+ * afterDate: '2026-01-01',
7607
+ * beforeDate: '2026-01-31',
7608
+ * status: 'completed',
7609
+ * limit: 50
7610
+ * })
7611
+ * ```
6979
7612
  */
6980
- listTranscripts(provider: TranscriptionProvider, options?: {
6981
- limit?: number;
6982
- offset?: number;
6983
- status?: string;
6984
- }): Promise<{
7613
+ listTranscripts(provider: TranscriptionProvider, options?: ListTranscriptsOptions): Promise<{
6985
7614
  transcripts: UnifiedTranscriptResponse[];
6986
7615
  total?: number;
6987
7616
  hasMore?: boolean;
@@ -7321,6 +7950,59 @@ declare class GladiaAdapter extends BaseAdapter {
7321
7950
  deleteTranscript(transcriptId: string, jobType?: "pre-recorded" | "streaming"): Promise<{
7322
7951
  success: boolean;
7323
7952
  }>;
7953
+ /**
7954
+ * List recent transcriptions with filtering
7955
+ *
7956
+ * Retrieves a list of transcription jobs (both pre-recorded and streaming)
7957
+ * with optional filtering by status, date, and custom metadata.
7958
+ *
7959
+ * @param options - Filtering and pagination options
7960
+ * @param options.limit - Maximum number of transcripts to return
7961
+ * @param options.offset - Pagination offset (skip N results)
7962
+ * @param options.status - Filter by status (queued, processing, done, error)
7963
+ * @param options.date - Filter by exact date (ISO format YYYY-MM-DD)
7964
+ * @param options.beforeDate - Filter for transcripts before this date
7965
+ * @param options.afterDate - Filter for transcripts after this date
7966
+ * @param options.gladia - Full Gladia-specific options (custom_metadata, kind, etc.)
7967
+ * @returns List of transcripts with pagination info
7968
+ *
7969
+ * @example List recent transcripts
7970
+ * ```typescript
7971
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
7972
+ * limit: 50,
7973
+ * status: 'done'
7974
+ * })
7975
+ * ```
7976
+ *
7977
+ * @example Filter by date range
7978
+ * ```typescript
7979
+ * const { transcripts } = await adapter.listTranscripts({
7980
+ * afterDate: '2026-01-01',
7981
+ * beforeDate: '2026-01-31',
7982
+ * limit: 100
7983
+ * })
7984
+ * ```
7985
+ *
7986
+ * @example Filter by custom metadata
7987
+ * ```typescript
7988
+ * const { transcripts } = await adapter.listTranscripts({
7989
+ * gladia: {
7990
+ * custom_metadata: { project: 'my-project' }
7991
+ * }
7992
+ * })
7993
+ * ```
7994
+ *
7995
+ * @see https://docs.gladia.io/
7996
+ */
7997
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
7998
+ transcripts: UnifiedTranscriptResponse[];
7999
+ total?: number;
8000
+ hasMore?: boolean;
8001
+ }>;
8002
+ /**
8003
+ * Normalize a transcript list item to unified format
8004
+ */
8005
+ private normalizeListItem;
7324
8006
  /**
7325
8007
  * Stream audio for real-time transcription
7326
8008
  *
@@ -7624,6 +8306,57 @@ declare class AssemblyAIAdapter extends BaseAdapter {
7624
8306
  deleteTranscript(transcriptId: string): Promise<{
7625
8307
  success: boolean;
7626
8308
  }>;
8309
+ /**
8310
+ * List recent transcriptions with filtering
8311
+ *
8312
+ * Retrieves a list of transcripts with optional filtering by status and date.
8313
+ * Transcripts are sorted from newest to oldest and can be retrieved for the
8314
+ * last 90 days of usage.
8315
+ *
8316
+ * @param options - Filtering and pagination options
8317
+ * @param options.limit - Maximum number of transcripts (max 200)
8318
+ * @param options.status - Filter by status (queued, processing, completed, error)
8319
+ * @param options.date - Filter by exact date (ISO format YYYY-MM-DD)
8320
+ * @param options.assemblyai - Full AssemblyAI-specific options (before_id, after_id, etc.)
8321
+ * @returns List of transcripts with pagination info
8322
+ *
8323
+ * @example List recent transcripts
8324
+ * ```typescript
8325
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
8326
+ * limit: 50,
8327
+ * status: 'completed'
8328
+ * })
8329
+ * ```
8330
+ *
8331
+ * @example Filter by date
8332
+ * ```typescript
8333
+ * const { transcripts } = await adapter.listTranscripts({
8334
+ * date: '2026-01-07',
8335
+ * limit: 100
8336
+ * })
8337
+ * ```
8338
+ *
8339
+ * @example Use cursor pagination
8340
+ * ```typescript
8341
+ * const { transcripts } = await adapter.listTranscripts({
8342
+ * assemblyai: {
8343
+ * after_id: 'abc123', // Get transcripts after this ID
8344
+ * limit: 50
8345
+ * }
8346
+ * })
8347
+ * ```
8348
+ *
8349
+ * @see https://www.assemblyai.com/docs/api-reference/transcripts/list
8350
+ */
8351
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
8352
+ transcripts: UnifiedTranscriptResponse[];
8353
+ total?: number;
8354
+ hasMore?: boolean;
8355
+ }>;
8356
+ /**
8357
+ * Normalize a transcript list item to unified format
8358
+ */
8359
+ private normalizeListItem;
7627
8360
  /**
7628
8361
  * Build AssemblyAI transcription request from unified options
7629
8362
  */
@@ -8147,6 +8880,48 @@ declare class AzureSTTAdapter extends BaseAdapter {
8147
8880
  deleteTranscript(transcriptId: string): Promise<{
8148
8881
  success: boolean;
8149
8882
  }>;
8883
+ /**
8884
+ * List recent transcriptions with filtering
8885
+ *
8886
+ * Retrieves a list of transcription jobs for the authenticated subscription.
8887
+ * Azure uses OData filtering for advanced queries.
8888
+ *
8889
+ * @param options - Filtering and pagination options
8890
+ * @param options.limit - Maximum number of transcripts (maps to 'top')
8891
+ * @param options.offset - Pagination offset (maps to 'skip')
8892
+ * @param options.status - Filter by status (uses OData filter)
8893
+ * @returns List of transcripts with pagination info
8894
+ *
8895
+ * @example List recent transcripts
8896
+ * ```typescript
8897
+ * const { transcripts, hasMore } = await adapter.listTranscripts({
8898
+ * limit: 50
8899
+ * })
8900
+ * ```
8901
+ *
8902
+ * @example Filter by status using OData
8903
+ * ```typescript
8904
+ * const { transcripts } = await adapter.listTranscripts({
8905
+ * status: 'Succeeded',
8906
+ * limit: 100
8907
+ * })
8908
+ * ```
8909
+ *
8910
+ * @see https://learn.microsoft.com/azure/cognitive-services/speech-service/batch-transcription
8911
+ */
8912
+ listTranscripts(options?: ListTranscriptsOptions): Promise<{
8913
+ transcripts: UnifiedTranscriptResponse[];
8914
+ total?: number;
8915
+ hasMore?: boolean;
8916
+ }>;
8917
+ /**
8918
+ * Map unified status to Azure status format using generated enum
8919
+ */
8920
+ private mapStatusToAzure;
8921
+ /**
8922
+ * Normalize a transcript list item to unified format
8923
+ */
8924
+ private normalizeListItem;
8150
8925
  /**
8151
8926
  * Build Azure-specific transcription properties
8152
8927
  */
@@ -11920,73 +12695,6 @@ type StreamingControllerInitStreamingSessionV2Params = {
11920
12695
 
11921
12696
  type TranscriptionControllerGetTranscriptV2200 = PreRecordedResponse | StreamingResponse;
11922
12697
 
11923
- /**
11924
- * Generated by orval v7.9.0 🍺
11925
- * Do not edit manually.
11926
- * Gladia Control API
11927
- * OpenAPI spec version: 1.0
11928
- */
11929
- type TranscriptionControllerListV2KindItem = (typeof TranscriptionControllerListV2KindItem)[keyof typeof TranscriptionControllerListV2KindItem];
11930
- declare const TranscriptionControllerListV2KindItem: {
11931
- readonly "pre-recorded": "pre-recorded";
11932
- readonly live: "live";
11933
- };
11934
-
11935
- /**
11936
- * Generated by orval v7.9.0 🍺
11937
- * Do not edit manually.
11938
- * Gladia Control API
11939
- * OpenAPI spec version: 1.0
11940
- */
11941
- type TranscriptionControllerListV2StatusItem = (typeof TranscriptionControllerListV2StatusItem)[keyof typeof TranscriptionControllerListV2StatusItem];
11942
- declare const TranscriptionControllerListV2StatusItem: {
11943
- readonly queued: "queued";
11944
- readonly processing: "processing";
11945
- readonly done: "done";
11946
- readonly error: "error";
11947
- };
11948
-
11949
- /**
11950
- * Generated by orval v7.9.0 🍺
11951
- * Do not edit manually.
11952
- * Gladia Control API
11953
- * OpenAPI spec version: 1.0
11954
- */
11955
-
11956
- type TranscriptionControllerListV2Params = {
11957
- /**
11958
- * The starting point for pagination. A value of 0 starts from the first item.
11959
- */
11960
- offset?: number;
11961
- /**
11962
- * The maximum number of items to return. Useful for pagination and controlling data payload size.
11963
- */
11964
- limit?: number;
11965
- /**
11966
- * Filter items relevant to a specific date in ISO format (YYYY-MM-DD).
11967
- */
11968
- date?: string;
11969
- /**
11970
- * Include items that occurred before the specified date in ISO format.
11971
- */
11972
- before_date?: string;
11973
- /**
11974
- * Filter for items after the specified date. Use with `before_date` for a range. Date in ISO format.
11975
- */
11976
- after_date?: string;
11977
- /**
11978
- * Filter the list based on item status. Accepts multiple values from the predefined list.
11979
- */
11980
- status?: TranscriptionControllerListV2StatusItem[];
11981
- custom_metadata?: {
11982
- [key: string]: unknown;
11983
- };
11984
- /**
11985
- * Filter the list based on the item type. Supports multiple values from the predefined list.
11986
- */
11987
- kind?: TranscriptionControllerListV2KindItem[];
11988
- };
11989
-
11990
12698
  /**
11991
12699
  * Generated by orval v7.9.0 🍺
11992
12700
  * Do not edit manually.
@@ -13226,101 +13934,6 @@ interface ListTranscriptParams {
13226
13934
  throttled_only?: boolean;
13227
13935
  }
13228
13936
 
13229
- /**
13230
- * Generated by orval v7.9.0 🍺
13231
- * Do not edit manually.
13232
- * AssemblyAI API
13233
- * AssemblyAI API
13234
- * OpenAPI spec version: 1.3.4
13235
- */
13236
- /**
13237
- * Get transcripts that were created after this transcript ID
13238
- */
13239
- type AfterId = string;
13240
-
13241
- /**
13242
- * Generated by orval v7.9.0 🍺
13243
- * Do not edit manually.
13244
- * AssemblyAI API
13245
- * AssemblyAI API
13246
- * OpenAPI spec version: 1.3.4
13247
- */
13248
- /**
13249
- * Get transcripts that were created before this transcript ID
13250
- */
13251
- type BeforeId = string;
13252
-
13253
- /**
13254
- * Generated by orval v7.9.0 🍺
13255
- * Do not edit manually.
13256
- * AssemblyAI API
13257
- * AssemblyAI API
13258
- * OpenAPI spec version: 1.3.4
13259
- */
13260
- /**
13261
- * Only get transcripts created on this date
13262
- */
13263
- type CreatedOn = string;
13264
-
13265
- /**
13266
- * Generated by orval v7.9.0 🍺
13267
- * Do not edit manually.
13268
- * AssemblyAI API
13269
- * AssemblyAI API
13270
- * OpenAPI spec version: 1.3.4
13271
- */
13272
- /**
13273
- * Maximum amount of transcripts to retrieve
13274
- */
13275
- type Limit = number;
13276
-
13277
- /**
13278
- * Generated by orval v7.9.0 🍺
13279
- * Do not edit manually.
13280
- * AssemblyAI API
13281
- * AssemblyAI API
13282
- * OpenAPI spec version: 1.3.4
13283
- */
13284
- /**
13285
- * Only get throttled transcripts, overrides the status filter
13286
- */
13287
- type ThrottledOnly = boolean;
13288
-
13289
- /**
13290
- * Generated by orval v7.9.0 🍺
13291
- * Do not edit manually.
13292
- * AssemblyAI API
13293
- * AssemblyAI API
13294
- * OpenAPI spec version: 1.3.4
13295
- */
13296
-
13297
- type ListTranscriptsParams = {
13298
- /**
13299
- * Maximum amount of transcripts to retrieve
13300
- */
13301
- limit?: Limit;
13302
- /**
13303
- * Filter by transcript status
13304
- */
13305
- status?: TranscriptStatus;
13306
- /**
13307
- * Only get transcripts created on this date
13308
- */
13309
- created_on?: CreatedOn;
13310
- /**
13311
- * Get transcripts that were created before this transcript ID
13312
- */
13313
- before_id?: BeforeId;
13314
- /**
13315
- * Get transcripts that were created after this transcript ID
13316
- */
13317
- after_id?: AfterId;
13318
- /**
13319
- * Only get throttled transcripts, overrides the status filter
13320
- */
13321
- throttled_only?: ThrottledOnly;
13322
- };
13323
-
13324
13937
  /**
13325
13938
  * Generated by orval v7.9.0 🍺
13326
13939
  * Do not edit manually.
@@ -13932,4 +14545,4 @@ declare namespace index {
13932
14545
  export { index_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, type index_AutoHighlightResult as AutoHighlightResult, type index_AutoHighlightsResult as AutoHighlightsResult, type index_BadRequestResponse as BadRequestResponse, type index_CannotAccessUploadedFileResponse as CannotAccessUploadedFileResponse, type index_Chapter as Chapter, type index_ContentSafetyLabel as ContentSafetyLabel, type index_ContentSafetyLabelResult as ContentSafetyLabelResult, type index_ContentSafetyLabelsResult as ContentSafetyLabelsResult, type index_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, type index_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, type index_CreateRealtimeTemporaryTokenParams as CreateRealtimeTemporaryTokenParams, type index_Entity as Entity, index_EntityType as EntityType, type Error$1 as Error, type index_GatewayTimeoutResponse as GatewayTimeoutResponse, type index_GetSubtitlesParams as GetSubtitlesParams, type index_InternalServerErrorResponse as InternalServerErrorResponse, type index_LemurActionItemsParams as LemurActionItemsParams, type index_LemurActionItemsParamsAllOf as LemurActionItemsParamsAllOf, type index_LemurActionItemsResponse as LemurActionItemsResponse, type index_LemurBaseParams as LemurBaseParams, type index_LemurBaseParamsContext as LemurBaseParamsContext, type index_LemurBaseParamsContextOneOf as LemurBaseParamsContextOneOf, type index_LemurBaseParamsFinalModel as LemurBaseParamsFinalModel, type index_LemurBaseResponse as LemurBaseResponse, index_LemurModel as LemurModel, type index_LemurQuestion as LemurQuestion, type index_LemurQuestionAnswer as LemurQuestionAnswer, type index_LemurQuestionAnswerParams as LemurQuestionAnswerParams, type index_LemurQuestionAnswerParamsAllOf as LemurQuestionAnswerParamsAllOf, type index_LemurQuestionAnswerResponse as LemurQuestionAnswerResponse, type index_LemurQuestionAnswerResponseAllOf as LemurQuestionAnswerResponseAllOf, type index_LemurQuestionContext as LemurQuestionContext, type index_LemurQuestionContextOneOf as LemurQuestionContextOneOf, type index_LemurResponse as LemurResponse, type index_LemurStringResponse as LemurStringResponse, type index_LemurStringResponseAllOf as LemurStringResponseAllOf, type index_LemurSummaryParams as LemurSummaryParams, type index_LemurSummaryParamsAllOf as LemurSummaryParamsAllOf, type index_LemurSummaryResponse as LemurSummaryResponse, type index_LemurTaskParams as LemurTaskParams, type index_LemurTaskParamsAllOf as LemurTaskParamsAllOf, type index_LemurTaskResponse as LemurTaskResponse, type index_LemurUsage as LemurUsage, type index_ListTranscriptParams as ListTranscriptParams, type index_ListTranscriptsParams as ListTranscriptsParams, type index_NotFoundResponse as NotFoundResponse, type index_PageDetails as PageDetails, type index_PageDetailsNextUrl as PageDetailsNextUrl, type index_PageDetailsPrevUrl as PageDetailsPrevUrl, type index_ParagraphsResponse as ParagraphsResponse, index_PiiPolicy as PiiPolicy, type index_PurgeLemurRequestDataResponse as PurgeLemurRequestDataResponse, type index_RealtimeTemporaryTokenResponse as RealtimeTemporaryTokenResponse, index_RedactPiiAudioQuality as RedactPiiAudioQuality, type index_RedactedAudioNotification as RedactedAudioNotification, type index_RedactedAudioResponse as RedactedAudioResponse, index_RedactedAudioStatus as RedactedAudioStatus, type index_SentencesResponse as SentencesResponse, index_Sentiment as Sentiment, type SentimentAnalysisResult$1 as SentimentAnalysisResult, type index_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, type index_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, type index_ServiceUnavailableResponse as ServiceUnavailableResponse, type index_SeverityScoreSummary as SeverityScoreSummary, index_SpeechModel as SpeechModel, index_SubstitutionPolicy as SubstitutionPolicy, index_SubtitleFormat as SubtitleFormat, index_SummaryModel as SummaryModel, index_SummaryType as SummaryType, type index_Timestamp as Timestamp, type index_TooManyRequestsResponse as TooManyRequestsResponse, type index_TopicDetectionModelResult as TopicDetectionModelResult, type index_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, type index_TopicDetectionResult as TopicDetectionResult, type index_TopicDetectionResultLabelsItem as TopicDetectionResultLabelsItem, type index_Transcript as Transcript, type index_TranscriptAudioDuration as TranscriptAudioDuration, type index_TranscriptAudioEndAt as TranscriptAudioEndAt, type index_TranscriptAudioStartFrom as TranscriptAudioStartFrom, type index_TranscriptAutoChapters as TranscriptAutoChapters, type index_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index_TranscriptBoostParam as TranscriptBoostParam, type index_TranscriptBoostParamProperty as TranscriptBoostParamProperty, type index_TranscriptChapters as TranscriptChapters, type index_TranscriptConfidence as TranscriptConfidence, type index_TranscriptContentSafety as TranscriptContentSafety, type index_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, type index_TranscriptCustomSpelling as TranscriptCustomSpelling, type index_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, type index_TranscriptCustomTopics as TranscriptCustomTopics, type index_TranscriptDisfluencies as TranscriptDisfluencies, type index_TranscriptEntities as TranscriptEntities, type index_TranscriptEntityDetection as TranscriptEntityDetection, type index_TranscriptFilterProfanity as TranscriptFilterProfanity, type index_TranscriptFormatText as TranscriptFormatText, type index_TranscriptIabCategories as TranscriptIabCategories, type index_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index_TranscriptLanguageCode as TranscriptLanguageCode, type index_TranscriptLanguageCodeProperty as TranscriptLanguageCodeProperty, type index_TranscriptLanguageConfidence as TranscriptLanguageConfidence, type index_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, type index_TranscriptLanguageDetection as TranscriptLanguageDetection, type index_TranscriptList as TranscriptList, type index_TranscriptListItem as TranscriptListItem, type index_TranscriptListItemCompleted as TranscriptListItemCompleted, type index_TranscriptListItemError as TranscriptListItemError, type index_TranscriptMultichannel as TranscriptMultichannel, type index_TranscriptOptionalParams as TranscriptOptionalParams, type index_TranscriptOptionalParamsLanguageCode as TranscriptOptionalParamsLanguageCode, type index_TranscriptOptionalParamsLanguageCodeOneOf as TranscriptOptionalParamsLanguageCodeOneOf, type index_TranscriptOptionalParamsRedactPiiSub as TranscriptOptionalParamsRedactPiiSub, type index_TranscriptOptionalParamsSpeakersExpected as TranscriptOptionalParamsSpeakersExpected, type index_TranscriptOptionalParamsSpeechModel as TranscriptOptionalParamsSpeechModel, type index_TranscriptOptionalParamsSpeechThreshold as TranscriptOptionalParamsSpeechThreshold, type index_TranscriptOptionalParamsWebhookAuthHeaderName as TranscriptOptionalParamsWebhookAuthHeaderName, type index_TranscriptOptionalParamsWebhookAuthHeaderValue as TranscriptOptionalParamsWebhookAuthHeaderValue, type index_TranscriptParagraph as TranscriptParagraph, type index_TranscriptParams as TranscriptParams, type index_TranscriptParamsAllOf as TranscriptParamsAllOf, type index_TranscriptPunctuate as TranscriptPunctuate, type index_TranscriptReadyNotification as TranscriptReadyNotification, index_TranscriptReadyStatus as TranscriptReadyStatus, type index_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, type index_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, type index_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, type index_TranscriptSentence as TranscriptSentence, type index_TranscriptSentenceChannel as TranscriptSentenceChannel, type index_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, type index_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, type index_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, type index_TranscriptSpeakerLabels as TranscriptSpeakerLabels, type index_TranscriptSpeakersExpected as TranscriptSpeakersExpected, type index_TranscriptSpeechModel as TranscriptSpeechModel, type index_TranscriptSpeechThreshold as TranscriptSpeechThreshold, type index_TranscriptSpeedBoost as TranscriptSpeedBoost, index_TranscriptStatus as TranscriptStatus, type index_TranscriptSummary as TranscriptSummary, type index_TranscriptSummaryModel as TranscriptSummaryModel, type index_TranscriptSummaryType as TranscriptSummaryType, type index_TranscriptText as TranscriptText, type index_TranscriptThrottled as TranscriptThrottled, type index_TranscriptUtterance as TranscriptUtterance, type index_TranscriptUtteranceChannel as TranscriptUtteranceChannel, type index_TranscriptUtterances as TranscriptUtterances, type index_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, type index_TranscriptWebhookNotification as TranscriptWebhookNotification, type index_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, type index_TranscriptWebhookUrl as TranscriptWebhookUrl, type index_TranscriptWord as TranscriptWord, type index_TranscriptWordChannel as TranscriptWordChannel, type index_TranscriptWordSpeaker as TranscriptWordSpeaker, type index_TranscriptWords as TranscriptWords, type index_UnauthorizedResponse as UnauthorizedResponse, type index_UploadedFile as UploadedFile, type index_WordSearchMatch as WordSearchMatch, type index_WordSearchParams as WordSearchParams, type index_WordSearchResponse as WordSearchResponse, type index_WordSearchTimestamp as WordSearchTimestamp };
13933
14546
  }
13934
14547
 
13935
- export { AssemblyAIAdapter, type Chapter as AssemblyAIChapter, type ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, AssemblyAIEncodingType, type Entity as AssemblyAIEntity, type AssemblyAIExtendedData, type AutoHighlightsResult as AssemblyAIHighlightsResult, type TranscriptOptionalParams as AssemblyAIOptions, AssemblyAISampleRateType, type SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAISpeechModelType, type AssemblyAIStreamingOptions, type TopicDetectionModelResult as AssemblyAITopicsResult, index as AssemblyAITypes, type AssemblyAIUpdateConfiguration, AssemblyAIWebhookHandler, type TranscriptWebhookNotification as AssemblyAIWebhookPayload, type AudioAckEvent, type AudioChunk, type AudioInput, AudioResponseFormat, AudioTranscriptionModel, AzureSTTAdapter, AzureWebhookHandler, BaseAdapter, BaseWebhookHandler, type BatchOnlyProvider, type ChapterizationEvent, DeepgramAdapter, ListenV1EncodingParameter as DeepgramEncoding, type DeepgramExtendedData, type ListenV1ResponseMetadata as DeepgramMetadata, DeepgramModelType, type ListenV1MediaTranscribeParams as DeepgramOptions, ListenV1RedactParameterOneOfItem as DeepgramRedact, ListenV1RedactParameterOneOfItem as DeepgramRedactType, type DeepgramStreamingOptions, SharedCustomTopicModeParameter as DeepgramTopicMode, SharedCustomTopicModeParameter as DeepgramTopicModeType, DeepgramWebhookHandler, type ListenV1Response as DeepgramWebhookPayload, type EntityEvent, GladiaAdapter, type AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, type AudioToLlmListDTO as GladiaAudioToLlmResult, StreamingSupportedBitDepthEnum as GladiaBitDepth, type ChapterizationDTO as GladiaChapters, type CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, StreamingSupportedEncodingEnum as GladiaEncoding, type NamedEntityRecognitionDTO as GladiaEntities, type GladiaExtendedData, TranscriptionLanguageCodeEnum as GladiaLanguage, StreamingSupportedModels as GladiaModel, type ModerationDTO as GladiaModeration, type InitTranscriptionRequest as GladiaOptions, StreamingSupportedSampleRateEnum as GladiaSampleRate, type SentimentAnalysisDTO as GladiaSentiment, type SpeakerReidentificationDTO as GladiaSpeakerReidentification, type GladiaStreamingOptions, type StreamingRequest as GladiaStreamingRequest, type StructuredDataExtractionDTO as GladiaStructuredData, type TranslationDTO as GladiaTranslation, TranslationLanguageCodeEnum as GladiaTranslationLanguage, index$1 as GladiaTypes, type CallbackTranscriptionErrorPayload as GladiaWebhookErrorPayload, GladiaWebhookHandler, type GladiaWebhookPayload, type CallbackTranscriptionSuccessPayload as GladiaWebhookSuccessPayload, type LifecycleEvent, ListenV1EncodingParameter, type ListenV1LanguageParameter, type ListenV1ModelParameter, type ListenV1VersionParameter, OpenAIWhisperAdapter, type CreateTranscriptionRequest as OpenAIWhisperOptions, type ProviderCapabilities, type ProviderConfig, type ProviderExtendedDataMap, type ProviderRawResponseMap, type ProviderStreamingOptions, type ProviderWebhookPayloadMap, type SentimentEvent, type SessionStatus, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type Speaker, type SpeechEvent, SpeechmaticsAdapter, type SpeechmaticsOperatingPoint, SpeechmaticsWebhookHandler, type StreamEvent, type StreamEventType, type StreamingCallbacks, type StreamingOptions, type StreamingOptionsForProvider, type StreamingProvider, type StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedSampleRateEnum, type SummarizationEvent, type TranscribeOptions, type TranscribeStreamParams, type TranscriptionAdapter, type TranscriptionLanguage, type TranscriptionModel, type TranscriptionProvider, type TranscriptionStatus, type TranslationEvent, type UnifiedTranscriptResponse, type UnifiedWebhookEvent, type Utterance, VoiceRouter, type VoiceRouterConfig, type WebhookEventType, WebhookRouter, type WebhookRouterOptions, type WebhookRouterResult, type WebhookValidation, type WebhookVerificationOptions, type Word, createAssemblyAIAdapter, createAssemblyAIWebhookHandler, createAzureSTTAdapter, createAzureWebhookHandler, createDeepgramAdapter, createDeepgramWebhookHandler, createGladiaAdapter, createGladiaWebhookHandler, createOpenAIWhisperAdapter, createSpeechmaticsAdapter, createVoiceRouter, createWebhookRouter };
14548
+ export { AssemblyAIAdapter, type Chapter as AssemblyAIChapter, type ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, AssemblyAIEncodingType, type Entity as AssemblyAIEntity, type AssemblyAIExtendedData, type AutoHighlightsResult as AssemblyAIHighlightsResult, type TranscriptOptionalParams as AssemblyAIOptions, AssemblyAISampleRateType, type SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAISpeechModelType, type AssemblyAIStreamingOptions, type TopicDetectionModelResult as AssemblyAITopicsResult, index as AssemblyAITypes, type AssemblyAIUpdateConfiguration, AssemblyAIWebhookHandler, type TranscriptWebhookNotification as AssemblyAIWebhookPayload, type AudioAckEvent, type AudioChunk, type AudioInput, AudioResponseFormat, AudioTranscriptionModel, AzureSTTAdapter, AzureWebhookHandler, BaseAdapter, BaseWebhookHandler, type BatchOnlyProvider, type ChapterizationEvent, DeepgramAdapter, type DeepgramExtendedData, type ListenV1ResponseMetadata as DeepgramMetadata, DeepgramModelType, type ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramRedactType, type DeepgramStreamingOptions, DeepgramTopicModeType, DeepgramWebhookHandler, type ListenV1Response as DeepgramWebhookPayload, type EntityEvent, GladiaAdapter, type AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, type AudioToLlmListDTO as GladiaAudioToLlmResult, type ChapterizationDTO as GladiaChapters, type CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, type NamedEntityRecognitionDTO as GladiaEntities, type GladiaExtendedData, type ModerationDTO as GladiaModeration, type InitTranscriptionRequest as GladiaOptions, type SentimentAnalysisDTO as GladiaSentiment, type SpeakerReidentificationDTO as GladiaSpeakerReidentification, type GladiaStreamingOptions, type StreamingRequest as GladiaStreamingRequest, type StructuredDataExtractionDTO as GladiaStructuredData, type TranslationDTO as GladiaTranslation, index$1 as GladiaTypes, type CallbackTranscriptionErrorPayload as GladiaWebhookErrorPayload, GladiaWebhookHandler, type GladiaWebhookPayload, type CallbackTranscriptionSuccessPayload as GladiaWebhookSuccessPayload, type LifecycleEvent, type ListTranscriptsOptions, ListenV1EncodingParameter, type ListenV1LanguageParameter, type ListenV1ModelParameter, type ListenV1VersionParameter, OpenAIWhisperAdapter, type CreateTranscriptionRequest as OpenAIWhisperOptions, type ProviderCapabilities, type ProviderConfig, type ProviderExtendedDataMap, type ProviderRawResponseMap, type ProviderStreamingOptions, type ProviderWebhookPayloadMap, type SentimentEvent, type SessionStatus, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type Speaker, type SpeechEvent, SpeechmaticsAdapter, type SpeechmaticsOperatingPoint, SpeechmaticsWebhookHandler, type StreamEvent, type StreamEventType, type StreamingCallbacks, type StreamingOptions, type StreamingOptionsForProvider, type StreamingProvider, type StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedSampleRateEnum, type SummarizationEvent, type TranscribeOptions, type TranscribeStreamParams, type TranscriptionAdapter, type TranscriptionLanguage, type TranscriptionModel, type TranscriptionProvider, type TranscriptionStatus, type TranslationEvent, type UnifiedTranscriptResponse, type UnifiedWebhookEvent, type Utterance, VoiceRouter, type VoiceRouterConfig, type WebhookEventType, WebhookRouter, type WebhookRouterOptions, type WebhookRouterResult, type WebhookValidation, type WebhookVerificationOptions, type Word, createAssemblyAIAdapter, createAssemblyAIWebhookHandler, createAzureSTTAdapter, createAzureWebhookHandler, createDeepgramAdapter, createDeepgramWebhookHandler, createGladiaAdapter, createGladiaWebhookHandler, createOpenAIWhisperAdapter, createSpeechmaticsAdapter, createVoiceRouter, createWebhookRouter };