voice-router-dev 0.3.2 → 0.3.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +478 -0
- package/dist/constants.d.mts +376 -246
- package/dist/constants.d.ts +376 -246
- package/dist/constants.js +87 -9
- package/dist/constants.js.map +1 -1
- package/dist/constants.mjs +80 -9
- package/dist/constants.mjs.map +1 -1
- package/dist/index.d.mts +1464 -352
- package/dist/index.d.ts +1464 -352
- package/dist/index.js +718 -46
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +718 -46
- package/dist/index.mjs.map +1 -1
- package/package.json +2 -1
package/dist/index.d.mts
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
import {
|
|
2
|
-
export { AssemblyAIEncoding, AssemblyAISampleRate, AssemblyAISpeechModel, DeepgramModel } from './constants.mjs';
|
|
1
|
+
import { DeepgramModelType, DeepgramTopicModeType, DeepgramRedactType, AssemblyAISampleRateType, AssemblyAIEncodingType, AssemblyAISpeechModelType } from './constants.mjs';
|
|
2
|
+
export { AssemblyAIEncoding, AssemblyAISampleRate, AssemblyAISpeechModel, DeepgramEncoding, DeepgramModel, DeepgramRedact, DeepgramTopicMode, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaSampleRate, GladiaTranslationLanguage } from './constants.mjs';
|
|
3
3
|
|
|
4
4
|
/**
|
|
5
5
|
* Unified audio encoding types for Voice Router SDK
|
|
@@ -700,6 +700,118 @@ interface CallbackConfigDto {
|
|
|
700
700
|
method?: CallbackMethodEnum;
|
|
701
701
|
}
|
|
702
702
|
|
|
703
|
+
/**
|
|
704
|
+
* Generated by orval v7.9.0 🍺
|
|
705
|
+
* Do not edit manually.
|
|
706
|
+
* Gladia Control API
|
|
707
|
+
* OpenAPI spec version: 1.0
|
|
708
|
+
*/
|
|
709
|
+
/**
|
|
710
|
+
* Specify the language in which it will be pronounced when sound comparison occurs. Default to transcription language.
|
|
711
|
+
*/
|
|
712
|
+
type TranscriptionLanguageCodeEnum = (typeof TranscriptionLanguageCodeEnum)[keyof typeof TranscriptionLanguageCodeEnum];
|
|
713
|
+
declare const TranscriptionLanguageCodeEnum: {
|
|
714
|
+
readonly af: "af";
|
|
715
|
+
readonly am: "am";
|
|
716
|
+
readonly ar: "ar";
|
|
717
|
+
readonly as: "as";
|
|
718
|
+
readonly az: "az";
|
|
719
|
+
readonly ba: "ba";
|
|
720
|
+
readonly be: "be";
|
|
721
|
+
readonly bg: "bg";
|
|
722
|
+
readonly bn: "bn";
|
|
723
|
+
readonly bo: "bo";
|
|
724
|
+
readonly br: "br";
|
|
725
|
+
readonly bs: "bs";
|
|
726
|
+
readonly ca: "ca";
|
|
727
|
+
readonly cs: "cs";
|
|
728
|
+
readonly cy: "cy";
|
|
729
|
+
readonly da: "da";
|
|
730
|
+
readonly de: "de";
|
|
731
|
+
readonly el: "el";
|
|
732
|
+
readonly en: "en";
|
|
733
|
+
readonly es: "es";
|
|
734
|
+
readonly et: "et";
|
|
735
|
+
readonly eu: "eu";
|
|
736
|
+
readonly fa: "fa";
|
|
737
|
+
readonly fi: "fi";
|
|
738
|
+
readonly fo: "fo";
|
|
739
|
+
readonly fr: "fr";
|
|
740
|
+
readonly gl: "gl";
|
|
741
|
+
readonly gu: "gu";
|
|
742
|
+
readonly ha: "ha";
|
|
743
|
+
readonly haw: "haw";
|
|
744
|
+
readonly he: "he";
|
|
745
|
+
readonly hi: "hi";
|
|
746
|
+
readonly hr: "hr";
|
|
747
|
+
readonly ht: "ht";
|
|
748
|
+
readonly hu: "hu";
|
|
749
|
+
readonly hy: "hy";
|
|
750
|
+
readonly id: "id";
|
|
751
|
+
readonly is: "is";
|
|
752
|
+
readonly it: "it";
|
|
753
|
+
readonly ja: "ja";
|
|
754
|
+
readonly jw: "jw";
|
|
755
|
+
readonly ka: "ka";
|
|
756
|
+
readonly kk: "kk";
|
|
757
|
+
readonly km: "km";
|
|
758
|
+
readonly kn: "kn";
|
|
759
|
+
readonly ko: "ko";
|
|
760
|
+
readonly la: "la";
|
|
761
|
+
readonly lb: "lb";
|
|
762
|
+
readonly ln: "ln";
|
|
763
|
+
readonly lo: "lo";
|
|
764
|
+
readonly lt: "lt";
|
|
765
|
+
readonly lv: "lv";
|
|
766
|
+
readonly mg: "mg";
|
|
767
|
+
readonly mi: "mi";
|
|
768
|
+
readonly mk: "mk";
|
|
769
|
+
readonly ml: "ml";
|
|
770
|
+
readonly mn: "mn";
|
|
771
|
+
readonly mr: "mr";
|
|
772
|
+
readonly ms: "ms";
|
|
773
|
+
readonly mt: "mt";
|
|
774
|
+
readonly my: "my";
|
|
775
|
+
readonly ne: "ne";
|
|
776
|
+
readonly nl: "nl";
|
|
777
|
+
readonly nn: "nn";
|
|
778
|
+
readonly no: "no";
|
|
779
|
+
readonly oc: "oc";
|
|
780
|
+
readonly pa: "pa";
|
|
781
|
+
readonly pl: "pl";
|
|
782
|
+
readonly ps: "ps";
|
|
783
|
+
readonly pt: "pt";
|
|
784
|
+
readonly ro: "ro";
|
|
785
|
+
readonly ru: "ru";
|
|
786
|
+
readonly sa: "sa";
|
|
787
|
+
readonly sd: "sd";
|
|
788
|
+
readonly si: "si";
|
|
789
|
+
readonly sk: "sk";
|
|
790
|
+
readonly sl: "sl";
|
|
791
|
+
readonly sn: "sn";
|
|
792
|
+
readonly so: "so";
|
|
793
|
+
readonly sq: "sq";
|
|
794
|
+
readonly sr: "sr";
|
|
795
|
+
readonly su: "su";
|
|
796
|
+
readonly sv: "sv";
|
|
797
|
+
readonly sw: "sw";
|
|
798
|
+
readonly ta: "ta";
|
|
799
|
+
readonly te: "te";
|
|
800
|
+
readonly tg: "tg";
|
|
801
|
+
readonly th: "th";
|
|
802
|
+
readonly tk: "tk";
|
|
803
|
+
readonly tl: "tl";
|
|
804
|
+
readonly tr: "tr";
|
|
805
|
+
readonly tt: "tt";
|
|
806
|
+
readonly uk: "uk";
|
|
807
|
+
readonly ur: "ur";
|
|
808
|
+
readonly uz: "uz";
|
|
809
|
+
readonly vi: "vi";
|
|
810
|
+
readonly yi: "yi";
|
|
811
|
+
readonly yo: "yo";
|
|
812
|
+
readonly zh: "zh";
|
|
813
|
+
};
|
|
814
|
+
|
|
703
815
|
/**
|
|
704
816
|
* Generated by orval v7.9.0 🍺
|
|
705
817
|
* Do not edit manually.
|
|
@@ -935,6 +1047,119 @@ interface SummarizationConfigDTO {
|
|
|
935
1047
|
type?: SummaryTypesEnum;
|
|
936
1048
|
}
|
|
937
1049
|
|
|
1050
|
+
/**
|
|
1051
|
+
* Generated by orval v7.9.0 🍺
|
|
1052
|
+
* Do not edit manually.
|
|
1053
|
+
* Gladia Control API
|
|
1054
|
+
* OpenAPI spec version: 1.0
|
|
1055
|
+
*/
|
|
1056
|
+
/**
|
|
1057
|
+
* Target language in `iso639-1` format you want the transcription translated to
|
|
1058
|
+
*/
|
|
1059
|
+
type TranslationLanguageCodeEnum = (typeof TranslationLanguageCodeEnum)[keyof typeof TranslationLanguageCodeEnum];
|
|
1060
|
+
declare const TranslationLanguageCodeEnum: {
|
|
1061
|
+
readonly af: "af";
|
|
1062
|
+
readonly am: "am";
|
|
1063
|
+
readonly ar: "ar";
|
|
1064
|
+
readonly as: "as";
|
|
1065
|
+
readonly az: "az";
|
|
1066
|
+
readonly ba: "ba";
|
|
1067
|
+
readonly be: "be";
|
|
1068
|
+
readonly bg: "bg";
|
|
1069
|
+
readonly bn: "bn";
|
|
1070
|
+
readonly bo: "bo";
|
|
1071
|
+
readonly br: "br";
|
|
1072
|
+
readonly bs: "bs";
|
|
1073
|
+
readonly ca: "ca";
|
|
1074
|
+
readonly cs: "cs";
|
|
1075
|
+
readonly cy: "cy";
|
|
1076
|
+
readonly da: "da";
|
|
1077
|
+
readonly de: "de";
|
|
1078
|
+
readonly el: "el";
|
|
1079
|
+
readonly en: "en";
|
|
1080
|
+
readonly es: "es";
|
|
1081
|
+
readonly et: "et";
|
|
1082
|
+
readonly eu: "eu";
|
|
1083
|
+
readonly fa: "fa";
|
|
1084
|
+
readonly fi: "fi";
|
|
1085
|
+
readonly fo: "fo";
|
|
1086
|
+
readonly fr: "fr";
|
|
1087
|
+
readonly gl: "gl";
|
|
1088
|
+
readonly gu: "gu";
|
|
1089
|
+
readonly ha: "ha";
|
|
1090
|
+
readonly haw: "haw";
|
|
1091
|
+
readonly he: "he";
|
|
1092
|
+
readonly hi: "hi";
|
|
1093
|
+
readonly hr: "hr";
|
|
1094
|
+
readonly ht: "ht";
|
|
1095
|
+
readonly hu: "hu";
|
|
1096
|
+
readonly hy: "hy";
|
|
1097
|
+
readonly id: "id";
|
|
1098
|
+
readonly is: "is";
|
|
1099
|
+
readonly it: "it";
|
|
1100
|
+
readonly ja: "ja";
|
|
1101
|
+
readonly jw: "jw";
|
|
1102
|
+
readonly ka: "ka";
|
|
1103
|
+
readonly kk: "kk";
|
|
1104
|
+
readonly km: "km";
|
|
1105
|
+
readonly kn: "kn";
|
|
1106
|
+
readonly ko: "ko";
|
|
1107
|
+
readonly la: "la";
|
|
1108
|
+
readonly lb: "lb";
|
|
1109
|
+
readonly ln: "ln";
|
|
1110
|
+
readonly lo: "lo";
|
|
1111
|
+
readonly lt: "lt";
|
|
1112
|
+
readonly lv: "lv";
|
|
1113
|
+
readonly mg: "mg";
|
|
1114
|
+
readonly mi: "mi";
|
|
1115
|
+
readonly mk: "mk";
|
|
1116
|
+
readonly ml: "ml";
|
|
1117
|
+
readonly mn: "mn";
|
|
1118
|
+
readonly mr: "mr";
|
|
1119
|
+
readonly ms: "ms";
|
|
1120
|
+
readonly mt: "mt";
|
|
1121
|
+
readonly my: "my";
|
|
1122
|
+
readonly ne: "ne";
|
|
1123
|
+
readonly nl: "nl";
|
|
1124
|
+
readonly nn: "nn";
|
|
1125
|
+
readonly no: "no";
|
|
1126
|
+
readonly oc: "oc";
|
|
1127
|
+
readonly pa: "pa";
|
|
1128
|
+
readonly pl: "pl";
|
|
1129
|
+
readonly ps: "ps";
|
|
1130
|
+
readonly pt: "pt";
|
|
1131
|
+
readonly ro: "ro";
|
|
1132
|
+
readonly ru: "ru";
|
|
1133
|
+
readonly sa: "sa";
|
|
1134
|
+
readonly sd: "sd";
|
|
1135
|
+
readonly si: "si";
|
|
1136
|
+
readonly sk: "sk";
|
|
1137
|
+
readonly sl: "sl";
|
|
1138
|
+
readonly sn: "sn";
|
|
1139
|
+
readonly so: "so";
|
|
1140
|
+
readonly sq: "sq";
|
|
1141
|
+
readonly sr: "sr";
|
|
1142
|
+
readonly su: "su";
|
|
1143
|
+
readonly sv: "sv";
|
|
1144
|
+
readonly sw: "sw";
|
|
1145
|
+
readonly ta: "ta";
|
|
1146
|
+
readonly te: "te";
|
|
1147
|
+
readonly tg: "tg";
|
|
1148
|
+
readonly th: "th";
|
|
1149
|
+
readonly tk: "tk";
|
|
1150
|
+
readonly tl: "tl";
|
|
1151
|
+
readonly tr: "tr";
|
|
1152
|
+
readonly tt: "tt";
|
|
1153
|
+
readonly uk: "uk";
|
|
1154
|
+
readonly ur: "ur";
|
|
1155
|
+
readonly uz: "uz";
|
|
1156
|
+
readonly vi: "vi";
|
|
1157
|
+
readonly wo: "wo";
|
|
1158
|
+
readonly yi: "yi";
|
|
1159
|
+
readonly yo: "yo";
|
|
1160
|
+
readonly zh: "zh";
|
|
1161
|
+
};
|
|
1162
|
+
|
|
938
1163
|
/**
|
|
939
1164
|
* Generated by orval v7.9.0 🍺
|
|
940
1165
|
* Do not edit manually.
|
|
@@ -4130,12 +4355,16 @@ interface Transcription {
|
|
|
4130
4355
|
/**
|
|
4131
4356
|
* Generated by orval v7.9.0 🍺
|
|
4132
4357
|
* Do not edit manually.
|
|
4133
|
-
*
|
|
4134
|
-
*
|
|
4135
|
-
|
|
4136
|
-
* OpenAPI spec version: 1.0.0
|
|
4358
|
+
* Gladia Control API
|
|
4359
|
+
* OpenAPI spec version: 1.0
|
|
4137
4360
|
*/
|
|
4138
|
-
|
|
4361
|
+
/**
|
|
4362
|
+
* The model used to process the audio. "solaria-1" is used by default.
|
|
4363
|
+
*/
|
|
4364
|
+
type StreamingSupportedModels = (typeof StreamingSupportedModels)[keyof typeof StreamingSupportedModels];
|
|
4365
|
+
declare const StreamingSupportedModels: {
|
|
4366
|
+
readonly "solaria-1": "solaria-1";
|
|
4367
|
+
};
|
|
4139
4368
|
|
|
4140
4369
|
/**
|
|
4141
4370
|
* Generated by orval v7.9.0 🍺
|
|
@@ -4254,6 +4483,44 @@ type SharedTopicsParameter = boolean;
|
|
|
4254
4483
|
*/
|
|
4255
4484
|
type SharedCustomTopicParameter = string | string[];
|
|
4256
4485
|
|
|
4486
|
+
/**
|
|
4487
|
+
* Generated by orval v7.9.0 🍺
|
|
4488
|
+
* Do not edit manually.
|
|
4489
|
+
* Deepgram API Specification
|
|
4490
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
4491
|
+
|
|
4492
|
+
* OpenAPI spec version: 1.0.0
|
|
4493
|
+
*/
|
|
4494
|
+
/**
|
|
4495
|
+
* SharedCustomTopicModeParameter type definition
|
|
4496
|
+
*/
|
|
4497
|
+
/**
|
|
4498
|
+
* SharedCustomTopicModeParameter type definition
|
|
4499
|
+
*/
|
|
4500
|
+
/**
|
|
4501
|
+
* SharedCustomTopicModeParameter type definition
|
|
4502
|
+
*/
|
|
4503
|
+
/**
|
|
4504
|
+
* SharedCustomTopicModeParameter type definition
|
|
4505
|
+
*/
|
|
4506
|
+
/**
|
|
4507
|
+
* SharedCustomTopicModeParameter type definition
|
|
4508
|
+
*/
|
|
4509
|
+
/**
|
|
4510
|
+
* SharedCustomTopicModeParameter type definition
|
|
4511
|
+
*/
|
|
4512
|
+
/**
|
|
4513
|
+
* SharedCustomTopicModeParameter type definition
|
|
4514
|
+
*/
|
|
4515
|
+
/**
|
|
4516
|
+
* SharedCustomTopicModeParameter type definition
|
|
4517
|
+
*/
|
|
4518
|
+
type SharedCustomTopicModeParameter = typeof SharedCustomTopicModeParameter[keyof typeof SharedCustomTopicModeParameter];
|
|
4519
|
+
declare const SharedCustomTopicModeParameter: {
|
|
4520
|
+
readonly extended: "extended";
|
|
4521
|
+
readonly strict: "strict";
|
|
4522
|
+
};
|
|
4523
|
+
|
|
4257
4524
|
/**
|
|
4258
4525
|
* Generated by orval v7.9.0 🍺
|
|
4259
4526
|
* Do not edit manually.
|
|
@@ -4364,6 +4631,48 @@ type ListenV1DiarizeParameter = boolean;
|
|
|
4364
4631
|
*/
|
|
4365
4632
|
type ListenV1DictationParameter = boolean;
|
|
4366
4633
|
|
|
4634
|
+
/**
|
|
4635
|
+
* Generated by orval v7.9.0 🍺
|
|
4636
|
+
* Do not edit manually.
|
|
4637
|
+
* Deepgram API Specification
|
|
4638
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
4639
|
+
|
|
4640
|
+
* OpenAPI spec version: 1.0.0
|
|
4641
|
+
*/
|
|
4642
|
+
/**
|
|
4643
|
+
* ListenV1EncodingParameter type definition
|
|
4644
|
+
*/
|
|
4645
|
+
/**
|
|
4646
|
+
* ListenV1EncodingParameter type definition
|
|
4647
|
+
*/
|
|
4648
|
+
/**
|
|
4649
|
+
* ListenV1EncodingParameter type definition
|
|
4650
|
+
*/
|
|
4651
|
+
/**
|
|
4652
|
+
* ListenV1EncodingParameter type definition
|
|
4653
|
+
*/
|
|
4654
|
+
/**
|
|
4655
|
+
* ListenV1EncodingParameter type definition
|
|
4656
|
+
*/
|
|
4657
|
+
/**
|
|
4658
|
+
* ListenV1EncodingParameter type definition
|
|
4659
|
+
*/
|
|
4660
|
+
/**
|
|
4661
|
+
* ListenV1EncodingParameter type definition
|
|
4662
|
+
*/
|
|
4663
|
+
/**
|
|
4664
|
+
* ListenV1EncodingParameter type definition
|
|
4665
|
+
*/
|
|
4666
|
+
type ListenV1EncodingParameter = typeof ListenV1EncodingParameter[keyof typeof ListenV1EncodingParameter];
|
|
4667
|
+
declare const ListenV1EncodingParameter: {
|
|
4668
|
+
readonly linear16: "linear16";
|
|
4669
|
+
readonly flac: "flac";
|
|
4670
|
+
readonly mulaw: "mulaw";
|
|
4671
|
+
readonly opus: "opus";
|
|
4672
|
+
readonly speex: "speex";
|
|
4673
|
+
readonly g729: "g729";
|
|
4674
|
+
};
|
|
4675
|
+
|
|
4367
4676
|
/**
|
|
4368
4677
|
* Generated by orval v7.9.0 🍺
|
|
4369
4678
|
* Do not edit manually.
|
|
@@ -4426,6 +4735,16 @@ type ListenV1LanguageParameter = string;
|
|
|
4426
4735
|
*/
|
|
4427
4736
|
type ListenV1MeasurementsParameter = boolean;
|
|
4428
4737
|
|
|
4738
|
+
/**
|
|
4739
|
+
* Generated by orval v7.9.0 🍺
|
|
4740
|
+
* Do not edit manually.
|
|
4741
|
+
* Deepgram API Specification
|
|
4742
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
4743
|
+
|
|
4744
|
+
* OpenAPI spec version: 1.0.0
|
|
4745
|
+
*/
|
|
4746
|
+
type ListenV1ModelParameter = "nova-3" | "nova-3-general" | "nova-3-medical" | "nova-2" | "nova-2-general" | "nova-2-meeting" | "nova-2-finance" | "nova-2-conversationalai" | "nova-2-voicemail" | "nova-2-video" | "nova-2-medical" | "nova-2-drivethru" | "nova-2-automotive" | "nova" | "nova-general" | "nova-phonecall" | "nova-medical" | "enhanced" | "enhanced-general" | "enhanced-meeting" | "enhanced-phonecall" | "enhanced-finance" | "base" | "meeting" | "phonecall" | "finance" | "conversationalai" | "voicemail" | "video" | string;
|
|
4747
|
+
|
|
4429
4748
|
/**
|
|
4430
4749
|
* Generated by orval v7.9.0 🍺
|
|
4431
4750
|
* Do not edit manually.
|
|
@@ -4491,6 +4810,21 @@ type ListenV1ProfanityFilterParameter = boolean;
|
|
|
4491
4810
|
*/
|
|
4492
4811
|
type ListenV1PunctuateParameter = boolean;
|
|
4493
4812
|
|
|
4813
|
+
/**
|
|
4814
|
+
* Generated by orval v7.9.0 🍺
|
|
4815
|
+
* Do not edit manually.
|
|
4816
|
+
* Deepgram API Specification
|
|
4817
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
4818
|
+
|
|
4819
|
+
* OpenAPI spec version: 1.0.0
|
|
4820
|
+
*/
|
|
4821
|
+
type ListenV1RedactParameterOneOfItem = (typeof ListenV1RedactParameterOneOfItem)[keyof typeof ListenV1RedactParameterOneOfItem];
|
|
4822
|
+
declare const ListenV1RedactParameterOneOfItem: {
|
|
4823
|
+
readonly pci: "pci";
|
|
4824
|
+
readonly pii: "pii";
|
|
4825
|
+
readonly numbers: "numbers";
|
|
4826
|
+
};
|
|
4827
|
+
|
|
4494
4828
|
/**
|
|
4495
4829
|
* Generated by orval v7.9.0 🍺
|
|
4496
4830
|
* Do not edit manually.
|
|
@@ -5445,6 +5779,62 @@ type StreamingRequestCustomMetadata = {
|
|
|
5445
5779
|
[key: string]: unknown;
|
|
5446
5780
|
};
|
|
5447
5781
|
|
|
5782
|
+
/**
|
|
5783
|
+
* Generated by orval v7.9.0 🍺
|
|
5784
|
+
* Do not edit manually.
|
|
5785
|
+
* Gladia Control API
|
|
5786
|
+
* OpenAPI spec version: 1.0
|
|
5787
|
+
*/
|
|
5788
|
+
/**
|
|
5789
|
+
* The bit depth of the audio stream
|
|
5790
|
+
*/
|
|
5791
|
+
type StreamingSupportedBitDepthEnum = (typeof StreamingSupportedBitDepthEnum)[keyof typeof StreamingSupportedBitDepthEnum];
|
|
5792
|
+
declare const StreamingSupportedBitDepthEnum: {
|
|
5793
|
+
readonly NUMBER_8: 8;
|
|
5794
|
+
readonly NUMBER_16: 16;
|
|
5795
|
+
readonly NUMBER_24: 24;
|
|
5796
|
+
readonly NUMBER_32: 32;
|
|
5797
|
+
};
|
|
5798
|
+
|
|
5799
|
+
/**
|
|
5800
|
+
* Generated by orval v7.9.0 🍺
|
|
5801
|
+
* Do not edit manually.
|
|
5802
|
+
* Gladia Control API
|
|
5803
|
+
* OpenAPI spec version: 1.0
|
|
5804
|
+
*/
|
|
5805
|
+
/**
|
|
5806
|
+
* The encoding format of the audio stream. Supported formats:
|
|
5807
|
+
- PCM: 8, 16, 24, and 32 bits
|
|
5808
|
+
- A-law: 8 bits
|
|
5809
|
+
- μ-law: 8 bits
|
|
5810
|
+
|
|
5811
|
+
Note: No need to add WAV headers to raw audio as the API supports both formats.
|
|
5812
|
+
*/
|
|
5813
|
+
type StreamingSupportedEncodingEnum = (typeof StreamingSupportedEncodingEnum)[keyof typeof StreamingSupportedEncodingEnum];
|
|
5814
|
+
declare const StreamingSupportedEncodingEnum: {
|
|
5815
|
+
readonly "wav/pcm": "wav/pcm";
|
|
5816
|
+
readonly "wav/alaw": "wav/alaw";
|
|
5817
|
+
readonly "wav/ulaw": "wav/ulaw";
|
|
5818
|
+
};
|
|
5819
|
+
|
|
5820
|
+
/**
|
|
5821
|
+
* Generated by orval v7.9.0 🍺
|
|
5822
|
+
* Do not edit manually.
|
|
5823
|
+
* Gladia Control API
|
|
5824
|
+
* OpenAPI spec version: 1.0
|
|
5825
|
+
*/
|
|
5826
|
+
/**
|
|
5827
|
+
* The sample rate of the audio stream
|
|
5828
|
+
*/
|
|
5829
|
+
type StreamingSupportedSampleRateEnum = (typeof StreamingSupportedSampleRateEnum)[keyof typeof StreamingSupportedSampleRateEnum];
|
|
5830
|
+
declare const StreamingSupportedSampleRateEnum: {
|
|
5831
|
+
readonly NUMBER_8000: 8000;
|
|
5832
|
+
readonly NUMBER_16000: 16000;
|
|
5833
|
+
readonly NUMBER_32000: 32000;
|
|
5834
|
+
readonly NUMBER_44100: 44100;
|
|
5835
|
+
readonly NUMBER_48000: 48000;
|
|
5836
|
+
};
|
|
5837
|
+
|
|
5448
5838
|
/**
|
|
5449
5839
|
* Generated by orval v7.9.0 🍺
|
|
5450
5840
|
* Do not edit manually.
|
|
@@ -5624,7 +6014,7 @@ interface DeepgramStreamingOptions {
|
|
|
5624
6014
|
* Custom topic detection mode
|
|
5625
6015
|
* Use `DeepgramTopicMode` const for autocomplete
|
|
5626
6016
|
*/
|
|
5627
|
-
customTopicMode?:
|
|
6017
|
+
customTopicMode?: DeepgramTopicModeType;
|
|
5628
6018
|
/** Enable intent recognition */
|
|
5629
6019
|
intents?: boolean;
|
|
5630
6020
|
/** Custom intent definitions */
|
|
@@ -5633,7 +6023,7 @@ interface DeepgramStreamingOptions {
|
|
|
5633
6023
|
* Custom intent detection mode
|
|
5634
6024
|
* Use `DeepgramTopicMode` const for autocomplete
|
|
5635
6025
|
*/
|
|
5636
|
-
customIntentMode?:
|
|
6026
|
+
customIntentMode?: DeepgramTopicModeType;
|
|
5637
6027
|
/** Enable summarization */
|
|
5638
6028
|
summarize?: boolean;
|
|
5639
6029
|
/** Custom vocabulary/keywords for boosting */
|
|
@@ -5652,7 +6042,7 @@ interface DeepgramStreamingOptions {
|
|
|
5652
6042
|
* { redact: [DeepgramRedact.pii, DeepgramRedact.pci] }
|
|
5653
6043
|
* ```
|
|
5654
6044
|
*/
|
|
5655
|
-
redact?: boolean |
|
|
6045
|
+
redact?: boolean | DeepgramRedactType[];
|
|
5656
6046
|
/** Callback URL for webhooks */
|
|
5657
6047
|
callback?: string;
|
|
5658
6048
|
/** Extra metadata to include in response */
|
|
@@ -5669,139 +6059,592 @@ interface DeepgramStreamingOptions {
|
|
|
5669
6059
|
vadThreshold?: number;
|
|
5670
6060
|
}
|
|
5671
6061
|
|
|
5672
|
-
/**
|
|
5673
|
-
* AssemblyAI streaming options
|
|
5674
|
-
*
|
|
5675
|
-
* Based on the v3 Universal Streaming API parameters from the AssemblyAI SDK.
|
|
5676
|
-
* Supports advanced features like VAD tuning, end-of-turn detection, and profanity filtering.
|
|
5677
|
-
*
|
|
5678
|
-
* @see https://www.assemblyai.com/docs/speech-to-text/streaming
|
|
6062
|
+
/**
|
|
6063
|
+
* AssemblyAI streaming options
|
|
6064
|
+
*
|
|
6065
|
+
* Based on the v3 Universal Streaming API parameters from the AssemblyAI SDK.
|
|
6066
|
+
* Supports advanced features like VAD tuning, end-of-turn detection, and profanity filtering.
|
|
6067
|
+
*
|
|
6068
|
+
* @see https://www.assemblyai.com/docs/speech-to-text/streaming
|
|
6069
|
+
*/
|
|
6070
|
+
interface AssemblyAIStreamingOptions {
|
|
6071
|
+
/**
|
|
6072
|
+
* Sample rate in Hz
|
|
6073
|
+
* Use `AssemblyAISampleRate` const for autocomplete:
|
|
6074
|
+
* @example
|
|
6075
|
+
* ```typescript
|
|
6076
|
+
* import { AssemblyAISampleRate } from '@meeting-baas/sdk'
|
|
6077
|
+
* { sampleRate: AssemblyAISampleRate.rate16000 }
|
|
6078
|
+
* ```
|
|
6079
|
+
*/
|
|
6080
|
+
sampleRate?: AssemblyAISampleRateType;
|
|
6081
|
+
/**
|
|
6082
|
+
* Audio encoding format
|
|
6083
|
+
* Use `AssemblyAIEncoding` const for autocomplete:
|
|
6084
|
+
* @example
|
|
6085
|
+
* ```typescript
|
|
6086
|
+
* import { AssemblyAIEncoding } from '@meeting-baas/sdk'
|
|
6087
|
+
* { encoding: AssemblyAIEncoding.pcmS16le }
|
|
6088
|
+
* ```
|
|
6089
|
+
*/
|
|
6090
|
+
encoding?: AssemblyAIEncodingType;
|
|
6091
|
+
/**
|
|
6092
|
+
* Speech model to use
|
|
6093
|
+
*
|
|
6094
|
+
* Strict union type - only accepts valid AssemblyAI streaming models.
|
|
6095
|
+
* Use `AssemblyAISpeechModel` const for autocomplete:
|
|
6096
|
+
* @example
|
|
6097
|
+
* ```typescript
|
|
6098
|
+
* import { AssemblyAISpeechModel } from 'voice-router-dev'
|
|
6099
|
+
* { speechModel: AssemblyAISpeechModel.english }
|
|
6100
|
+
* { speechModel: AssemblyAISpeechModel.multilingual }
|
|
6101
|
+
* // Or use string literals directly:
|
|
6102
|
+
* { speechModel: "universal-streaming-english" }
|
|
6103
|
+
* ```
|
|
6104
|
+
*/
|
|
6105
|
+
speechModel?: AssemblyAISpeechModelType;
|
|
6106
|
+
/** Enable automatic language detection */
|
|
6107
|
+
languageDetection?: boolean;
|
|
6108
|
+
/**
|
|
6109
|
+
* Confidence threshold for end-of-turn detection (0-1)
|
|
6110
|
+
* Higher values require more confidence before ending a turn
|
|
6111
|
+
* @default 0.5
|
|
6112
|
+
*/
|
|
6113
|
+
endOfTurnConfidenceThreshold?: number;
|
|
6114
|
+
/**
|
|
6115
|
+
* Minimum silence duration (ms) to trigger end-of-turn when confident
|
|
6116
|
+
* Only applies when confidence is above threshold
|
|
6117
|
+
* @default 1000
|
|
6118
|
+
*/
|
|
6119
|
+
minEndOfTurnSilenceWhenConfident?: number;
|
|
6120
|
+
/**
|
|
6121
|
+
* Maximum silence duration (ms) before forcing end-of-turn
|
|
6122
|
+
* Regardless of confidence level
|
|
6123
|
+
* @default 20000
|
|
6124
|
+
*/
|
|
6125
|
+
maxTurnSilence?: number;
|
|
6126
|
+
/**
|
|
6127
|
+
* VAD sensitivity threshold (0-1)
|
|
6128
|
+
* Lower values are more sensitive to quiet speech
|
|
6129
|
+
*/
|
|
6130
|
+
vadThreshold?: number;
|
|
6131
|
+
/**
|
|
6132
|
+
* Enable real-time text formatting of turns
|
|
6133
|
+
* Applies punctuation, capitalization, and formatting
|
|
6134
|
+
*/
|
|
6135
|
+
formatTurns?: boolean;
|
|
6136
|
+
/** Filter profanity in real-time transcription */
|
|
6137
|
+
filterProfanity?: boolean;
|
|
6138
|
+
/**
|
|
6139
|
+
* Key terms to boost in recognition
|
|
6140
|
+
* Increases recognition accuracy for specific words/phrases
|
|
6141
|
+
*/
|
|
6142
|
+
keyterms?: string[];
|
|
6143
|
+
/**
|
|
6144
|
+
* Key term prompting for context
|
|
6145
|
+
* Provides additional context about the terms to improve recognition
|
|
6146
|
+
*/
|
|
6147
|
+
keytermsPrompt?: string[];
|
|
6148
|
+
/**
|
|
6149
|
+
* Inactivity timeout in milliseconds
|
|
6150
|
+
* Session will close if no audio is received for this duration
|
|
6151
|
+
*/
|
|
6152
|
+
inactivityTimeout?: number;
|
|
6153
|
+
/**
|
|
6154
|
+
* Use token-based authentication
|
|
6155
|
+
* If true, will create a temporary token before connecting
|
|
6156
|
+
*/
|
|
6157
|
+
useToken?: boolean;
|
|
6158
|
+
/**
|
|
6159
|
+
* Token expiration time in seconds (minimum 60)
|
|
6160
|
+
* Only used if useToken is true
|
|
6161
|
+
* @default 3600
|
|
6162
|
+
*/
|
|
6163
|
+
tokenExpiresIn?: number;
|
|
6164
|
+
}
|
|
6165
|
+
/**
|
|
6166
|
+
* AssemblyAI dynamic configuration update
|
|
6167
|
+
* Can be sent mid-stream to adjust parameters
|
|
6168
|
+
*/
|
|
6169
|
+
type AssemblyAIUpdateConfiguration = Omit<StreamingUpdateConfiguration, "type">;
|
|
6170
|
+
/**
|
|
6171
|
+
* Union of all provider-specific streaming options
|
|
6172
|
+
*/
|
|
6173
|
+
type ProviderStreamingOptions = ({
|
|
6174
|
+
provider: "gladia";
|
|
6175
|
+
} & GladiaStreamingOptions) | ({
|
|
6176
|
+
provider: "deepgram";
|
|
6177
|
+
} & DeepgramStreamingOptions) | ({
|
|
6178
|
+
provider: "assemblyai";
|
|
6179
|
+
} & AssemblyAIStreamingOptions);
|
|
6180
|
+
/**
|
|
6181
|
+
* Type-safe streaming options for a specific provider
|
|
6182
|
+
*/
|
|
6183
|
+
type StreamingOptionsForProvider<P extends StreamingProvider> = P extends "gladia" ? GladiaStreamingOptions : P extends "deepgram" ? DeepgramStreamingOptions : P extends "assemblyai" ? AssemblyAIStreamingOptions : never;
|
|
6184
|
+
/**
|
|
6185
|
+
* Type-safe transcribeStream parameters for a specific provider
|
|
6186
|
+
*/
|
|
6187
|
+
interface TranscribeStreamParams<P extends StreamingProvider> {
|
|
6188
|
+
/** Streaming options specific to this provider */
|
|
6189
|
+
options?: StreamingOptionsForProvider<P> & {
|
|
6190
|
+
provider: P;
|
|
6191
|
+
};
|
|
6192
|
+
/** Event callbacks */
|
|
6193
|
+
callbacks?: StreamingCallbacks;
|
|
6194
|
+
}
|
|
6195
|
+
|
|
6196
|
+
/**
|
|
6197
|
+
* Generated by orval v7.9.0 🍺
|
|
6198
|
+
* Do not edit manually.
|
|
6199
|
+
* AssemblyAI API
|
|
6200
|
+
* AssemblyAI API
|
|
6201
|
+
* OpenAPI spec version: 1.3.4
|
|
6202
|
+
*/
|
|
6203
|
+
/**
|
|
6204
|
+
* Get transcripts that were created after this transcript ID
|
|
6205
|
+
*/
|
|
6206
|
+
type AfterId = string;
|
|
6207
|
+
|
|
6208
|
+
/**
|
|
6209
|
+
* Generated by orval v7.9.0 🍺
|
|
6210
|
+
* Do not edit manually.
|
|
6211
|
+
* AssemblyAI API
|
|
6212
|
+
* AssemblyAI API
|
|
6213
|
+
* OpenAPI spec version: 1.3.4
|
|
6214
|
+
*/
|
|
6215
|
+
/**
|
|
6216
|
+
* Get transcripts that were created before this transcript ID
|
|
6217
|
+
*/
|
|
6218
|
+
type BeforeId = string;
|
|
6219
|
+
|
|
6220
|
+
/**
|
|
6221
|
+
* Generated by orval v7.9.0 🍺
|
|
6222
|
+
* Do not edit manually.
|
|
6223
|
+
* AssemblyAI API
|
|
6224
|
+
* AssemblyAI API
|
|
6225
|
+
* OpenAPI spec version: 1.3.4
|
|
6226
|
+
*/
|
|
6227
|
+
/**
|
|
6228
|
+
* Only get transcripts created on this date
|
|
6229
|
+
*/
|
|
6230
|
+
type CreatedOn = string;
|
|
6231
|
+
|
|
6232
|
+
/**
|
|
6233
|
+
* Generated by orval v7.9.0 🍺
|
|
6234
|
+
* Do not edit manually.
|
|
6235
|
+
* AssemblyAI API
|
|
6236
|
+
* AssemblyAI API
|
|
6237
|
+
* OpenAPI spec version: 1.3.4
|
|
6238
|
+
*/
|
|
6239
|
+
/**
|
|
6240
|
+
* Maximum amount of transcripts to retrieve
|
|
6241
|
+
*/
|
|
6242
|
+
type Limit = number;
|
|
6243
|
+
|
|
6244
|
+
/**
|
|
6245
|
+
* Generated by orval v7.9.0 🍺
|
|
6246
|
+
* Do not edit manually.
|
|
6247
|
+
* AssemblyAI API
|
|
6248
|
+
* AssemblyAI API
|
|
6249
|
+
* OpenAPI spec version: 1.3.4
|
|
6250
|
+
*/
|
|
6251
|
+
/**
|
|
6252
|
+
* Only get throttled transcripts, overrides the status filter
|
|
6253
|
+
*/
|
|
6254
|
+
type ThrottledOnly = boolean;
|
|
6255
|
+
|
|
6256
|
+
/**
|
|
6257
|
+
* Generated by orval v7.9.0 🍺
|
|
6258
|
+
* Do not edit manually.
|
|
6259
|
+
* AssemblyAI API
|
|
6260
|
+
* AssemblyAI API
|
|
6261
|
+
* OpenAPI spec version: 1.3.4
|
|
6262
|
+
*/
|
|
6263
|
+
|
|
6264
|
+
type ListTranscriptsParams = {
|
|
6265
|
+
/**
|
|
6266
|
+
* Maximum amount of transcripts to retrieve
|
|
6267
|
+
*/
|
|
6268
|
+
limit?: Limit;
|
|
6269
|
+
/**
|
|
6270
|
+
* Filter by transcript status
|
|
6271
|
+
*/
|
|
6272
|
+
status?: TranscriptStatus;
|
|
6273
|
+
/**
|
|
6274
|
+
* Only get transcripts created on this date
|
|
6275
|
+
*/
|
|
6276
|
+
created_on?: CreatedOn;
|
|
6277
|
+
/**
|
|
6278
|
+
* Get transcripts that were created before this transcript ID
|
|
6279
|
+
*/
|
|
6280
|
+
before_id?: BeforeId;
|
|
6281
|
+
/**
|
|
6282
|
+
* Get transcripts that were created after this transcript ID
|
|
6283
|
+
*/
|
|
6284
|
+
after_id?: AfterId;
|
|
6285
|
+
/**
|
|
6286
|
+
* Only get throttled transcripts, overrides the status filter
|
|
6287
|
+
*/
|
|
6288
|
+
throttled_only?: ThrottledOnly;
|
|
6289
|
+
};
|
|
6290
|
+
|
|
6291
|
+
/**
|
|
6292
|
+
* Generated by orval v7.9.0 🍺
|
|
6293
|
+
* Do not edit manually.
|
|
6294
|
+
* Gladia Control API
|
|
6295
|
+
* OpenAPI spec version: 1.0
|
|
6296
|
+
*/
|
|
6297
|
+
type TranscriptionControllerListV2KindItem = (typeof TranscriptionControllerListV2KindItem)[keyof typeof TranscriptionControllerListV2KindItem];
|
|
6298
|
+
declare const TranscriptionControllerListV2KindItem: {
|
|
6299
|
+
readonly "pre-recorded": "pre-recorded";
|
|
6300
|
+
readonly live: "live";
|
|
6301
|
+
};
|
|
6302
|
+
|
|
6303
|
+
/**
|
|
6304
|
+
* Generated by orval v7.9.0 🍺
|
|
6305
|
+
* Do not edit manually.
|
|
6306
|
+
* Gladia Control API
|
|
6307
|
+
* OpenAPI spec version: 1.0
|
|
6308
|
+
*/
|
|
6309
|
+
type TranscriptionControllerListV2StatusItem = (typeof TranscriptionControllerListV2StatusItem)[keyof typeof TranscriptionControllerListV2StatusItem];
|
|
6310
|
+
declare const TranscriptionControllerListV2StatusItem: {
|
|
6311
|
+
readonly queued: "queued";
|
|
6312
|
+
readonly processing: "processing";
|
|
6313
|
+
readonly done: "done";
|
|
6314
|
+
readonly error: "error";
|
|
6315
|
+
};
|
|
6316
|
+
|
|
6317
|
+
/**
|
|
6318
|
+
* Generated by orval v7.9.0 🍺
|
|
6319
|
+
* Do not edit manually.
|
|
6320
|
+
* Gladia Control API
|
|
6321
|
+
* OpenAPI spec version: 1.0
|
|
6322
|
+
*/
|
|
6323
|
+
|
|
6324
|
+
type TranscriptionControllerListV2Params = {
|
|
6325
|
+
/**
|
|
6326
|
+
* The starting point for pagination. A value of 0 starts from the first item.
|
|
6327
|
+
*/
|
|
6328
|
+
offset?: number;
|
|
6329
|
+
/**
|
|
6330
|
+
* The maximum number of items to return. Useful for pagination and controlling data payload size.
|
|
6331
|
+
*/
|
|
6332
|
+
limit?: number;
|
|
6333
|
+
/**
|
|
6334
|
+
* Filter items relevant to a specific date in ISO format (YYYY-MM-DD).
|
|
6335
|
+
*/
|
|
6336
|
+
date?: string;
|
|
6337
|
+
/**
|
|
6338
|
+
* Include items that occurred before the specified date in ISO format.
|
|
6339
|
+
*/
|
|
6340
|
+
before_date?: string;
|
|
6341
|
+
/**
|
|
6342
|
+
* Filter for items after the specified date. Use with `before_date` for a range. Date in ISO format.
|
|
6343
|
+
*/
|
|
6344
|
+
after_date?: string;
|
|
6345
|
+
/**
|
|
6346
|
+
* Filter the list based on item status. Accepts multiple values from the predefined list.
|
|
6347
|
+
*/
|
|
6348
|
+
status?: TranscriptionControllerListV2StatusItem[];
|
|
6349
|
+
custom_metadata?: {
|
|
6350
|
+
[key: string]: unknown;
|
|
6351
|
+
};
|
|
6352
|
+
/**
|
|
6353
|
+
* Filter the list based on the item type. Supports multiple values from the predefined list.
|
|
6354
|
+
*/
|
|
6355
|
+
kind?: TranscriptionControllerListV2KindItem[];
|
|
6356
|
+
};
|
|
6357
|
+
|
|
6358
|
+
/**
|
|
6359
|
+
* Generated by orval v7.9.0 🍺
|
|
6360
|
+
* Do not edit manually.
|
|
6361
|
+
* Deepgram API Specification
|
|
6362
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6363
|
+
|
|
6364
|
+
* OpenAPI spec version: 1.0.0
|
|
6365
|
+
*/
|
|
6366
|
+
/**
|
|
6367
|
+
* Start date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
|
|
6368
|
+
*/
|
|
6369
|
+
type ManageV1StartDateTimeParameter = string;
|
|
6370
|
+
|
|
6371
|
+
/**
|
|
6372
|
+
* Generated by orval v7.9.0 🍺
|
|
6373
|
+
* Do not edit manually.
|
|
6374
|
+
* Deepgram API Specification
|
|
6375
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6376
|
+
|
|
6377
|
+
* OpenAPI spec version: 1.0.0
|
|
6378
|
+
*/
|
|
6379
|
+
/**
|
|
6380
|
+
* End date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
|
|
6381
|
+
*/
|
|
6382
|
+
type ManageV1EndDateTimeParameter = string;
|
|
6383
|
+
|
|
6384
|
+
/**
|
|
6385
|
+
* Generated by orval v7.9.0 🍺
|
|
6386
|
+
* Do not edit manually.
|
|
6387
|
+
* Deepgram API Specification
|
|
6388
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6389
|
+
|
|
6390
|
+
* OpenAPI spec version: 1.0.0
|
|
6391
|
+
*/
|
|
6392
|
+
/**
|
|
6393
|
+
* Number of results to return per page. Default 10. Range [1,1000]
|
|
6394
|
+
*/
|
|
6395
|
+
type ManageV1LimitParameter = number;
|
|
6396
|
+
|
|
6397
|
+
/**
|
|
6398
|
+
* Generated by orval v7.9.0 🍺
|
|
6399
|
+
* Do not edit manually.
|
|
6400
|
+
* Deepgram API Specification
|
|
6401
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6402
|
+
|
|
6403
|
+
* OpenAPI spec version: 1.0.0
|
|
6404
|
+
*/
|
|
6405
|
+
/**
|
|
6406
|
+
* Navigate and return the results to retrieve specific portions of information of the response
|
|
6407
|
+
*/
|
|
6408
|
+
type ManageV1PageParameter = number;
|
|
6409
|
+
|
|
6410
|
+
/**
|
|
6411
|
+
* Generated by orval v7.9.0 🍺
|
|
6412
|
+
* Do not edit manually.
|
|
6413
|
+
* Deepgram API Specification
|
|
6414
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6415
|
+
|
|
6416
|
+
* OpenAPI spec version: 1.0.0
|
|
6417
|
+
*/
|
|
6418
|
+
/**
|
|
6419
|
+
* Filter for requests where a specific accessor was used
|
|
6420
|
+
*/
|
|
6421
|
+
type ManageV1FilterAccessorParameter = string;
|
|
6422
|
+
|
|
6423
|
+
/**
|
|
6424
|
+
* Generated by orval v7.9.0 🍺
|
|
6425
|
+
* Do not edit manually.
|
|
6426
|
+
* Deepgram API Specification
|
|
6427
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6428
|
+
|
|
6429
|
+
* OpenAPI spec version: 1.0.0
|
|
6430
|
+
*/
|
|
6431
|
+
/**
|
|
6432
|
+
* Filter for a specific request id
|
|
6433
|
+
*/
|
|
6434
|
+
type ManageV1FilterRequestIdParameter = string;
|
|
6435
|
+
|
|
6436
|
+
/**
|
|
6437
|
+
* Generated by orval v7.9.0 🍺
|
|
6438
|
+
* Do not edit manually.
|
|
6439
|
+
* Deepgram API Specification
|
|
6440
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6441
|
+
|
|
6442
|
+
* OpenAPI spec version: 1.0.0
|
|
6443
|
+
*/
|
|
6444
|
+
/**
|
|
6445
|
+
* Deployment type for the requests
|
|
6446
|
+
*/
|
|
6447
|
+
/**
|
|
6448
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6449
|
+
*/
|
|
6450
|
+
/**
|
|
6451
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6452
|
+
*/
|
|
6453
|
+
/**
|
|
6454
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6455
|
+
*/
|
|
6456
|
+
/**
|
|
6457
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6458
|
+
*/
|
|
6459
|
+
/**
|
|
6460
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6461
|
+
*/
|
|
6462
|
+
/**
|
|
6463
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6464
|
+
*/
|
|
6465
|
+
/**
|
|
6466
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6467
|
+
*/
|
|
6468
|
+
/**
|
|
6469
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
6470
|
+
*/
|
|
6471
|
+
type ManageV1FilterDeploymentParameter = typeof ManageV1FilterDeploymentParameter[keyof typeof ManageV1FilterDeploymentParameter];
|
|
6472
|
+
declare const ManageV1FilterDeploymentParameter: {
|
|
6473
|
+
readonly hosted: "hosted";
|
|
6474
|
+
readonly beta: "beta";
|
|
6475
|
+
};
|
|
6476
|
+
|
|
6477
|
+
/**
|
|
6478
|
+
* Generated by orval v7.9.0 🍺
|
|
6479
|
+
* Do not edit manually.
|
|
6480
|
+
* Deepgram API Specification
|
|
6481
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6482
|
+
|
|
6483
|
+
* OpenAPI spec version: 1.0.0
|
|
6484
|
+
*/
|
|
6485
|
+
/**
|
|
6486
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6487
|
+
*/
|
|
6488
|
+
/**
|
|
6489
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6490
|
+
*/
|
|
6491
|
+
/**
|
|
6492
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6493
|
+
*/
|
|
6494
|
+
/**
|
|
6495
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6496
|
+
*/
|
|
6497
|
+
/**
|
|
6498
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6499
|
+
*/
|
|
6500
|
+
/**
|
|
6501
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6502
|
+
*/
|
|
6503
|
+
/**
|
|
6504
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6505
|
+
*/
|
|
6506
|
+
/**
|
|
6507
|
+
* ManageV1FilterEndpointParameter type definition
|
|
6508
|
+
*/
|
|
6509
|
+
type ManageV1FilterEndpointParameter = typeof ManageV1FilterEndpointParameter[keyof typeof ManageV1FilterEndpointParameter];
|
|
6510
|
+
declare const ManageV1FilterEndpointParameter: {
|
|
6511
|
+
readonly listen: "listen";
|
|
6512
|
+
readonly read: "read";
|
|
6513
|
+
readonly speak: "speak";
|
|
6514
|
+
readonly agent: "agent";
|
|
6515
|
+
};
|
|
6516
|
+
|
|
6517
|
+
/**
|
|
6518
|
+
* Generated by orval v7.9.0 🍺
|
|
6519
|
+
* Do not edit manually.
|
|
6520
|
+
* Deepgram API Specification
|
|
6521
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6522
|
+
|
|
6523
|
+
* OpenAPI spec version: 1.0.0
|
|
6524
|
+
*/
|
|
6525
|
+
/**
|
|
6526
|
+
* Method type for the request
|
|
6527
|
+
*/
|
|
6528
|
+
/**
|
|
6529
|
+
* ManageV1FilterMethodParameter type definition
|
|
6530
|
+
*/
|
|
6531
|
+
/**
|
|
6532
|
+
* ManageV1FilterMethodParameter type definition
|
|
6533
|
+
*/
|
|
6534
|
+
/**
|
|
6535
|
+
* ManageV1FilterMethodParameter type definition
|
|
6536
|
+
*/
|
|
6537
|
+
/**
|
|
6538
|
+
* ManageV1FilterMethodParameter type definition
|
|
6539
|
+
*/
|
|
6540
|
+
/**
|
|
6541
|
+
* ManageV1FilterMethodParameter type definition
|
|
6542
|
+
*/
|
|
6543
|
+
/**
|
|
6544
|
+
* ManageV1FilterMethodParameter type definition
|
|
6545
|
+
*/
|
|
6546
|
+
/**
|
|
6547
|
+
* ManageV1FilterMethodParameter type definition
|
|
6548
|
+
*/
|
|
6549
|
+
/**
|
|
6550
|
+
* ManageV1FilterMethodParameter type definition
|
|
6551
|
+
*/
|
|
6552
|
+
type ManageV1FilterMethodParameter = typeof ManageV1FilterMethodParameter[keyof typeof ManageV1FilterMethodParameter];
|
|
6553
|
+
declare const ManageV1FilterMethodParameter: {
|
|
6554
|
+
readonly sync: "sync";
|
|
6555
|
+
readonly async: "async";
|
|
6556
|
+
readonly streaming: "streaming";
|
|
6557
|
+
};
|
|
6558
|
+
|
|
6559
|
+
/**
|
|
6560
|
+
* Generated by orval v7.9.0 🍺
|
|
6561
|
+
* Do not edit manually.
|
|
6562
|
+
* Deepgram API Specification
|
|
6563
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6564
|
+
|
|
6565
|
+
* OpenAPI spec version: 1.0.0
|
|
6566
|
+
*/
|
|
6567
|
+
/**
|
|
6568
|
+
* ManageV1FilterStatusParameter type definition
|
|
6569
|
+
*/
|
|
6570
|
+
/**
|
|
6571
|
+
* ManageV1FilterStatusParameter type definition
|
|
6572
|
+
*/
|
|
6573
|
+
/**
|
|
6574
|
+
* ManageV1FilterStatusParameter type definition
|
|
6575
|
+
*/
|
|
6576
|
+
/**
|
|
6577
|
+
* ManageV1FilterStatusParameter type definition
|
|
6578
|
+
*/
|
|
6579
|
+
/**
|
|
6580
|
+
* ManageV1FilterStatusParameter type definition
|
|
6581
|
+
*/
|
|
6582
|
+
/**
|
|
6583
|
+
* ManageV1FilterStatusParameter type definition
|
|
6584
|
+
*/
|
|
6585
|
+
/**
|
|
6586
|
+
* ManageV1FilterStatusParameter type definition
|
|
6587
|
+
*/
|
|
6588
|
+
/**
|
|
6589
|
+
* ManageV1FilterStatusParameter type definition
|
|
6590
|
+
*/
|
|
6591
|
+
type ManageV1FilterStatusParameter = typeof ManageV1FilterStatusParameter[keyof typeof ManageV1FilterStatusParameter];
|
|
6592
|
+
declare const ManageV1FilterStatusParameter: {
|
|
6593
|
+
readonly succeeded: "succeeded";
|
|
6594
|
+
readonly failed: "failed";
|
|
6595
|
+
};
|
|
6596
|
+
|
|
6597
|
+
/**
|
|
6598
|
+
* Generated by orval v7.9.0 🍺
|
|
6599
|
+
* Do not edit manually.
|
|
6600
|
+
* Deepgram API Specification
|
|
6601
|
+
* APIs for speech-to-text transcription, text-to-speech synthesis, language understanding, and account management.
|
|
6602
|
+
|
|
6603
|
+
* OpenAPI spec version: 1.0.0
|
|
5679
6604
|
*/
|
|
5680
|
-
|
|
5681
|
-
|
|
5682
|
-
* Sample rate in Hz
|
|
5683
|
-
* Use `AssemblyAISampleRate` const for autocomplete:
|
|
5684
|
-
* @example
|
|
5685
|
-
* ```typescript
|
|
5686
|
-
* import { AssemblyAISampleRate } from '@meeting-baas/sdk'
|
|
5687
|
-
* { sampleRate: AssemblyAISampleRate.rate16000 }
|
|
5688
|
-
* ```
|
|
5689
|
-
*/
|
|
5690
|
-
sampleRate?: AssemblyAISampleRateType;
|
|
5691
|
-
/**
|
|
5692
|
-
* Audio encoding format
|
|
5693
|
-
* Use `AssemblyAIEncoding` const for autocomplete:
|
|
5694
|
-
* @example
|
|
5695
|
-
* ```typescript
|
|
5696
|
-
* import { AssemblyAIEncoding } from '@meeting-baas/sdk'
|
|
5697
|
-
* { encoding: AssemblyAIEncoding.pcmS16le }
|
|
5698
|
-
* ```
|
|
5699
|
-
*/
|
|
5700
|
-
encoding?: AssemblyAIEncodingType;
|
|
5701
|
-
/**
|
|
5702
|
-
* Speech model to use
|
|
5703
|
-
*
|
|
5704
|
-
* Strict union type - only accepts valid AssemblyAI streaming models.
|
|
5705
|
-
* Use `AssemblyAISpeechModel` const for autocomplete:
|
|
5706
|
-
* @example
|
|
5707
|
-
* ```typescript
|
|
5708
|
-
* import { AssemblyAISpeechModel } from 'voice-router-dev'
|
|
5709
|
-
* { speechModel: AssemblyAISpeechModel.english }
|
|
5710
|
-
* { speechModel: AssemblyAISpeechModel.multilingual }
|
|
5711
|
-
* // Or use string literals directly:
|
|
5712
|
-
* { speechModel: "universal-streaming-english" }
|
|
5713
|
-
* ```
|
|
5714
|
-
*/
|
|
5715
|
-
speechModel?: AssemblyAISpeechModelType;
|
|
5716
|
-
/** Enable automatic language detection */
|
|
5717
|
-
languageDetection?: boolean;
|
|
6605
|
+
|
|
6606
|
+
type ManageV1ProjectsRequestsListParams = {
|
|
5718
6607
|
/**
|
|
5719
|
-
*
|
|
5720
|
-
* Higher values require more confidence before ending a turn
|
|
5721
|
-
* @default 0.5
|
|
6608
|
+
* Start date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
|
|
5722
6609
|
*/
|
|
5723
|
-
|
|
6610
|
+
start?: ManageV1StartDateTimeParameter;
|
|
5724
6611
|
/**
|
|
5725
|
-
*
|
|
5726
|
-
* Only applies when confidence is above threshold
|
|
5727
|
-
* @default 1000
|
|
6612
|
+
* End date of the requested date range. Formats accepted are YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, or YYYY-MM-DDTHH:MM:SS+HH:MM
|
|
5728
6613
|
*/
|
|
5729
|
-
|
|
6614
|
+
end?: ManageV1EndDateTimeParameter;
|
|
5730
6615
|
/**
|
|
5731
|
-
*
|
|
5732
|
-
* Regardless of confidence level
|
|
5733
|
-
* @default 20000
|
|
6616
|
+
* Number of results to return per page. Default 10. Range [1,1000]
|
|
5734
6617
|
*/
|
|
5735
|
-
|
|
6618
|
+
limit?: ManageV1LimitParameter;
|
|
5736
6619
|
/**
|
|
5737
|
-
*
|
|
5738
|
-
* Lower values are more sensitive to quiet speech
|
|
6620
|
+
* Navigate and return the results to retrieve specific portions of information of the response
|
|
5739
6621
|
*/
|
|
5740
|
-
|
|
6622
|
+
page?: ManageV1PageParameter;
|
|
5741
6623
|
/**
|
|
5742
|
-
*
|
|
5743
|
-
* Applies punctuation, capitalization, and formatting
|
|
6624
|
+
* Filter for requests where a specific accessor was used
|
|
5744
6625
|
*/
|
|
5745
|
-
|
|
5746
|
-
/** Filter profanity in real-time transcription */
|
|
5747
|
-
filterProfanity?: boolean;
|
|
6626
|
+
accessor?: ManageV1FilterAccessorParameter;
|
|
5748
6627
|
/**
|
|
5749
|
-
*
|
|
5750
|
-
* Increases recognition accuracy for specific words/phrases
|
|
6628
|
+
* Filter for a specific request id
|
|
5751
6629
|
*/
|
|
5752
|
-
|
|
6630
|
+
request_id?: ManageV1FilterRequestIdParameter;
|
|
5753
6631
|
/**
|
|
5754
|
-
*
|
|
5755
|
-
* Provides additional context about the terms to improve recognition
|
|
6632
|
+
* Filter for requests where a specific deployment was used
|
|
5756
6633
|
*/
|
|
5757
|
-
|
|
6634
|
+
deployment?: ManageV1FilterDeploymentParameter;
|
|
5758
6635
|
/**
|
|
5759
|
-
*
|
|
5760
|
-
* Session will close if no audio is received for this duration
|
|
6636
|
+
* Filter for requests where a specific endpoint was used
|
|
5761
6637
|
*/
|
|
5762
|
-
|
|
6638
|
+
endpoint?: ManageV1FilterEndpointParameter;
|
|
5763
6639
|
/**
|
|
5764
|
-
*
|
|
5765
|
-
* If true, will create a temporary token before connecting
|
|
6640
|
+
* Filter for requests where a specific method was used
|
|
5766
6641
|
*/
|
|
5767
|
-
|
|
6642
|
+
method?: ManageV1FilterMethodParameter;
|
|
5768
6643
|
/**
|
|
5769
|
-
*
|
|
5770
|
-
* Only used if useToken is true
|
|
5771
|
-
* @default 3600
|
|
6644
|
+
* Filter for requests that succeeded (status code < 300) or failed (status code >=400)
|
|
5772
6645
|
*/
|
|
5773
|
-
|
|
5774
|
-
}
|
|
5775
|
-
/**
|
|
5776
|
-
* AssemblyAI dynamic configuration update
|
|
5777
|
-
* Can be sent mid-stream to adjust parameters
|
|
5778
|
-
*/
|
|
5779
|
-
type AssemblyAIUpdateConfiguration = Omit<StreamingUpdateConfiguration, "type">;
|
|
5780
|
-
/**
|
|
5781
|
-
* Union of all provider-specific streaming options
|
|
5782
|
-
*/
|
|
5783
|
-
type ProviderStreamingOptions = ({
|
|
5784
|
-
provider: "gladia";
|
|
5785
|
-
} & GladiaStreamingOptions) | ({
|
|
5786
|
-
provider: "deepgram";
|
|
5787
|
-
} & DeepgramStreamingOptions) | ({
|
|
5788
|
-
provider: "assemblyai";
|
|
5789
|
-
} & AssemblyAIStreamingOptions);
|
|
5790
|
-
/**
|
|
5791
|
-
* Type-safe streaming options for a specific provider
|
|
5792
|
-
*/
|
|
5793
|
-
type StreamingOptionsForProvider<P extends StreamingProvider> = P extends "gladia" ? GladiaStreamingOptions : P extends "deepgram" ? DeepgramStreamingOptions : P extends "assemblyai" ? AssemblyAIStreamingOptions : never;
|
|
5794
|
-
/**
|
|
5795
|
-
* Type-safe transcribeStream parameters for a specific provider
|
|
5796
|
-
*/
|
|
5797
|
-
interface TranscribeStreamParams<P extends StreamingProvider> {
|
|
5798
|
-
/** Streaming options specific to this provider */
|
|
5799
|
-
options?: StreamingOptionsForProvider<P> & {
|
|
5800
|
-
provider: P;
|
|
5801
|
-
};
|
|
5802
|
-
/** Event callbacks */
|
|
5803
|
-
callbacks?: StreamingCallbacks;
|
|
5804
|
-
}
|
|
6646
|
+
status?: ManageV1FilterStatusParameter;
|
|
6647
|
+
};
|
|
5805
6648
|
|
|
5806
6649
|
/**
|
|
5807
6650
|
* Unified types for the Voice Router SDK
|
|
@@ -5816,15 +6659,20 @@ type SpeechmaticsOperatingPoint = "standard" | "enhanced";
|
|
|
5816
6659
|
/**
|
|
5817
6660
|
* Unified transcription model type with autocomplete for all providers
|
|
5818
6661
|
*
|
|
5819
|
-
*
|
|
5820
|
-
* - Deepgram: nova-3, nova-2, enhanced, base,
|
|
6662
|
+
* Strict union type - only accepts valid models from each provider:
|
|
6663
|
+
* - Deepgram: nova-3, nova-2, enhanced, base, etc.
|
|
5821
6664
|
* - AssemblyAI: best, slam-1, universal
|
|
5822
6665
|
* - Gladia: solaria-1
|
|
5823
6666
|
* - Speechmatics: standard, enhanced
|
|
5824
6667
|
*
|
|
5825
|
-
*
|
|
6668
|
+
* Use provider const objects for autocomplete:
|
|
6669
|
+
* @example
|
|
6670
|
+
* ```typescript
|
|
6671
|
+
* import { DeepgramModel } from 'voice-router-dev'
|
|
6672
|
+
* { model: DeepgramModel["nova-3"] }
|
|
6673
|
+
* ```
|
|
5826
6674
|
*/
|
|
5827
|
-
type TranscriptionModel =
|
|
6675
|
+
type TranscriptionModel = DeepgramModelType | StreamingSupportedModels | SpeechModel | SpeechmaticsOperatingPoint;
|
|
5828
6676
|
/**
|
|
5829
6677
|
* Unified transcription language type with autocomplete for all providers
|
|
5830
6678
|
*
|
|
@@ -5944,6 +6792,60 @@ interface ProviderCapabilities {
|
|
|
5944
6792
|
entityDetection: boolean;
|
|
5945
6793
|
/** PII redaction */
|
|
5946
6794
|
piiRedaction: boolean;
|
|
6795
|
+
/** List/fetch previous transcriptions */
|
|
6796
|
+
listTranscripts: boolean;
|
|
6797
|
+
/** Delete existing transcriptions */
|
|
6798
|
+
deleteTranscript: boolean;
|
|
6799
|
+
/** Download original audio file from transcription (only Gladia supports this) */
|
|
6800
|
+
getAudioFile?: boolean;
|
|
6801
|
+
}
|
|
6802
|
+
|
|
6803
|
+
/**
|
|
6804
|
+
* Options for listing transcripts with date/time filtering
|
|
6805
|
+
*
|
|
6806
|
+
* Providers support different filtering capabilities:
|
|
6807
|
+
* - AssemblyAI: status, created_on, before_id, after_id, throttled_only
|
|
6808
|
+
* - Gladia: status, date, before_date, after_date, custom_metadata
|
|
6809
|
+
* - Azure: status, skip, top, filter (OData)
|
|
6810
|
+
* - Deepgram: start, end, status, page, request_id, endpoint (requires projectId)
|
|
6811
|
+
*
|
|
6812
|
+
* @example Filter by date
|
|
6813
|
+
* ```typescript
|
|
6814
|
+
* await adapter.listTranscripts({
|
|
6815
|
+
* date: '2026-01-07', // Exact date (ISO format)
|
|
6816
|
+
* status: 'completed',
|
|
6817
|
+
* limit: 50
|
|
6818
|
+
* })
|
|
6819
|
+
* ```
|
|
6820
|
+
*
|
|
6821
|
+
* @example Filter by date range
|
|
6822
|
+
* ```typescript
|
|
6823
|
+
* await adapter.listTranscripts({
|
|
6824
|
+
* afterDate: '2026-01-01',
|
|
6825
|
+
* beforeDate: '2026-01-31',
|
|
6826
|
+
* limit: 100
|
|
6827
|
+
* })
|
|
6828
|
+
* ```
|
|
6829
|
+
*/
|
|
6830
|
+
interface ListTranscriptsOptions {
|
|
6831
|
+
/** Maximum number of transcripts to retrieve */
|
|
6832
|
+
limit?: number;
|
|
6833
|
+
/** Pagination offset (skip N results) */
|
|
6834
|
+
offset?: number;
|
|
6835
|
+
/** Filter by transcript status */
|
|
6836
|
+
status?: "queued" | "processing" | "completed" | "error" | string;
|
|
6837
|
+
/** Filter by exact date (ISO format: YYYY-MM-DD) */
|
|
6838
|
+
date?: string;
|
|
6839
|
+
/** Filter for transcripts created before this date (ISO format) */
|
|
6840
|
+
beforeDate?: string;
|
|
6841
|
+
/** Filter for transcripts created after this date (ISO format) */
|
|
6842
|
+
afterDate?: string;
|
|
6843
|
+
/** AssemblyAI-specific list options */
|
|
6844
|
+
assemblyai?: Partial<ListTranscriptsParams>;
|
|
6845
|
+
/** Gladia-specific list options */
|
|
6846
|
+
gladia?: Partial<TranscriptionControllerListV2Params>;
|
|
6847
|
+
/** Deepgram-specific list options (request history) */
|
|
6848
|
+
deepgram?: Partial<ManageV1ProjectsRequestsListParams>;
|
|
5947
6849
|
}
|
|
5948
6850
|
/**
|
|
5949
6851
|
* Audio input for transcription
|
|
@@ -6091,6 +6993,135 @@ interface Utterance {
|
|
|
6091
6993
|
* Transcription status
|
|
6092
6994
|
*/
|
|
6093
6995
|
type TranscriptionStatus = "queued" | "processing" | "completed" | "error";
|
|
6996
|
+
/**
|
|
6997
|
+
* Transcript metadata with typed common fields
|
|
6998
|
+
*
|
|
6999
|
+
* Contains provider-agnostic metadata fields that are commonly available.
|
|
7000
|
+
* Provider-specific fields can be accessed via the index signature.
|
|
7001
|
+
*
|
|
7002
|
+
* @example
|
|
7003
|
+
* ```typescript
|
|
7004
|
+
* const { transcripts } = await router.listTranscripts('assemblyai', { limit: 20 });
|
|
7005
|
+
* transcripts.forEach(item => {
|
|
7006
|
+
* console.log(item.data?.metadata?.audioUrl); // string | undefined
|
|
7007
|
+
* console.log(item.data?.metadata?.createdAt); // string | undefined
|
|
7008
|
+
* console.log(item.data?.metadata?.audioDuration); // number | undefined
|
|
7009
|
+
* });
|
|
7010
|
+
* ```
|
|
7011
|
+
*/
|
|
7012
|
+
interface TranscriptMetadata {
|
|
7013
|
+
/**
|
|
7014
|
+
* Original audio URL/source you provided to the API (echoed back).
|
|
7015
|
+
* This is NOT a provider-hosted URL - it's what you sent when creating the transcription.
|
|
7016
|
+
*/
|
|
7017
|
+
sourceAudioUrl?: string;
|
|
7018
|
+
/**
|
|
7019
|
+
* True if the provider stored the audio and it can be downloaded via adapter.getAudioFile().
|
|
7020
|
+
* Currently only Gladia supports this - other providers discard audio after processing.
|
|
7021
|
+
*
|
|
7022
|
+
* @example
|
|
7023
|
+
* ```typescript
|
|
7024
|
+
* if (item.data?.metadata?.audioFileAvailable) {
|
|
7025
|
+
* const audio = await gladiaAdapter.getAudioFile(item.data.id)
|
|
7026
|
+
* // audio.data is a Blob
|
|
7027
|
+
* }
|
|
7028
|
+
* ```
|
|
7029
|
+
*/
|
|
7030
|
+
audioFileAvailable?: boolean;
|
|
7031
|
+
/** Resource URL for the transcript */
|
|
7032
|
+
resourceUrl?: string;
|
|
7033
|
+
/** Creation timestamp (ISO 8601) */
|
|
7034
|
+
createdAt?: string;
|
|
7035
|
+
/** Completion timestamp (ISO 8601) */
|
|
7036
|
+
completedAt?: string;
|
|
7037
|
+
/** Last action timestamp (Azure) */
|
|
7038
|
+
lastActionAt?: string;
|
|
7039
|
+
/** Audio duration in seconds */
|
|
7040
|
+
audioDuration?: number;
|
|
7041
|
+
/** Transcript type */
|
|
7042
|
+
kind?: "batch" | "pre-recorded" | "live" | "streaming";
|
|
7043
|
+
/** Display name (Azure) */
|
|
7044
|
+
displayName?: string;
|
|
7045
|
+
/** Files URL (Azure) */
|
|
7046
|
+
filesUrl?: string;
|
|
7047
|
+
/** Custom metadata (Gladia) */
|
|
7048
|
+
customMetadata?: Record<string, unknown>;
|
|
7049
|
+
/** Provider-specific fields */
|
|
7050
|
+
[key: string]: unknown;
|
|
7051
|
+
}
|
|
7052
|
+
/**
|
|
7053
|
+
* Transcript data structure
|
|
7054
|
+
*
|
|
7055
|
+
* Contains the core transcript information returned by getTranscript and listTranscripts.
|
|
7056
|
+
*
|
|
7057
|
+
* @example
|
|
7058
|
+
* ```typescript
|
|
7059
|
+
* const result = await router.getTranscript('abc123', 'assemblyai');
|
|
7060
|
+
* if (result.success && result.data) {
|
|
7061
|
+
* console.log(result.data.id); // string
|
|
7062
|
+
* console.log(result.data.text); // string
|
|
7063
|
+
* console.log(result.data.status); // TranscriptionStatus
|
|
7064
|
+
* console.log(result.data.metadata); // TranscriptMetadata
|
|
7065
|
+
* }
|
|
7066
|
+
* ```
|
|
7067
|
+
*/
|
|
7068
|
+
interface TranscriptData {
|
|
7069
|
+
/** Unique transcript ID */
|
|
7070
|
+
id: string;
|
|
7071
|
+
/** Full transcribed text (empty for list items) */
|
|
7072
|
+
text: string;
|
|
7073
|
+
/** Overall confidence score (0-1) */
|
|
7074
|
+
confidence?: number;
|
|
7075
|
+
/** Transcription status */
|
|
7076
|
+
status: TranscriptionStatus;
|
|
7077
|
+
/** Detected or specified language code */
|
|
7078
|
+
language?: string;
|
|
7079
|
+
/** Audio duration in seconds */
|
|
7080
|
+
duration?: number;
|
|
7081
|
+
/** Speaker diarization results */
|
|
7082
|
+
speakers?: Speaker[];
|
|
7083
|
+
/** Word-level transcription with timestamps */
|
|
7084
|
+
words?: Word[];
|
|
7085
|
+
/** Utterances (speaker turns) */
|
|
7086
|
+
utterances?: Utterance[];
|
|
7087
|
+
/** Summary of the content (if summarization enabled) */
|
|
7088
|
+
summary?: string;
|
|
7089
|
+
/** Transcript metadata */
|
|
7090
|
+
metadata?: TranscriptMetadata;
|
|
7091
|
+
/** Creation timestamp (shorthand for metadata.createdAt) */
|
|
7092
|
+
createdAt?: string;
|
|
7093
|
+
/** Completion timestamp (shorthand for metadata.completedAt) */
|
|
7094
|
+
completedAt?: string;
|
|
7095
|
+
}
|
|
7096
|
+
/**
|
|
7097
|
+
* Response from listTranscripts
|
|
7098
|
+
*
|
|
7099
|
+
* @example
|
|
7100
|
+
* ```typescript
|
|
7101
|
+
* import type { ListTranscriptsResponse } from 'voice-router-dev';
|
|
7102
|
+
*
|
|
7103
|
+
* const response: ListTranscriptsResponse = await router.listTranscripts('assemblyai', {
|
|
7104
|
+
* status: 'completed',
|
|
7105
|
+
* limit: 50
|
|
7106
|
+
* });
|
|
7107
|
+
*
|
|
7108
|
+
* response.transcripts.forEach(item => {
|
|
7109
|
+
* console.log(item.data?.id, item.data?.status);
|
|
7110
|
+
* });
|
|
7111
|
+
*
|
|
7112
|
+
* if (response.hasMore) {
|
|
7113
|
+
* // Fetch next page
|
|
7114
|
+
* }
|
|
7115
|
+
* ```
|
|
7116
|
+
*/
|
|
7117
|
+
interface ListTranscriptsResponse {
|
|
7118
|
+
/** List of transcripts */
|
|
7119
|
+
transcripts: UnifiedTranscriptResponse[];
|
|
7120
|
+
/** Total count (if available from provider) */
|
|
7121
|
+
total?: number;
|
|
7122
|
+
/** Whether more results are available */
|
|
7123
|
+
hasMore?: boolean;
|
|
7124
|
+
}
|
|
6094
7125
|
/**
|
|
6095
7126
|
* Map of provider names to their raw response types
|
|
6096
7127
|
* Enables type-safe access to provider-specific raw responses
|
|
@@ -6140,34 +7171,7 @@ interface UnifiedTranscriptResponse<P extends TranscriptionProvider = Transcript
|
|
|
6140
7171
|
/** Provider that performed the transcription */
|
|
6141
7172
|
provider: P;
|
|
6142
7173
|
/** Transcription data (only present on success) */
|
|
6143
|
-
data?:
|
|
6144
|
-
/** Unique transcription ID */
|
|
6145
|
-
id: string;
|
|
6146
|
-
/** Full transcribed text */
|
|
6147
|
-
text: string;
|
|
6148
|
-
/** Overall confidence score (0-1) */
|
|
6149
|
-
confidence?: number;
|
|
6150
|
-
/** Transcription status */
|
|
6151
|
-
status: TranscriptionStatus;
|
|
6152
|
-
/** Detected or specified language code */
|
|
6153
|
-
language?: string;
|
|
6154
|
-
/** Audio duration in seconds */
|
|
6155
|
-
duration?: number;
|
|
6156
|
-
/** Speaker diarization results */
|
|
6157
|
-
speakers?: Speaker[];
|
|
6158
|
-
/** Word-level transcription with timestamps */
|
|
6159
|
-
words?: Word[];
|
|
6160
|
-
/** Utterances (speaker turns) */
|
|
6161
|
-
utterances?: Utterance[];
|
|
6162
|
-
/** Summary of the content (if summarization enabled) */
|
|
6163
|
-
summary?: string;
|
|
6164
|
-
/** Additional provider-specific metadata */
|
|
6165
|
-
metadata?: Record<string, unknown>;
|
|
6166
|
-
/** Creation timestamp */
|
|
6167
|
-
createdAt?: string;
|
|
6168
|
-
/** Completion timestamp */
|
|
6169
|
-
completedAt?: string;
|
|
6170
|
-
};
|
|
7174
|
+
data?: TranscriptData;
|
|
6171
7175
|
/**
|
|
6172
7176
|
* Extended provider-specific data (fully typed from OpenAPI specs)
|
|
6173
7177
|
*
|
|
@@ -6678,14 +7682,22 @@ interface TranscriptionAdapter {
|
|
|
6678
7682
|
success: boolean;
|
|
6679
7683
|
}>;
|
|
6680
7684
|
/**
|
|
6681
|
-
* List recent transcriptions
|
|
6682
|
-
*
|
|
7685
|
+
* List recent transcriptions with filtering
|
|
7686
|
+
*
|
|
7687
|
+
* Supports date/time filtering, status filtering, and pagination.
|
|
7688
|
+
* Not all providers support all options - use provider-specific passthrough for full control.
|
|
7689
|
+
*
|
|
7690
|
+
* @example Filter by date range
|
|
7691
|
+
* ```typescript
|
|
7692
|
+
* const { transcripts } = await adapter.listTranscripts({
|
|
7693
|
+
* afterDate: '2026-01-01',
|
|
7694
|
+
* beforeDate: '2026-01-31',
|
|
7695
|
+
* status: 'completed',
|
|
7696
|
+
* limit: 50
|
|
7697
|
+
* })
|
|
7698
|
+
* ```
|
|
6683
7699
|
*/
|
|
6684
|
-
listTranscripts?(options?: {
|
|
6685
|
-
limit?: number;
|
|
6686
|
-
offset?: number;
|
|
6687
|
-
status?: string;
|
|
6688
|
-
}): Promise<{
|
|
7700
|
+
listTranscripts?(options?: ListTranscriptsOptions): Promise<{
|
|
6689
7701
|
transcripts: UnifiedTranscriptResponse[];
|
|
6690
7702
|
total?: number;
|
|
6691
7703
|
hasMore?: boolean;
|
|
@@ -6981,14 +7993,22 @@ declare class VoiceRouter {
|
|
|
6981
7993
|
success: boolean;
|
|
6982
7994
|
}>;
|
|
6983
7995
|
/**
|
|
6984
|
-
* List recent transcriptions
|
|
6985
|
-
*
|
|
7996
|
+
* List recent transcriptions with filtering
|
|
7997
|
+
*
|
|
7998
|
+
* Supports date/time filtering, status filtering, and pagination.
|
|
7999
|
+
* Not all providers support this operation.
|
|
8000
|
+
*
|
|
8001
|
+
* @example Filter by date range
|
|
8002
|
+
* ```typescript
|
|
8003
|
+
* const { transcripts } = await router.listTranscripts('assemblyai', {
|
|
8004
|
+
* afterDate: '2026-01-01',
|
|
8005
|
+
* beforeDate: '2026-01-31',
|
|
8006
|
+
* status: 'completed',
|
|
8007
|
+
* limit: 50
|
|
8008
|
+
* })
|
|
8009
|
+
* ```
|
|
6986
8010
|
*/
|
|
6987
|
-
listTranscripts(provider: TranscriptionProvider, options?: {
|
|
6988
|
-
limit?: number;
|
|
6989
|
-
offset?: number;
|
|
6990
|
-
status?: string;
|
|
6991
|
-
}): Promise<{
|
|
8011
|
+
listTranscripts(provider: TranscriptionProvider, options?: ListTranscriptsOptions): Promise<{
|
|
6992
8012
|
transcripts: UnifiedTranscriptResponse[];
|
|
6993
8013
|
total?: number;
|
|
6994
8014
|
hasMore?: boolean;
|
|
@@ -7301,33 +8321,127 @@ declare class GladiaAdapter extends BaseAdapter {
|
|
|
7301
8321
|
*/
|
|
7302
8322
|
private extractUtterances;
|
|
7303
8323
|
/**
|
|
7304
|
-
* Delete a transcription job and its associated data
|
|
8324
|
+
* Delete a transcription job and its associated data
|
|
8325
|
+
*
|
|
8326
|
+
* Removes the transcription data from Gladia's servers. This action is
|
|
8327
|
+
* irreversible. Supports both pre-recorded and streaming job IDs.
|
|
8328
|
+
*
|
|
8329
|
+
* @param transcriptId - The ID of the transcript/job to delete
|
|
8330
|
+
* @param jobType - Type of job: 'pre-recorded' or 'streaming' (defaults to 'pre-recorded')
|
|
8331
|
+
* @returns Promise with success status
|
|
8332
|
+
*
|
|
8333
|
+
* @example Delete a pre-recorded transcript
|
|
8334
|
+
* ```typescript
|
|
8335
|
+
* const result = await adapter.deleteTranscript('abc123');
|
|
8336
|
+
* if (result.success) {
|
|
8337
|
+
* console.log('Transcript deleted successfully');
|
|
8338
|
+
* }
|
|
8339
|
+
* ```
|
|
8340
|
+
*
|
|
8341
|
+
* @example Delete a streaming job
|
|
8342
|
+
* ```typescript
|
|
8343
|
+
* const result = await adapter.deleteTranscript('stream-456', 'streaming');
|
|
8344
|
+
* ```
|
|
8345
|
+
*
|
|
8346
|
+
* @see https://docs.gladia.io/
|
|
8347
|
+
*/
|
|
8348
|
+
deleteTranscript(transcriptId: string, jobType?: "pre-recorded" | "streaming"): Promise<{
|
|
8349
|
+
success: boolean;
|
|
8350
|
+
}>;
|
|
8351
|
+
/**
|
|
8352
|
+
* Download the original audio file from a transcription
|
|
8353
|
+
*
|
|
8354
|
+
* Gladia stores the audio files used for transcription and allows downloading them.
|
|
8355
|
+
* This works for both pre-recorded and streaming (live) transcriptions.
|
|
8356
|
+
*
|
|
8357
|
+
* @param transcriptId - The ID of the transcript/job
|
|
8358
|
+
* @param jobType - Type of job: 'pre-recorded' or 'streaming' (defaults to 'pre-recorded')
|
|
8359
|
+
* @returns Promise with the audio file as a Blob, or error
|
|
8360
|
+
*
|
|
8361
|
+
* @example Download audio from a pre-recorded transcription
|
|
8362
|
+
* ```typescript
|
|
8363
|
+
* const result = await adapter.getAudioFile('abc123');
|
|
8364
|
+
* if (result.success && result.data) {
|
|
8365
|
+
* // Save to file (Node.js)
|
|
8366
|
+
* const buffer = Buffer.from(await result.data.arrayBuffer());
|
|
8367
|
+
* fs.writeFileSync('audio.mp3', buffer);
|
|
8368
|
+
*
|
|
8369
|
+
* // Or create download URL (browser)
|
|
8370
|
+
* const url = URL.createObjectURL(result.data);
|
|
8371
|
+
* }
|
|
8372
|
+
* ```
|
|
8373
|
+
*
|
|
8374
|
+
* @example Download audio from a live/streaming session
|
|
8375
|
+
* ```typescript
|
|
8376
|
+
* const result = await adapter.getAudioFile('stream-456', 'streaming');
|
|
8377
|
+
* if (result.success) {
|
|
8378
|
+
* console.log('Audio file size:', result.data?.size);
|
|
8379
|
+
* }
|
|
8380
|
+
* ```
|
|
8381
|
+
*
|
|
8382
|
+
* @see https://docs.gladia.io/
|
|
8383
|
+
*/
|
|
8384
|
+
getAudioFile(transcriptId: string, jobType?: "pre-recorded" | "streaming"): Promise<{
|
|
8385
|
+
success: boolean;
|
|
8386
|
+
data?: Blob;
|
|
8387
|
+
error?: {
|
|
8388
|
+
code: string;
|
|
8389
|
+
message: string;
|
|
8390
|
+
};
|
|
8391
|
+
}>;
|
|
8392
|
+
/**
|
|
8393
|
+
* List recent transcriptions with filtering
|
|
7305
8394
|
*
|
|
7306
|
-
*
|
|
7307
|
-
*
|
|
8395
|
+
* Retrieves a list of transcription jobs (both pre-recorded and streaming)
|
|
8396
|
+
* with optional filtering by status, date, and custom metadata.
|
|
7308
8397
|
*
|
|
7309
|
-
* @param
|
|
7310
|
-
* @param
|
|
7311
|
-
* @
|
|
8398
|
+
* @param options - Filtering and pagination options
|
|
8399
|
+
* @param options.limit - Maximum number of transcripts to return
|
|
8400
|
+
* @param options.offset - Pagination offset (skip N results)
|
|
8401
|
+
* @param options.status - Filter by status (queued, processing, done, error)
|
|
8402
|
+
* @param options.date - Filter by exact date (ISO format YYYY-MM-DD)
|
|
8403
|
+
* @param options.beforeDate - Filter for transcripts before this date
|
|
8404
|
+
* @param options.afterDate - Filter for transcripts after this date
|
|
8405
|
+
* @param options.gladia - Full Gladia-specific options (custom_metadata, kind, etc.)
|
|
8406
|
+
* @returns List of transcripts with pagination info
|
|
7312
8407
|
*
|
|
7313
|
-
* @example
|
|
8408
|
+
* @example List recent transcripts
|
|
7314
8409
|
* ```typescript
|
|
7315
|
-
* const
|
|
7316
|
-
*
|
|
7317
|
-
*
|
|
7318
|
-
* }
|
|
8410
|
+
* const { transcripts, hasMore } = await adapter.listTranscripts({
|
|
8411
|
+
* limit: 50,
|
|
8412
|
+
* status: 'done'
|
|
8413
|
+
* })
|
|
7319
8414
|
* ```
|
|
7320
8415
|
*
|
|
7321
|
-
* @example
|
|
8416
|
+
* @example Filter by date range
|
|
7322
8417
|
* ```typescript
|
|
7323
|
-
* const
|
|
8418
|
+
* const { transcripts } = await adapter.listTranscripts({
|
|
8419
|
+
* afterDate: '2026-01-01',
|
|
8420
|
+
* beforeDate: '2026-01-31',
|
|
8421
|
+
* limit: 100
|
|
8422
|
+
* })
|
|
8423
|
+
* ```
|
|
8424
|
+
*
|
|
8425
|
+
* @example Filter by custom metadata
|
|
8426
|
+
* ```typescript
|
|
8427
|
+
* const { transcripts } = await adapter.listTranscripts({
|
|
8428
|
+
* gladia: {
|
|
8429
|
+
* custom_metadata: { project: 'my-project' }
|
|
8430
|
+
* }
|
|
8431
|
+
* })
|
|
7324
8432
|
* ```
|
|
7325
8433
|
*
|
|
7326
8434
|
* @see https://docs.gladia.io/
|
|
7327
8435
|
*/
|
|
7328
|
-
|
|
7329
|
-
|
|
8436
|
+
listTranscripts(options?: ListTranscriptsOptions): Promise<{
|
|
8437
|
+
transcripts: UnifiedTranscriptResponse[];
|
|
8438
|
+
total?: number;
|
|
8439
|
+
hasMore?: boolean;
|
|
7330
8440
|
}>;
|
|
8441
|
+
/**
|
|
8442
|
+
* Normalize a transcript list item to unified format
|
|
8443
|
+
*/
|
|
8444
|
+
private normalizeListItem;
|
|
7331
8445
|
/**
|
|
7332
8446
|
* Stream audio for real-time transcription
|
|
7333
8447
|
*
|
|
@@ -7631,6 +8745,57 @@ declare class AssemblyAIAdapter extends BaseAdapter {
|
|
|
7631
8745
|
deleteTranscript(transcriptId: string): Promise<{
|
|
7632
8746
|
success: boolean;
|
|
7633
8747
|
}>;
|
|
8748
|
+
/**
|
|
8749
|
+
* List recent transcriptions with filtering
|
|
8750
|
+
*
|
|
8751
|
+
* Retrieves a list of transcripts with optional filtering by status and date.
|
|
8752
|
+
* Transcripts are sorted from newest to oldest and can be retrieved for the
|
|
8753
|
+
* last 90 days of usage.
|
|
8754
|
+
*
|
|
8755
|
+
* @param options - Filtering and pagination options
|
|
8756
|
+
* @param options.limit - Maximum number of transcripts (max 200)
|
|
8757
|
+
* @param options.status - Filter by status (queued, processing, completed, error)
|
|
8758
|
+
* @param options.date - Filter by exact date (ISO format YYYY-MM-DD)
|
|
8759
|
+
* @param options.assemblyai - Full AssemblyAI-specific options (before_id, after_id, etc.)
|
|
8760
|
+
* @returns List of transcripts with pagination info
|
|
8761
|
+
*
|
|
8762
|
+
* @example List recent transcripts
|
|
8763
|
+
* ```typescript
|
|
8764
|
+
* const { transcripts, hasMore } = await adapter.listTranscripts({
|
|
8765
|
+
* limit: 50,
|
|
8766
|
+
* status: 'completed'
|
|
8767
|
+
* })
|
|
8768
|
+
* ```
|
|
8769
|
+
*
|
|
8770
|
+
* @example Filter by date
|
|
8771
|
+
* ```typescript
|
|
8772
|
+
* const { transcripts } = await adapter.listTranscripts({
|
|
8773
|
+
* date: '2026-01-07',
|
|
8774
|
+
* limit: 100
|
|
8775
|
+
* })
|
|
8776
|
+
* ```
|
|
8777
|
+
*
|
|
8778
|
+
* @example Use cursor pagination
|
|
8779
|
+
* ```typescript
|
|
8780
|
+
* const { transcripts } = await adapter.listTranscripts({
|
|
8781
|
+
* assemblyai: {
|
|
8782
|
+
* after_id: 'abc123', // Get transcripts after this ID
|
|
8783
|
+
* limit: 50
|
|
8784
|
+
* }
|
|
8785
|
+
* })
|
|
8786
|
+
* ```
|
|
8787
|
+
*
|
|
8788
|
+
* @see https://www.assemblyai.com/docs/api-reference/transcripts/list
|
|
8789
|
+
*/
|
|
8790
|
+
listTranscripts(options?: ListTranscriptsOptions): Promise<{
|
|
8791
|
+
transcripts: UnifiedTranscriptResponse[];
|
|
8792
|
+
total?: number;
|
|
8793
|
+
hasMore?: boolean;
|
|
8794
|
+
}>;
|
|
8795
|
+
/**
|
|
8796
|
+
* Normalize a transcript list item to unified format
|
|
8797
|
+
*/
|
|
8798
|
+
private normalizeListItem;
|
|
7634
8799
|
/**
|
|
7635
8800
|
* Build AssemblyAI transcription request from unified options
|
|
7636
8801
|
*/
|
|
@@ -7819,7 +8984,10 @@ declare class DeepgramAdapter extends BaseAdapter {
|
|
|
7819
8984
|
private client?;
|
|
7820
8985
|
protected baseUrl: string;
|
|
7821
8986
|
private wsBaseUrl;
|
|
7822
|
-
|
|
8987
|
+
private projectId?;
|
|
8988
|
+
initialize(config: ProviderConfig & {
|
|
8989
|
+
projectId?: string;
|
|
8990
|
+
}): void;
|
|
7823
8991
|
/**
|
|
7824
8992
|
* Submit audio for transcription
|
|
7825
8993
|
*
|
|
@@ -7867,12 +9035,30 @@ declare class DeepgramAdapter extends BaseAdapter {
|
|
|
7867
9035
|
/**
|
|
7868
9036
|
* Get transcription result by ID
|
|
7869
9037
|
*
|
|
7870
|
-
*
|
|
7871
|
-
* for retrieving cached results if you've stored the request ID.
|
|
7872
|
-
* The initial transcribe() call already returns complete results.
|
|
9038
|
+
* Retrieves a previous transcription from Deepgram's request history.
|
|
7873
9039
|
*
|
|
7874
|
-
*
|
|
7875
|
-
*
|
|
9040
|
+
* Unlike the list endpoint, getting a single request DOES include the full
|
|
9041
|
+
* transcript response. Requires `projectId` to be set during initialization.
|
|
9042
|
+
*
|
|
9043
|
+
* @param transcriptId - Request ID from a previous transcription
|
|
9044
|
+
* @returns Full transcript response including text, words, and metadata
|
|
9045
|
+
*
|
|
9046
|
+
* @example Get a transcript by request ID
|
|
9047
|
+
* ```typescript
|
|
9048
|
+
* const adapter = new DeepgramAdapter()
|
|
9049
|
+
* adapter.initialize({
|
|
9050
|
+
* apiKey: process.env.DEEPGRAM_API_KEY,
|
|
9051
|
+
* projectId: process.env.DEEPGRAM_PROJECT_ID
|
|
9052
|
+
* })
|
|
9053
|
+
*
|
|
9054
|
+
* const result = await adapter.getTranscript('abc123-request-id')
|
|
9055
|
+
* if (result.success) {
|
|
9056
|
+
* console.log(result.data?.text)
|
|
9057
|
+
* console.log(result.data?.words)
|
|
9058
|
+
* }
|
|
9059
|
+
* ```
|
|
9060
|
+
*
|
|
9061
|
+
* @see https://developers.deepgram.com/reference/get-request
|
|
7876
9062
|
*/
|
|
7877
9063
|
getTranscript(transcriptId: string): Promise<UnifiedTranscriptResponse>;
|
|
7878
9064
|
/**
|
|
@@ -8012,6 +9198,52 @@ declare class DeepgramAdapter extends BaseAdapter {
|
|
|
8012
9198
|
* Handle all WebSocket message types from Deepgram streaming
|
|
8013
9199
|
*/
|
|
8014
9200
|
private handleWebSocketMessage;
|
|
9201
|
+
/**
|
|
9202
|
+
* List recent transcription requests from Deepgram's request history
|
|
9203
|
+
*
|
|
9204
|
+
* **Important:** Deepgram processes synchronously and doesn't store transcript content.
|
|
9205
|
+
* This method returns request metadata (IDs, timestamps, status) but NOT the actual
|
|
9206
|
+
* transcript text. Use this for auditing, billing analysis, or request tracking.
|
|
9207
|
+
*
|
|
9208
|
+
* Requires `projectId` to be set during initialization.
|
|
9209
|
+
*
|
|
9210
|
+
* @param options - Filtering and pagination options
|
|
9211
|
+
* @param options.limit - Maximum number of requests (default 10, max 1000)
|
|
9212
|
+
* @param options.status - Filter by status: 'succeeded' or 'failed'
|
|
9213
|
+
* @param options.afterDate - Start date (YYYY-MM-DD or ISO 8601)
|
|
9214
|
+
* @param options.beforeDate - End date (YYYY-MM-DD or ISO 8601)
|
|
9215
|
+
* @param options.deepgram - Provider-specific params (page, request_id, etc.)
|
|
9216
|
+
* @returns List of transcription request metadata
|
|
9217
|
+
*
|
|
9218
|
+
* @example List recent transcription requests
|
|
9219
|
+
* ```typescript
|
|
9220
|
+
* const adapter = new DeepgramAdapter()
|
|
9221
|
+
* adapter.initialize({
|
|
9222
|
+
* apiKey: process.env.DEEPGRAM_API_KEY,
|
|
9223
|
+
* projectId: process.env.DEEPGRAM_PROJECT_ID
|
|
9224
|
+
* })
|
|
9225
|
+
*
|
|
9226
|
+
* const { transcripts, hasMore } = await adapter.listTranscripts({
|
|
9227
|
+
* limit: 50,
|
|
9228
|
+
* status: 'succeeded'
|
|
9229
|
+
* })
|
|
9230
|
+
*
|
|
9231
|
+
* transcripts.forEach(t => {
|
|
9232
|
+
* console.log(t.data?.id, t.data?.metadata?.createdAt)
|
|
9233
|
+
* })
|
|
9234
|
+
* ```
|
|
9235
|
+
*
|
|
9236
|
+
* @see https://developers.deepgram.com/reference/get-all-requests
|
|
9237
|
+
*/
|
|
9238
|
+
listTranscripts(options?: ListTranscriptsOptions): Promise<{
|
|
9239
|
+
transcripts: UnifiedTranscriptResponse[];
|
|
9240
|
+
total?: number;
|
|
9241
|
+
hasMore?: boolean;
|
|
9242
|
+
}>;
|
|
9243
|
+
/**
|
|
9244
|
+
* Normalize a Deepgram request history item to unified format
|
|
9245
|
+
*/
|
|
9246
|
+
private normalizeRequestItem;
|
|
8015
9247
|
}
|
|
8016
9248
|
/**
|
|
8017
9249
|
* Factory function to create a Deepgram adapter
|
|
@@ -8154,6 +9386,48 @@ declare class AzureSTTAdapter extends BaseAdapter {
|
|
|
8154
9386
|
deleteTranscript(transcriptId: string): Promise<{
|
|
8155
9387
|
success: boolean;
|
|
8156
9388
|
}>;
|
|
9389
|
+
/**
|
|
9390
|
+
* List recent transcriptions with filtering
|
|
9391
|
+
*
|
|
9392
|
+
* Retrieves a list of transcription jobs for the authenticated subscription.
|
|
9393
|
+
* Azure uses OData filtering for advanced queries.
|
|
9394
|
+
*
|
|
9395
|
+
* @param options - Filtering and pagination options
|
|
9396
|
+
* @param options.limit - Maximum number of transcripts (maps to 'top')
|
|
9397
|
+
* @param options.offset - Pagination offset (maps to 'skip')
|
|
9398
|
+
* @param options.status - Filter by status (uses OData filter)
|
|
9399
|
+
* @returns List of transcripts with pagination info
|
|
9400
|
+
*
|
|
9401
|
+
* @example List recent transcripts
|
|
9402
|
+
* ```typescript
|
|
9403
|
+
* const { transcripts, hasMore } = await adapter.listTranscripts({
|
|
9404
|
+
* limit: 50
|
|
9405
|
+
* })
|
|
9406
|
+
* ```
|
|
9407
|
+
*
|
|
9408
|
+
* @example Filter by status using OData
|
|
9409
|
+
* ```typescript
|
|
9410
|
+
* const { transcripts } = await adapter.listTranscripts({
|
|
9411
|
+
* status: 'Succeeded',
|
|
9412
|
+
* limit: 100
|
|
9413
|
+
* })
|
|
9414
|
+
* ```
|
|
9415
|
+
*
|
|
9416
|
+
* @see https://learn.microsoft.com/azure/cognitive-services/speech-service/batch-transcription
|
|
9417
|
+
*/
|
|
9418
|
+
listTranscripts(options?: ListTranscriptsOptions): Promise<{
|
|
9419
|
+
transcripts: UnifiedTranscriptResponse[];
|
|
9420
|
+
total?: number;
|
|
9421
|
+
hasMore?: boolean;
|
|
9422
|
+
}>;
|
|
9423
|
+
/**
|
|
9424
|
+
* Map unified status to Azure status format using generated enum
|
|
9425
|
+
*/
|
|
9426
|
+
private mapStatusToAzure;
|
|
9427
|
+
/**
|
|
9428
|
+
* Normalize a transcript list item to unified format
|
|
9429
|
+
*/
|
|
9430
|
+
private normalizeListItem;
|
|
8157
9431
|
/**
|
|
8158
9432
|
* Build Azure-specific transcription properties
|
|
8159
9433
|
*/
|
|
@@ -11927,73 +13201,6 @@ type StreamingControllerInitStreamingSessionV2Params = {
|
|
|
11927
13201
|
|
|
11928
13202
|
type TranscriptionControllerGetTranscriptV2200 = PreRecordedResponse | StreamingResponse;
|
|
11929
13203
|
|
|
11930
|
-
/**
|
|
11931
|
-
* Generated by orval v7.9.0 🍺
|
|
11932
|
-
* Do not edit manually.
|
|
11933
|
-
* Gladia Control API
|
|
11934
|
-
* OpenAPI spec version: 1.0
|
|
11935
|
-
*/
|
|
11936
|
-
type TranscriptionControllerListV2KindItem = (typeof TranscriptionControllerListV2KindItem)[keyof typeof TranscriptionControllerListV2KindItem];
|
|
11937
|
-
declare const TranscriptionControllerListV2KindItem: {
|
|
11938
|
-
readonly "pre-recorded": "pre-recorded";
|
|
11939
|
-
readonly live: "live";
|
|
11940
|
-
};
|
|
11941
|
-
|
|
11942
|
-
/**
|
|
11943
|
-
* Generated by orval v7.9.0 🍺
|
|
11944
|
-
* Do not edit manually.
|
|
11945
|
-
* Gladia Control API
|
|
11946
|
-
* OpenAPI spec version: 1.0
|
|
11947
|
-
*/
|
|
11948
|
-
type TranscriptionControllerListV2StatusItem = (typeof TranscriptionControllerListV2StatusItem)[keyof typeof TranscriptionControllerListV2StatusItem];
|
|
11949
|
-
declare const TranscriptionControllerListV2StatusItem: {
|
|
11950
|
-
readonly queued: "queued";
|
|
11951
|
-
readonly processing: "processing";
|
|
11952
|
-
readonly done: "done";
|
|
11953
|
-
readonly error: "error";
|
|
11954
|
-
};
|
|
11955
|
-
|
|
11956
|
-
/**
|
|
11957
|
-
* Generated by orval v7.9.0 🍺
|
|
11958
|
-
* Do not edit manually.
|
|
11959
|
-
* Gladia Control API
|
|
11960
|
-
* OpenAPI spec version: 1.0
|
|
11961
|
-
*/
|
|
11962
|
-
|
|
11963
|
-
type TranscriptionControllerListV2Params = {
|
|
11964
|
-
/**
|
|
11965
|
-
* The starting point for pagination. A value of 0 starts from the first item.
|
|
11966
|
-
*/
|
|
11967
|
-
offset?: number;
|
|
11968
|
-
/**
|
|
11969
|
-
* The maximum number of items to return. Useful for pagination and controlling data payload size.
|
|
11970
|
-
*/
|
|
11971
|
-
limit?: number;
|
|
11972
|
-
/**
|
|
11973
|
-
* Filter items relevant to a specific date in ISO format (YYYY-MM-DD).
|
|
11974
|
-
*/
|
|
11975
|
-
date?: string;
|
|
11976
|
-
/**
|
|
11977
|
-
* Include items that occurred before the specified date in ISO format.
|
|
11978
|
-
*/
|
|
11979
|
-
before_date?: string;
|
|
11980
|
-
/**
|
|
11981
|
-
* Filter for items after the specified date. Use with `before_date` for a range. Date in ISO format.
|
|
11982
|
-
*/
|
|
11983
|
-
after_date?: string;
|
|
11984
|
-
/**
|
|
11985
|
-
* Filter the list based on item status. Accepts multiple values from the predefined list.
|
|
11986
|
-
*/
|
|
11987
|
-
status?: TranscriptionControllerListV2StatusItem[];
|
|
11988
|
-
custom_metadata?: {
|
|
11989
|
-
[key: string]: unknown;
|
|
11990
|
-
};
|
|
11991
|
-
/**
|
|
11992
|
-
* Filter the list based on the item type. Supports multiple values from the predefined list.
|
|
11993
|
-
*/
|
|
11994
|
-
kind?: TranscriptionControllerListV2KindItem[];
|
|
11995
|
-
};
|
|
11996
|
-
|
|
11997
13204
|
/**
|
|
11998
13205
|
* Generated by orval v7.9.0 🍺
|
|
11999
13206
|
* Do not edit manually.
|
|
@@ -13233,101 +14440,6 @@ interface ListTranscriptParams {
|
|
|
13233
14440
|
throttled_only?: boolean;
|
|
13234
14441
|
}
|
|
13235
14442
|
|
|
13236
|
-
/**
|
|
13237
|
-
* Generated by orval v7.9.0 🍺
|
|
13238
|
-
* Do not edit manually.
|
|
13239
|
-
* AssemblyAI API
|
|
13240
|
-
* AssemblyAI API
|
|
13241
|
-
* OpenAPI spec version: 1.3.4
|
|
13242
|
-
*/
|
|
13243
|
-
/**
|
|
13244
|
-
* Get transcripts that were created after this transcript ID
|
|
13245
|
-
*/
|
|
13246
|
-
type AfterId = string;
|
|
13247
|
-
|
|
13248
|
-
/**
|
|
13249
|
-
* Generated by orval v7.9.0 🍺
|
|
13250
|
-
* Do not edit manually.
|
|
13251
|
-
* AssemblyAI API
|
|
13252
|
-
* AssemblyAI API
|
|
13253
|
-
* OpenAPI spec version: 1.3.4
|
|
13254
|
-
*/
|
|
13255
|
-
/**
|
|
13256
|
-
* Get transcripts that were created before this transcript ID
|
|
13257
|
-
*/
|
|
13258
|
-
type BeforeId = string;
|
|
13259
|
-
|
|
13260
|
-
/**
|
|
13261
|
-
* Generated by orval v7.9.0 🍺
|
|
13262
|
-
* Do not edit manually.
|
|
13263
|
-
* AssemblyAI API
|
|
13264
|
-
* AssemblyAI API
|
|
13265
|
-
* OpenAPI spec version: 1.3.4
|
|
13266
|
-
*/
|
|
13267
|
-
/**
|
|
13268
|
-
* Only get transcripts created on this date
|
|
13269
|
-
*/
|
|
13270
|
-
type CreatedOn = string;
|
|
13271
|
-
|
|
13272
|
-
/**
|
|
13273
|
-
* Generated by orval v7.9.0 🍺
|
|
13274
|
-
* Do not edit manually.
|
|
13275
|
-
* AssemblyAI API
|
|
13276
|
-
* AssemblyAI API
|
|
13277
|
-
* OpenAPI spec version: 1.3.4
|
|
13278
|
-
*/
|
|
13279
|
-
/**
|
|
13280
|
-
* Maximum amount of transcripts to retrieve
|
|
13281
|
-
*/
|
|
13282
|
-
type Limit = number;
|
|
13283
|
-
|
|
13284
|
-
/**
|
|
13285
|
-
* Generated by orval v7.9.0 🍺
|
|
13286
|
-
* Do not edit manually.
|
|
13287
|
-
* AssemblyAI API
|
|
13288
|
-
* AssemblyAI API
|
|
13289
|
-
* OpenAPI spec version: 1.3.4
|
|
13290
|
-
*/
|
|
13291
|
-
/**
|
|
13292
|
-
* Only get throttled transcripts, overrides the status filter
|
|
13293
|
-
*/
|
|
13294
|
-
type ThrottledOnly = boolean;
|
|
13295
|
-
|
|
13296
|
-
/**
|
|
13297
|
-
* Generated by orval v7.9.0 🍺
|
|
13298
|
-
* Do not edit manually.
|
|
13299
|
-
* AssemblyAI API
|
|
13300
|
-
* AssemblyAI API
|
|
13301
|
-
* OpenAPI spec version: 1.3.4
|
|
13302
|
-
*/
|
|
13303
|
-
|
|
13304
|
-
type ListTranscriptsParams = {
|
|
13305
|
-
/**
|
|
13306
|
-
* Maximum amount of transcripts to retrieve
|
|
13307
|
-
*/
|
|
13308
|
-
limit?: Limit;
|
|
13309
|
-
/**
|
|
13310
|
-
* Filter by transcript status
|
|
13311
|
-
*/
|
|
13312
|
-
status?: TranscriptStatus;
|
|
13313
|
-
/**
|
|
13314
|
-
* Only get transcripts created on this date
|
|
13315
|
-
*/
|
|
13316
|
-
created_on?: CreatedOn;
|
|
13317
|
-
/**
|
|
13318
|
-
* Get transcripts that were created before this transcript ID
|
|
13319
|
-
*/
|
|
13320
|
-
before_id?: BeforeId;
|
|
13321
|
-
/**
|
|
13322
|
-
* Get transcripts that were created after this transcript ID
|
|
13323
|
-
*/
|
|
13324
|
-
after_id?: AfterId;
|
|
13325
|
-
/**
|
|
13326
|
-
* Only get throttled transcripts, overrides the status filter
|
|
13327
|
-
*/
|
|
13328
|
-
throttled_only?: ThrottledOnly;
|
|
13329
|
-
};
|
|
13330
|
-
|
|
13331
14443
|
/**
|
|
13332
14444
|
* Generated by orval v7.9.0 🍺
|
|
13333
14445
|
* Do not edit manually.
|
|
@@ -13939,4 +15051,4 @@ declare namespace index {
|
|
|
13939
15051
|
export { index_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, type index_AutoHighlightResult as AutoHighlightResult, type index_AutoHighlightsResult as AutoHighlightsResult, type index_BadRequestResponse as BadRequestResponse, type index_CannotAccessUploadedFileResponse as CannotAccessUploadedFileResponse, type index_Chapter as Chapter, type index_ContentSafetyLabel as ContentSafetyLabel, type index_ContentSafetyLabelResult as ContentSafetyLabelResult, type index_ContentSafetyLabelsResult as ContentSafetyLabelsResult, type index_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, type index_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, type index_CreateRealtimeTemporaryTokenParams as CreateRealtimeTemporaryTokenParams, type index_Entity as Entity, index_EntityType as EntityType, type Error$1 as Error, type index_GatewayTimeoutResponse as GatewayTimeoutResponse, type index_GetSubtitlesParams as GetSubtitlesParams, type index_InternalServerErrorResponse as InternalServerErrorResponse, type index_LemurActionItemsParams as LemurActionItemsParams, type index_LemurActionItemsParamsAllOf as LemurActionItemsParamsAllOf, type index_LemurActionItemsResponse as LemurActionItemsResponse, type index_LemurBaseParams as LemurBaseParams, type index_LemurBaseParamsContext as LemurBaseParamsContext, type index_LemurBaseParamsContextOneOf as LemurBaseParamsContextOneOf, type index_LemurBaseParamsFinalModel as LemurBaseParamsFinalModel, type index_LemurBaseResponse as LemurBaseResponse, index_LemurModel as LemurModel, type index_LemurQuestion as LemurQuestion, type index_LemurQuestionAnswer as LemurQuestionAnswer, type index_LemurQuestionAnswerParams as LemurQuestionAnswerParams, type index_LemurQuestionAnswerParamsAllOf as LemurQuestionAnswerParamsAllOf, type index_LemurQuestionAnswerResponse as LemurQuestionAnswerResponse, type index_LemurQuestionAnswerResponseAllOf as LemurQuestionAnswerResponseAllOf, type index_LemurQuestionContext as LemurQuestionContext, type index_LemurQuestionContextOneOf as LemurQuestionContextOneOf, type index_LemurResponse as LemurResponse, type index_LemurStringResponse as LemurStringResponse, type index_LemurStringResponseAllOf as LemurStringResponseAllOf, type index_LemurSummaryParams as LemurSummaryParams, type index_LemurSummaryParamsAllOf as LemurSummaryParamsAllOf, type index_LemurSummaryResponse as LemurSummaryResponse, type index_LemurTaskParams as LemurTaskParams, type index_LemurTaskParamsAllOf as LemurTaskParamsAllOf, type index_LemurTaskResponse as LemurTaskResponse, type index_LemurUsage as LemurUsage, type index_ListTranscriptParams as ListTranscriptParams, type index_ListTranscriptsParams as ListTranscriptsParams, type index_NotFoundResponse as NotFoundResponse, type index_PageDetails as PageDetails, type index_PageDetailsNextUrl as PageDetailsNextUrl, type index_PageDetailsPrevUrl as PageDetailsPrevUrl, type index_ParagraphsResponse as ParagraphsResponse, index_PiiPolicy as PiiPolicy, type index_PurgeLemurRequestDataResponse as PurgeLemurRequestDataResponse, type index_RealtimeTemporaryTokenResponse as RealtimeTemporaryTokenResponse, index_RedactPiiAudioQuality as RedactPiiAudioQuality, type index_RedactedAudioNotification as RedactedAudioNotification, type index_RedactedAudioResponse as RedactedAudioResponse, index_RedactedAudioStatus as RedactedAudioStatus, type index_SentencesResponse as SentencesResponse, index_Sentiment as Sentiment, type SentimentAnalysisResult$1 as SentimentAnalysisResult, type index_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, type index_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, type index_ServiceUnavailableResponse as ServiceUnavailableResponse, type index_SeverityScoreSummary as SeverityScoreSummary, index_SpeechModel as SpeechModel, index_SubstitutionPolicy as SubstitutionPolicy, index_SubtitleFormat as SubtitleFormat, index_SummaryModel as SummaryModel, index_SummaryType as SummaryType, type index_Timestamp as Timestamp, type index_TooManyRequestsResponse as TooManyRequestsResponse, type index_TopicDetectionModelResult as TopicDetectionModelResult, type index_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, type index_TopicDetectionResult as TopicDetectionResult, type index_TopicDetectionResultLabelsItem as TopicDetectionResultLabelsItem, type index_Transcript as Transcript, type index_TranscriptAudioDuration as TranscriptAudioDuration, type index_TranscriptAudioEndAt as TranscriptAudioEndAt, type index_TranscriptAudioStartFrom as TranscriptAudioStartFrom, type index_TranscriptAutoChapters as TranscriptAutoChapters, type index_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index_TranscriptBoostParam as TranscriptBoostParam, type index_TranscriptBoostParamProperty as TranscriptBoostParamProperty, type index_TranscriptChapters as TranscriptChapters, type index_TranscriptConfidence as TranscriptConfidence, type index_TranscriptContentSafety as TranscriptContentSafety, type index_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, type index_TranscriptCustomSpelling as TranscriptCustomSpelling, type index_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, type index_TranscriptCustomTopics as TranscriptCustomTopics, type index_TranscriptDisfluencies as TranscriptDisfluencies, type index_TranscriptEntities as TranscriptEntities, type index_TranscriptEntityDetection as TranscriptEntityDetection, type index_TranscriptFilterProfanity as TranscriptFilterProfanity, type index_TranscriptFormatText as TranscriptFormatText, type index_TranscriptIabCategories as TranscriptIabCategories, type index_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index_TranscriptLanguageCode as TranscriptLanguageCode, type index_TranscriptLanguageCodeProperty as TranscriptLanguageCodeProperty, type index_TranscriptLanguageConfidence as TranscriptLanguageConfidence, type index_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, type index_TranscriptLanguageDetection as TranscriptLanguageDetection, type index_TranscriptList as TranscriptList, type index_TranscriptListItem as TranscriptListItem, type index_TranscriptListItemCompleted as TranscriptListItemCompleted, type index_TranscriptListItemError as TranscriptListItemError, type index_TranscriptMultichannel as TranscriptMultichannel, type index_TranscriptOptionalParams as TranscriptOptionalParams, type index_TranscriptOptionalParamsLanguageCode as TranscriptOptionalParamsLanguageCode, type index_TranscriptOptionalParamsLanguageCodeOneOf as TranscriptOptionalParamsLanguageCodeOneOf, type index_TranscriptOptionalParamsRedactPiiSub as TranscriptOptionalParamsRedactPiiSub, type index_TranscriptOptionalParamsSpeakersExpected as TranscriptOptionalParamsSpeakersExpected, type index_TranscriptOptionalParamsSpeechModel as TranscriptOptionalParamsSpeechModel, type index_TranscriptOptionalParamsSpeechThreshold as TranscriptOptionalParamsSpeechThreshold, type index_TranscriptOptionalParamsWebhookAuthHeaderName as TranscriptOptionalParamsWebhookAuthHeaderName, type index_TranscriptOptionalParamsWebhookAuthHeaderValue as TranscriptOptionalParamsWebhookAuthHeaderValue, type index_TranscriptParagraph as TranscriptParagraph, type index_TranscriptParams as TranscriptParams, type index_TranscriptParamsAllOf as TranscriptParamsAllOf, type index_TranscriptPunctuate as TranscriptPunctuate, type index_TranscriptReadyNotification as TranscriptReadyNotification, index_TranscriptReadyStatus as TranscriptReadyStatus, type index_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, type index_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, type index_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, type index_TranscriptSentence as TranscriptSentence, type index_TranscriptSentenceChannel as TranscriptSentenceChannel, type index_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, type index_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, type index_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, type index_TranscriptSpeakerLabels as TranscriptSpeakerLabels, type index_TranscriptSpeakersExpected as TranscriptSpeakersExpected, type index_TranscriptSpeechModel as TranscriptSpeechModel, type index_TranscriptSpeechThreshold as TranscriptSpeechThreshold, type index_TranscriptSpeedBoost as TranscriptSpeedBoost, index_TranscriptStatus as TranscriptStatus, type index_TranscriptSummary as TranscriptSummary, type index_TranscriptSummaryModel as TranscriptSummaryModel, type index_TranscriptSummaryType as TranscriptSummaryType, type index_TranscriptText as TranscriptText, type index_TranscriptThrottled as TranscriptThrottled, type index_TranscriptUtterance as TranscriptUtterance, type index_TranscriptUtteranceChannel as TranscriptUtteranceChannel, type index_TranscriptUtterances as TranscriptUtterances, type index_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, type index_TranscriptWebhookNotification as TranscriptWebhookNotification, type index_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, type index_TranscriptWebhookUrl as TranscriptWebhookUrl, type index_TranscriptWord as TranscriptWord, type index_TranscriptWordChannel as TranscriptWordChannel, type index_TranscriptWordSpeaker as TranscriptWordSpeaker, type index_TranscriptWords as TranscriptWords, type index_UnauthorizedResponse as UnauthorizedResponse, type index_UploadedFile as UploadedFile, type index_WordSearchMatch as WordSearchMatch, type index_WordSearchParams as WordSearchParams, type index_WordSearchResponse as WordSearchResponse, type index_WordSearchTimestamp as WordSearchTimestamp };
|
|
13940
15052
|
}
|
|
13941
15053
|
|
|
13942
|
-
export { AssemblyAIAdapter, type Chapter as AssemblyAIChapter, type ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, AssemblyAIEncodingType, type Entity as AssemblyAIEntity, type AssemblyAIExtendedData, type AutoHighlightsResult as AssemblyAIHighlightsResult, type TranscriptOptionalParams as AssemblyAIOptions, AssemblyAISampleRateType, type SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAISpeechModelType, type AssemblyAIStreamingOptions, type TopicDetectionModelResult as AssemblyAITopicsResult, index as AssemblyAITypes, type AssemblyAIUpdateConfiguration, AssemblyAIWebhookHandler, type TranscriptWebhookNotification as AssemblyAIWebhookPayload, type AudioAckEvent, type AudioChunk, type AudioInput, AudioResponseFormat, AudioTranscriptionModel, AzureSTTAdapter, AzureWebhookHandler, BaseAdapter, BaseWebhookHandler, type BatchOnlyProvider, type ChapterizationEvent, DeepgramAdapter,
|
|
15054
|
+
export { AssemblyAIAdapter, type Chapter as AssemblyAIChapter, type ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, AssemblyAIEncodingType, type Entity as AssemblyAIEntity, type AssemblyAIExtendedData, type AutoHighlightsResult as AssemblyAIHighlightsResult, type TranscriptOptionalParams as AssemblyAIOptions, AssemblyAISampleRateType, type SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAISpeechModelType, type AssemblyAIStreamingOptions, type TopicDetectionModelResult as AssemblyAITopicsResult, index as AssemblyAITypes, type AssemblyAIUpdateConfiguration, AssemblyAIWebhookHandler, type TranscriptWebhookNotification as AssemblyAIWebhookPayload, type AudioAckEvent, type AudioChunk, type AudioInput, AudioResponseFormat, AudioTranscriptionModel, AzureSTTAdapter, AzureWebhookHandler, BaseAdapter, BaseWebhookHandler, type BatchOnlyProvider, type ChapterizationEvent, DeepgramAdapter, type DeepgramExtendedData, type ListenV1ResponseMetadata as DeepgramMetadata, DeepgramModelType, type ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramRedactType, type DeepgramStreamingOptions, DeepgramTopicModeType, DeepgramWebhookHandler, type ListenV1Response as DeepgramWebhookPayload, type EntityEvent, GladiaAdapter, type AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, type AudioToLlmListDTO as GladiaAudioToLlmResult, type ChapterizationDTO as GladiaChapters, type CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, type NamedEntityRecognitionDTO as GladiaEntities, type GladiaExtendedData, type ModerationDTO as GladiaModeration, type InitTranscriptionRequest as GladiaOptions, type SentimentAnalysisDTO as GladiaSentiment, type SpeakerReidentificationDTO as GladiaSpeakerReidentification, type GladiaStreamingOptions, type StreamingRequest as GladiaStreamingRequest, type StructuredDataExtractionDTO as GladiaStructuredData, type TranslationDTO as GladiaTranslation, index$1 as GladiaTypes, type CallbackTranscriptionErrorPayload as GladiaWebhookErrorPayload, GladiaWebhookHandler, type GladiaWebhookPayload, type CallbackTranscriptionSuccessPayload as GladiaWebhookSuccessPayload, type LifecycleEvent, type ListTranscriptsOptions, type ListTranscriptsResponse, ListenV1EncodingParameter, type ListenV1LanguageParameter, type ListenV1ModelParameter, type ListenV1VersionParameter, OpenAIWhisperAdapter, type CreateTranscriptionRequest as OpenAIWhisperOptions, type ProviderCapabilities, type ProviderConfig, type ProviderExtendedDataMap, type ProviderRawResponseMap, type ProviderStreamingOptions, type ProviderWebhookPayloadMap, type SentimentEvent, type SessionStatus, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type Speaker, type SpeechEvent, SpeechmaticsAdapter, type SpeechmaticsOperatingPoint, SpeechmaticsWebhookHandler, type StreamEvent, type StreamEventType, type StreamingCallbacks, type StreamingOptions, type StreamingOptionsForProvider, type StreamingProvider, type StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedSampleRateEnum, type SummarizationEvent, type TranscribeOptions, type TranscribeStreamParams, type TranscriptData, type TranscriptMetadata, type TranscriptionAdapter, type TranscriptionLanguage, type TranscriptionModel, type TranscriptionProvider, type TranscriptionStatus, type TranslationEvent, type UnifiedTranscriptResponse, type UnifiedWebhookEvent, type Utterance, VoiceRouter, type VoiceRouterConfig, type WebhookEventType, WebhookRouter, type WebhookRouterOptions, type WebhookRouterResult, type WebhookValidation, type WebhookVerificationOptions, type Word, createAssemblyAIAdapter, createAssemblyAIWebhookHandler, createAzureSTTAdapter, createAzureWebhookHandler, createDeepgramAdapter, createDeepgramWebhookHandler, createGladiaAdapter, createGladiaWebhookHandler, createOpenAIWhisperAdapter, createSpeechmaticsAdapter, createVoiceRouter, createWebhookRouter };
|