voice-router-dev 0.8.3 → 0.8.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2501,648 +2501,878 @@ type CreateTranscription200One = CreateTranscriptionResponseJson | CreateTranscr
2501
2501
  * Generated by orval v7.9.0 🍺
2502
2502
  * Do not edit manually.
2503
2503
  * AssemblyAI API
2504
- * AssemblyAI API
2505
- * OpenAPI spec version: 1.3.4
2504
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2505
+ * OpenAPI spec version: 1.0.0
2506
2506
  */
2507
2507
  /**
2508
- * The status of your transcript. Possible values are queued, processing, completed, or error.
2508
+ * The duration of this transcript object's media file, in seconds
2509
2509
  */
2510
- type TranscriptStatus = (typeof TranscriptStatus)[keyof typeof TranscriptStatus];
2511
- declare const TranscriptStatus: {
2512
- readonly queued: "queued";
2513
- readonly processing: "processing";
2514
- readonly completed: "completed";
2515
- readonly error: "error";
2516
- };
2510
+ type TranscriptAudioDuration = number | null;
2517
2511
 
2518
2512
  /**
2519
2513
  * Generated by orval v7.9.0 🍺
2520
2514
  * Do not edit manually.
2521
2515
  * AssemblyAI API
2522
- * AssemblyAI API
2523
- * OpenAPI spec version: 1.3.4
2516
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2517
+ * OpenAPI spec version: 1.0.0
2524
2518
  */
2525
2519
  /**
2526
- * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
2527
- The default value is 'en_us'.
2528
-
2520
+ * The point in time, in milliseconds, in the file at which the transcription was terminated. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details.
2529
2521
  */
2530
- type TranscriptLanguageCode = (typeof TranscriptLanguageCode)[keyof typeof TranscriptLanguageCode];
2531
- declare const TranscriptLanguageCode: {
2532
- readonly en: "en";
2533
- readonly en_au: "en_au";
2534
- readonly en_uk: "en_uk";
2535
- readonly en_us: "en_us";
2536
- readonly es: "es";
2537
- readonly fr: "fr";
2538
- readonly de: "de";
2539
- readonly it: "it";
2540
- readonly pt: "pt";
2541
- readonly nl: "nl";
2542
- readonly af: "af";
2543
- readonly sq: "sq";
2544
- readonly am: "am";
2545
- readonly ar: "ar";
2546
- readonly hy: "hy";
2547
- readonly as: "as";
2548
- readonly az: "az";
2549
- readonly ba: "ba";
2550
- readonly eu: "eu";
2551
- readonly be: "be";
2552
- readonly bn: "bn";
2553
- readonly bs: "bs";
2554
- readonly br: "br";
2555
- readonly bg: "bg";
2556
- readonly my: "my";
2557
- readonly ca: "ca";
2558
- readonly zh: "zh";
2559
- readonly hr: "hr";
2560
- readonly cs: "cs";
2561
- readonly da: "da";
2562
- readonly et: "et";
2563
- readonly fo: "fo";
2564
- readonly fi: "fi";
2565
- readonly gl: "gl";
2566
- readonly ka: "ka";
2567
- readonly el: "el";
2568
- readonly gu: "gu";
2569
- readonly ht: "ht";
2570
- readonly ha: "ha";
2571
- readonly haw: "haw";
2572
- readonly he: "he";
2573
- readonly hi: "hi";
2574
- readonly hu: "hu";
2575
- readonly is: "is";
2576
- readonly id: "id";
2577
- readonly ja: "ja";
2578
- readonly jw: "jw";
2579
- readonly kn: "kn";
2580
- readonly kk: "kk";
2581
- readonly km: "km";
2582
- readonly ko: "ko";
2583
- readonly lo: "lo";
2584
- readonly la: "la";
2585
- readonly lv: "lv";
2586
- readonly ln: "ln";
2587
- readonly lt: "lt";
2588
- readonly lb: "lb";
2589
- readonly mk: "mk";
2590
- readonly mg: "mg";
2591
- readonly ms: "ms";
2592
- readonly ml: "ml";
2593
- readonly mt: "mt";
2594
- readonly mi: "mi";
2595
- readonly mr: "mr";
2596
- readonly mn: "mn";
2597
- readonly ne: "ne";
2598
- readonly no: "no";
2599
- readonly nn: "nn";
2600
- readonly oc: "oc";
2601
- readonly pa: "pa";
2602
- readonly ps: "ps";
2603
- readonly fa: "fa";
2604
- readonly pl: "pl";
2605
- readonly ro: "ro";
2606
- readonly ru: "ru";
2607
- readonly sa: "sa";
2608
- readonly sr: "sr";
2609
- readonly sn: "sn";
2610
- readonly sd: "sd";
2611
- readonly si: "si";
2612
- readonly sk: "sk";
2613
- readonly sl: "sl";
2614
- readonly so: "so";
2615
- readonly su: "su";
2616
- readonly sw: "sw";
2617
- readonly sv: "sv";
2618
- readonly tl: "tl";
2619
- readonly tg: "tg";
2620
- readonly ta: "ta";
2621
- readonly tt: "tt";
2622
- readonly te: "te";
2623
- readonly th: "th";
2624
- readonly bo: "bo";
2625
- readonly tr: "tr";
2626
- readonly tk: "tk";
2627
- readonly uk: "uk";
2628
- readonly ur: "ur";
2629
- readonly uz: "uz";
2630
- readonly vi: "vi";
2631
- readonly cy: "cy";
2632
- readonly yi: "yi";
2633
- readonly yo: "yo";
2634
- };
2522
+ type TranscriptAudioEndAt = number | null;
2635
2523
 
2636
2524
  /**
2637
2525
  * Generated by orval v7.9.0 🍺
2638
2526
  * Do not edit manually.
2639
2527
  * AssemblyAI API
2640
- * AssemblyAI API
2641
- * OpenAPI spec version: 1.3.4
2528
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2529
+ * OpenAPI spec version: 1.0.0
2642
2530
  */
2643
-
2644
2531
  /**
2645
- * The language of your audio file.
2646
- Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
2647
- The default value is 'en_us'.
2648
-
2532
+ * The point in time, in milliseconds, in the file at which the transcription was started. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details.
2649
2533
  */
2650
- type TranscriptLanguageCodeProperty = TranscriptLanguageCode | string;
2534
+ type TranscriptAudioStartFrom = number | null;
2651
2535
 
2652
2536
  /**
2653
2537
  * Generated by orval v7.9.0 🍺
2654
2538
  * Do not edit manually.
2655
2539
  * AssemblyAI API
2656
- * AssemblyAI API
2657
- * OpenAPI spec version: 1.3.4
2540
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2541
+ * OpenAPI spec version: 1.0.0
2658
2542
  */
2659
2543
  /**
2660
- * Whether [Automatic language detection](https://www.assemblyai.com/docs/models/speech-recognition#automatic-language-detection) is enabled, either true or false
2544
+ * Whether [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) is enabled, can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible chapter summaries. See the [updated Auto Chapters page](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for details.
2545
+
2546
+ Note: This parameter is only supported for the Universal-2 model.
2547
+
2661
2548
  */
2662
- type TranscriptLanguageDetection = boolean | null;
2549
+ type TranscriptAutoChapters = boolean | null;
2663
2550
 
2664
2551
  /**
2665
2552
  * Generated by orval v7.9.0 🍺
2666
2553
  * Do not edit manually.
2667
2554
  * AssemblyAI API
2668
- * AssemblyAI API
2669
- * OpenAPI spec version: 1.3.4
2555
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2556
+ * OpenAPI spec version: 1.0.0
2670
2557
  */
2671
2558
  /**
2672
- * The confidence threshold for the automatically detected language.
2673
- An error will be returned if the language confidence is below this threshold.
2674
-
2675
- * @minimum 0
2676
- * @maximum 1
2559
+ * Either success, or unavailable in the rare case that the model failed
2677
2560
  */
2678
- type TranscriptLanguageConfidenceThreshold = number | null;
2561
+ type AudioIntelligenceModelStatus = (typeof AudioIntelligenceModelStatus)[keyof typeof AudioIntelligenceModelStatus];
2562
+ declare const AudioIntelligenceModelStatus: {
2563
+ readonly success: "success";
2564
+ readonly unavailable: "unavailable";
2565
+ };
2679
2566
 
2680
2567
  /**
2681
2568
  * Generated by orval v7.9.0 🍺
2682
2569
  * Do not edit manually.
2683
2570
  * AssemblyAI API
2684
- * AssemblyAI API
2685
- * OpenAPI spec version: 1.3.4
2571
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2572
+ * OpenAPI spec version: 1.0.0
2686
2573
  */
2687
2574
  /**
2688
- * The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence)
2689
- * @minimum 0
2690
- * @maximum 1
2575
+ * Timestamp containing a start and end property in milliseconds
2691
2576
  */
2692
- type TranscriptLanguageConfidence = number | null;
2577
+ interface Timestamp {
2578
+ /** The start time in milliseconds */
2579
+ start: number;
2580
+ /** The end time in milliseconds */
2581
+ end: number;
2582
+ }
2693
2583
 
2694
2584
  /**
2695
2585
  * Generated by orval v7.9.0 🍺
2696
2586
  * Do not edit manually.
2697
2587
  * AssemblyAI API
2698
- * AssemblyAI API
2699
- * OpenAPI spec version: 1.3.4
2700
- */
2701
- /**
2702
- * The speech model to use for the transcription.
2588
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2589
+ * OpenAPI spec version: 1.0.0
2703
2590
  */
2704
- type SpeechModel = (typeof SpeechModel)[keyof typeof SpeechModel];
2705
- declare const SpeechModel: {
2706
- readonly best: "best";
2707
- readonly "slam-1": "slam-1";
2708
- readonly universal: "universal";
2709
- };
2591
+
2592
+ interface AutoHighlightResult {
2593
+ /** The total number of times the key phrase appears in the audio file */
2594
+ count: number;
2595
+ /** The total relevancy to the overall audio file of this key phrase - a greater number means more relevant */
2596
+ rank: number;
2597
+ /** The text itself of the key phrase */
2598
+ text: string;
2599
+ /** The timestamp of the of the key phrase */
2600
+ timestamps: Timestamp[];
2601
+ }
2710
2602
 
2711
2603
  /**
2712
2604
  * Generated by orval v7.9.0 🍺
2713
2605
  * Do not edit manually.
2714
2606
  * AssemblyAI API
2715
- * AssemblyAI API
2716
- * OpenAPI spec version: 1.3.4
2607
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2608
+ * OpenAPI spec version: 1.0.0
2717
2609
  */
2718
2610
 
2719
2611
  /**
2720
- * The speech model used for the transcription. When `null`, the default model is used.
2612
+ * An array of results for the Key Phrases model, if it is enabled.
2613
+ See [Key phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) for more information.
2614
+
2721
2615
  */
2722
- type TranscriptSpeechModel = SpeechModel | null;
2616
+ interface AutoHighlightsResult {
2617
+ /** The status of the Key Phrases model. Either success, or unavailable in the rare case that the model failed. */
2618
+ status: AudioIntelligenceModelStatus;
2619
+ /** A temporally-sequential array of Key Phrases */
2620
+ results: AutoHighlightResult[];
2621
+ }
2723
2622
 
2724
2623
  /**
2725
2624
  * Generated by orval v7.9.0 🍺
2726
2625
  * Do not edit manually.
2727
2626
  * AssemblyAI API
2728
- * AssemblyAI API
2729
- * OpenAPI spec version: 1.3.4
2627
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2628
+ * OpenAPI spec version: 1.0.0
2730
2629
  */
2630
+
2731
2631
  /**
2732
- * The textual transcript of your media file
2632
+ * An array of results for the Key Phrases model, if it is enabled.
2633
+ See [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) for more information.
2634
+
2733
2635
  */
2734
- type TranscriptText = string | null;
2636
+ type TranscriptAutoHighlightsResult = AutoHighlightsResult | null;
2735
2637
 
2736
2638
  /**
2737
2639
  * Generated by orval v7.9.0 🍺
2738
2640
  * Do not edit manually.
2739
2641
  * AssemblyAI API
2740
- * AssemblyAI API
2741
- * OpenAPI spec version: 1.3.4
2642
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2643
+ * OpenAPI spec version: 1.0.0
2742
2644
  */
2743
2645
  /**
2744
- * The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
2646
+ * Chapter of the audio file
2745
2647
  */
2746
- type TranscriptWordChannel = string | null;
2648
+ interface Chapter$1 {
2649
+ /** An ultra-short summary (just a few words) of the content spoken in the chapter */
2650
+ gist: string;
2651
+ /** A single sentence summary of the content spoken during the chapter */
2652
+ headline: string;
2653
+ /** A one paragraph summary of the content spoken during the chapter */
2654
+ summary: string;
2655
+ /** The starting time, in milliseconds, for the chapter */
2656
+ start: number;
2657
+ /** The starting time, in milliseconds, for the chapter */
2658
+ end: number;
2659
+ }
2747
2660
 
2748
2661
  /**
2749
2662
  * Generated by orval v7.9.0 🍺
2750
2663
  * Do not edit manually.
2751
2664
  * AssemblyAI API
2752
- * AssemblyAI API
2753
- * OpenAPI spec version: 1.3.4
2665
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2666
+ * OpenAPI spec version: 1.0.0
2754
2667
  */
2668
+
2755
2669
  /**
2756
- * The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null
2670
+ * An array of temporally sequential chapters for the audio file. See [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for more information.
2757
2671
  */
2758
- type TranscriptWordSpeaker = string | null;
2672
+ type TranscriptChapters = Chapter$1[] | null;
2759
2673
 
2760
2674
  /**
2761
2675
  * Generated by orval v7.9.0 🍺
2762
2676
  * Do not edit manually.
2763
2677
  * AssemblyAI API
2764
- * AssemblyAI API
2765
- * OpenAPI spec version: 1.3.4
2678
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2679
+ * OpenAPI spec version: 1.0.0
2766
2680
  */
2767
-
2768
- interface TranscriptWord {
2769
- /**
2770
- * The confidence score for the transcript of this word
2771
- * @minimum 0
2772
- * @maximum 1
2773
- */
2774
- confidence: number;
2775
- /** The starting time, in milliseconds, for the word */
2776
- start: number;
2777
- /** The ending time, in milliseconds, for the word */
2778
- end: number;
2779
- /** The text of the word */
2780
- text: string;
2781
- /** The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
2782
- channel?: TranscriptWordChannel;
2783
- /** The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null */
2784
- speaker: TranscriptWordSpeaker;
2785
- }
2681
+ /**
2682
+ * The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence)
2683
+ */
2684
+ type TranscriptConfidence = number | null;
2786
2685
 
2787
2686
  /**
2788
2687
  * Generated by orval v7.9.0 🍺
2789
2688
  * Do not edit manually.
2790
2689
  * AssemblyAI API
2791
- * AssemblyAI API
2792
- * OpenAPI spec version: 1.3.4
2690
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2691
+ * OpenAPI spec version: 1.0.0
2793
2692
  */
2794
-
2795
2693
  /**
2796
- * An array of temporally-sequential word objects, one for each word in the transcript.
2797
- See [Speech recognition](https://www.assemblyai.com/docs/models/speech-recognition) for more information.
2798
-
2694
+ * Whether [Content Moderation](https://www.assemblyai.com/docs/content-moderation) is enabled, can be true or false
2799
2695
  */
2800
- type TranscriptWords = TranscriptWord[] | null;
2696
+ type TranscriptContentSafety = boolean | null;
2801
2697
 
2802
2698
  /**
2803
2699
  * Generated by orval v7.9.0 🍺
2804
2700
  * Do not edit manually.
2805
2701
  * AssemblyAI API
2806
- * AssemblyAI API
2807
- * OpenAPI spec version: 1.3.4
2808
- */
2809
- /**
2810
- * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
2702
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2703
+ * OpenAPI spec version: 1.0.0
2811
2704
  */
2812
- type TranscriptUtteranceChannel = string | null;
2705
+ interface ContentSafetyLabel {
2706
+ /** The label of the sensitive topic */
2707
+ label: string;
2708
+ /** The confidence score for the topic being discussed, from 0 to 1 */
2709
+ confidence: number;
2710
+ /** How severely the topic is discussed in the section, from 0 to 1 */
2711
+ severity: number;
2712
+ }
2813
2713
 
2814
2714
  /**
2815
2715
  * Generated by orval v7.9.0 🍺
2816
2716
  * Do not edit manually.
2817
2717
  * AssemblyAI API
2818
- * AssemblyAI API
2819
- * OpenAPI spec version: 1.3.4
2718
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2719
+ * OpenAPI spec version: 1.0.0
2820
2720
  */
2821
2721
 
2822
- interface TranscriptUtterance {
2823
- /**
2824
- * The confidence score for the transcript of this utterance
2825
- * @minimum 0
2826
- * @maximum 1
2827
- */
2828
- confidence: number;
2829
- /** The starting time, in milliseconds, of the utterance in the audio file */
2830
- start: number;
2831
- /** The ending time, in milliseconds, of the utterance in the audio file */
2832
- end: number;
2833
- /** The text for this utterance */
2722
+ interface ContentSafetyLabelResult {
2723
+ /** The transcript of the section flagged by the Content Moderation model */
2834
2724
  text: string;
2835
- /** The words in the utterance. */
2836
- words: TranscriptWord[];
2837
- /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
2838
- channel?: TranscriptUtteranceChannel;
2839
- /** The speaker of this utterance, where each speaker is assigned a sequential capital letter - e.g. "A" for Speaker A, "B" for Speaker B, etc. */
2840
- speaker: string;
2725
+ /** An array of safety labels, one per sensitive topic that was detected in the section */
2726
+ labels: ContentSafetyLabel[];
2727
+ /** The sentence index at which the section begins */
2728
+ sentences_idx_start: number;
2729
+ /** The sentence index at which the section ends */
2730
+ sentences_idx_end: number;
2731
+ /** Timestamp information for the section */
2732
+ timestamp: Timestamp;
2841
2733
  }
2842
2734
 
2843
2735
  /**
2844
2736
  * Generated by orval v7.9.0 🍺
2845
2737
  * Do not edit manually.
2846
2738
  * AssemblyAI API
2739
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2740
+ * OpenAPI spec version: 1.0.0
2741
+ */
2742
+ /**
2743
+ * A summary of the Content Moderation confidence results for the entire audio file
2744
+ */
2745
+ type ContentSafetyLabelsResultSummary = {
2746
+ [key: string]: number;
2747
+ };
2748
+
2749
+ /**
2750
+ * Generated by orval v7.9.0 🍺
2751
+ * Do not edit manually.
2847
2752
  * AssemblyAI API
2848
- * OpenAPI spec version: 1.3.4
2753
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2754
+ * OpenAPI spec version: 1.0.0
2849
2755
  */
2756
+ interface SeverityScoreSummary {
2757
+ low: number;
2758
+ medium: number;
2759
+ high: number;
2760
+ }
2850
2761
 
2851
2762
  /**
2852
- * When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
2853
- See [Speaker diarization](https://www.assemblyai.com/docs/speech-to-text/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/speech-to-text/speech-recognition#multichannel-transcription) for more information.
2763
+ * Generated by orval v7.9.0 🍺
2764
+ * Do not edit manually.
2765
+ * AssemblyAI API
2766
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2767
+ * OpenAPI spec version: 1.0.0
2768
+ */
2854
2769
 
2770
+ /**
2771
+ * A summary of the Content Moderation severity results for the entire audio file
2855
2772
  */
2856
- type TranscriptUtterances = TranscriptUtterance[] | null;
2773
+ type ContentSafetyLabelsResultSeverityScoreSummary = {
2774
+ [key: string]: SeverityScoreSummary;
2775
+ };
2857
2776
 
2858
2777
  /**
2859
2778
  * Generated by orval v7.9.0 🍺
2860
2779
  * Do not edit manually.
2861
2780
  * AssemblyAI API
2862
- * AssemblyAI API
2863
- * OpenAPI spec version: 1.3.4
2781
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2782
+ * OpenAPI spec version: 1.0.0
2864
2783
  */
2784
+
2865
2785
  /**
2866
- * The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence)
2867
- * @minimum 0
2868
- * @maximum 1
2786
+ * An array of results for the Content Moderation model, if it is enabled.
2787
+ See [Content moderation](https://www.assemblyai.com/docs/content-moderation) for more information.
2788
+
2869
2789
  */
2870
- type TranscriptConfidence = number | null;
2790
+ interface ContentSafetyLabelsResult {
2791
+ /** The status of the Content Moderation model. Either success, or unavailable in the rare case that the model failed. */
2792
+ status: AudioIntelligenceModelStatus;
2793
+ /** An array of results for the Content Moderation model */
2794
+ results: ContentSafetyLabelResult[];
2795
+ /** A summary of the Content Moderation confidence results for the entire audio file */
2796
+ summary: ContentSafetyLabelsResultSummary;
2797
+ /** A summary of the Content Moderation severity results for the entire audio file */
2798
+ severity_score_summary: ContentSafetyLabelsResultSeverityScoreSummary;
2799
+ }
2871
2800
 
2872
2801
  /**
2873
2802
  * Generated by orval v7.9.0 🍺
2874
2803
  * Do not edit manually.
2875
2804
  * AssemblyAI API
2876
- * AssemblyAI API
2877
- * OpenAPI spec version: 1.3.4
2805
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2806
+ * OpenAPI spec version: 1.0.0
2878
2807
  */
2808
+
2879
2809
  /**
2880
- * The duration of this transcript object's media file, in seconds
2810
+ * An array of results for the Content Moderation model, if it is enabled.
2811
+ See [Content moderation](https://www.assemblyai.com/docs/content-moderation) for more information.
2812
+
2881
2813
  */
2882
- type TranscriptAudioDuration = number | null;
2814
+ type TranscriptContentSafetyLabels = ContentSafetyLabelsResult | null;
2883
2815
 
2884
2816
  /**
2885
2817
  * Generated by orval v7.9.0 🍺
2886
2818
  * Do not edit manually.
2887
2819
  * AssemblyAI API
2820
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2821
+ * OpenAPI spec version: 1.0.0
2822
+ */
2823
+ /**
2824
+ * Object containing words or phrases to replace, and the word or phrase to replace with
2825
+ */
2826
+ interface TranscriptCustomSpelling {
2827
+ /** Words or phrases to replace */
2828
+ from: string[];
2829
+ /** Word to replace with */
2830
+ to: string;
2831
+ }
2832
+
2833
+ /**
2834
+ * Generated by orval v7.9.0 🍺
2835
+ * Do not edit manually.
2888
2836
  * AssemblyAI API
2889
- * OpenAPI spec version: 1.3.4
2837
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2838
+ * OpenAPI spec version: 1.0.0
2890
2839
  */
2840
+
2891
2841
  /**
2892
- * Whether Automatic Punctuation is enabled, either true or false
2842
+ * Customize how words are spelled and formatted using to and from values. See [Custom Spelling](https://www.assemblyai.com/docs/pre-recorded-audio/custom-spelling) for more details.
2893
2843
  */
2894
- type TranscriptPunctuate = boolean | null;
2844
+ type TranscriptCustomSpellingProperty = TranscriptCustomSpelling[] | null;
2895
2845
 
2896
2846
  /**
2897
2847
  * Generated by orval v7.9.0 🍺
2898
2848
  * Do not edit manually.
2899
2849
  * AssemblyAI API
2850
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2851
+ * OpenAPI spec version: 1.0.0
2852
+ */
2853
+ /**
2854
+ * Transcribe [Filler Words](https://www.assemblyai.com/docs/pre-recorded-audio/filler-words), like "umm", in your media file; can be true or false
2855
+ */
2856
+ type TranscriptDisfluencies = boolean | null;
2857
+
2858
+ /**
2859
+ * Generated by orval v7.9.0 🍺
2860
+ * Do not edit manually.
2900
2861
  * AssemblyAI API
2901
- * OpenAPI spec version: 1.3.4
2862
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2863
+ * OpenAPI spec version: 1.0.0
2902
2864
  */
2903
2865
  /**
2904
- * Whether Text Formatting is enabled, either true or false
2866
+ * The domain-specific model applied to the transcript. When set to `"medical-v1"`, [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) was used to improve accuracy for medical terminology.
2867
+
2905
2868
  */
2906
- type TranscriptFormatText = boolean | null;
2869
+ type TranscriptDomain = string | null;
2907
2870
 
2908
2871
  /**
2909
2872
  * Generated by orval v7.9.0 🍺
2910
2873
  * Do not edit manually.
2911
2874
  * AssemblyAI API
2875
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2876
+ * OpenAPI spec version: 1.0.0
2877
+ */
2878
+ /**
2879
+ * The type of entity for the detected entity
2880
+ */
2881
+ type EntityType = (typeof EntityType)[keyof typeof EntityType];
2882
+ declare const EntityType: {
2883
+ readonly account_number: "account_number";
2884
+ readonly banking_information: "banking_information";
2885
+ readonly blood_type: "blood_type";
2886
+ readonly credit_card_cvv: "credit_card_cvv";
2887
+ readonly credit_card_expiration: "credit_card_expiration";
2888
+ readonly credit_card_number: "credit_card_number";
2889
+ readonly date: "date";
2890
+ readonly date_interval: "date_interval";
2891
+ readonly date_of_birth: "date_of_birth";
2892
+ readonly drivers_license: "drivers_license";
2893
+ readonly drug: "drug";
2894
+ readonly duration: "duration";
2895
+ readonly email_address: "email_address";
2896
+ readonly event: "event";
2897
+ readonly filename: "filename";
2898
+ readonly gender_sexuality: "gender_sexuality";
2899
+ readonly healthcare_number: "healthcare_number";
2900
+ readonly injury: "injury";
2901
+ readonly ip_address: "ip_address";
2902
+ readonly language: "language";
2903
+ readonly location: "location";
2904
+ readonly marital_status: "marital_status";
2905
+ readonly medical_condition: "medical_condition";
2906
+ readonly medical_process: "medical_process";
2907
+ readonly money_amount: "money_amount";
2908
+ readonly nationality: "nationality";
2909
+ readonly number_sequence: "number_sequence";
2910
+ readonly occupation: "occupation";
2911
+ readonly organization: "organization";
2912
+ readonly passport_number: "passport_number";
2913
+ readonly password: "password";
2914
+ readonly person_age: "person_age";
2915
+ readonly person_name: "person_name";
2916
+ readonly phone_number: "phone_number";
2917
+ readonly physical_attribute: "physical_attribute";
2918
+ readonly political_affiliation: "political_affiliation";
2919
+ readonly religion: "religion";
2920
+ readonly statistics: "statistics";
2921
+ readonly time: "time";
2922
+ readonly url: "url";
2923
+ readonly us_social_security_number: "us_social_security_number";
2924
+ readonly username: "username";
2925
+ readonly vehicle_id: "vehicle_id";
2926
+ readonly zodiac_sign: "zodiac_sign";
2927
+ };
2928
+
2929
+ /**
2930
+ * Generated by orval v7.9.0 🍺
2931
+ * Do not edit manually.
2912
2932
  * AssemblyAI API
2913
- * OpenAPI spec version: 1.3.4
2933
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2934
+ * OpenAPI spec version: 1.0.0
2914
2935
  */
2936
+
2915
2937
  /**
2916
- * Transcribe Filler Words, like "umm", in your media file; can be true or false
2938
+ * A detected entity
2917
2939
  */
2918
- type TranscriptDisfluencies = boolean | null;
2940
+ interface Entity {
2941
+ /** The type of entity for the detected entity */
2942
+ entity_type: EntityType;
2943
+ /** The text for the detected entity */
2944
+ text: string;
2945
+ /** The starting time, in milliseconds, at which the detected entity appears in the audio file */
2946
+ start: number;
2947
+ /** The ending time, in milliseconds, for the detected entity in the audio file */
2948
+ end: number;
2949
+ }
2919
2950
 
2920
2951
  /**
2921
2952
  * Generated by orval v7.9.0 🍺
2922
2953
  * Do not edit manually.
2923
2954
  * AssemblyAI API
2955
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2956
+ * OpenAPI spec version: 1.0.0
2957
+ */
2958
+
2959
+ /**
2960
+ * An array of results for the Entity Detection model, if it is enabled.
2961
+ See [Entity detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) for more information.
2962
+
2963
+ */
2964
+ type TranscriptEntities = Entity[] | null;
2965
+
2966
+ /**
2967
+ * Generated by orval v7.9.0 🍺
2968
+ * Do not edit manually.
2924
2969
  * AssemblyAI API
2925
- * OpenAPI spec version: 1.3.4
2970
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2971
+ * OpenAPI spec version: 1.0.0
2926
2972
  */
2927
2973
  /**
2928
- * Whether [Multichannel transcription](https://www.assemblyai.com/docs/models/speech-recognition#multichannel-transcription) was enabled in the transcription request, either true or false
2974
+ * Whether [Entity Detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) is enabled, can be true or false
2929
2975
  */
2930
- type TranscriptMultichannel = boolean | null;
2976
+ type TranscriptEntityDetection = boolean | null;
2931
2977
 
2932
2978
  /**
2933
2979
  * Generated by orval v7.9.0 🍺
2934
2980
  * Do not edit manually.
2935
2981
  * AssemblyAI API
2982
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2983
+ * OpenAPI spec version: 1.0.0
2984
+ */
2985
+ /**
2986
+ * Whether [Profanity Filtering](https://www.assemblyai.com/docs/profanity-filtering) is enabled, either true or false
2987
+ */
2988
+ type TranscriptFilterProfanity = boolean | null;
2989
+
2990
+ /**
2991
+ * Generated by orval v7.9.0 🍺
2992
+ * Do not edit manually.
2936
2993
  * AssemblyAI API
2937
- * OpenAPI spec version: 1.3.4
2994
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2995
+ * OpenAPI spec version: 1.0.0
2938
2996
  */
2939
2997
  /**
2940
- * The URL to which we send webhook requests.
2941
- We sends two different types of webhook requests.
2942
- One request when a transcript is completed or failed, and one request when the redacted audio is ready if redact_pii_audio is enabled.
2998
+ * Whether [Text Formatting](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false
2999
+ */
3000
+ type TranscriptFormatText = boolean | null;
2943
3001
 
3002
+ /**
3003
+ * Generated by orval v7.9.0 🍺
3004
+ * Do not edit manually.
3005
+ * AssemblyAI API
3006
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3007
+ * OpenAPI spec version: 1.0.0
2944
3008
  */
2945
- type TranscriptWebhookUrl = string | null;
3009
+ /**
3010
+ * Whether [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) is enabled, can be true or false
3011
+ */
3012
+ type TranscriptIabCategories = boolean | null;
2946
3013
 
2947
3014
  /**
2948
3015
  * Generated by orval v7.9.0 🍺
2949
3016
  * Do not edit manually.
2950
3017
  * AssemblyAI API
3018
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3019
+ * OpenAPI spec version: 1.0.0
3020
+ */
3021
+ interface TopicDetectionResultLabelsItems {
3022
+ /** How relevant the detected topic is of a detected topic */
3023
+ relevance: number;
3024
+ /** The IAB taxonomical label for the label of the detected topic, where > denotes supertopic/subtopic relationship */
3025
+ label: string;
3026
+ }
3027
+
3028
+ /**
3029
+ * Generated by orval v7.9.0 🍺
3030
+ * Do not edit manually.
2951
3031
  * AssemblyAI API
2952
- * OpenAPI spec version: 1.3.4
3032
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3033
+ * OpenAPI spec version: 1.0.0
2953
3034
  */
3035
+
2954
3036
  /**
2955
- * The status code we received from your server when delivering the transcript completed or failed webhook request, if a webhook URL was provided
3037
+ * The result of the topic detection model
2956
3038
  */
2957
- type TranscriptWebhookStatusCode = number | null;
3039
+ interface TopicDetectionResult$1 {
3040
+ /** The text in the transcript in which a detected topic occurs */
3041
+ text: string;
3042
+ /** An array of detected topics in the text */
3043
+ labels?: TopicDetectionResultLabelsItems[];
3044
+ timestamp?: Timestamp;
3045
+ }
2958
3046
 
2959
3047
  /**
2960
3048
  * Generated by orval v7.9.0 🍺
2961
3049
  * Do not edit manually.
2962
3050
  * AssemblyAI API
2963
- * AssemblyAI API
2964
- * OpenAPI spec version: 1.3.4
3051
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3052
+ * OpenAPI spec version: 1.0.0
2965
3053
  */
2966
3054
  /**
2967
- * The header name to be sent with the transcript completed or failed webhook requests
3055
+ * The overall relevance of topic to the entire audio file
2968
3056
  */
2969
- type TranscriptWebhookAuthHeaderName = string | null;
3057
+ type TopicDetectionModelResultSummary = {
3058
+ [key: string]: number;
3059
+ };
2970
3060
 
2971
3061
  /**
2972
3062
  * Generated by orval v7.9.0 🍺
2973
3063
  * Do not edit manually.
2974
3064
  * AssemblyAI API
3065
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3066
+ * OpenAPI spec version: 1.0.0
3067
+ */
3068
+
3069
+ /**
3070
+ * The result of the Topic Detection model, if it is enabled.
3071
+ See [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) for more information.
3072
+
3073
+ */
3074
+ interface TopicDetectionModelResult {
3075
+ /** The status of the Topic Detection model. Either success, or unavailable in the rare case that the model failed. */
3076
+ status: AudioIntelligenceModelStatus;
3077
+ /** An array of results for the Topic Detection model */
3078
+ results: TopicDetectionResult$1[];
3079
+ /** The overall relevance of topic to the entire audio file */
3080
+ summary: TopicDetectionModelResultSummary;
3081
+ }
3082
+
3083
+ /**
3084
+ * Generated by orval v7.9.0 🍺
3085
+ * Do not edit manually.
2975
3086
  * AssemblyAI API
2976
- * OpenAPI spec version: 1.3.4
3087
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3088
+ * OpenAPI spec version: 1.0.0
2977
3089
  */
3090
+
2978
3091
  /**
2979
- * Whether speed boost is enabled
2980
- * @deprecated
3092
+ * The result of the Topic Detection model, if it is enabled.
3093
+ See [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) for more information.
3094
+
2981
3095
  */
2982
- type TranscriptSpeedBoost = boolean | null;
3096
+ type TranscriptIabCategoriesResult = TopicDetectionModelResult | null;
2983
3097
 
2984
3098
  /**
2985
3099
  * Generated by orval v7.9.0 🍺
2986
3100
  * Do not edit manually.
2987
3101
  * AssemblyAI API
3102
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3103
+ * OpenAPI spec version: 1.0.0
3104
+ */
3105
+ /**
3106
+ * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
3107
+ The default value is 'en_us'.
3108
+
3109
+ */
3110
+ type TranscriptLanguageCode = (typeof TranscriptLanguageCode)[keyof typeof TranscriptLanguageCode];
3111
+ declare const TranscriptLanguageCode: {
3112
+ readonly en: "en";
3113
+ readonly en_au: "en_au";
3114
+ readonly en_uk: "en_uk";
3115
+ readonly en_us: "en_us";
3116
+ readonly es: "es";
3117
+ readonly fr: "fr";
3118
+ readonly de: "de";
3119
+ readonly it: "it";
3120
+ readonly pt: "pt";
3121
+ readonly nl: "nl";
3122
+ readonly af: "af";
3123
+ readonly sq: "sq";
3124
+ readonly am: "am";
3125
+ readonly ar: "ar";
3126
+ readonly hy: "hy";
3127
+ readonly as: "as";
3128
+ readonly az: "az";
3129
+ readonly ba: "ba";
3130
+ readonly eu: "eu";
3131
+ readonly be: "be";
3132
+ readonly bn: "bn";
3133
+ readonly bs: "bs";
3134
+ readonly br: "br";
3135
+ readonly bg: "bg";
3136
+ readonly my: "my";
3137
+ readonly ca: "ca";
3138
+ readonly zh: "zh";
3139
+ readonly hr: "hr";
3140
+ readonly cs: "cs";
3141
+ readonly da: "da";
3142
+ readonly et: "et";
3143
+ readonly fo: "fo";
3144
+ readonly fi: "fi";
3145
+ readonly gl: "gl";
3146
+ readonly ka: "ka";
3147
+ readonly el: "el";
3148
+ readonly gu: "gu";
3149
+ readonly ht: "ht";
3150
+ readonly ha: "ha";
3151
+ readonly haw: "haw";
3152
+ readonly he: "he";
3153
+ readonly hi: "hi";
3154
+ readonly hu: "hu";
3155
+ readonly is: "is";
3156
+ readonly id: "id";
3157
+ readonly ja: "ja";
3158
+ readonly jw: "jw";
3159
+ readonly kn: "kn";
3160
+ readonly kk: "kk";
3161
+ readonly km: "km";
3162
+ readonly ko: "ko";
3163
+ readonly lo: "lo";
3164
+ readonly la: "la";
3165
+ readonly lv: "lv";
3166
+ readonly ln: "ln";
3167
+ readonly lt: "lt";
3168
+ readonly lb: "lb";
3169
+ readonly mk: "mk";
3170
+ readonly mg: "mg";
3171
+ readonly ms: "ms";
3172
+ readonly ml: "ml";
3173
+ readonly mt: "mt";
3174
+ readonly mi: "mi";
3175
+ readonly mr: "mr";
3176
+ readonly mn: "mn";
3177
+ readonly ne: "ne";
3178
+ readonly no: "no";
3179
+ readonly nn: "nn";
3180
+ readonly oc: "oc";
3181
+ readonly pa: "pa";
3182
+ readonly ps: "ps";
3183
+ readonly fa: "fa";
3184
+ readonly pl: "pl";
3185
+ readonly ro: "ro";
3186
+ readonly ru: "ru";
3187
+ readonly sa: "sa";
3188
+ readonly sr: "sr";
3189
+ readonly sn: "sn";
3190
+ readonly sd: "sd";
3191
+ readonly si: "si";
3192
+ readonly sk: "sk";
3193
+ readonly sl: "sl";
3194
+ readonly so: "so";
3195
+ readonly su: "su";
3196
+ readonly sw: "sw";
3197
+ readonly sv: "sv";
3198
+ readonly tl: "tl";
3199
+ readonly tg: "tg";
3200
+ readonly ta: "ta";
3201
+ readonly tt: "tt";
3202
+ readonly te: "te";
3203
+ readonly th: "th";
3204
+ readonly bo: "bo";
3205
+ readonly tr: "tr";
3206
+ readonly tk: "tk";
3207
+ readonly uk: "uk";
3208
+ readonly ur: "ur";
3209
+ readonly uz: "uz";
3210
+ readonly vi: "vi";
3211
+ readonly cy: "cy";
3212
+ readonly yi: "yi";
3213
+ readonly yo: "yo";
3214
+ };
3215
+
3216
+ /**
3217
+ * Generated by orval v7.9.0 🍺
3218
+ * Do not edit manually.
2988
3219
  * AssemblyAI API
2989
- * OpenAPI spec version: 1.3.4
3220
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3221
+ * OpenAPI spec version: 1.0.0
2990
3222
  */
3223
+
2991
3224
  /**
2992
- * Either success, or unavailable in the rare case that the model failed
3225
+ * The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
3226
+ One of the values specified must be `en`.
3227
+
2993
3228
  */
2994
- type AudioIntelligenceModelStatus = (typeof AudioIntelligenceModelStatus)[keyof typeof AudioIntelligenceModelStatus];
2995
- declare const AudioIntelligenceModelStatus: {
2996
- readonly success: "success";
2997
- readonly unavailable: "unavailable";
2998
- };
3229
+ type TranscriptLanguageCodes = TranscriptLanguageCode[] | null;
2999
3230
 
3000
3231
  /**
3001
3232
  * Generated by orval v7.9.0 🍺
3002
3233
  * Do not edit manually.
3003
3234
  * AssemblyAI API
3004
- * AssemblyAI API
3005
- * OpenAPI spec version: 1.3.4
3235
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3236
+ * OpenAPI spec version: 1.0.0
3006
3237
  */
3007
3238
  /**
3008
- * Timestamp containing a start and end property in milliseconds
3239
+ * The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence). See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
3009
3240
  */
3010
- interface Timestamp {
3011
- /** The start time in milliseconds */
3012
- start: number;
3013
- /** The end time in milliseconds */
3014
- end: number;
3015
- }
3241
+ type TranscriptLanguageConfidence = number | null;
3016
3242
 
3017
3243
  /**
3018
3244
  * Generated by orval v7.9.0 🍺
3019
3245
  * Do not edit manually.
3020
3246
  * AssemblyAI API
3021
- * AssemblyAI API
3022
- * OpenAPI spec version: 1.3.4
3247
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3248
+ * OpenAPI spec version: 1.0.0
3023
3249
  */
3250
+ /**
3251
+ * The confidence threshold for the automatically detected language.
3252
+ An error will be returned if the language confidence is below this threshold.
3253
+ See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
3024
3254
 
3025
- interface AutoHighlightResult {
3026
- /** The total number of times the key phrase appears in the audio file */
3027
- count: number;
3028
- /**
3029
- * The total relevancy to the overall audio file of this key phrase - a greater number means more relevant
3030
- * @minimum 0
3031
- * @maximum 1
3032
- */
3033
- rank: number;
3034
- /** The text itself of the key phrase */
3035
- text: string;
3036
- /** The timestamp of the of the key phrase */
3037
- timestamps: Timestamp[];
3038
- }
3255
+ */
3256
+ type TranscriptLanguageConfidenceThreshold = number | null;
3039
3257
 
3040
3258
  /**
3041
3259
  * Generated by orval v7.9.0 🍺
3042
3260
  * Do not edit manually.
3043
3261
  * AssemblyAI API
3044
- * AssemblyAI API
3045
- * OpenAPI spec version: 1.3.4
3262
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3263
+ * OpenAPI spec version: 1.0.0
3046
3264
  */
3047
-
3048
3265
  /**
3049
- * An array of results for the Key Phrases model, if it is enabled.
3050
- See [Key phrases](https://www.assemblyai.com/docs/models/key-phrases) for more information.
3051
-
3266
+ * Whether [Automatic language detection](/docs/pre-recorded-audio/automatic-language-detection) is enabled, either true or false
3052
3267
  */
3053
- interface AutoHighlightsResult {
3054
- /** The status of the Key Phrases model. Either success, or unavailable in the rare case that the model failed. */
3055
- status: AudioIntelligenceModelStatus;
3056
- /** A temporally-sequential array of Key Phrases */
3057
- results: AutoHighlightResult[];
3058
- }
3268
+ type TranscriptLanguageDetection = boolean | null;
3059
3269
 
3060
3270
  /**
3061
3271
  * Generated by orval v7.9.0 🍺
3062
3272
  * Do not edit manually.
3063
3273
  * AssemblyAI API
3064
- * AssemblyAI API
3065
- * OpenAPI spec version: 1.3.4
3274
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3275
+ * OpenAPI spec version: 1.0.0
3066
3276
  */
3067
-
3068
3277
  /**
3069
- * An array of results for the Key Phrases model, if it is enabled.
3070
- See [Key Phrases](https://www.assemblyai.com/docs/models/key-phrases) for more information.
3071
-
3278
+ * Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection).
3072
3279
  */
3073
- type TranscriptAutoHighlightsResult = AutoHighlightsResult | null;
3280
+ interface TranscriptLanguageDetectionOptions {
3281
+ /** List of languages expected in the audio file. Defaults to `["all"]` when unspecified. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details. */
3282
+ expected_languages?: string[];
3283
+ /** If the detected language of the audio file is not in the list of expected languages, the `fallback_language` is used. Specify `["auto"]` to let our model choose the fallback language from `expected_languages` with the highest confidence score. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
3284
+ */
3285
+ fallback_language?: string;
3286
+ /** Whether [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) should be detected.
3287
+ */
3288
+ code_switching?: boolean;
3289
+ /** The confidence threshold for [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) detection. If the code switching confidence is below this threshold, the transcript will be processed in the language with the highest `language_detection_confidence` score.
3290
+ */
3291
+ code_switching_confidence_threshold?: number;
3292
+ }
3074
3293
 
3075
3294
  /**
3076
3295
  * Generated by orval v7.9.0 🍺
3077
3296
  * Do not edit manually.
3078
3297
  * AssemblyAI API
3079
- * AssemblyAI API
3080
- * OpenAPI spec version: 1.3.4
3298
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3299
+ * OpenAPI spec version: 1.0.0
3081
3300
  */
3082
3301
  /**
3083
- * The point in time, in milliseconds, in the file at which the transcription was started
3302
+ * Whether [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) was enabled in the transcription request, either true or false
3084
3303
  */
3085
- type TranscriptAudioStartFrom = number | null;
3304
+ type TranscriptMultichannel = boolean | null;
3086
3305
 
3087
3306
  /**
3088
3307
  * Generated by orval v7.9.0 🍺
3089
3308
  * Do not edit manually.
3090
3309
  * AssemblyAI API
3091
- * AssemblyAI API
3092
- * OpenAPI spec version: 1.3.4
3310
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3311
+ * OpenAPI spec version: 1.0.0
3093
3312
  */
3094
3313
  /**
3095
- * The point in time, in milliseconds, in the file at which the transcription was terminated
3314
+ * Whether [Automatic Punctuation](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false
3096
3315
  */
3097
- type TranscriptAudioEndAt = number | null;
3316
+ type TranscriptPunctuate = boolean | null;
3098
3317
 
3099
3318
  /**
3100
3319
  * Generated by orval v7.9.0 🍺
3101
3320
  * Do not edit manually.
3102
3321
  * AssemblyAI API
3103
- * AssemblyAI API
3104
- * OpenAPI spec version: 1.3.4
3322
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3323
+ * OpenAPI spec version: 1.0.0
3105
3324
  */
3106
3325
  /**
3107
- * The word boost parameter value
3326
+ * Whether a redacted version of the audio file was generated,
3327
+ either true or false. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
3328
+
3108
3329
  */
3109
- type TranscriptBoostParamProperty = string | null;
3330
+ type TranscriptRedactPiiAudio = boolean | null;
3110
3331
 
3111
3332
  /**
3112
3333
  * Generated by orval v7.9.0 🍺
3113
3334
  * Do not edit manually.
3114
3335
  * AssemblyAI API
3115
- * AssemblyAI API
3116
- * OpenAPI spec version: 1.3.4
3336
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3337
+ * OpenAPI spec version: 1.0.0
3117
3338
  */
3118
3339
  /**
3119
- * Whether [Profanity Filtering](https://www.assemblyai.com/docs/models/speech-recognition#profanity-filtering) is enabled, either true or false
3340
+ * Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep.
3120
3341
  */
3121
- type TranscriptFilterProfanity = boolean | null;
3342
+ type TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod = (typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod)[keyof typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod];
3343
+ declare const TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod: {
3344
+ readonly silence: "silence";
3345
+ };
3122
3346
 
3123
3347
  /**
3124
3348
  * Generated by orval v7.9.0 🍺
3125
3349
  * Do not edit manually.
3126
3350
  * AssemblyAI API
3127
- * AssemblyAI API
3128
- * OpenAPI spec version: 1.3.4
3351
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3352
+ * OpenAPI spec version: 1.0.0
3129
3353
  */
3354
+
3130
3355
  /**
3131
- * Whether a redacted version of the audio file was generated,
3132
- either true or false. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
3356
+ * The options for PII-redacted audio, if redact_pii_audio is enabled.
3357
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
3133
3358
 
3134
3359
  */
3135
- type TranscriptRedactPiiAudio = boolean | null;
3360
+ interface TranscriptRedactPiiAudioOptions {
3361
+ /** By default, audio redaction provides redacted audio URLs only when speech is detected. However, if your use-case specifically requires redacted audio files even for silent audio files without any dialogue, you can opt to receive these URLs by setting this parameter to `true`. */
3362
+ return_redacted_no_speech_audio?: boolean;
3363
+ /** Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep. */
3364
+ override_audio_redaction_method?: TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod;
3365
+ }
3136
3366
 
3137
3367
  /**
3138
3368
  * Generated by orval v7.9.0 🍺
3139
3369
  * Do not edit manually.
3140
3370
  * AssemblyAI API
3141
- * AssemblyAI API
3142
- * OpenAPI spec version: 1.3.4
3371
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3372
+ * OpenAPI spec version: 1.0.0
3143
3373
  */
3144
3374
  /**
3145
- * Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details.
3375
+ * Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more details.
3146
3376
  */
3147
3377
  type RedactPiiAudioQuality = (typeof RedactPiiAudioQuality)[keyof typeof RedactPiiAudioQuality];
3148
3378
  declare const RedactPiiAudioQuality: {
@@ -3154,13 +3384,13 @@ declare const RedactPiiAudioQuality: {
3154
3384
  * Generated by orval v7.9.0 🍺
3155
3385
  * Do not edit manually.
3156
3386
  * AssemblyAI API
3157
- * AssemblyAI API
3158
- * OpenAPI spec version: 1.3.4
3387
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3388
+ * OpenAPI spec version: 1.0.0
3159
3389
  */
3160
3390
 
3161
3391
  /**
3162
3392
  * The audio quality of the PII-redacted audio file, if redact_pii_audio is enabled.
3163
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
3393
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
3164
3394
 
3165
3395
  */
3166
3396
  type TranscriptRedactPiiAudioQuality = RedactPiiAudioQuality | null;
@@ -3169,8 +3399,8 @@ type TranscriptRedactPiiAudioQuality = RedactPiiAudioQuality | null;
3169
3399
  * Generated by orval v7.9.0 🍺
3170
3400
  * Do not edit manually.
3171
3401
  * AssemblyAI API
3172
- * AssemblyAI API
3173
- * OpenAPI spec version: 1.3.4
3402
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3403
+ * OpenAPI spec version: 1.0.0
3174
3404
  */
3175
3405
  /**
3176
3406
  * The type of PII to redact
@@ -3227,869 +3457,1004 @@ declare const PiiPolicy: {
3227
3457
  * Generated by orval v7.9.0 🍺
3228
3458
  * Do not edit manually.
3229
3459
  * AssemblyAI API
3230
- * AssemblyAI API
3231
- * OpenAPI spec version: 1.3.4
3460
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3461
+ * OpenAPI spec version: 1.0.0
3462
+ */
3463
+
3464
+ /**
3465
+ * The list of PII Redaction policies that were enabled, if PII Redaction is enabled.
3466
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more information.
3467
+
3468
+ */
3469
+ type TranscriptRedactPiiPolicies = PiiPolicy[] | null;
3470
+
3471
+ /**
3472
+ * Generated by orval v7.9.0 🍺
3473
+ * Do not edit manually.
3474
+ * AssemblyAI API
3475
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3476
+ * OpenAPI spec version: 1.0.0
3477
+ */
3478
+ /**
3479
+ * The replacement logic for detected PII, can be `entity_name` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details.
3480
+ */
3481
+ type SubstitutionPolicy = (typeof SubstitutionPolicy)[keyof typeof SubstitutionPolicy];
3482
+ declare const SubstitutionPolicy: {
3483
+ readonly entity_name: "entity_name";
3484
+ readonly hash: "hash";
3485
+ };
3486
+
3487
+ /**
3488
+ * Generated by orval v7.9.0 🍺
3489
+ * Do not edit manually.
3490
+ * AssemblyAI API
3491
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3492
+ * OpenAPI spec version: 1.0.0
3493
+ */
3494
+ /**
3495
+ * Whether [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) is enabled, can be true or false
3496
+ */
3497
+ type TranscriptSentimentAnalysis = boolean | null;
3498
+
3499
+ /**
3500
+ * Generated by orval v7.9.0 🍺
3501
+ * Do not edit manually.
3502
+ * AssemblyAI API
3503
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3504
+ * OpenAPI spec version: 1.0.0
3505
+ */
3506
+ type Sentiment = (typeof Sentiment)[keyof typeof Sentiment];
3507
+ declare const Sentiment: {
3508
+ readonly POSITIVE: "POSITIVE";
3509
+ readonly NEUTRAL: "NEUTRAL";
3510
+ readonly NEGATIVE: "NEGATIVE";
3511
+ };
3512
+
3513
+ /**
3514
+ * Generated by orval v7.9.0 🍺
3515
+ * Do not edit manually.
3516
+ * AssemblyAI API
3517
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3518
+ * OpenAPI spec version: 1.0.0
3519
+ */
3520
+ /**
3521
+ * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3522
+ */
3523
+ type SentimentAnalysisResultChannel = string | null;
3524
+
3525
+ /**
3526
+ * Generated by orval v7.9.0 🍺
3527
+ * Do not edit manually.
3528
+ * AssemblyAI API
3529
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3530
+ * OpenAPI spec version: 1.0.0
3531
+ */
3532
+ /**
3533
+ * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
3534
+ */
3535
+ type SentimentAnalysisResultSpeaker = string | null;
3536
+
3537
+ /**
3538
+ * Generated by orval v7.9.0 🍺
3539
+ * Do not edit manually.
3540
+ * AssemblyAI API
3541
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3542
+ * OpenAPI spec version: 1.0.0
3543
+ */
3544
+
3545
+ /**
3546
+ * The result of the Sentiment Analysis model
3547
+ */
3548
+ interface SentimentAnalysisResult$1 {
3549
+ /** The transcript of the sentence */
3550
+ text: string;
3551
+ /** The starting time, in milliseconds, of the sentence */
3552
+ start: number;
3553
+ /** The ending time, in milliseconds, of the sentence */
3554
+ end: number;
3555
+ /** The detected sentiment for the sentence, one of POSITIVE, NEUTRAL, NEGATIVE */
3556
+ sentiment: Sentiment;
3557
+ /** The confidence score for the detected sentiment of the sentence, from 0 to 1 */
3558
+ confidence: number;
3559
+ /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
3560
+ channel?: SentimentAnalysisResultChannel;
3561
+ /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
3562
+ speaker: SentimentAnalysisResultSpeaker;
3563
+ }
3564
+
3565
+ /**
3566
+ * Generated by orval v7.9.0 🍺
3567
+ * Do not edit manually.
3568
+ * AssemblyAI API
3569
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3570
+ * OpenAPI spec version: 1.0.0
3571
+ */
3572
+
3573
+ /**
3574
+ * An array of results for the Sentiment Analysis model, if it is enabled.
3575
+ See [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) for more information.
3576
+
3577
+ */
3578
+ type TranscriptSentimentAnalysisResults = SentimentAnalysisResult$1[] | null;
3579
+
3580
+ /**
3581
+ * Generated by orval v7.9.0 🍺
3582
+ * Do not edit manually.
3583
+ * AssemblyAI API
3584
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3585
+ * OpenAPI spec version: 1.0.0
3586
+ */
3587
+ /**
3588
+ * Whether [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, can be true or false
3589
+ */
3590
+ type TranscriptSpeakerLabels = boolean | null;
3591
+
3592
+ /**
3593
+ * Generated by orval v7.9.0 🍺
3594
+ * Do not edit manually.
3595
+ * AssemblyAI API
3596
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3597
+ * OpenAPI spec version: 1.0.0
3598
+ */
3599
+ /**
3600
+ * Tell the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details.
3601
+ */
3602
+ type TranscriptSpeakersExpected = number | null;
3603
+
3604
+ /**
3605
+ * Generated by orval v7.9.0 🍺
3606
+ * Do not edit manually.
3607
+ * AssemblyAI API
3608
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3609
+ * OpenAPI spec version: 1.0.0
3610
+ */
3611
+ /**
3612
+ * The speech model to use for the transcription. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models.
3613
+ */
3614
+ type SpeechModel = string;
3615
+
3616
+ /**
3617
+ * Generated by orval v7.9.0 🍺
3618
+ * Do not edit manually.
3619
+ * AssemblyAI API
3620
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3621
+ * OpenAPI spec version: 1.0.0
3622
+ */
3623
+
3624
+ /**
3625
+ * List multiple speech models in priority order, allowing our system to automatically route your audio to the best available option. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models and routing behavior.
3626
+
3627
+ */
3628
+ type TranscriptSpeechModels = SpeechModel[] | null;
3629
+
3630
+ /**
3631
+ * Generated by orval v7.9.0 🍺
3632
+ * Do not edit manually.
3633
+ * AssemblyAI API
3634
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3635
+ * OpenAPI spec version: 1.0.0
3636
+ */
3637
+ /**
3638
+ * Defaults to null. Reject audio files that contain less than this fraction of speech.
3639
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
3640
+
3641
+ */
3642
+ type TranscriptSpeechThreshold = number | null;
3643
+
3644
+ /**
3645
+ * Generated by orval v7.9.0 🍺
3646
+ * Do not edit manually.
3647
+ * AssemblyAI API
3648
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3649
+ * OpenAPI spec version: 1.0.0
3650
+ */
3651
+ interface TranslationRequestBodyTranslation {
3652
+ /** List of target language codes (e.g., `["es", "de"]`). See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for supported languages. */
3653
+ target_languages: string[];
3654
+ /** Use formal language style. See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for more details. */
3655
+ formal?: boolean;
3656
+ /** When enabled with Speaker Labels, returns translated text in the utterances array. Each utterance will include a `translated_texts` key containing translations for each target language. */
3657
+ match_original_utterance?: boolean;
3658
+ }
3659
+
3660
+ /**
3661
+ * Generated by orval v7.9.0 🍺
3662
+ * Do not edit manually.
3663
+ * AssemblyAI API
3664
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3665
+ * OpenAPI spec version: 1.0.0
3232
3666
  */
3233
3667
 
3234
3668
  /**
3235
- * The list of PII Redaction policies that were enabled, if PII Redaction is enabled.
3236
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
3237
-
3669
+ * Request body for [Translation](https://www.assemblyai.com/docs/speech-understanding/translation).
3238
3670
  */
3239
- type TranscriptRedactPiiPolicies = PiiPolicy[] | null;
3671
+ interface TranslationRequestBody {
3672
+ translation: TranslationRequestBodyTranslation;
3673
+ }
3240
3674
 
3241
3675
  /**
3242
3676
  * Generated by orval v7.9.0 🍺
3243
3677
  * Do not edit manually.
3244
3678
  * AssemblyAI API
3245
- * AssemblyAI API
3246
- * OpenAPI spec version: 1.3.4
3679
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3680
+ * OpenAPI spec version: 1.0.0
3247
3681
  */
3248
3682
  /**
3249
- * The replacement logic for detected PII, can be "entity_name" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details.
3683
+ * Type of speaker identification. See [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification) for details on each type.
3250
3684
  */
3251
- type SubstitutionPolicy = (typeof SubstitutionPolicy)[keyof typeof SubstitutionPolicy];
3252
- declare const SubstitutionPolicy: {
3253
- readonly entity_name: "entity_name";
3254
- readonly hash: "hash";
3685
+ type SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType = (typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType)[keyof typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType];
3686
+ declare const SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType: {
3687
+ readonly role: "role";
3688
+ readonly name: "name";
3255
3689
  };
3256
3690
 
3257
3691
  /**
3258
3692
  * Generated by orval v7.9.0 🍺
3259
3693
  * Do not edit manually.
3260
3694
  * AssemblyAI API
3261
- * AssemblyAI API
3262
- * OpenAPI spec version: 1.3.4
3263
- */
3264
- /**
3265
- * Whether [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, can be true or false
3695
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3696
+ * OpenAPI spec version: 1.0.0
3266
3697
  */
3267
- type TranscriptSpeakerLabels = boolean | null;
3698
+ interface SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems {
3699
+ /** The role of the speaker. Required when `speaker_type` is "role". */
3700
+ role?: string;
3701
+ /** The name of the speaker. Required when `speaker_type` is "name". */
3702
+ name?: string;
3703
+ /** A description of the speaker to help the model identify them based on conversational context. */
3704
+ description?: string;
3705
+ }
3268
3706
 
3269
3707
  /**
3270
3708
  * Generated by orval v7.9.0 🍺
3271
3709
  * Do not edit manually.
3272
3710
  * AssemblyAI API
3273
- * AssemblyAI API
3274
- * OpenAPI spec version: 1.3.4
3275
- */
3276
- /**
3277
- * Tell the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details.
3711
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3712
+ * OpenAPI spec version: 1.0.0
3278
3713
  */
3279
- type TranscriptSpeakersExpected = number | null;
3714
+
3715
+ interface SpeakerIdentificationRequestBodySpeakerIdentification {
3716
+ /** Type of speaker identification. See [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification) for details on each type. */
3717
+ speaker_type: SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType;
3718
+ /** Required if speaker_type is "role". Each value must be 35 characters or less. */
3719
+ known_values?: string[];
3720
+ /** An array of speaker objects with metadata to improve identification accuracy. Each object should include a `role` or `name` (depending on `speaker_type`) and an optional `description` to help the model identify the speaker. You can also include any additional custom properties (e.g., `company`, `title`) to provide more context. Use this as an alternative to `known_values` when you want to provide additional context about each speaker. */
3721
+ speakers?: SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems[];
3722
+ }
3280
3723
 
3281
3724
  /**
3282
3725
  * Generated by orval v7.9.0 🍺
3283
3726
  * Do not edit manually.
3284
3727
  * AssemblyAI API
3285
- * AssemblyAI API
3286
- * OpenAPI spec version: 1.3.4
3728
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3729
+ * OpenAPI spec version: 1.0.0
3287
3730
  */
3731
+
3288
3732
  /**
3289
- * Whether [Content Moderation](https://www.assemblyai.com/docs/models/content-moderation) is enabled, can be true or false
3733
+ * Request body for [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification).
3290
3734
  */
3291
- type TranscriptContentSafety = boolean | null;
3735
+ interface SpeakerIdentificationRequestBody {
3736
+ speaker_identification: SpeakerIdentificationRequestBodySpeakerIdentification;
3737
+ }
3292
3738
 
3293
3739
  /**
3294
3740
  * Generated by orval v7.9.0 🍺
3295
3741
  * Do not edit manually.
3296
3742
  * AssemblyAI API
3297
- * AssemblyAI API
3298
- * OpenAPI spec version: 1.3.4
3743
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3744
+ * OpenAPI spec version: 1.0.0
3299
3745
  */
3300
- interface ContentSafetyLabel {
3301
- /** The label of the sensitive topic */
3302
- label: string;
3303
- /**
3304
- * The confidence score for the topic being discussed, from 0 to 1
3305
- * @minimum 0
3306
- * @maximum 1
3307
- */
3308
- confidence: number;
3309
- /**
3310
- * How severely the topic is discussed in the section, from 0 to 1
3311
- * @minimum 0
3312
- * @maximum 1
3313
- */
3314
- severity: number;
3746
+ interface CustomFormattingRequestBodyCustomFormatting {
3747
+ /** Date format pattern (e.g., `"mm/dd/yyyy"`). See [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting) for more details. */
3748
+ date?: string;
3749
+ /** Phone number format pattern (e.g., `"(xxx)xxx-xxxx"`). See [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting) for more details. */
3750
+ phone_number?: string;
3751
+ /** Email format pattern (e.g., `"username@domain.com"`). See [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting) for more details. */
3752
+ email?: string;
3315
3753
  }
3316
3754
 
3317
3755
  /**
3318
3756
  * Generated by orval v7.9.0 🍺
3319
3757
  * Do not edit manually.
3320
3758
  * AssemblyAI API
3321
- * AssemblyAI API
3322
- * OpenAPI spec version: 1.3.4
3759
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3760
+ * OpenAPI spec version: 1.0.0
3323
3761
  */
3324
3762
 
3325
- interface ContentSafetyLabelResult {
3326
- /** The transcript of the section flagged by the Content Moderation model */
3327
- text: string;
3328
- /** An array of safety labels, one per sensitive topic that was detected in the section */
3329
- labels: ContentSafetyLabel[];
3330
- /** The sentence index at which the section begins */
3331
- sentences_idx_start: number;
3332
- /** The sentence index at which the section ends */
3333
- sentences_idx_end: number;
3334
- /** Timestamp information for the section */
3335
- timestamp: Timestamp;
3763
+ /**
3764
+ * Request body for [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting).
3765
+ */
3766
+ interface CustomFormattingRequestBody {
3767
+ custom_formatting: CustomFormattingRequestBodyCustomFormatting;
3336
3768
  }
3337
3769
 
3338
3770
  /**
3339
3771
  * Generated by orval v7.9.0 🍺
3340
3772
  * Do not edit manually.
3341
3773
  * AssemblyAI API
3342
- * AssemblyAI API
3343
- * OpenAPI spec version: 1.3.4
3774
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3775
+ * OpenAPI spec version: 1.0.0
3344
3776
  */
3777
+
3778
+ type TranscriptSpeechUnderstandingRequest = TranslationRequestBody | SpeakerIdentificationRequestBody | CustomFormattingRequestBody;
3779
+
3345
3780
  /**
3346
- * A summary of the Content Moderation confidence results for the entire audio file
3781
+ * Generated by orval v7.9.0 🍺
3782
+ * Do not edit manually.
3783
+ * AssemblyAI API
3784
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3785
+ * OpenAPI spec version: 1.0.0
3347
3786
  */
3348
- type ContentSafetyLabelsResultSummary = {
3349
- [key: string]: number;
3350
- };
3787
+ interface TranslationResponseTranslation {
3788
+ status?: string;
3789
+ }
3351
3790
 
3352
3791
  /**
3353
3792
  * Generated by orval v7.9.0 🍺
3354
3793
  * Do not edit manually.
3355
3794
  * AssemblyAI API
3356
- * AssemblyAI API
3357
- * OpenAPI spec version: 1.3.4
3795
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3796
+ * OpenAPI spec version: 1.0.0
3358
3797
  */
3359
- interface SeverityScoreSummary {
3360
- /**
3361
- * @minimum 0
3362
- * @maximum 1
3363
- */
3364
- low: number;
3365
- /**
3366
- * @minimum 0
3367
- * @maximum 1
3368
- */
3369
- medium: number;
3370
- /**
3371
- * @minimum 0
3372
- * @maximum 1
3373
- */
3374
- high: number;
3798
+
3799
+ interface TranslationResponse {
3800
+ translation?: TranslationResponseTranslation;
3375
3801
  }
3376
3802
 
3377
3803
  /**
3378
3804
  * Generated by orval v7.9.0 🍺
3379
3805
  * Do not edit manually.
3380
3806
  * AssemblyAI API
3381
- * AssemblyAI API
3382
- * OpenAPI spec version: 1.3.4
3807
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3808
+ * OpenAPI spec version: 1.0.0
3383
3809
  */
3384
-
3385
3810
  /**
3386
- * A summary of the Content Moderation severity results for the entire audio file
3811
+ * A mapping of the original generic speaker labels (e.g., "A", "B") to the identified speaker names or roles.
3387
3812
  */
3388
- type ContentSafetyLabelsResultSeverityScoreSummary = {
3389
- [key: string]: SeverityScoreSummary;
3813
+ type SpeakerIdentificationResponseSpeakerIdentificationMapping = {
3814
+ [key: string]: string;
3390
3815
  };
3391
3816
 
3392
3817
  /**
3393
3818
  * Generated by orval v7.9.0 🍺
3394
3819
  * Do not edit manually.
3395
3820
  * AssemblyAI API
3396
- * AssemblyAI API
3397
- * OpenAPI spec version: 1.3.4
3821
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3822
+ * OpenAPI spec version: 1.0.0
3398
3823
  */
3399
3824
 
3400
- /**
3401
- * An array of results for the Content Moderation model, if it is enabled.
3402
- See [Content moderation](https://www.assemblyai.com/docs/models/content-moderation) for more information.
3403
-
3404
- */
3405
- interface ContentSafetyLabelsResult {
3406
- /** The status of the Content Moderation model. Either success, or unavailable in the rare case that the model failed. */
3407
- status: AudioIntelligenceModelStatus;
3408
- /** An array of results for the Content Moderation model */
3409
- results: ContentSafetyLabelResult[];
3410
- /** A summary of the Content Moderation confidence results for the entire audio file */
3411
- summary: ContentSafetyLabelsResultSummary;
3412
- /** A summary of the Content Moderation severity results for the entire audio file */
3413
- severity_score_summary: ContentSafetyLabelsResultSeverityScoreSummary;
3825
+ interface SpeakerIdentificationResponseSpeakerIdentification {
3826
+ /** A mapping of the original generic speaker labels (e.g., "A", "B") to the identified speaker names or roles. */
3827
+ mapping?: SpeakerIdentificationResponseSpeakerIdentificationMapping;
3828
+ status?: string;
3414
3829
  }
3415
3830
 
3416
3831
  /**
3417
3832
  * Generated by orval v7.9.0 🍺
3418
3833
  * Do not edit manually.
3419
3834
  * AssemblyAI API
3420
- * AssemblyAI API
3421
- * OpenAPI spec version: 1.3.4
3835
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3836
+ * OpenAPI spec version: 1.0.0
3422
3837
  */
3423
3838
 
3424
- /**
3425
- * An array of results for the Content Moderation model, if it is enabled.
3426
- See [Content moderation](https://www.assemblyai.com/docs/models/content-moderation) for more information.
3427
-
3428
- */
3429
- type TranscriptContentSafetyLabels = ContentSafetyLabelsResult | null;
3839
+ interface SpeakerIdentificationResponse {
3840
+ speaker_identification?: SpeakerIdentificationResponseSpeakerIdentification;
3841
+ }
3430
3842
 
3431
3843
  /**
3432
3844
  * Generated by orval v7.9.0 🍺
3433
3845
  * Do not edit manually.
3434
3846
  * AssemblyAI API
3435
- * AssemblyAI API
3436
- * OpenAPI spec version: 1.3.4
3437
- */
3438
- /**
3439
- * Whether [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) is enabled, can be true or false
3847
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3848
+ * OpenAPI spec version: 1.0.0
3440
3849
  */
3441
- type TranscriptIabCategories = boolean | null;
3850
+ type CustomFormattingResponseCustomFormattingMapping = {
3851
+ [key: string]: string;
3852
+ };
3442
3853
 
3443
3854
  /**
3444
3855
  * Generated by orval v7.9.0 🍺
3445
3856
  * Do not edit manually.
3446
3857
  * AssemblyAI API
3447
- * AssemblyAI API
3448
- * OpenAPI spec version: 1.3.4
3858
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3859
+ * OpenAPI spec version: 1.0.0
3449
3860
  */
3450
- type TopicDetectionResultLabelsItem = {
3451
- /**
3452
- * How relevant the detected topic is of a detected topic
3453
- * @minimum 0
3454
- * @maximum 1
3455
- */
3456
- relevance: number;
3457
- /** The IAB taxonomical label for the label of the detected topic, where > denotes supertopic/subtopic relationship */
3458
- label: string;
3459
- };
3861
+
3862
+ interface CustomFormattingResponseCustomFormatting {
3863
+ mapping?: CustomFormattingResponseCustomFormattingMapping;
3864
+ formatted_text?: string;
3865
+ }
3460
3866
 
3461
3867
  /**
3462
3868
  * Generated by orval v7.9.0 🍺
3463
3869
  * Do not edit manually.
3464
3870
  * AssemblyAI API
3465
- * AssemblyAI API
3466
- * OpenAPI spec version: 1.3.4
3871
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3872
+ * OpenAPI spec version: 1.0.0
3467
3873
  */
3468
3874
 
3469
- /**
3470
- * The result of the topic detection model
3471
- */
3472
- interface TopicDetectionResult$1 {
3473
- /** The text in the transcript in which a detected topic occurs */
3474
- text: string;
3475
- /** An array of detected topics in the text */
3476
- labels?: TopicDetectionResultLabelsItem[];
3477
- timestamp?: Timestamp;
3875
+ interface CustomFormattingResponse {
3876
+ custom_formatting?: CustomFormattingResponseCustomFormatting;
3478
3877
  }
3479
3878
 
3480
3879
  /**
3481
3880
  * Generated by orval v7.9.0 🍺
3482
3881
  * Do not edit manually.
3483
3882
  * AssemblyAI API
3484
- * AssemblyAI API
3485
- * OpenAPI spec version: 1.3.4
3486
- */
3487
- /**
3488
- * The overall relevance of topic to the entire audio file
3883
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3884
+ * OpenAPI spec version: 1.0.0
3489
3885
  */
3490
- type TopicDetectionModelResultSummary = {
3491
- [key: string]: number;
3492
- };
3886
+
3887
+ type TranscriptSpeechUnderstandingResponse = TranslationResponse | SpeakerIdentificationResponse | CustomFormattingResponse;
3493
3888
 
3494
3889
  /**
3495
3890
  * Generated by orval v7.9.0 🍺
3496
3891
  * Do not edit manually.
3497
3892
  * AssemblyAI API
3498
- * AssemblyAI API
3499
- * OpenAPI spec version: 1.3.4
3893
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3894
+ * OpenAPI spec version: 1.0.0
3500
3895
  */
3501
3896
 
3502
3897
  /**
3503
- * The result of the Topic Detection model, if it is enabled.
3504
- See [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) for more information.
3898
+ * Speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
3505
3899
 
3506
3900
  */
3507
- interface TopicDetectionModelResult {
3508
- /** The status of the Topic Detection model. Either success, or unavailable in the rare case that the model failed. */
3509
- status: AudioIntelligenceModelStatus;
3510
- /** An array of results for the Topic Detection model */
3511
- results: TopicDetectionResult$1[];
3512
- /** The overall relevance of topic to the entire audio file */
3513
- summary: TopicDetectionModelResultSummary;
3901
+ interface TranscriptSpeechUnderstanding {
3902
+ request?: TranscriptSpeechUnderstandingRequest;
3903
+ response?: TranscriptSpeechUnderstandingResponse;
3514
3904
  }
3515
3905
 
3516
3906
  /**
3517
3907
  * Generated by orval v7.9.0 🍺
3518
3908
  * Do not edit manually.
3519
3909
  * AssemblyAI API
3520
- * AssemblyAI API
3521
- * OpenAPI spec version: 1.3.4
3910
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3911
+ * OpenAPI spec version: 1.0.0
3522
3912
  */
3523
-
3524
3913
  /**
3525
- * The result of the Topic Detection model, if it is enabled.
3526
- See [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) for more information.
3527
-
3914
+ * The status of your transcript. Possible values are queued, processing, completed, or error.
3528
3915
  */
3529
- type TranscriptIabCategoriesResult = TopicDetectionModelResult | null;
3916
+ type TranscriptStatus = (typeof TranscriptStatus)[keyof typeof TranscriptStatus];
3917
+ declare const TranscriptStatus: {
3918
+ readonly queued: "queued";
3919
+ readonly processing: "processing";
3920
+ readonly completed: "completed";
3921
+ readonly error: "error";
3922
+ };
3530
3923
 
3531
3924
  /**
3532
3925
  * Generated by orval v7.9.0 🍺
3533
3926
  * Do not edit manually.
3534
3927
  * AssemblyAI API
3535
- * AssemblyAI API
3536
- * OpenAPI spec version: 1.3.4
3928
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3929
+ * OpenAPI spec version: 1.0.0
3537
3930
  */
3538
3931
  /**
3539
- * Object containing words or phrases to replace, and the word or phrase to replace with
3932
+ * The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
3540
3933
  */
3541
- interface TranscriptCustomSpelling {
3542
- /** Words or phrases to replace */
3543
- from: string[];
3544
- /** Word to replace with */
3545
- to: string;
3546
- }
3934
+ type TranscriptSummary = string | null;
3547
3935
 
3548
3936
  /**
3549
3937
  * Generated by orval v7.9.0 🍺
3550
3938
  * Do not edit manually.
3551
3939
  * AssemblyAI API
3552
- * AssemblyAI API
3553
- * OpenAPI spec version: 1.3.4
3940
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3941
+ * OpenAPI spec version: 1.0.0
3554
3942
  */
3555
-
3556
3943
  /**
3557
- * Customize how words are spelled and formatted using to and from values
3944
+ * The Summarization model used to generate the summary,
3945
+ if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-models) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
3946
+
3558
3947
  */
3559
- type TranscriptCustomSpellingProperty = TranscriptCustomSpelling[] | null;
3948
+ type TranscriptSummaryModel = string | null;
3560
3949
 
3561
3950
  /**
3562
3951
  * Generated by orval v7.9.0 🍺
3563
3952
  * Do not edit manually.
3564
3953
  * AssemblyAI API
3565
- * AssemblyAI API
3566
- * OpenAPI spec version: 1.3.4
3954
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3955
+ * OpenAPI spec version: 1.0.0
3567
3956
  */
3568
3957
  /**
3569
- * Whether [Auto Chapters](https://www.assemblyai.com/docs/models/auto-chapters) is enabled, can be true or false
3958
+ * The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-types) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
3570
3959
  */
3571
- type TranscriptAutoChapters = boolean | null;
3960
+ type TranscriptSummaryType = string | null;
3572
3961
 
3573
3962
  /**
3574
3963
  * Generated by orval v7.9.0 🍺
3575
3964
  * Do not edit manually.
3576
3965
  * AssemblyAI API
3577
- * AssemblyAI API
3578
- * OpenAPI spec version: 1.3.4
3966
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3967
+ * OpenAPI spec version: 1.0.0
3579
3968
  */
3580
3969
  /**
3581
- * Chapter of the audio file
3970
+ * Whether [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) were removed from the transcript text.
3971
+
3972
+ Note: This parameter is only supported for the Universal-3 Pro model.
3973
+
3582
3974
  */
3583
- interface Chapter$1 {
3584
- /** An ultra-short summary (just a few words) of the content spoken in the chapter */
3585
- gist: string;
3586
- /** A single sentence summary of the content spoken during the chapter */
3587
- headline: string;
3588
- /** A one paragraph summary of the content spoken during the chapter */
3589
- summary: string;
3590
- /** The starting time, in milliseconds, for the chapter */
3591
- start: number;
3592
- /** The starting time, in milliseconds, for the chapter */
3593
- end: number;
3594
- }
3975
+ type TranscriptRemoveAudioTags = (typeof TranscriptRemoveAudioTags)[keyof typeof TranscriptRemoveAudioTags];
3976
+ declare const TranscriptRemoveAudioTags: {
3977
+ readonly all: "all";
3978
+ };
3595
3979
 
3596
3980
  /**
3597
3981
  * Generated by orval v7.9.0 🍺
3598
3982
  * Do not edit manually.
3599
3983
  * AssemblyAI API
3600
- * AssemblyAI API
3601
- * OpenAPI spec version: 1.3.4
3984
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3985
+ * OpenAPI spec version: 1.0.0
3602
3986
  */
3603
3987
 
3604
3988
  /**
3605
- * An array of temporally sequential chapters for the audio file
3989
+ * Whether [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) were removed from the transcript text.
3990
+
3991
+ Note: This parameter is only supported for the Universal-3 Pro model.
3992
+
3606
3993
  */
3607
- type TranscriptChapters = Chapter$1[] | null;
3994
+ type TranscriptRemoveAudioTagsProperty = TranscriptRemoveAudioTags | null;
3608
3995
 
3609
3996
  /**
3610
3997
  * Generated by orval v7.9.0 🍺
3611
3998
  * Do not edit manually.
3612
3999
  * AssemblyAI API
3613
- * AssemblyAI API
3614
- * OpenAPI spec version: 1.3.4
4000
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4001
+ * OpenAPI spec version: 1.0.0
3615
4002
  */
3616
4003
  /**
3617
- * The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
4004
+ * The temperature that was used for the model's response. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for more details.
4005
+
4006
+ Note: This parameter can only be used with the Universal-3 Pro model.
4007
+
3618
4008
  */
3619
- type TranscriptSummaryType = string | null;
4009
+ type TranscriptTemperature = number | null;
3620
4010
 
3621
4011
  /**
3622
4012
  * Generated by orval v7.9.0 🍺
3623
4013
  * Do not edit manually.
3624
4014
  * AssemblyAI API
3625
- * AssemblyAI API
3626
- * OpenAPI spec version: 1.3.4
4015
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4016
+ * OpenAPI spec version: 1.0.0
3627
4017
  */
3628
4018
  /**
3629
- * The Summarization model used to generate the summary,
3630
- if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
3631
-
4019
+ * The textual transcript of your media file
3632
4020
  */
3633
- type TranscriptSummaryModel = string | null;
4021
+ type TranscriptText = string | null;
3634
4022
 
3635
4023
  /**
3636
4024
  * Generated by orval v7.9.0 🍺
3637
4025
  * Do not edit manually.
3638
4026
  * AssemblyAI API
3639
- * AssemblyAI API
3640
- * OpenAPI spec version: 1.3.4
4027
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4028
+ * OpenAPI spec version: 1.0.0
3641
4029
  */
3642
4030
  /**
3643
- * The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
4031
+ * True while a request is throttled and false when a request is no longer throttled
3644
4032
  */
3645
- type TranscriptSummary = string | null;
4033
+ type TranscriptThrottled = boolean | null;
3646
4034
 
3647
4035
  /**
3648
4036
  * Generated by orval v7.9.0 🍺
3649
4037
  * Do not edit manually.
3650
4038
  * AssemblyAI API
3651
- * AssemblyAI API
3652
- * OpenAPI spec version: 1.3.4
4039
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4040
+ * OpenAPI spec version: 1.0.0
3653
4041
  */
3654
4042
  /**
3655
- * Whether custom topics is enabled, either true or false
3656
- * @deprecated
4043
+ * The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3657
4044
  */
3658
- type TranscriptCustomTopics = boolean | null;
4045
+ type TranscriptWordChannel = string | null;
3659
4046
 
3660
4047
  /**
3661
4048
  * Generated by orval v7.9.0 🍺
3662
4049
  * Do not edit manually.
3663
4050
  * AssemblyAI API
3664
- * AssemblyAI API
3665
- * OpenAPI spec version: 1.3.4
4051
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4052
+ * OpenAPI spec version: 1.0.0
3666
4053
  */
3667
4054
  /**
3668
- * Whether [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) is enabled, can be true or false
4055
+ * The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
3669
4056
  */
3670
- type TranscriptSentimentAnalysis = boolean | null;
4057
+ type TranscriptWordSpeaker = string | null;
3671
4058
 
3672
4059
  /**
3673
4060
  * Generated by orval v7.9.0 🍺
3674
4061
  * Do not edit manually.
3675
4062
  * AssemblyAI API
3676
- * AssemblyAI API
3677
- * OpenAPI spec version: 1.3.4
4063
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4064
+ * OpenAPI spec version: 1.0.0
3678
4065
  */
3679
- type Sentiment = (typeof Sentiment)[keyof typeof Sentiment];
3680
- declare const Sentiment: {
3681
- readonly POSITIVE: "POSITIVE";
3682
- readonly NEUTRAL: "NEUTRAL";
3683
- readonly NEGATIVE: "NEGATIVE";
3684
- };
4066
+
4067
+ interface TranscriptWord {
4068
+ /** The confidence score for the transcript of this word */
4069
+ confidence: number;
4070
+ /** The starting time, in milliseconds, for the word */
4071
+ start: number;
4072
+ /** The ending time, in milliseconds, for the word */
4073
+ end: number;
4074
+ /** The text of the word */
4075
+ text: string;
4076
+ /** The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
4077
+ channel?: TranscriptWordChannel;
4078
+ /** The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
4079
+ speaker: TranscriptWordSpeaker;
4080
+ }
3685
4081
 
3686
4082
  /**
3687
4083
  * Generated by orval v7.9.0 🍺
3688
4084
  * Do not edit manually.
3689
4085
  * AssemblyAI API
3690
- * AssemblyAI API
3691
- * OpenAPI spec version: 1.3.4
4086
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4087
+ * OpenAPI spec version: 1.0.0
3692
4088
  */
3693
4089
  /**
3694
4090
  * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3695
4091
  */
3696
- type SentimentAnalysisResultChannel = string | null;
4092
+ type TranscriptUtteranceChannel = string | null;
3697
4093
 
3698
4094
  /**
3699
4095
  * Generated by orval v7.9.0 🍺
3700
4096
  * Do not edit manually.
3701
4097
  * AssemblyAI API
3702
- * AssemblyAI API
3703
- * OpenAPI spec version: 1.3.4
4098
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4099
+ * OpenAPI spec version: 1.0.0
3704
4100
  */
3705
4101
  /**
3706
- * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null
4102
+ * Translations keyed by language code (e.g., `{"es": "Texto traducido", "de": "Übersetzter Text"}`). Only present when `match_original_utterance` is enabled with translation.
3707
4103
  */
3708
- type SentimentAnalysisResultSpeaker = string | null;
4104
+ type TranscriptUtteranceTranslatedTexts = {
4105
+ [key: string]: string;
4106
+ };
3709
4107
 
3710
4108
  /**
3711
4109
  * Generated by orval v7.9.0 🍺
3712
4110
  * Do not edit manually.
3713
4111
  * AssemblyAI API
3714
- * AssemblyAI API
3715
- * OpenAPI spec version: 1.3.4
4112
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4113
+ * OpenAPI spec version: 1.0.0
3716
4114
  */
3717
4115
 
3718
- /**
3719
- * The result of the Sentiment Analysis model
3720
- */
3721
- interface SentimentAnalysisResult$1 {
3722
- /** The transcript of the sentence */
3723
- text: string;
3724
- /** The starting time, in milliseconds, of the sentence */
4116
+ interface TranscriptUtterance {
4117
+ /** The confidence score for the transcript of this utterance */
4118
+ confidence: number;
4119
+ /** The starting time, in milliseconds, of the utterance in the audio file */
3725
4120
  start: number;
3726
- /** The ending time, in milliseconds, of the sentence */
4121
+ /** The ending time, in milliseconds, of the utterance in the audio file */
3727
4122
  end: number;
3728
- /** The detected sentiment for the sentence, one of POSITIVE, NEUTRAL, NEGATIVE */
3729
- sentiment: Sentiment;
3730
- /**
3731
- * The confidence score for the detected sentiment of the sentence, from 0 to 1
3732
- * @minimum 0
3733
- * @maximum 1
3734
- */
3735
- confidence: number;
4123
+ /** The text for this utterance */
4124
+ text: string;
4125
+ /** The words in the utterance. */
4126
+ words: TranscriptWord[];
3736
4127
  /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
3737
- channel?: SentimentAnalysisResultChannel;
3738
- /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null */
3739
- speaker: SentimentAnalysisResultSpeaker;
4128
+ channel?: TranscriptUtteranceChannel;
4129
+ /** The speaker of this utterance, where each speaker is assigned a sequential capital letter - e.g. "A" for Speaker A, "B" for Speaker B, etc. */
4130
+ speaker: string;
4131
+ /** Translations keyed by language code (e.g., `{"es": "Texto traducido", "de": "Übersetzter Text"}`). Only present when `match_original_utterance` is enabled with translation. */
4132
+ translated_texts?: TranscriptUtteranceTranslatedTexts;
3740
4133
  }
3741
4134
 
3742
4135
  /**
3743
4136
  * Generated by orval v7.9.0 🍺
3744
4137
  * Do not edit manually.
3745
4138
  * AssemblyAI API
3746
- * AssemblyAI API
3747
- * OpenAPI spec version: 1.3.4
4139
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4140
+ * OpenAPI spec version: 1.0.0
3748
4141
  */
3749
4142
 
3750
4143
  /**
3751
- * An array of results for the Sentiment Analysis model, if it is enabled.
3752
- See [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) for more information.
4144
+ * When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
4145
+ See [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) for more information.
3753
4146
 
3754
4147
  */
3755
- type TranscriptSentimentAnalysisResults = SentimentAnalysisResult$1[] | null;
4148
+ type TranscriptUtterances = TranscriptUtterance[] | null;
3756
4149
 
3757
4150
  /**
3758
4151
  * Generated by orval v7.9.0 🍺
3759
4152
  * Do not edit manually.
3760
4153
  * AssemblyAI API
3761
- * AssemblyAI API
3762
- * OpenAPI spec version: 1.3.4
4154
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4155
+ * OpenAPI spec version: 1.0.0
3763
4156
  */
3764
4157
  /**
3765
- * Whether [Entity Detection](https://www.assemblyai.com/docs/models/entity-detection) is enabled, can be true or false
4158
+ * The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests
3766
4159
  */
3767
- type TranscriptEntityDetection = boolean | null;
4160
+ type TranscriptWebhookAuthHeaderName = string | null;
3768
4161
 
3769
4162
  /**
3770
4163
  * Generated by orval v7.9.0 🍺
3771
4164
  * Do not edit manually.
3772
4165
  * AssemblyAI API
3773
- * AssemblyAI API
3774
- * OpenAPI spec version: 1.3.4
4166
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4167
+ * OpenAPI spec version: 1.0.0
3775
4168
  */
3776
4169
  /**
3777
- * The type of entity for the detected entity
4170
+ * The status code we received from your server when delivering the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) request, if a webhook URL was provided
3778
4171
  */
3779
- type EntityType = (typeof EntityType)[keyof typeof EntityType];
3780
- declare const EntityType: {
3781
- readonly account_number: "account_number";
3782
- readonly banking_information: "banking_information";
3783
- readonly blood_type: "blood_type";
3784
- readonly credit_card_cvv: "credit_card_cvv";
3785
- readonly credit_card_expiration: "credit_card_expiration";
3786
- readonly credit_card_number: "credit_card_number";
3787
- readonly date: "date";
3788
- readonly date_interval: "date_interval";
3789
- readonly date_of_birth: "date_of_birth";
3790
- readonly drivers_license: "drivers_license";
3791
- readonly drug: "drug";
3792
- readonly duration: "duration";
3793
- readonly email_address: "email_address";
3794
- readonly event: "event";
3795
- readonly filename: "filename";
3796
- readonly gender_sexuality: "gender_sexuality";
3797
- readonly healthcare_number: "healthcare_number";
3798
- readonly injury: "injury";
3799
- readonly ip_address: "ip_address";
3800
- readonly language: "language";
3801
- readonly location: "location";
3802
- readonly marital_status: "marital_status";
3803
- readonly medical_condition: "medical_condition";
3804
- readonly medical_process: "medical_process";
3805
- readonly money_amount: "money_amount";
3806
- readonly nationality: "nationality";
3807
- readonly number_sequence: "number_sequence";
3808
- readonly occupation: "occupation";
3809
- readonly organization: "organization";
3810
- readonly passport_number: "passport_number";
3811
- readonly password: "password";
3812
- readonly person_age: "person_age";
3813
- readonly person_name: "person_name";
3814
- readonly phone_number: "phone_number";
3815
- readonly physical_attribute: "physical_attribute";
3816
- readonly political_affiliation: "political_affiliation";
3817
- readonly religion: "religion";
3818
- readonly statistics: "statistics";
3819
- readonly time: "time";
3820
- readonly url: "url";
3821
- readonly us_social_security_number: "us_social_security_number";
3822
- readonly username: "username";
3823
- readonly vehicle_id: "vehicle_id";
3824
- readonly zodiac_sign: "zodiac_sign";
3825
- };
4172
+ type TranscriptWebhookStatusCode = number | null;
3826
4173
 
3827
4174
  /**
3828
4175
  * Generated by orval v7.9.0 🍺
3829
4176
  * Do not edit manually.
3830
4177
  * AssemblyAI API
3831
- * AssemblyAI API
3832
- * OpenAPI spec version: 1.3.4
4178
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4179
+ * OpenAPI spec version: 1.0.0
3833
4180
  */
3834
-
3835
4181
  /**
3836
- * A detected entity
4182
+ * The URL to which we send [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests.
4183
+
3837
4184
  */
3838
- interface Entity {
3839
- /** The type of entity for the detected entity */
3840
- entity_type: EntityType;
3841
- /** The text for the detected entity */
3842
- text: string;
3843
- /** The starting time, in milliseconds, at which the detected entity appears in the audio file */
3844
- start: number;
3845
- /** The ending time, in milliseconds, for the detected entity in the audio file */
3846
- end: number;
3847
- }
4185
+ type TranscriptWebhookUrl = string | null;
3848
4186
 
3849
4187
  /**
3850
4188
  * Generated by orval v7.9.0 🍺
3851
4189
  * Do not edit manually.
3852
4190
  * AssemblyAI API
3853
- * AssemblyAI API
3854
- * OpenAPI spec version: 1.3.4
4191
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4192
+ * OpenAPI spec version: 1.0.0
3855
4193
  */
3856
4194
 
3857
4195
  /**
3858
- * An array of results for the Entity Detection model, if it is enabled.
3859
- See [Entity detection](https://www.assemblyai.com/docs/models/entity-detection) for more information.
4196
+ * An array of temporally-sequential word objects, one for each word in the transcript.
3860
4197
 
3861
4198
  */
3862
- type TranscriptEntities = Entity[] | null;
4199
+ type TranscriptWords = TranscriptWord[] | null;
3863
4200
 
3864
4201
  /**
3865
4202
  * Generated by orval v7.9.0 🍺
3866
4203
  * Do not edit manually.
3867
4204
  * AssemblyAI API
4205
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4206
+ * OpenAPI spec version: 1.0.0
4207
+ */
4208
+ /**
4209
+ * This parameter does not currently have any functionality attached to it.
4210
+ */
4211
+ type TranscriptCustomTopics = boolean | null;
4212
+
4213
+ /**
4214
+ * Generated by orval v7.9.0 🍺
4215
+ * Do not edit manually.
3868
4216
  * AssemblyAI API
3869
- * OpenAPI spec version: 1.3.4
4217
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4218
+ * OpenAPI spec version: 1.0.0
3870
4219
  */
4220
+
3871
4221
  /**
3872
- * Defaults to null. Reject audio files that contain less than this fraction of speech.
3873
- Valid values are in the range [0, 1] inclusive.
4222
+ * This parameter has been replaced with the `speech_models` parameter, learn more about the `speech_models` parameter [here](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model).
3874
4223
 
3875
- * @minimum 0
3876
- * @maximum 1
3877
4224
  */
3878
- type TranscriptSpeechThreshold = number | null;
4225
+ type TranscriptSpeechModel = SpeechModel | null;
3879
4226
 
3880
4227
  /**
3881
4228
  * Generated by orval v7.9.0 🍺
3882
4229
  * Do not edit manually.
3883
4230
  * AssemblyAI API
3884
- * AssemblyAI API
3885
- * OpenAPI spec version: 1.3.4
4231
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4232
+ * OpenAPI spec version: 1.0.0
3886
4233
  */
3887
4234
  /**
3888
- * True while a request is throttled and false when a request is no longer throttled
4235
+ * This parameter does not currently have any functionality attached to it.
3889
4236
  */
3890
- type TranscriptThrottled = boolean | null;
4237
+ type TranscriptSpeedBoost = boolean | null;
3891
4238
 
3892
4239
  /**
3893
4240
  * Generated by orval v7.9.0 🍺
3894
4241
  * Do not edit manually.
3895
4242
  * AssemblyAI API
4243
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4244
+ * OpenAPI spec version: 1.0.0
4245
+ */
4246
+ /**
4247
+ * Translated text keyed by language code. See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for more details.
4248
+ */
4249
+ interface TranscriptTranslatedTexts {
4250
+ /** Translated text for this language code */
4251
+ language_code?: string;
4252
+ }
4253
+
4254
+ /**
4255
+ * Generated by orval v7.9.0 🍺
4256
+ * Do not edit manually.
3896
4257
  * AssemblyAI API
3897
- * OpenAPI spec version: 1.3.4
4258
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4259
+ * OpenAPI spec version: 1.0.0
3898
4260
  */
3899
4261
 
3900
4262
  /**
3901
4263
  * A transcript object
3902
4264
  */
3903
4265
  interface Transcript {
3904
- /** The unique identifier of your transcript */
3905
- id: string;
4266
+ /** The number of audio channels in the audio file. This is only present when [multichannel](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) is enabled. */
4267
+ audio_channels?: number;
4268
+ /** The duration of this transcript object's media file, in seconds */
4269
+ audio_duration?: TranscriptAudioDuration;
4270
+ /** The point in time, in milliseconds, in the file at which the transcription was terminated. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
4271
+ audio_end_at?: TranscriptAudioEndAt;
4272
+ /** The point in time, in milliseconds, in the file at which the transcription was started. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
4273
+ audio_start_from?: TranscriptAudioStartFrom;
3906
4274
  /** The URL of the media that was transcribed */
3907
4275
  audio_url: string;
3908
- /** The status of your transcript. Possible values are queued, processing, completed, or error. */
3909
- status: TranscriptStatus;
3910
- /** The language of your audio file.
3911
- Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
3912
- The default value is 'en_us'.
3913
- */
3914
- language_code?: TranscriptLanguageCodeProperty;
3915
- /** Whether [Automatic language detection](https://www.assemblyai.com/docs/models/speech-recognition#automatic-language-detection) is enabled, either true or false */
3916
- language_detection?: TranscriptLanguageDetection;
3917
- /**
3918
- * The confidence threshold for the automatically detected language.
3919
- An error will be returned if the language confidence is below this threshold.
4276
+ /** Whether [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) is enabled, can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible chapter summaries. See the [updated Auto Chapters page](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for details.
3920
4277
 
3921
- * @minimum 0
3922
- * @maximum 1
3923
- */
3924
- language_confidence_threshold: TranscriptLanguageConfidenceThreshold;
3925
- /**
3926
- * The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence)
3927
- * @minimum 0
3928
- * @maximum 1
3929
- */
3930
- language_confidence: TranscriptLanguageConfidence;
3931
- /** The speech model used for the transcription. When `null`, the default model is used. */
3932
- speech_model: TranscriptSpeechModel;
3933
- /** The textual transcript of your media file */
3934
- text?: TranscriptText;
3935
- /** An array of temporally-sequential word objects, one for each word in the transcript.
3936
- See [Speech recognition](https://www.assemblyai.com/docs/models/speech-recognition) for more information.
4278
+ Note: This parameter is only supported for the Universal-2 model.
3937
4279
  */
3938
- words?: TranscriptWords;
3939
- /** When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
3940
- See [Speaker diarization](https://www.assemblyai.com/docs/speech-to-text/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/speech-to-text/speech-recognition#multichannel-transcription) for more information.
4280
+ auto_chapters?: TranscriptAutoChapters;
4281
+ /** Whether [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) is enabled, either true or false */
4282
+ auto_highlights: boolean;
4283
+ /** An array of results for the Key Phrases model, if it is enabled.
4284
+ See [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) for more information.
3941
4285
  */
3942
- utterances?: TranscriptUtterances;
3943
- /**
3944
- * The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence)
3945
- * @minimum 0
3946
- * @maximum 1
3947
- */
4286
+ auto_highlights_result?: TranscriptAutoHighlightsResult;
4287
+ /** An array of temporally sequential chapters for the audio file. See [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for more information. */
4288
+ chapters?: TranscriptChapters;
4289
+ /** The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence) */
3948
4290
  confidence?: TranscriptConfidence;
3949
- /** The duration of this transcript object's media file, in seconds */
3950
- audio_duration?: TranscriptAudioDuration;
3951
- /** Whether Automatic Punctuation is enabled, either true or false */
3952
- punctuate?: TranscriptPunctuate;
3953
- /** Whether Text Formatting is enabled, either true or false */
3954
- format_text?: TranscriptFormatText;
3955
- /** Transcribe Filler Words, like "umm", in your media file; can be true or false */
3956
- disfluencies?: TranscriptDisfluencies;
3957
- /** Whether [Multichannel transcription](https://www.assemblyai.com/docs/models/speech-recognition#multichannel-transcription) was enabled in the transcription request, either true or false */
3958
- multichannel?: TranscriptMultichannel;
3959
- /** The number of audio channels in the audio file. This is only present when multichannel is enabled. */
3960
- audio_channels?: number;
3961
- /** The URL to which we send webhook requests.
3962
- We sends two different types of webhook requests.
3963
- One request when a transcript is completed or failed, and one request when the redacted audio is ready if redact_pii_audio is enabled.
4291
+ /** Whether [Content Moderation](https://www.assemblyai.com/docs/content-moderation) is enabled, can be true or false */
4292
+ content_safety?: TranscriptContentSafety;
4293
+ /** An array of results for the Content Moderation model, if it is enabled.
4294
+ See [Content moderation](https://www.assemblyai.com/docs/content-moderation) for more information.
3964
4295
  */
3965
- webhook_url?: TranscriptWebhookUrl;
3966
- /** The status code we received from your server when delivering the transcript completed or failed webhook request, if a webhook URL was provided */
3967
- webhook_status_code?: TranscriptWebhookStatusCode;
3968
- /** Whether webhook authentication details were provided */
3969
- webhook_auth: boolean;
3970
- /** The header name to be sent with the transcript completed or failed webhook requests */
3971
- webhook_auth_header_name?: TranscriptWebhookAuthHeaderName;
3972
- /**
3973
- * Whether speed boost is enabled
3974
- * @deprecated
4296
+ content_safety_labels?: TranscriptContentSafetyLabels;
4297
+ /** Customize how words are spelled and formatted using to and from values. See [Custom Spelling](https://www.assemblyai.com/docs/pre-recorded-audio/custom-spelling) for more details. */
4298
+ custom_spelling?: TranscriptCustomSpellingProperty;
4299
+ /** Transcribe [Filler Words](https://www.assemblyai.com/docs/pre-recorded-audio/filler-words), like "umm", in your media file; can be true or false */
4300
+ disfluencies?: TranscriptDisfluencies;
4301
+ /** The domain-specific model applied to the transcript. When set to `"medical-v1"`, [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) was used to improve accuracy for medical terminology.
3975
4302
  */
3976
- speed_boost?: TranscriptSpeedBoost;
3977
- /** Whether Key Phrases is enabled, either true or false */
3978
- auto_highlights: boolean;
3979
- /** An array of results for the Key Phrases model, if it is enabled.
3980
- See [Key Phrases](https://www.assemblyai.com/docs/models/key-phrases) for more information.
4303
+ domain?: TranscriptDomain;
4304
+ /** An array of results for the Entity Detection model, if it is enabled.
4305
+ See [Entity detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) for more information.
3981
4306
  */
3982
- auto_highlights_result?: TranscriptAutoHighlightsResult;
3983
- /** The point in time, in milliseconds, in the file at which the transcription was started */
3984
- audio_start_from?: TranscriptAudioStartFrom;
3985
- /** The point in time, in milliseconds, in the file at which the transcription was terminated */
3986
- audio_end_at?: TranscriptAudioEndAt;
3987
- /**
3988
- * The list of custom vocabulary to boost transcription probability for
3989
- * @deprecated
4307
+ entities?: TranscriptEntities;
4308
+ /** Whether [Entity Detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) is enabled, can be true or false */
4309
+ entity_detection?: TranscriptEntityDetection;
4310
+ /** Error message of why the transcript failed */
4311
+ error?: string;
4312
+ /** Whether [Profanity Filtering](https://www.assemblyai.com/docs/profanity-filtering) is enabled, either true or false */
4313
+ filter_profanity?: TranscriptFilterProfanity;
4314
+ /** Whether [Text Formatting](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false */
4315
+ format_text?: TranscriptFormatText;
4316
+ /** Whether [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) is enabled, can be true or false */
4317
+ iab_categories?: TranscriptIabCategories;
4318
+ /** The result of the Topic Detection model, if it is enabled.
4319
+ See [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) for more information.
4320
+ */
4321
+ iab_categories_result?: TranscriptIabCategoriesResult;
4322
+ /** The unique identifier of your transcript */
4323
+ id: string;
4324
+ /** Improve accuracy with up to 200 (for Universal-2) or 1000 (for Universal-3 Pro) domain-specific words or phrases (maximum 6 words per phrase). See [Keyterms Prompting](https://www.assemblyai.com/docs/pre-recorded-audio/keyterms-prompting) for more details.
3990
4325
  */
3991
- word_boost?: string[];
3992
- /** The word boost parameter value */
3993
- boost_param?: TranscriptBoostParamProperty;
3994
- /** Whether [Profanity Filtering](https://www.assemblyai.com/docs/models/speech-recognition#profanity-filtering) is enabled, either true or false */
3995
- filter_profanity?: TranscriptFilterProfanity;
3996
- /** Whether [PII Redaction](https://www.assemblyai.com/docs/models/pii-redaction) is enabled, either true or false */
4326
+ keyterms_prompt?: string[];
4327
+ /** The language of your audio file.
4328
+ Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
4329
+ The default value is 'en_us'.
4330
+ */
4331
+ language_code?: TranscriptLanguageCode;
4332
+ /** The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
4333
+ One of the values specified must be `en`.
4334
+ */
4335
+ language_codes?: TranscriptLanguageCodes;
4336
+ /** The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence). See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details. */
4337
+ language_confidence: TranscriptLanguageConfidence;
4338
+ /** The confidence threshold for the automatically detected language.
4339
+ An error will be returned if the language confidence is below this threshold.
4340
+ See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
4341
+ */
4342
+ language_confidence_threshold: TranscriptLanguageConfidenceThreshold;
4343
+ /** Whether [Automatic language detection](/docs/pre-recorded-audio/automatic-language-detection) is enabled, either true or false */
4344
+ language_detection?: TranscriptLanguageDetection;
4345
+ /** Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection). */
4346
+ language_detection_options?: TranscriptLanguageDetectionOptions;
4347
+ /** Whether [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) was enabled in the transcription request, either true or false */
4348
+ multichannel?: TranscriptMultichannel;
4349
+ /** Provide natural language prompting of up to 1,500 words of contextual information to the model. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for best practices.
4350
+
4351
+ Note: This parameter is only supported for the Universal-3 Pro model.
4352
+ */
4353
+ prompt?: string;
4354
+ /** Whether [Automatic Punctuation](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false */
4355
+ punctuate?: TranscriptPunctuate;
4356
+ /** Whether [PII Redaction](https://www.assemblyai.com/docs/pii-redaction) is enabled, either true or false */
3997
4357
  redact_pii: boolean;
3998
4358
  /** Whether a redacted version of the audio file was generated,
3999
- either true or false. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
4359
+ either true or false. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
4000
4360
  */
4001
4361
  redact_pii_audio?: TranscriptRedactPiiAudio;
4362
+ /** The options for PII-redacted audio, if redact_pii_audio is enabled.
4363
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
4364
+ */
4365
+ redact_pii_audio_options?: TranscriptRedactPiiAudioOptions;
4002
4366
  /** The audio quality of the PII-redacted audio file, if redact_pii_audio is enabled.
4003
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
4367
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
4004
4368
  */
4005
4369
  redact_pii_audio_quality?: TranscriptRedactPiiAudioQuality;
4006
4370
  /** The list of PII Redaction policies that were enabled, if PII Redaction is enabled.
4007
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
4371
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more information.
4008
4372
  */
4009
4373
  redact_pii_policies?: TranscriptRedactPiiPolicies;
4010
- /** The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
4374
+ /** The replacement logic for detected PII, can be `entity_type` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
4011
4375
  redact_pii_sub?: SubstitutionPolicy;
4012
- /** Whether [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, can be true or false */
4376
+ /** Whether [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) is enabled, can be true or false */
4377
+ sentiment_analysis?: TranscriptSentimentAnalysis;
4378
+ /** An array of results for the Sentiment Analysis model, if it is enabled.
4379
+ See [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) for more information.
4380
+ */
4381
+ sentiment_analysis_results?: TranscriptSentimentAnalysisResults;
4382
+ /** Whether [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, can be true or false */
4013
4383
  speaker_labels?: TranscriptSpeakerLabels;
4014
- /** Tell the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details. */
4384
+ /** Tell the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details. */
4015
4385
  speakers_expected?: TranscriptSpeakersExpected;
4016
- /** Whether [Content Moderation](https://www.assemblyai.com/docs/models/content-moderation) is enabled, can be true or false */
4017
- content_safety?: TranscriptContentSafety;
4018
- /** An array of results for the Content Moderation model, if it is enabled.
4019
- See [Content moderation](https://www.assemblyai.com/docs/models/content-moderation) for more information.
4020
- */
4021
- content_safety_labels?: TranscriptContentSafetyLabels;
4022
- /** Whether [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) is enabled, can be true or false */
4023
- iab_categories?: TranscriptIabCategories;
4024
- /** The result of the Topic Detection model, if it is enabled.
4025
- See [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) for more information.
4026
- */
4027
- iab_categories_result?: TranscriptIabCategoriesResult;
4028
- /** Customize how words are spelled and formatted using to and from values */
4029
- custom_spelling?: TranscriptCustomSpellingProperty;
4030
- /** Improve accuracy with up to 1000 domain-specific words or phrases (maximum 6 words per phrase).
4386
+ /** The speech model that was actually used for the transcription. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models. */
4387
+ speech_model_used?: SpeechModel;
4388
+ /** List multiple speech models in priority order, allowing our system to automatically route your audio to the best available option. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models and routing behavior.
4031
4389
  */
4032
- keyterms_prompt?: string[];
4033
- /**
4034
- * This parameter does not currently have any functionality attached to it.
4035
- * @deprecated
4390
+ speech_models?: TranscriptSpeechModels;
4391
+ /** Defaults to null. Reject audio files that contain less than this fraction of speech.
4392
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
4393
+ */
4394
+ speech_threshold?: TranscriptSpeechThreshold;
4395
+ /** Speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
4036
4396
  */
4037
- prompt?: string;
4038
- /** Whether [Auto Chapters](https://www.assemblyai.com/docs/models/auto-chapters) is enabled, can be true or false */
4039
- auto_chapters?: TranscriptAutoChapters;
4040
- /** An array of temporally sequential chapters for the audio file */
4041
- chapters?: TranscriptChapters;
4042
- /** Whether [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled, either true or false */
4397
+ speech_understanding?: TranscriptSpeechUnderstanding;
4398
+ /** The status of your transcript. Possible values are queued, processing, completed, or error. */
4399
+ status: TranscriptStatus;
4400
+ /** Whether [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization) is enabled, either true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
4401
+
4402
+ Note: This parameter is only supported for the Universal-2 model.
4403
+ */
4043
4404
  summarization: boolean;
4044
- /** The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled */
4045
- summary_type?: TranscriptSummaryType;
4405
+ /** The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
4406
+ summary?: TranscriptSummary;
4046
4407
  /** The Summarization model used to generate the summary,
4047
- if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
4408
+ if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-models) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
4048
4409
  */
4049
4410
  summary_model?: TranscriptSummaryModel;
4050
- /** The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled */
4051
- summary?: TranscriptSummary;
4052
- /**
4053
- * Whether custom topics is enabled, either true or false
4054
- * @deprecated
4055
- */
4056
- custom_topics?: TranscriptCustomTopics;
4057
- /** The list of custom topics provided if custom topics is enabled */
4058
- topics?: string[];
4059
- /** Whether [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) is enabled, can be true or false */
4060
- sentiment_analysis?: TranscriptSentimentAnalysis;
4061
- /** An array of results for the Sentiment Analysis model, if it is enabled.
4062
- See [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) for more information.
4063
- */
4064
- sentiment_analysis_results?: TranscriptSentimentAnalysisResults;
4065
- /** Whether [Entity Detection](https://www.assemblyai.com/docs/models/entity-detection) is enabled, can be true or false */
4066
- entity_detection?: TranscriptEntityDetection;
4067
- /** An array of results for the Entity Detection model, if it is enabled.
4068
- See [Entity detection](https://www.assemblyai.com/docs/models/entity-detection) for more information.
4411
+ /** The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-types) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
4412
+ summary_type?: TranscriptSummaryType;
4413
+ /** Whether [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) were removed from the transcript text.
4414
+
4415
+ Note: This parameter is only supported for the Universal-3 Pro model.
4069
4416
  */
4070
- entities?: TranscriptEntities;
4071
- /**
4072
- * Defaults to null. Reject audio files that contain less than this fraction of speech.
4073
- Valid values are in the range [0, 1] inclusive.
4417
+ remove_audio_tags?: TranscriptRemoveAudioTagsProperty;
4418
+ /** The temperature that was used for the model's response. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for more details.
4074
4419
 
4075
- * @minimum 0
4076
- * @maximum 1
4077
- */
4078
- speech_threshold?: TranscriptSpeechThreshold;
4420
+ Note: This parameter can only be used with the Universal-3 Pro model.
4421
+ */
4422
+ temperature?: TranscriptTemperature;
4423
+ /** The textual transcript of your media file */
4424
+ text?: TranscriptText;
4079
4425
  /** True while a request is throttled and false when a request is no longer throttled */
4080
4426
  throttled?: TranscriptThrottled;
4081
- /** Error message of why the transcript failed */
4082
- error?: string;
4083
- /**
4084
- * The language model that was used for the transcript
4085
- * @deprecated
4427
+ /** When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
4428
+ See [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) for more information.
4429
+ */
4430
+ utterances?: TranscriptUtterances;
4431
+ /** Whether [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) authentication details were provided */
4432
+ webhook_auth: boolean;
4433
+ /** The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests */
4434
+ webhook_auth_header_name?: TranscriptWebhookAuthHeaderName;
4435
+ /** The status code we received from your server when delivering the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) request, if a webhook URL was provided */
4436
+ webhook_status_code?: TranscriptWebhookStatusCode;
4437
+ /** The URL to which we send [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests.
4086
4438
  */
4087
- language_model: string;
4088
- /**
4089
- * The acoustic model that was used for the transcript
4090
- * @deprecated
4439
+ webhook_url?: TranscriptWebhookUrl;
4440
+ /** An array of temporally-sequential word objects, one for each word in the transcript.
4091
4441
  */
4442
+ words?: TranscriptWords;
4443
+ /** This parameter does not currently have any functionality attached to it. */
4092
4444
  acoustic_model: string;
4445
+ /** This parameter does not currently have any functionality attached to it. */
4446
+ custom_topics?: TranscriptCustomTopics;
4447
+ /** This parameter does not currently have any functionality attached to it. */
4448
+ language_model: string;
4449
+ /** This parameter has been replaced with the `speech_models` parameter, learn more about the `speech_models` parameter [here](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model).
4450
+ */
4451
+ speech_model: TranscriptSpeechModel;
4452
+ /** This parameter does not currently have any functionality attached to it. */
4453
+ speed_boost?: TranscriptSpeedBoost;
4454
+ /** This parameter does not currently have any functionality attached to it. */
4455
+ topics?: string[];
4456
+ /** Translated text keyed by language code. See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for more details. */
4457
+ translated_texts?: TranscriptTranslatedTexts;
4093
4458
  }
4094
4459
 
4095
4460
  /**
@@ -4446,6 +4811,18 @@ type SharedCallbackParameter = string;
4446
4811
  /**
4447
4812
  * SharedCallbackMethodParameter type definition
4448
4813
  */
4814
+ /**
4815
+ * SharedCallbackMethodParameter type definition
4816
+ */
4817
+ /**
4818
+ * SharedCallbackMethodParameter type definition
4819
+ */
4820
+ /**
4821
+ * SharedCallbackMethodParameter type definition
4822
+ */
4823
+ /**
4824
+ * SharedCallbackMethodParameter type definition
4825
+ */
4449
4826
  type SharedCallbackMethodParameter = typeof SharedCallbackMethodParameter[keyof typeof SharedCallbackMethodParameter];
4450
4827
  declare const SharedCallbackMethodParameter: {
4451
4828
  readonly POST: "POST";
@@ -4550,6 +4927,18 @@ type SharedCustomTopicParameter = string | string[];
4550
4927
  /**
4551
4928
  * SharedCustomTopicModeParameter type definition
4552
4929
  */
4930
+ /**
4931
+ * SharedCustomTopicModeParameter type definition
4932
+ */
4933
+ /**
4934
+ * SharedCustomTopicModeParameter type definition
4935
+ */
4936
+ /**
4937
+ * SharedCustomTopicModeParameter type definition
4938
+ */
4939
+ /**
4940
+ * SharedCustomTopicModeParameter type definition
4941
+ */
4553
4942
  type SharedCustomTopicModeParameter = typeof SharedCustomTopicModeParameter[keyof typeof SharedCustomTopicModeParameter];
4554
4943
  declare const SharedCustomTopicModeParameter: {
4555
4944
  readonly extended: "extended";
@@ -4611,6 +5000,18 @@ type SharedCustomIntentParameter = string | string[];
4611
5000
  /**
4612
5001
  * SharedCustomIntentModeParameter type definition
4613
5002
  */
5003
+ /**
5004
+ * SharedCustomIntentModeParameter type definition
5005
+ */
5006
+ /**
5007
+ * SharedCustomIntentModeParameter type definition
5008
+ */
5009
+ /**
5010
+ * SharedCustomIntentModeParameter type definition
5011
+ */
5012
+ /**
5013
+ * SharedCustomIntentModeParameter type definition
5014
+ */
4614
5015
  type SharedCustomIntentModeParameter = typeof SharedCustomIntentModeParameter[keyof typeof SharedCustomIntentModeParameter];
4615
5016
  declare const SharedCustomIntentModeParameter: {
4616
5017
  readonly extended: "extended";
@@ -4698,6 +5099,18 @@ type ListenV1DictationParameter = boolean;
4698
5099
  /**
4699
5100
  * ListenV1EncodingParameter type definition
4700
5101
  */
5102
+ /**
5103
+ * ListenV1EncodingParameter type definition
5104
+ */
5105
+ /**
5106
+ * ListenV1EncodingParameter type definition
5107
+ */
5108
+ /**
5109
+ * ListenV1EncodingParameter type definition
5110
+ */
5111
+ /**
5112
+ * ListenV1EncodingParameter type definition
5113
+ */
4701
5114
  type ListenV1EncodingParameter = typeof ListenV1EncodingParameter[keyof typeof ListenV1EncodingParameter];
4702
5115
  declare const ListenV1EncodingParameter: {
4703
5116
  readonly linear16: "linear16";
@@ -5113,281 +5526,434 @@ type ListenV1MediaTranscribeParams = {
5113
5526
  * Generated by orval v7.9.0 🍺
5114
5527
  * Do not edit manually.
5115
5528
  * AssemblyAI API
5116
- * AssemblyAI API
5117
- * OpenAPI spec version: 1.3.4
5529
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5530
+ * OpenAPI spec version: 1.0.0
5118
5531
  */
5532
+ /**
5533
+ * Enable domain-specific transcription models to improve accuracy for specialized terminology. Set to `"medical-v1"` to enable [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) for improved accuracy of medical terms such as medications, procedures, conditions, and dosages.
5119
5534
 
5120
- type TranscriptOptionalParamsLanguageCodeOneOf = TranscriptLanguageCode | string;
5535
+ Supported languages: English (`en`), Spanish (`es`), German (`de`), French (`fr`). If used with an unsupported language, the parameter is ignored and a warning is returned.
5536
+
5537
+ */
5538
+ type TranscriptParamsDomain = string | null;
5121
5539
 
5122
5540
  /**
5123
5541
  * Generated by orval v7.9.0 🍺
5124
5542
  * Do not edit manually.
5125
5543
  * AssemblyAI API
5126
- * AssemblyAI API
5127
- * OpenAPI spec version: 1.3.4
5544
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5545
+ * OpenAPI spec version: 1.0.0
5128
5546
  */
5129
5547
 
5130
5548
  /**
5131
- * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
5549
+ * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
5132
5550
  The default value is 'en_us'.
5133
5551
 
5134
5552
  */
5135
- type TranscriptOptionalParamsLanguageCode = TranscriptOptionalParamsLanguageCodeOneOf | null;
5553
+ type TranscriptParamsLanguageCode = TranscriptLanguageCode | null;
5136
5554
 
5137
5555
  /**
5138
5556
  * Generated by orval v7.9.0 🍺
5139
5557
  * Do not edit manually.
5140
5558
  * AssemblyAI API
5141
- * AssemblyAI API
5142
- * OpenAPI spec version: 1.3.4
5559
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5560
+ * OpenAPI spec version: 1.0.0
5143
5561
  */
5144
5562
 
5145
5563
  /**
5146
- * The speech model to use for the transcription. When `null`, the "best" model is used.
5564
+ * The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
5565
+ One of the values specified must be `en`.
5566
+
5147
5567
  */
5148
- type TranscriptOptionalParamsSpeechModel = SpeechModel | null;
5568
+ type TranscriptParamsLanguageCodes = TranscriptLanguageCode[] | null;
5149
5569
 
5150
5570
  /**
5151
5571
  * Generated by orval v7.9.0 🍺
5152
5572
  * Do not edit manually.
5153
5573
  * AssemblyAI API
5574
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5575
+ * OpenAPI spec version: 1.0.0
5576
+ */
5577
+ /**
5578
+ * Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection).
5579
+ */
5580
+ interface TranscriptOptionalParamsLanguageDetectionOptions {
5581
+ /** List of languages expected in the audio file. Defaults to `["all"]` when unspecified. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details. */
5582
+ expected_languages?: string[];
5583
+ /** If the detected language of the audio file is not in the list of expected languages, the `fallback_language` is used. Specify `["auto"]` to let our model choose the fallback language from `expected_languages` with the highest confidence score. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
5584
+ */
5585
+ fallback_language?: string;
5586
+ /** Whether [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) should be detected.
5587
+ */
5588
+ code_switching?: boolean;
5589
+ /** The confidence threshold for [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) detection. If the code switching confidence is below this threshold, the transcript will be processed in the language with the highest `language_detection_confidence` score.
5590
+ */
5591
+ code_switching_confidence_threshold?: number;
5592
+ }
5593
+
5594
+ /**
5595
+ * Generated by orval v7.9.0 🍺
5596
+ * Do not edit manually.
5154
5597
  * AssemblyAI API
5155
- * OpenAPI spec version: 1.3.4
5598
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5599
+ * OpenAPI spec version: 1.0.0
5156
5600
  */
5157
5601
  /**
5158
- * The header name to be sent with the transcript completed or failed webhook requests
5602
+ * Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep.
5159
5603
  */
5160
- type TranscriptOptionalParamsWebhookAuthHeaderName = string | null;
5604
+ type TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod = (typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod)[keyof typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod];
5605
+ declare const TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod: {
5606
+ readonly silence: "silence";
5607
+ };
5161
5608
 
5162
5609
  /**
5163
5610
  * Generated by orval v7.9.0 🍺
5164
5611
  * Do not edit manually.
5165
5612
  * AssemblyAI API
5613
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5614
+ * OpenAPI spec version: 1.0.0
5615
+ */
5616
+
5617
+ /**
5618
+ * Specify options for [PII redacted audio](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) files.
5619
+ */
5620
+ interface TranscriptOptionalParamsRedactPiiAudioOptions {
5621
+ /** By default, audio redaction provides redacted audio URLs only when speech is detected. However, if your use-case specifically requires redacted audio files even for silent audio files without any dialogue, you can opt to receive these URLs by setting this parameter to `true`. */
5622
+ return_redacted_no_speech_audio?: boolean;
5623
+ /** Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep. */
5624
+ override_audio_redaction_method?: TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod;
5625
+ }
5626
+
5627
+ /**
5628
+ * Generated by orval v7.9.0 🍺
5629
+ * Do not edit manually.
5166
5630
  * AssemblyAI API
5167
- * OpenAPI spec version: 1.3.4
5631
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5632
+ * OpenAPI spec version: 1.0.0
5168
5633
  */
5634
+
5169
5635
  /**
5170
- * The header value to send back with the transcript completed or failed webhook requests for added security
5636
+ * The replacement logic for detected PII, can be `entity_type` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details.
5171
5637
  */
5172
- type TranscriptOptionalParamsWebhookAuthHeaderValue = string | null;
5638
+ type TranscriptParamsRedactPiiSub = SubstitutionPolicy | null;
5173
5639
 
5174
5640
  /**
5175
5641
  * Generated by orval v7.9.0 🍺
5176
5642
  * Do not edit manually.
5177
5643
  * AssemblyAI API
5644
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5645
+ * OpenAPI spec version: 1.0.0
5646
+ */
5647
+ /**
5648
+ * Specify options for [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers). Use this to set a range of possible speakers.
5649
+ */
5650
+ interface TranscriptOptionalParamsSpeakerOptions {
5651
+ /** The minimum number of speakers expected in the audio file. See [Set a range of possible speakers](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers) for more details. */
5652
+ min_speakers_expected?: number;
5653
+ /** <Warning>Setting this parameter too high may hurt model accuracy</Warning>
5654
+ The maximum number of speakers expected in the audio file. The default depends on audio duration: no limit for 0-2 minutes, 10 for 2-10 minutes, and 30 for 10+ minutes. See [Set a range of possible speakers](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers) for more details.
5655
+ */
5656
+ max_speakers_expected?: number;
5657
+ }
5658
+
5659
+ /**
5660
+ * Generated by orval v7.9.0 🍺
5661
+ * Do not edit manually.
5178
5662
  * AssemblyAI API
5179
- * OpenAPI spec version: 1.3.4
5663
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5664
+ * OpenAPI spec version: 1.0.0
5180
5665
  */
5181
5666
  /**
5182
- * How much to boost specified words
5667
+ * Tells the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details.
5183
5668
  */
5184
- type TranscriptBoostParam = (typeof TranscriptBoostParam)[keyof typeof TranscriptBoostParam];
5185
- declare const TranscriptBoostParam: {
5186
- readonly low: "low";
5187
- readonly default: "default";
5188
- readonly high: "high";
5189
- };
5669
+ type TranscriptParamsSpeakersExpected = number | null;
5670
+
5671
+ /**
5672
+ * Generated by orval v7.9.0 🍺
5673
+ * Do not edit manually.
5674
+ * AssemblyAI API
5675
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5676
+ * OpenAPI spec version: 1.0.0
5677
+ */
5678
+ /**
5679
+ * Reject audio files that contain less than this fraction of speech.
5680
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
5681
+
5682
+ */
5683
+ type TranscriptParamsSpeechThreshold = number | null;
5190
5684
 
5191
5685
  /**
5192
5686
  * Generated by orval v7.9.0 🍺
5193
5687
  * Do not edit manually.
5194
5688
  * AssemblyAI API
5689
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5690
+ * OpenAPI spec version: 1.0.0
5691
+ */
5692
+
5693
+ type TranscriptOptionalParamsSpeechUnderstandingRequest = TranslationRequestBody | SpeakerIdentificationRequestBody | CustomFormattingRequestBody;
5694
+
5695
+ /**
5696
+ * Generated by orval v7.9.0 🍺
5697
+ * Do not edit manually.
5195
5698
  * AssemblyAI API
5196
- * OpenAPI spec version: 1.3.4
5699
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5700
+ * OpenAPI spec version: 1.0.0
5701
+ */
5702
+
5703
+ /**
5704
+ * Enable speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
5705
+
5197
5706
  */
5707
+ interface TranscriptOptionalParamsSpeechUnderstanding {
5708
+ request: TranscriptOptionalParamsSpeechUnderstandingRequest;
5709
+ }
5198
5710
 
5199
5711
  /**
5200
- * The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details.
5712
+ * Generated by orval v7.9.0 🍺
5713
+ * Do not edit manually.
5714
+ * AssemblyAI API
5715
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5716
+ * OpenAPI spec version: 1.0.0
5717
+ */
5718
+ /**
5719
+ * The model to summarize the transcript
5201
5720
  */
5202
- type TranscriptOptionalParamsRedactPiiSub = SubstitutionPolicy | null;
5721
+ type SummaryModel = (typeof SummaryModel)[keyof typeof SummaryModel];
5722
+ declare const SummaryModel: {
5723
+ readonly informative: "informative";
5724
+ readonly conversational: "conversational";
5725
+ readonly catchy: "catchy";
5726
+ };
5203
5727
 
5204
5728
  /**
5205
5729
  * Generated by orval v7.9.0 🍺
5206
5730
  * Do not edit manually.
5207
5731
  * AssemblyAI API
5732
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5733
+ * OpenAPI spec version: 1.0.0
5734
+ */
5735
+ /**
5736
+ * The type of summary
5737
+ */
5738
+ type SummaryType = (typeof SummaryType)[keyof typeof SummaryType];
5739
+ declare const SummaryType: {
5740
+ readonly bullets: "bullets";
5741
+ readonly bullets_verbose: "bullets_verbose";
5742
+ readonly gist: "gist";
5743
+ readonly headline: "headline";
5744
+ readonly paragraph: "paragraph";
5745
+ };
5746
+
5747
+ /**
5748
+ * Generated by orval v7.9.0 🍺
5749
+ * Do not edit manually.
5208
5750
  * AssemblyAI API
5209
- * OpenAPI spec version: 1.3.4
5751
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5752
+ * OpenAPI spec version: 1.0.0
5210
5753
  */
5211
5754
  /**
5212
- * Tells the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details.
5755
+ * Remove [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) from the transcript text. Set to `"all"` to remove all audio tags.
5756
+
5757
+ Note: This parameter is only supported for the Universal-3 Pro model.
5758
+
5213
5759
  */
5214
- type TranscriptOptionalParamsSpeakersExpected = number | null;
5760
+ type TranscriptOptionalParamsRemoveAudioTags = (typeof TranscriptOptionalParamsRemoveAudioTags)[keyof typeof TranscriptOptionalParamsRemoveAudioTags];
5761
+ declare const TranscriptOptionalParamsRemoveAudioTags: {
5762
+ readonly all: "all";
5763
+ };
5215
5764
 
5216
5765
  /**
5217
5766
  * Generated by orval v7.9.0 🍺
5218
5767
  * Do not edit manually.
5219
5768
  * AssemblyAI API
5769
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5770
+ * OpenAPI spec version: 1.0.0
5771
+ */
5772
+
5773
+ /**
5774
+ * Remove [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) from the transcript text. Set to `"all"` to remove all audio tags.
5775
+
5776
+ Note: This parameter is only supported for the Universal-3 Pro model.
5777
+
5778
+ */
5779
+ type TranscriptParamsRemoveAudioTags = TranscriptOptionalParamsRemoveAudioTags | null;
5780
+
5781
+ /**
5782
+ * Generated by orval v7.9.0 🍺
5783
+ * Do not edit manually.
5220
5784
  * AssemblyAI API
5221
- * OpenAPI spec version: 1.3.4
5785
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5786
+ * OpenAPI spec version: 1.0.0
5222
5787
  */
5223
5788
  /**
5224
- * Reject audio files that contain less than this fraction of speech.
5225
- Valid values are in the range [0, 1] inclusive.
5226
-
5227
- * @minimum 0
5228
- * @maximum 1
5789
+ * The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests
5229
5790
  */
5230
- type TranscriptOptionalParamsSpeechThreshold = number | null;
5791
+ type TranscriptParamsWebhookAuthHeaderName = string | null;
5231
5792
 
5232
5793
  /**
5233
5794
  * Generated by orval v7.9.0 🍺
5234
5795
  * Do not edit manually.
5235
5796
  * AssemblyAI API
5236
- * AssemblyAI API
5237
- * OpenAPI spec version: 1.3.4
5797
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5798
+ * OpenAPI spec version: 1.0.0
5238
5799
  */
5239
5800
  /**
5240
- * The model to summarize the transcript
5801
+ * The header value to send back with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests for added security
5241
5802
  */
5242
- type SummaryModel = (typeof SummaryModel)[keyof typeof SummaryModel];
5243
- declare const SummaryModel: {
5244
- readonly informative: "informative";
5245
- readonly conversational: "conversational";
5246
- readonly catchy: "catchy";
5247
- };
5803
+ type TranscriptParamsWebhookAuthHeaderValue = string | null;
5248
5804
 
5249
5805
  /**
5250
5806
  * Generated by orval v7.9.0 🍺
5251
5807
  * Do not edit manually.
5252
5808
  * AssemblyAI API
5253
- * AssemblyAI API
5254
- * OpenAPI spec version: 1.3.4
5809
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5810
+ * OpenAPI spec version: 1.0.0
5255
5811
  */
5812
+
5256
5813
  /**
5257
- * The type of summary
5814
+ * This parameter has been replaced with the `speech_models` parameter, learn more about the `speech_models` parameter [here](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model).
5815
+
5258
5816
  */
5259
- type SummaryType = (typeof SummaryType)[keyof typeof SummaryType];
5260
- declare const SummaryType: {
5261
- readonly bullets: "bullets";
5262
- readonly bullets_verbose: "bullets_verbose";
5263
- readonly gist: "gist";
5264
- readonly headline: "headline";
5265
- readonly paragraph: "paragraph";
5266
- };
5817
+ type TranscriptParamsSpeechModel = SpeechModel | null;
5267
5818
 
5268
5819
  /**
5269
5820
  * Generated by orval v7.9.0 🍺
5270
5821
  * Do not edit manually.
5271
5822
  * AssemblyAI API
5272
- * AssemblyAI API
5273
- * OpenAPI spec version: 1.3.4
5823
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5824
+ * OpenAPI spec version: 1.0.0
5274
5825
  */
5275
5826
 
5276
5827
  /**
5277
5828
  * The parameters for creating a transcript
5278
5829
  */
5279
- interface TranscriptOptionalParams {
5280
- /** The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
5281
- The default value is 'en_us'.
5830
+ interface TranscriptParams {
5831
+ /** The point in time, in milliseconds, to stop transcribing in your media file. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
5832
+ audio_end_at?: number;
5833
+ /** The point in time, in milliseconds, to begin transcribing in your media file. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
5834
+ audio_start_from?: number;
5835
+ /** Enable [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters), can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible chapter summaries. See the [updated Auto Chapters page](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for details.
5836
+
5837
+ Note: This parameter is only supported for the Universal-2 model.
5282
5838
  */
5283
- language_code?: TranscriptOptionalParamsLanguageCode;
5284
- /** Enable [Automatic language detection](https://www.assemblyai.com/docs/models/speech-recognition#automatic-language-detection), either true or false. */
5285
- language_detection?: boolean;
5286
- /**
5287
- * The confidence threshold for the automatically detected language.
5288
- An error will be returned if the language confidence is below this threshold.
5289
- Defaults to 0.
5839
+ auto_chapters?: boolean;
5840
+ /** Enable [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases), either true or false */
5841
+ auto_highlights?: boolean;
5842
+ /** Enable [Content Moderation](https://www.assemblyai.com/docs/content-moderation), can be true or false */
5843
+ content_safety?: boolean;
5844
+ /** The confidence threshold for the [Content Moderation](https://www.assemblyai.com/docs/content-moderation) model. Values must be between 25 and 100. */
5845
+ content_safety_confidence?: number;
5846
+ /** Customize how words are spelled and formatted using to and from values. See [Custom Spelling](https://www.assemblyai.com/docs/pre-recorded-audio/custom-spelling) for more details. */
5847
+ custom_spelling?: TranscriptCustomSpelling[];
5848
+ /** Transcribe [Filler Words](https://www.assemblyai.com/docs/pre-recorded-audio/filler-words), like "umm", in your media file; can be true or false */
5849
+ disfluencies?: boolean;
5850
+ /** Enable domain-specific transcription models to improve accuracy for specialized terminology. Set to `"medical-v1"` to enable [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) for improved accuracy of medical terms such as medications, procedures, conditions, and dosages.
5290
5851
 
5291
- * @minimum 0
5292
- * @maximum 1
5852
+ Supported languages: English (`en`), Spanish (`es`), German (`de`), French (`fr`). If used with an unsupported language, the parameter is ignored and a warning is returned.
5853
+ */
5854
+ domain?: TranscriptParamsDomain;
5855
+ /** Enable [Entity Detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection), can be true or false */
5856
+ entity_detection?: boolean;
5857
+ /** Filter profanity from the transcribed text, can be true or false. See [Profanity Filtering](https://www.assemblyai.com/docs/profanity-filtering) for more details. */
5858
+ filter_profanity?: boolean;
5859
+ /** Enable [Text Formatting](https://www.assemblyai.com/docs/pre-recorded-audio), can be true or false */
5860
+ format_text?: boolean;
5861
+ /** Enable [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection), can be true or false */
5862
+ iab_categories?: boolean;
5863
+ /** Improve accuracy with up to 200 (for Universal-2) or 1000 (for Universal-3 Pro) domain-specific words or phrases (maximum 6 words per phrase). See [Keyterms Prompting](https://www.assemblyai.com/docs/pre-recorded-audio/keyterms-prompting) for more details.
5293
5864
  */
5865
+ keyterms_prompt?: string[];
5866
+ /** The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
5867
+ The default value is 'en_us'.
5868
+ */
5869
+ language_code?: TranscriptParamsLanguageCode;
5870
+ /** The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
5871
+ One of the values specified must be `en`.
5872
+ */
5873
+ language_codes?: TranscriptParamsLanguageCodes;
5874
+ /** The confidence threshold for the automatically detected language.
5875
+ An error will be returned if the language confidence is below this threshold.
5876
+ Defaults to 0. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
5877
+ */
5294
5878
  language_confidence_threshold?: number;
5295
- /** The speech model to use for the transcription. When `null`, the "best" model is used. */
5296
- speech_model?: TranscriptOptionalParamsSpeechModel;
5297
- /** Enable Automatic Punctuation, can be true or false */
5298
- punctuate?: boolean;
5299
- /** Enable Text Formatting, can be true or false */
5300
- format_text?: boolean;
5301
- /** Transcribe Filler Words, like "umm", in your media file; can be true or false */
5302
- disfluencies?: boolean;
5303
- /** Enable [Multichannel](https://www.assemblyai.com/docs/models/speech-recognition#multichannel-transcription) transcription, can be true or false. */
5879
+ /** Enable [Automatic language detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection), either true or false. */
5880
+ language_detection?: boolean;
5881
+ /** Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection). */
5882
+ language_detection_options?: TranscriptOptionalParamsLanguageDetectionOptions;
5883
+ /** Enable [Multichannel](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) transcription, can be true or false. */
5304
5884
  multichannel?: boolean;
5305
- /** The URL to which we send webhook requests.
5306
- We sends two different types of webhook requests.
5307
- One request when a transcript is completed or failed, and one request when the redacted audio is ready if redact_pii_audio is enabled.
5885
+ /** Provide natural language prompting of up to 1,500 words of contextual information to the model. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for best practices.
5886
+
5887
+ Note: This parameter is only supported for the Universal-3 Pro model.
5308
5888
  */
5309
- webhook_url?: string;
5310
- /** The header name to be sent with the transcript completed or failed webhook requests */
5311
- webhook_auth_header_name?: TranscriptOptionalParamsWebhookAuthHeaderName;
5312
- /** The header value to send back with the transcript completed or failed webhook requests for added security */
5313
- webhook_auth_header_value?: TranscriptOptionalParamsWebhookAuthHeaderValue;
5314
- /** Enable Key Phrases, either true or false */
5315
- auto_highlights?: boolean;
5316
- /** The point in time, in milliseconds, to begin transcribing in your media file */
5317
- audio_start_from?: number;
5318
- /** The point in time, in milliseconds, to stop transcribing in your media file */
5319
- audio_end_at?: number;
5320
- /**
5321
- * The list of custom vocabulary to boost transcription probability for
5322
- * @deprecated
5323
- */
5324
- word_boost?: string[];
5325
- /** How much to boost specified words */
5326
- boost_param?: TranscriptBoostParam;
5327
- /** Filter profanity from the transcribed text, can be true or false */
5328
- filter_profanity?: boolean;
5329
- /** Redact PII from the transcribed text using the Redact PII model, can be true or false */
5889
+ prompt?: string;
5890
+ /** Enable [Automatic Punctuation](https://www.assemblyai.com/docs/pre-recorded-audio), can be true or false */
5891
+ punctuate?: boolean;
5892
+ /** Redact PII from the transcribed text using the Redact PII model, can be true or false. See [PII Redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
5330
5893
  redact_pii?: boolean;
5331
- /** Generate a copy of the original media file with spoken PII "beeped" out, can be true or false. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5894
+ /** Generate a copy of the original media file with spoken PII "beeped" out, can be true or false. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more details. */
5332
5895
  redact_pii_audio?: boolean;
5333
- /** Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5896
+ /** Specify options for [PII redacted audio](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) files. */
5897
+ redact_pii_audio_options?: TranscriptOptionalParamsRedactPiiAudioOptions;
5898
+ /** Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more details. */
5334
5899
  redact_pii_audio_quality?: RedactPiiAudioQuality;
5335
- /** The list of PII Redaction policies to enable. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5900
+ /** The list of PII Redaction policies to enable. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
5336
5901
  redact_pii_policies?: PiiPolicy[];
5337
- /** The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5338
- redact_pii_sub?: TranscriptOptionalParamsRedactPiiSub;
5339
- /** Enable [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization), can be true or false */
5902
+ /** The replacement logic for detected PII, can be `entity_type` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
5903
+ redact_pii_sub?: TranscriptParamsRedactPiiSub;
5904
+ /** Enable [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis), can be true or false */
5905
+ sentiment_analysis?: boolean;
5906
+ /** Enable [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization), can be true or false */
5340
5907
  speaker_labels?: boolean;
5341
- /** Tells the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details. */
5342
- speakers_expected?: TranscriptOptionalParamsSpeakersExpected;
5343
- /** Enable [Content Moderation](https://www.assemblyai.com/docs/models/content-moderation), can be true or false */
5344
- content_safety?: boolean;
5345
- /**
5346
- * The confidence threshold for the Content Moderation model. Values must be between 25 and 100.
5347
- * @minimum 25
5348
- * @maximum 100
5349
- */
5350
- content_safety_confidence?: number;
5351
- /** Enable [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection), can be true or false */
5352
- iab_categories?: boolean;
5353
- /** Customize how words are spelled and formatted using to and from values */
5354
- custom_spelling?: TranscriptCustomSpelling[];
5355
- /** <Warning>`keyterms_prompt` is only supported when the `speech_model` is specified as `slam-1`</Warning>
5356
- Improve accuracy with up to 1000 domain-specific words or phrases (maximum 6 words per phrase).
5908
+ /** Specify options for [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers). Use this to set a range of possible speakers. */
5909
+ speaker_options?: TranscriptOptionalParamsSpeakerOptions;
5910
+ /** Tells the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details. */
5911
+ speakers_expected?: TranscriptParamsSpeakersExpected;
5912
+ /** List multiple speech models in priority order, allowing our system to automatically route your audio to the best available option. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models and routing behavior.
5913
+ */
5914
+ speech_models: SpeechModel[];
5915
+ /** Reject audio files that contain less than this fraction of speech.
5916
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
5357
5917
  */
5358
- keyterms_prompt?: string[];
5359
- /**
5360
- * This parameter does not currently have any functionality attached to it.
5361
- * @deprecated
5918
+ speech_threshold?: TranscriptParamsSpeechThreshold;
5919
+ /** Enable speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
5362
5920
  */
5363
- prompt?: string;
5364
- /** Enable [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis), can be true or false */
5365
- sentiment_analysis?: boolean;
5366
- /** Enable [Auto Chapters](https://www.assemblyai.com/docs/models/auto-chapters), can be true or false */
5367
- auto_chapters?: boolean;
5368
- /** Enable [Entity Detection](https://www.assemblyai.com/docs/models/entity-detection), can be true or false */
5369
- entity_detection?: boolean;
5370
- /**
5371
- * Reject audio files that contain less than this fraction of speech.
5372
- Valid values are in the range [0, 1] inclusive.
5921
+ speech_understanding?: TranscriptOptionalParamsSpeechUnderstanding;
5922
+ /** Enable [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization), can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
5373
5923
 
5374
- * @minimum 0
5375
- * @maximum 1
5376
- */
5377
- speech_threshold?: TranscriptOptionalParamsSpeechThreshold;
5378
- /** Enable [Summarization](https://www.assemblyai.com/docs/models/summarization), can be true or false */
5924
+ Note: This parameter is only supported for the Universal-2 model.
5925
+ */
5379
5926
  summarization?: boolean;
5380
- /** The model to summarize the transcript */
5927
+ /** The model to summarize the transcript. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
5381
5928
  summary_model?: SummaryModel;
5382
- /** The type of summary */
5929
+ /** The type of summary. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
5383
5930
  summary_type?: SummaryType;
5384
- /**
5385
- * Enable custom topics, either true or false
5386
- * @deprecated
5931
+ /** Remove [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) from the transcript text. Set to `"all"` to remove all audio tags.
5932
+
5933
+ Note: This parameter is only supported for the Universal-3 Pro model.
5934
+ */
5935
+ remove_audio_tags?: TranscriptParamsRemoveAudioTags;
5936
+ /** Control the amount of randomness injected into the model's response. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for more details.
5937
+
5938
+ Note: This parameter can only be used with the Universal-3 Pro model.
5939
+ */
5940
+ temperature?: number;
5941
+ /** The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests */
5942
+ webhook_auth_header_name?: TranscriptParamsWebhookAuthHeaderName;
5943
+ /** The header value to send back with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests for added security */
5944
+ webhook_auth_header_value?: TranscriptParamsWebhookAuthHeaderValue;
5945
+ /** The URL to which we send [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests.
5387
5946
  */
5947
+ webhook_url?: string;
5948
+ /** This parameter does not currently have any functionality attached to it. */
5388
5949
  custom_topics?: boolean;
5389
- /** The list of custom topics */
5950
+ /** This parameter has been replaced with the `speech_models` parameter, learn more about the `speech_models` parameter [here](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model).
5951
+ */
5952
+ speech_model?: TranscriptParamsSpeechModel;
5953
+ /** This parameter does not currently have any functionality attached to it. */
5390
5954
  topics?: string[];
5955
+ /** The URL of the audio or video file to transcribe. */
5956
+ audio_url: string;
5391
5957
  }
5392
5958
 
5393
5959
  /**
@@ -7204,75 +7770,15 @@ interface TranscribeStreamParams<P extends StreamingProvider> {
7204
7770
  * Generated by orval v7.9.0 🍺
7205
7771
  * Do not edit manually.
7206
7772
  * AssemblyAI API
7207
- * AssemblyAI API
7208
- * OpenAPI spec version: 1.3.4
7209
- */
7210
- /**
7211
- * Maximum amount of transcripts to retrieve
7212
- */
7213
- type Limit = number;
7214
-
7215
- /**
7216
- * Generated by orval v7.9.0 🍺
7217
- * Do not edit manually.
7218
- * AssemblyAI API
7219
- * AssemblyAI API
7220
- * OpenAPI spec version: 1.3.4
7221
- */
7222
- /**
7223
- * Only get transcripts created on this date
7224
- */
7225
- type CreatedOn = string;
7226
-
7227
- /**
7228
- * Generated by orval v7.9.0 🍺
7229
- * Do not edit manually.
7230
- * AssemblyAI API
7231
- * AssemblyAI API
7232
- * OpenAPI spec version: 1.3.4
7233
- */
7234
- /**
7235
- * Get transcripts that were created before this transcript ID
7236
- */
7237
- type BeforeId = string;
7238
-
7239
- /**
7240
- * Generated by orval v7.9.0 🍺
7241
- * Do not edit manually.
7242
- * AssemblyAI API
7243
- * AssemblyAI API
7244
- * OpenAPI spec version: 1.3.4
7245
- */
7246
- /**
7247
- * Get transcripts that were created after this transcript ID
7248
- */
7249
- type AfterId = string;
7250
-
7251
- /**
7252
- * Generated by orval v7.9.0 🍺
7253
- * Do not edit manually.
7254
- * AssemblyAI API
7255
- * AssemblyAI API
7256
- * OpenAPI spec version: 1.3.4
7257
- */
7258
- /**
7259
- * Only get throttled transcripts, overrides the status filter
7260
- */
7261
- type ThrottledOnly = boolean;
7262
-
7263
- /**
7264
- * Generated by orval v7.9.0 🍺
7265
- * Do not edit manually.
7266
- * AssemblyAI API
7267
- * AssemblyAI API
7268
- * OpenAPI spec version: 1.3.4
7773
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
7774
+ * OpenAPI spec version: 1.0.0
7269
7775
  */
7270
7776
 
7271
7777
  type ListTranscriptsParams = {
7272
7778
  /**
7273
7779
  * Maximum amount of transcripts to retrieve
7274
7780
  */
7275
- limit?: Limit;
7781
+ limit?: number;
7276
7782
  /**
7277
7783
  * Filter by transcript status
7278
7784
  */
@@ -7280,19 +7786,19 @@ type ListTranscriptsParams = {
7280
7786
  /**
7281
7787
  * Only get transcripts created on this date
7282
7788
  */
7283
- created_on?: CreatedOn;
7789
+ created_on?: string;
7284
7790
  /**
7285
7791
  * Get transcripts that were created before this transcript ID
7286
7792
  */
7287
- before_id?: BeforeId;
7793
+ before_id?: string;
7288
7794
  /**
7289
7795
  * Get transcripts that were created after this transcript ID
7290
7796
  */
7291
- after_id?: AfterId;
7797
+ after_id?: string;
7292
7798
  /**
7293
7799
  * Only get throttled transcripts, overrides the status filter
7294
7800
  */
7295
- throttled_only?: ThrottledOnly;
7801
+ throttled_only?: boolean;
7296
7802
  };
7297
7803
 
7298
7804
  /**
@@ -7475,6 +7981,18 @@ type ManageV1FilterRequestIdParameter = string;
7475
7981
  /**
7476
7982
  * ManageV1FilterDeploymentParameter type definition
7477
7983
  */
7984
+ /**
7985
+ * ManageV1FilterDeploymentParameter type definition
7986
+ */
7987
+ /**
7988
+ * ManageV1FilterDeploymentParameter type definition
7989
+ */
7990
+ /**
7991
+ * ManageV1FilterDeploymentParameter type definition
7992
+ */
7993
+ /**
7994
+ * ManageV1FilterDeploymentParameter type definition
7995
+ */
7478
7996
  type ManageV1FilterDeploymentParameter = typeof ManageV1FilterDeploymentParameter[keyof typeof ManageV1FilterDeploymentParameter];
7479
7997
  declare const ManageV1FilterDeploymentParameter: {
7480
7998
  readonly hosted: "hosted";
@@ -7513,6 +8031,18 @@ declare const ManageV1FilterDeploymentParameter: {
7513
8031
  /**
7514
8032
  * ManageV1FilterEndpointParameter type definition
7515
8033
  */
8034
+ /**
8035
+ * ManageV1FilterEndpointParameter type definition
8036
+ */
8037
+ /**
8038
+ * ManageV1FilterEndpointParameter type definition
8039
+ */
8040
+ /**
8041
+ * ManageV1FilterEndpointParameter type definition
8042
+ */
8043
+ /**
8044
+ * ManageV1FilterEndpointParameter type definition
8045
+ */
7516
8046
  type ManageV1FilterEndpointParameter = typeof ManageV1FilterEndpointParameter[keyof typeof ManageV1FilterEndpointParameter];
7517
8047
  declare const ManageV1FilterEndpointParameter: {
7518
8048
  readonly listen: "listen";
@@ -7556,6 +8086,18 @@ declare const ManageV1FilterEndpointParameter: {
7556
8086
  /**
7557
8087
  * ManageV1FilterMethodParameter type definition
7558
8088
  */
8089
+ /**
8090
+ * ManageV1FilterMethodParameter type definition
8091
+ */
8092
+ /**
8093
+ * ManageV1FilterMethodParameter type definition
8094
+ */
8095
+ /**
8096
+ * ManageV1FilterMethodParameter type definition
8097
+ */
8098
+ /**
8099
+ * ManageV1FilterMethodParameter type definition
8100
+ */
7559
8101
  type ManageV1FilterMethodParameter = typeof ManageV1FilterMethodParameter[keyof typeof ManageV1FilterMethodParameter];
7560
8102
  declare const ManageV1FilterMethodParameter: {
7561
8103
  readonly sync: "sync";
@@ -7595,6 +8137,18 @@ declare const ManageV1FilterMethodParameter: {
7595
8137
  /**
7596
8138
  * ManageV1FilterStatusParameter type definition
7597
8139
  */
8140
+ /**
8141
+ * ManageV1FilterStatusParameter type definition
8142
+ */
8143
+ /**
8144
+ * ManageV1FilterStatusParameter type definition
8145
+ */
8146
+ /**
8147
+ * ManageV1FilterStatusParameter type definition
8148
+ */
8149
+ /**
8150
+ * ManageV1FilterStatusParameter type definition
8151
+ */
7598
8152
  type ManageV1FilterStatusParameter = typeof ManageV1FilterStatusParameter[keyof typeof ManageV1FilterStatusParameter];
7599
8153
  declare const ManageV1FilterStatusParameter: {
7600
8154
  readonly succeeded: "succeeded";
@@ -7939,7 +8493,7 @@ interface TranscribeOptions {
7939
8493
  * AssemblyAI-specific options (passed directly to API)
7940
8494
  * @see https://www.assemblyai.com/docs/api-reference/transcripts/submit
7941
8495
  */
7942
- assemblyai?: Partial<TranscriptOptionalParams>;
8496
+ assemblyai?: Partial<TranscriptParams>;
7943
8497
  /**
7944
8498
  * Gladia-specific options (passed directly to API)
7945
8499
  * @see https://docs.gladia.io/
@@ -8884,96 +9438,6 @@ interface CallbackTranscriptionSuccessPayload {
8884
9438
  custom_metadata?: CallbackTranscriptionSuccessPayloadCustomMetadata;
8885
9439
  }
8886
9440
 
8887
- /**
8888
- * Generated by orval v7.9.0 🍺
8889
- * Do not edit manually.
8890
- * AssemblyAI API
8891
- * AssemblyAI API
8892
- * OpenAPI spec version: 1.3.4
8893
- */
8894
- /**
8895
- * The status of the redacted audio
8896
- */
8897
- type RedactedAudioStatus = (typeof RedactedAudioStatus)[keyof typeof RedactedAudioStatus];
8898
- declare const RedactedAudioStatus: {
8899
- readonly redacted_audio_ready: "redacted_audio_ready";
8900
- };
8901
-
8902
- /**
8903
- * Generated by orval v7.9.0 🍺
8904
- * Do not edit manually.
8905
- * AssemblyAI API
8906
- * AssemblyAI API
8907
- * OpenAPI spec version: 1.3.4
8908
- */
8909
-
8910
- interface RedactedAudioResponse {
8911
- /** The status of the redacted audio */
8912
- status: RedactedAudioStatus;
8913
- /** The URL of the redacted audio file */
8914
- redacted_audio_url: string;
8915
- }
8916
-
8917
- /**
8918
- * Generated by orval v7.9.0 🍺
8919
- * Do not edit manually.
8920
- * AssemblyAI API
8921
- * AssemblyAI API
8922
- * OpenAPI spec version: 1.3.4
8923
- */
8924
-
8925
- /**
8926
- * The notification when the redacted audio is ready.
8927
- */
8928
- type RedactedAudioNotification = RedactedAudioResponse;
8929
-
8930
- /**
8931
- * Generated by orval v7.9.0 🍺
8932
- * Do not edit manually.
8933
- * AssemblyAI API
8934
- * AssemblyAI API
8935
- * OpenAPI spec version: 1.3.4
8936
- */
8937
- /**
8938
- * The status of the transcript. Either completed or error.
8939
- */
8940
- type TranscriptReadyStatus = (typeof TranscriptReadyStatus)[keyof typeof TranscriptReadyStatus];
8941
- declare const TranscriptReadyStatus: {
8942
- readonly completed: "completed";
8943
- readonly error: "error";
8944
- };
8945
-
8946
- /**
8947
- * Generated by orval v7.9.0 🍺
8948
- * Do not edit manually.
8949
- * AssemblyAI API
8950
- * AssemblyAI API
8951
- * OpenAPI spec version: 1.3.4
8952
- */
8953
-
8954
- /**
8955
- * The notification when the transcript status is completed or error.
8956
- */
8957
- interface TranscriptReadyNotification {
8958
- /** The ID of the transcript */
8959
- transcript_id: string;
8960
- /** The status of the transcript. Either completed or error. */
8961
- status: TranscriptReadyStatus;
8962
- }
8963
-
8964
- /**
8965
- * Generated by orval v7.9.0 🍺
8966
- * Do not edit manually.
8967
- * AssemblyAI API
8968
- * AssemblyAI API
8969
- * OpenAPI spec version: 1.3.4
8970
- */
8971
-
8972
- /**
8973
- * The notifications sent to the webhook URL.
8974
- */
8975
- type TranscriptWebhookNotification = TranscriptReadyNotification | RedactedAudioNotification;
8976
-
8977
9441
  /**
8978
9442
  * Generated by orval v7.9.0 🍺
8979
9443
  * Do not edit manually.
@@ -10350,4 +10814,4 @@ interface SpeechToTextChunkResponseModel {
10350
10814
  entities?: SpeechToTextChunkResponseModelEntities;
10351
10815
  }
10352
10816
 
10353
- export { type CustomSpellingConfigDTO as $, type AssemblyAIStreamingOptions as A, type AudioToLlmListDTOError as B, type CallbackConfig as C, type DeepgramStreamingOptions as D, type AudioToLlmResultDTO as E, type FileResponse as F, type GladiaStreamingOptions as G, type CallbackConfigDto as H, CallbackMethodEnum as I, type CallbackTranscriptionErrorPayload as J, type CallbackTranscriptionErrorPayloadCustomMetadata as K, type ListTranscriptsOptions as L, type MessagesConfig as M, type NamedEntityRecognitionDTO as N, CallbackTranscriptionErrorPayloadEvent as O, type PreProcessingConfig as P, type CallbackTranscriptionSuccessPayload as Q, type RealtimeProcessingConfig as R, type StreamingOptions as S, type TranscribeOptions as T, type UnifiedTranscriptResponse as U, type CallbackTranscriptionSuccessPayloadCustomMetadata as V, type WordDTO as W, CallbackTranscriptionSuccessPayloadEvent as X, type ChapterizationDTOError as Y, type ChapterizationDTOResults as Z, type CodeSwitchingConfigDTO as _, type StreamingCallbacks as a, type Entity as a$, type CustomSpellingConfigDTOSpellingDictionary as a0, type CustomVocabularyConfigDTO as a1, type CustomVocabularyConfigDTOVocabularyItem as a2, type CustomVocabularyEntryDTO as a3, type DiarizationConfigDTO as a4, type DiarizationDTO as a5, type DiarizationDTOError as a6, type DisplayModeDTO as a7, type DisplayModeDTOError as a8, type ErrorDTO as a9, type SubtitlesConfigDTO as aA, SubtitlesFormatEnum as aB, SubtitlesStyleEnum as aC, type SummarizationConfigDTO as aD, type SummarizationDTOError as aE, SummaryTypesEnum as aF, TranscriptionControllerListV2KindItem as aG, type TranscriptionControllerListV2Params as aH, TranscriptionControllerListV2StatusItem as aI, type TranscriptionResultDTO as aJ, type TranslationConfigDTO as aK, type TranslationDTOError as aL, TranslationModelEnum as aM, type TranslationResultDTO as aN, type TranslationResultDTOError as aO, type TranscriptOptionalParams as aP, TranscriptStatus as aQ, type TranscriptWord as aR, AudioIntelligenceModelStatus as aS, type AutoHighlightResult as aT, type AutoHighlightsResult as aU, type Chapter$1 as aV, type ContentSafetyLabel as aW, type ContentSafetyLabelResult as aX, type ContentSafetyLabelsResult as aY, type ContentSafetyLabelsResultSeverityScoreSummary as aZ, type ContentSafetyLabelsResultSummary as a_, type InitTranscriptionRequest as aa, type InitTranscriptionRequestCustomMetadata as ab, type ModerationDTO as ac, type ModerationDTOError as ad, type NamedEntityRecognitionDTOError as ae, type NamesConsistencyDTO as af, type NamesConsistencyDTOError as ag, type PreRecordedRequestParamsResponse as ah, type PreRecordedResponseCustomMetadata as ai, type PreRecordedResponseFile as aj, PreRecordedResponseKind as ak, type PreRecordedResponsePostSessionMetadata as al, type PreRecordedResponseRequestParams as am, type PreRecordedResponseResult as an, PreRecordedResponseStatus as ao, type SentencesDTO as ap, type SentencesDTOError as aq, type SentimentAnalysisDTOError as ar, type SpeakerReidentificationDTO as as, type SpeakerReidentificationDTOError as at, type StreamingRequest as au, type StreamingRequestCustomMetadata as av, type StructuredDataExtractionConfigDTO as aw, type StructuredDataExtractionDTO as ax, type StructuredDataExtractionDTOError as ay, type SubtitleDTO as az, type StreamingSession as b, type TranscriptSentimentAnalysis as b$, EntityType as b0, type ListTranscriptsParams as b1, PiiPolicy as b2, RedactPiiAudioQuality as b3, type RedactedAudioNotification as b4, type RedactedAudioResponse as b5, RedactedAudioStatus as b6, Sentiment as b7, type SentimentAnalysisResult$1 as b8, type SentimentAnalysisResultChannel as b9, type TranscriptDisfluencies as bA, type TranscriptEntities as bB, type TranscriptEntityDetection as bC, type TranscriptFilterProfanity as bD, type TranscriptFormatText as bE, type TranscriptIabCategories as bF, type TranscriptIabCategoriesResult as bG, TranscriptLanguageCode as bH, type TranscriptLanguageCodeProperty as bI, type TranscriptLanguageConfidence as bJ, type TranscriptLanguageConfidenceThreshold as bK, type TranscriptLanguageDetection as bL, type TranscriptMultichannel as bM, type TranscriptOptionalParamsLanguageCode as bN, type TranscriptOptionalParamsLanguageCodeOneOf as bO, type TranscriptOptionalParamsRedactPiiSub as bP, type TranscriptOptionalParamsSpeakersExpected as bQ, type TranscriptOptionalParamsSpeechModel as bR, type TranscriptOptionalParamsSpeechThreshold as bS, type TranscriptOptionalParamsWebhookAuthHeaderName as bT, type TranscriptOptionalParamsWebhookAuthHeaderValue as bU, type TranscriptPunctuate as bV, type TranscriptReadyNotification as bW, TranscriptReadyStatus as bX, type TranscriptRedactPiiAudio as bY, type TranscriptRedactPiiAudioQuality as bZ, type TranscriptRedactPiiPolicies as b_, type SentimentAnalysisResultSpeaker as ba, type SeverityScoreSummary as bb, SpeechModel as bc, SubstitutionPolicy as bd, SummaryModel as be, SummaryType as bf, type Timestamp as bg, type TopicDetectionModelResult as bh, type TopicDetectionModelResultSummary as bi, type TopicDetectionResult$1 as bj, type TopicDetectionResultLabelsItem as bk, type Transcript as bl, type TranscriptAudioDuration as bm, type TranscriptAudioEndAt as bn, type TranscriptAudioStartFrom as bo, type TranscriptAutoChapters as bp, type TranscriptAutoHighlightsResult as bq, TranscriptBoostParam as br, type TranscriptBoostParamProperty as bs, type TranscriptChapters as bt, type TranscriptConfidence as bu, type TranscriptContentSafety as bv, type TranscriptContentSafetyLabels as bw, type TranscriptCustomSpelling as bx, type TranscriptCustomSpellingProperty as by, type TranscriptCustomTopics as bz, type StreamEvent as c, type TranscriptionWord as c$, type TranscriptSentimentAnalysisResults as c0, type TranscriptSpeakerLabels as c1, type TranscriptSpeakersExpected as c2, type TranscriptSpeechModel as c3, type TranscriptSpeechThreshold as c4, type TranscriptSpeedBoost as c5, type TranscriptSummary as c6, type TranscriptSummaryModel as c7, type TranscriptSummaryType as c8, type TranscriptText as c9, type TranscriptTextUsageTokens as cA, type TranscriptionSegment as cB, type RealtimeSessionCreateRequestGAModel as cC, RealtimeTranscriptionSessionCreateRequestTurnDetectionType as cD, RealtimeTranscriptionSessionCreateRequestInputAudioFormat as cE, AudioResponseFormat as cF, type CreateTranscription200One as cG, type CreateTranscriptionRequest as cH, type CreateTranscriptionRequestModel as cI, type CreateTranscriptionRequestStream as cJ, CreateTranscriptionRequestTimestampGranularitiesItem as cK, type CreateTranscriptionResponseDiarizedJson as cL, CreateTranscriptionResponseDiarizedJsonTask as cM, type CreateTranscriptionResponseDiarizedJsonUsage as cN, type CreateTranscriptionResponseJson as cO, type CreateTranscriptionResponseJsonLogprobsItem as cP, type CreateTranscriptionResponseJsonUsage as cQ, type CreateTranscriptionResponseVerboseJson as cR, type TranscriptTextUsageDuration as cS, TranscriptTextUsageDurationType as cT, type TranscriptTextUsageTokensInputTokenDetails as cU, TranscriptTextUsageTokensType as cV, type TranscriptionChunkingStrategy as cW, type TranscriptionChunkingStrategyAnyOf as cX, type TranscriptionDiarizedSegment as cY, TranscriptionDiarizedSegmentType as cZ, TranscriptionInclude as c_, type TranscriptThrottled as ca, type TranscriptUtterance as cb, type TranscriptUtteranceChannel as cc, type TranscriptUtterances as cd, type TranscriptWebhookAuthHeaderName as ce, type TranscriptWebhookNotification as cf, type TranscriptWebhookStatusCode as cg, type TranscriptWebhookUrl as ch, type TranscriptWordChannel as ci, type TranscriptWordSpeaker as cj, type TranscriptWords as ck, type StreamingUpdateConfiguration as cl, type Transcription as cm, type EntityError as cn, Status as co, type EntityReference as cp, type DiarizationProperties as cq, type DiarizationSpeakersProperties as cr, LanguageIdentificationMode as cs, type LanguageIdentificationProperties as ct, type LanguageIdentificationPropertiesSpeechModelMapping as cu, ProfanityFilterMode as cv, PunctuationMode as cw, type TranscriptionCustomProperties as cx, type TranscriptionLinks as cy, type TranscriptionProperties as cz, StreamingSupportedEncodingEnum as d, type ListenV1ResponseResultsChannelsItemSearchItemHitsItem as d$, type VadConfig as d0, VadConfigType as d1, type ListenV1Response as d2, type ManageV1FilterAccessorParameter as d3, ManageV1FilterDeploymentParameter as d4, type ManageV1LimitParameter as d5, type ManageV1PageParameter as d6, ManageV1FilterEndpointParameter as d7, ManageV1FilterMethodParameter as d8, type SharedTopics as d9, type ListenV1ModelParameter as dA, type ListenV1MultichannelParameter as dB, type ListenV1NumeralsParameter as dC, type ListenV1ParagraphsParameter as dD, type ListenV1ProfanityFilterParameter as dE, type ListenV1PunctuateParameter as dF, type ListenV1RedactParameter as dG, ListenV1RedactParameterOneOfItem as dH, type ListenV1ReplaceParameter as dI, type ListenV1ResponseMetadata as dJ, type ListenV1ResponseMetadataIntentsInfo as dK, type ListenV1ResponseMetadataModelInfo as dL, type ListenV1ResponseMetadataSentimentInfo as dM, type ListenV1ResponseMetadataSummaryInfo as dN, type ListenV1ResponseMetadataTopicsInfo as dO, type ListenV1ResponseResults as dP, type ListenV1ResponseResultsChannels as dQ, type ListenV1ResponseResultsChannelsItem as dR, type ListenV1ResponseResultsChannelsItemAlternativesItem as dS, type ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem as dT, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs as dU, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem as dV, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem as dW, type ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem as dX, type ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem as dY, type ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem as dZ, type ListenV1ResponseResultsChannelsItemSearchItem as d_, type SharedIntents as da, type SharedSentiments as db, type SharedCallbackParameter as dc, SharedCallbackMethodParameter as dd, type SharedSentimentParameter as de, type SharedSummarizeParameter as df, type SharedTagParameter as dg, type SharedTopicsParameter as dh, type SharedCustomTopicParameter as di, SharedCustomTopicModeParameter as dj, type SharedIntentsParameter as dk, type SharedCustomIntentParameter as dl, SharedCustomIntentModeParameter as dm, type SharedMipOptOutParameter as dn, type ListenV1DetectEntitiesParameter as dp, type ListenV1DetectLanguageParameter as dq, type ListenV1DiarizeParameter as dr, type ListenV1DictationParameter as ds, ListenV1EncodingParameter as dt, type ListenV1FillerWordsParameter as du, type ListenV1KeytermParameter as dv, type ListenV1KeywordsParameter as dw, type ListenV1LanguageParameter as dx, type ListenV1MeasurementsParameter as dy, type ListenV1MediaTranscribeParams as dz, StreamingSupportedBitDepthEnum as e, type SentimentSpeakerSummary as e$, type ListenV1ResponseResultsSummary as e0, type ListenV1ResponseResultsUtterances as e1, type ListenV1ResponseResultsUtterancesItem as e2, type ListenV1ResponseResultsUtterancesItemWordsItem as e3, type ListenV1SearchParameter as e4, type ListenV1SmartFormatParameter as e5, type ListenV1UttSplitParameter as e6, type ListenV1UtterancesParameter as e7, type ListenV1VersionParameter as e8, type ManageV1EndDateTimeParameter as e9, AutoChaptersResultErrorType as eA, type Chapter as eB, type JobInfo as eC, type LanguageIdentificationResult as eD, type LanguageIdentificationResultAlternative as eE, LanguageIdentificationResultError as eF, type LanguageIdentificationResultItem as eG, type LanguagePackInfo as eH, LanguagePackInfoWritingDirection as eI, type OutputConfigSrtOverrides as eJ, type RecognitionAlternative as eK, type RecognitionDisplay as eL, RecognitionDisplayDirection as eM, type RecognitionMetadata as eN, type RecognitionResult as eO, RecognitionResultAttachesTo as eP, RecognitionResultType as eQ, type RetrieveTranscriptResponse as eR, type RetrieveTranscriptResponseAudioEventSummary as eS, type RetrieveTranscriptResponseAudioEventSummaryChannels as eT, type RetrieveTranscriptResponseTranslations as eU, type SentimentAnalysisError as eV, SentimentAnalysisErrorType as eW, type SentimentAnalysisResult as eX, type SentimentAnalysisResultSentimentAnalysis as eY, type SentimentChannelSummary as eZ, type SentimentSegment as e_, type ManageV1FilterRequestIdParameter as ea, ManageV1FilterStatusParameter as eb, type ManageV1ProjectsRequestsListParams as ec, type ManageV1StartDateTimeParameter as ed, type SharedExtraParameter as ee, type SharedIntentsResults as ef, type SharedIntentsResultsIntents as eg, type SharedIntentsResultsIntentsSegmentsItem as eh, type SharedIntentsResultsIntentsSegmentsItemIntentsItem as ei, type SharedSentimentsAverage as ej, type SharedSentimentsSegmentsItem as ek, type SharedTopicsResults as el, type SharedTopicsResultsTopics as em, type SharedTopicsResultsTopicsSegmentsItem as en, type SharedTopicsResultsTopicsSegmentsItemTopicsItem as eo, JobType as ep, type AlignmentConfig as eq, type TranscriptionConfig as er, type TrackingData as es, type OutputConfig as et, OperatingPoint as eu, type AudioEventItem as ev, type AudioEventSummary as ew, type AudioEventSummaryItem as ex, type AutoChaptersResult as ey, type AutoChaptersResultError as ez, StreamingSupportedSampleRateEnum as f, SegmentedJsonExportOptionsFormat as f$, type SentimentSummary as f0, type SentimentSummaryDetail as f1, type SpokenFormRecognitionResult as f2, SpokenFormRecognitionResultType as f3, type SummarizationError as f4, SummarizationErrorType as f5, type SummarizationResult as f6, type TopicDetectionError as f7, TopicDetectionErrorType as f8, type TopicDetectionResult as f9, type BodySpeechToTextV1SpeechToTextPostLanguageCode as fA, BodySpeechToTextV1SpeechToTextPostModelId as fB, type BodySpeechToTextV1SpeechToTextPostNumSpeakers as fC, type BodySpeechToTextV1SpeechToTextPostSeed as fD, type BodySpeechToTextV1SpeechToTextPostTemperature as fE, BodySpeechToTextV1SpeechToTextPostTimestampsGranularity as fF, type BodySpeechToTextV1SpeechToTextPostWebhookId as fG, type BodySpeechToTextV1SpeechToTextPostWebhookMetadata as fH, type BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf as fI, type DetectedEntity as fJ, type DocxExportOptions as fK, DocxExportOptionsFormat as fL, type DocxExportOptionsMaxSegmentChars as fM, type DocxExportOptionsMaxSegmentDurationS as fN, type DocxExportOptionsSegmentOnSilenceLongerThanS as fO, type ExportOptions as fP, type HtmlExportOptions as fQ, HtmlExportOptionsFormat as fR, type HtmlExportOptionsMaxSegmentChars as fS, type HtmlExportOptionsMaxSegmentDurationS as fT, type HtmlExportOptionsSegmentOnSilenceLongerThanS as fU, type PdfExportOptions as fV, PdfExportOptionsFormat as fW, type PdfExportOptionsMaxSegmentChars as fX, type PdfExportOptionsMaxSegmentDurationS as fY, type PdfExportOptionsSegmentOnSilenceLongerThanS as fZ, type SegmentedJsonExportOptions as f_, type TopicDetectionSegment as fa, type TopicDetectionSegmentTopic as fb, type TopicDetectionSummary as fc, type TopicDetectionSummaryOverall as fd, type TrackingDataDetails as fe, type TranscriptionConfigAdditionalVocabItem as ff, TranscriptionConfigDiarization as fg, TranscriptionConfigMaxDelayMode as fh, type TranscriptionConfigPunctuationOverrides as fi, type TranscriptionConfigSpeakerDiarizationConfig as fj, type TranscriptionConfigTranscriptFilteringConfig as fk, type TranscriptionConfigTranscriptFilteringConfigReplacementsItem as fl, type TranslationError as fm, TranslationErrorType as fn, type TranslationSentence as fo, type WrittenFormRecognitionResult as fp, WrittenFormRecognitionResultType as fq, type SpeechToTextChunkResponseModel as fr, type AdditionalFormatResponseModel as fs, type AdditionalFormats as ft, type BodySpeechToTextV1SpeechToTextPost as fu, type BodySpeechToTextV1SpeechToTextPostCloudStorageUrl as fv, type BodySpeechToTextV1SpeechToTextPostDiarizationThreshold as fw, type BodySpeechToTextV1SpeechToTextPostEntityDetection as fx, type BodySpeechToTextV1SpeechToTextPostFile as fy, BodySpeechToTextV1SpeechToTextPostFileFormat as fz, StreamingSupportedModels as g, type ProviderStreamingOptions as g$, type SegmentedJsonExportOptionsMaxSegmentChars as g0, type SegmentedJsonExportOptionsMaxSegmentDurationS as g1, type SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS as g2, type SpeechToTextCharacterResponseModel as g3, type SpeechToTextCharacterResponseModelEnd as g4, type SpeechToTextCharacterResponseModelStart as g5, type SpeechToTextChunkResponseModelAdditionalFormats as g6, type SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem as g7, type SpeechToTextChunkResponseModelChannelIndex as g8, type SpeechToTextChunkResponseModelEntities as g9, type ProviderExtendedDataMap as gA, type StreamingProvider as gB, type BatchOnlyProvider as gC, type SessionStatus as gD, type Speaker as gE, type Word as gF, type Utterance as gG, type TranscriptionStatus as gH, type TranscriptMetadata as gI, type TranscriptData as gJ, type ListTranscriptsResponse as gK, type ProviderRawResponseMap as gL, type StreamEventType as gM, type SpeechEvent as gN, type TranslationEvent as gO, type SentimentEvent as gP, type EntityEvent as gQ, type SummarizationEvent as gR, type ChapterizationEvent as gS, type AudioAckEvent as gT, type LifecycleEvent as gU, type AudioChunk as gV, type RawWebSocketMessage as gW, type AssemblyAIUpdateConfiguration as gX, type OpenAIStreamingOptions as gY, type SonioxStreamingOptions as gZ, type ElevenLabsStreamingOptions as g_, type SpeechToTextChunkResponseModelTranscriptionId as ga, type SpeechToTextWordResponseModel as gb, type SpeechToTextWordResponseModelCharacters as gc, type SpeechToTextWordResponseModelEnd as gd, type SpeechToTextWordResponseModelSpeakerId as ge, type SpeechToTextWordResponseModelStart as gf, SpeechToTextWordResponseModelType as gg, type SrtExportOptions as gh, SrtExportOptionsFormat as gi, type SrtExportOptionsMaxCharactersPerLine as gj, type SrtExportOptionsMaxSegmentChars as gk, type SrtExportOptionsMaxSegmentDurationS as gl, type SrtExportOptionsSegmentOnSilenceLongerThanS as gm, type TxtExportOptions as gn, TxtExportOptionsFormat as go, type TxtExportOptionsMaxCharactersPerLine as gp, type TxtExportOptionsMaxSegmentChars as gq, type TxtExportOptionsMaxSegmentDurationS as gr, type TxtExportOptionsSegmentOnSilenceLongerThanS as gs, type SpeechmaticsOperatingPoint as gt, type TranscriptionModel as gu, type TranscriptionLanguage as gv, type AssemblyAIExtendedData as gw, type GladiaExtendedData as gx, type DeepgramExtendedData as gy, type ElevenLabsExtendedData as gz, type LanguageConfig as h, type StreamingOptionsForProvider as h0, type TranscribeStreamParams as h1, type BeginEvent as h2, type TurnEvent as h3, type TerminationEvent as h4, type ErrorEvent as h5, type StreamingEventMessage as h6, type StreamingWord as h7, type StreamingForceEndpoint as h8, type PostProcessingConfig as i, type TranscriptionMetadataDTO as j, type TranscriptionDTO as k, type TranslationDTO as l, type SummarizationDTO as m, type SentimentAnalysisDTO as n, type ChapterizationDTO as o, type PreRecordedResponse as p, type UtteranceDTO as q, TranscriptionLanguageCodeEnum as r, TranslationLanguageCodeEnum as s, StreamingSupportedRegions as t, type AddonErrorDTO as u, type AudioToLlmDTO as v, type AudioToLlmDTOError as w, type AudioToLlmDTOResults as x, type AudioToLlmListConfigDTO as y, type AudioToLlmListDTO as z };
10817
+ export { type CustomSpellingConfigDTO as $, type AssemblyAIStreamingOptions as A, type AudioToLlmListDTOError as B, type CallbackConfig as C, type DeepgramStreamingOptions as D, type AudioToLlmResultDTO as E, type FileResponse as F, type GladiaStreamingOptions as G, type CallbackConfigDto as H, CallbackMethodEnum as I, type CallbackTranscriptionErrorPayload as J, type CallbackTranscriptionErrorPayloadCustomMetadata as K, type ListTranscriptsOptions as L, type MessagesConfig as M, type NamedEntityRecognitionDTO as N, CallbackTranscriptionErrorPayloadEvent as O, type PreProcessingConfig as P, type CallbackTranscriptionSuccessPayload as Q, type RealtimeProcessingConfig as R, type StreamingOptions as S, type TranscribeOptions as T, type UnifiedTranscriptResponse as U, type CallbackTranscriptionSuccessPayloadCustomMetadata as V, type WordDTO as W, CallbackTranscriptionSuccessPayloadEvent as X, type ChapterizationDTOError as Y, type ChapterizationDTOResults as Z, type CodeSwitchingConfigDTO as _, type StreamingCallbacks as a, type CustomFormattingRequestBodyCustomFormatting as a$, type CustomSpellingConfigDTOSpellingDictionary as a0, type CustomVocabularyConfigDTO as a1, type CustomVocabularyConfigDTOVocabularyItem as a2, type CustomVocabularyEntryDTO as a3, type DiarizationConfigDTO as a4, type DiarizationDTO as a5, type DiarizationDTOError as a6, type DisplayModeDTO as a7, type DisplayModeDTOError as a8, type ErrorDTO as a9, type SubtitlesConfigDTO as aA, SubtitlesFormatEnum as aB, SubtitlesStyleEnum as aC, type SummarizationConfigDTO as aD, type SummarizationDTOError as aE, SummaryTypesEnum as aF, TranscriptionControllerListV2KindItem as aG, type TranscriptionControllerListV2Params as aH, TranscriptionControllerListV2StatusItem as aI, type TranscriptionResultDTO as aJ, type TranslationConfigDTO as aK, type TranslationDTOError as aL, TranslationModelEnum as aM, type TranslationResultDTO as aN, type TranslationResultDTOError as aO, TranscriptStatus as aP, type TranscriptWord as aQ, AudioIntelligenceModelStatus as aR, type AutoHighlightResult as aS, type AutoHighlightsResult as aT, type Chapter$1 as aU, type ContentSafetyLabel as aV, type ContentSafetyLabelResult as aW, type ContentSafetyLabelsResult as aX, type ContentSafetyLabelsResultSeverityScoreSummary as aY, type ContentSafetyLabelsResultSummary as aZ, type CustomFormattingRequestBody as a_, type InitTranscriptionRequest as aa, type InitTranscriptionRequestCustomMetadata as ab, type ModerationDTO as ac, type ModerationDTOError as ad, type NamedEntityRecognitionDTOError as ae, type NamesConsistencyDTO as af, type NamesConsistencyDTOError as ag, type PreRecordedRequestParamsResponse as ah, type PreRecordedResponseCustomMetadata as ai, type PreRecordedResponseFile as aj, PreRecordedResponseKind as ak, type PreRecordedResponsePostSessionMetadata as al, type PreRecordedResponseRequestParams as am, type PreRecordedResponseResult as an, PreRecordedResponseStatus as ao, type SentencesDTO as ap, type SentencesDTOError as aq, type SentimentAnalysisDTOError as ar, type SpeakerReidentificationDTO as as, type SpeakerReidentificationDTOError as at, type StreamingRequest as au, type StreamingRequestCustomMetadata as av, type StructuredDataExtractionConfigDTO as aw, type StructuredDataExtractionDTO as ax, type StructuredDataExtractionDTOError as ay, type SubtitleDTO as az, type StreamingSession as b, type TranscriptOptionalParamsSpeechUnderstandingRequest as b$, type CustomFormattingResponse as b0, type CustomFormattingResponseCustomFormatting as b1, type CustomFormattingResponseCustomFormattingMapping as b2, type Entity as b3, EntityType as b4, type ListTranscriptsParams as b5, PiiPolicy as b6, RedactPiiAudioQuality as b7, Sentiment as b8, type SentimentAnalysisResult$1 as b9, type TranscriptConfidence as bA, type TranscriptContentSafety as bB, type TranscriptContentSafetyLabels as bC, type TranscriptCustomSpelling as bD, type TranscriptCustomSpellingProperty as bE, type TranscriptCustomTopics as bF, type TranscriptDisfluencies as bG, type TranscriptDomain as bH, type TranscriptEntities as bI, type TranscriptEntityDetection as bJ, type TranscriptFilterProfanity as bK, type TranscriptFormatText as bL, type TranscriptIabCategories as bM, type TranscriptIabCategoriesResult as bN, TranscriptLanguageCode as bO, type TranscriptLanguageCodes as bP, type TranscriptLanguageConfidence as bQ, type TranscriptLanguageConfidenceThreshold as bR, type TranscriptLanguageDetection as bS, type TranscriptLanguageDetectionOptions as bT, type TranscriptMultichannel as bU, type TranscriptOptionalParamsLanguageDetectionOptions as bV, type TranscriptOptionalParamsRedactPiiAudioOptions as bW, TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod as bX, TranscriptOptionalParamsRemoveAudioTags as bY, type TranscriptOptionalParamsSpeakerOptions as bZ, type TranscriptOptionalParamsSpeechUnderstanding as b_, type SentimentAnalysisResultChannel as ba, type SentimentAnalysisResultSpeaker as bb, type SeverityScoreSummary as bc, type SpeakerIdentificationRequestBody as bd, type SpeakerIdentificationRequestBodySpeakerIdentification as be, SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType as bf, type SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems as bg, type SpeakerIdentificationResponse as bh, type SpeakerIdentificationResponseSpeakerIdentification as bi, type SpeakerIdentificationResponseSpeakerIdentificationMapping as bj, type SpeechModel as bk, SubstitutionPolicy as bl, SummaryModel as bm, SummaryType as bn, type Timestamp as bo, type TopicDetectionModelResult as bp, type TopicDetectionModelResultSummary as bq, type TopicDetectionResult$1 as br, type TopicDetectionResultLabelsItems as bs, type Transcript as bt, type TranscriptAudioDuration as bu, type TranscriptAudioEndAt as bv, type TranscriptAudioStartFrom as bw, type TranscriptAutoChapters as bx, type TranscriptAutoHighlightsResult as by, type TranscriptChapters as bz, type StreamEvent as c, type TranscriptionCustomProperties as c$, type TranscriptParams as c0, type TranscriptParamsDomain as c1, type TranscriptParamsLanguageCode as c2, type TranscriptParamsLanguageCodes as c3, type TranscriptParamsRedactPiiSub as c4, type TranscriptParamsRemoveAudioTags as c5, type TranscriptParamsSpeakersExpected as c6, type TranscriptParamsSpeechModel as c7, type TranscriptParamsSpeechThreshold as c8, type TranscriptParamsWebhookAuthHeaderName as c9, type TranscriptTranslatedTexts as cA, type TranscriptUtterance as cB, type TranscriptUtteranceChannel as cC, type TranscriptUtteranceTranslatedTexts as cD, type TranscriptUtterances as cE, type TranscriptWebhookAuthHeaderName as cF, type TranscriptWebhookStatusCode as cG, type TranscriptWebhookUrl as cH, type TranscriptWordChannel as cI, type TranscriptWordSpeaker as cJ, type TranscriptWords as cK, type TranslationRequestBody as cL, type TranslationRequestBodyTranslation as cM, type TranslationResponse as cN, type TranslationResponseTranslation as cO, type StreamingUpdateConfiguration as cP, type Transcription as cQ, type EntityError as cR, Status as cS, type EntityReference as cT, type DiarizationProperties as cU, type DiarizationSpeakersProperties as cV, LanguageIdentificationMode as cW, type LanguageIdentificationProperties as cX, type LanguageIdentificationPropertiesSpeechModelMapping as cY, ProfanityFilterMode as cZ, PunctuationMode as c_, type TranscriptParamsWebhookAuthHeaderValue as ca, type TranscriptPunctuate as cb, type TranscriptRedactPiiAudio as cc, type TranscriptRedactPiiAudioOptions as cd, TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod as ce, type TranscriptRedactPiiAudioQuality as cf, type TranscriptRedactPiiPolicies as cg, TranscriptRemoveAudioTags as ch, type TranscriptRemoveAudioTagsProperty as ci, type TranscriptSentimentAnalysis as cj, type TranscriptSentimentAnalysisResults as ck, type TranscriptSpeakerLabels as cl, type TranscriptSpeakersExpected as cm, type TranscriptSpeechModel as cn, type TranscriptSpeechModels as co, type TranscriptSpeechThreshold as cp, type TranscriptSpeechUnderstanding as cq, type TranscriptSpeechUnderstandingRequest as cr, type TranscriptSpeechUnderstandingResponse as cs, type TranscriptSpeedBoost as ct, type TranscriptSummary as cu, type TranscriptSummaryModel as cv, type TranscriptSummaryType as cw, type TranscriptTemperature as cx, type TranscriptText as cy, type TranscriptThrottled as cz, StreamingSupportedEncodingEnum as d, type ListenV1LanguageParameter as d$, type TranscriptionLinks as d0, type TranscriptionProperties as d1, type TranscriptTextUsageTokens as d2, type TranscriptionSegment as d3, type RealtimeSessionCreateRequestGAModel as d4, RealtimeTranscriptionSessionCreateRequestTurnDetectionType as d5, RealtimeTranscriptionSessionCreateRequestInputAudioFormat as d6, AudioResponseFormat as d7, type CreateTranscription200One as d8, type CreateTranscriptionRequest as d9, type ManageV1LimitParameter as dA, type ManageV1PageParameter as dB, ManageV1FilterEndpointParameter as dC, ManageV1FilterMethodParameter as dD, type SharedTopics as dE, type SharedIntents as dF, type SharedSentiments as dG, type SharedCallbackParameter as dH, SharedCallbackMethodParameter as dI, type SharedSentimentParameter as dJ, type SharedSummarizeParameter as dK, type SharedTagParameter as dL, type SharedTopicsParameter as dM, type SharedCustomTopicParameter as dN, SharedCustomTopicModeParameter as dO, type SharedIntentsParameter as dP, type SharedCustomIntentParameter as dQ, SharedCustomIntentModeParameter as dR, type SharedMipOptOutParameter as dS, type ListenV1DetectEntitiesParameter as dT, type ListenV1DetectLanguageParameter as dU, type ListenV1DiarizeParameter as dV, type ListenV1DictationParameter as dW, ListenV1EncodingParameter as dX, type ListenV1FillerWordsParameter as dY, type ListenV1KeytermParameter as dZ, type ListenV1KeywordsParameter as d_, type CreateTranscriptionRequestModel as da, type CreateTranscriptionRequestStream as db, CreateTranscriptionRequestTimestampGranularitiesItem as dc, type CreateTranscriptionResponseDiarizedJson as dd, CreateTranscriptionResponseDiarizedJsonTask as de, type CreateTranscriptionResponseDiarizedJsonUsage as df, type CreateTranscriptionResponseJson as dg, type CreateTranscriptionResponseJsonLogprobsItem as dh, type CreateTranscriptionResponseJsonUsage as di, type CreateTranscriptionResponseVerboseJson as dj, type TranscriptTextUsageDuration as dk, TranscriptTextUsageDurationType as dl, type TranscriptTextUsageTokensInputTokenDetails as dm, TranscriptTextUsageTokensType as dn, type TranscriptionChunkingStrategy as dp, type TranscriptionChunkingStrategyAnyOf as dq, type TranscriptionDiarizedSegment as dr, TranscriptionDiarizedSegmentType as ds, TranscriptionInclude as dt, type TranscriptionWord as du, type VadConfig as dv, VadConfigType as dw, type ListenV1Response as dx, type ManageV1FilterAccessorParameter as dy, ManageV1FilterDeploymentParameter as dz, StreamingSupportedBitDepthEnum as e, type AudioEventSummaryItem as e$, type ListenV1MeasurementsParameter as e0, type ListenV1MediaTranscribeParams as e1, type ListenV1ModelParameter as e2, type ListenV1MultichannelParameter as e3, type ListenV1NumeralsParameter as e4, type ListenV1ParagraphsParameter as e5, type ListenV1ProfanityFilterParameter as e6, type ListenV1PunctuateParameter as e7, type ListenV1RedactParameter as e8, ListenV1RedactParameterOneOfItem as e9, type ListenV1UttSplitParameter as eA, type ListenV1UtterancesParameter as eB, type ListenV1VersionParameter as eC, type ManageV1EndDateTimeParameter as eD, type ManageV1FilterRequestIdParameter as eE, ManageV1FilterStatusParameter as eF, type ManageV1ProjectsRequestsListParams as eG, type ManageV1StartDateTimeParameter as eH, type SharedExtraParameter as eI, type SharedIntentsResults as eJ, type SharedIntentsResultsIntents as eK, type SharedIntentsResultsIntentsSegmentsItem as eL, type SharedIntentsResultsIntentsSegmentsItemIntentsItem as eM, type SharedSentimentsAverage as eN, type SharedSentimentsSegmentsItem as eO, type SharedTopicsResults as eP, type SharedTopicsResultsTopics as eQ, type SharedTopicsResultsTopicsSegmentsItem as eR, type SharedTopicsResultsTopicsSegmentsItemTopicsItem as eS, JobType as eT, type AlignmentConfig as eU, type TranscriptionConfig as eV, type TrackingData as eW, type OutputConfig as eX, OperatingPoint as eY, type AudioEventItem as eZ, type AudioEventSummary as e_, type ListenV1ReplaceParameter as ea, type ListenV1ResponseMetadata as eb, type ListenV1ResponseMetadataIntentsInfo as ec, type ListenV1ResponseMetadataModelInfo as ed, type ListenV1ResponseMetadataSentimentInfo as ee, type ListenV1ResponseMetadataSummaryInfo as ef, type ListenV1ResponseMetadataTopicsInfo as eg, type ListenV1ResponseResults as eh, type ListenV1ResponseResultsChannels as ei, type ListenV1ResponseResultsChannelsItem as ej, type ListenV1ResponseResultsChannelsItemAlternativesItem as ek, type ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem as el, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs as em, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem as en, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem as eo, type ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem as ep, type ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem as eq, type ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem as er, type ListenV1ResponseResultsChannelsItemSearchItem as es, type ListenV1ResponseResultsChannelsItemSearchItemHitsItem as et, type ListenV1ResponseResultsSummary as eu, type ListenV1ResponseResultsUtterances as ev, type ListenV1ResponseResultsUtterancesItem as ew, type ListenV1ResponseResultsUtterancesItemWordsItem as ex, type ListenV1SearchParameter as ey, type ListenV1SmartFormatParameter as ez, StreamingSupportedSampleRateEnum as f, type BodySpeechToTextV1SpeechToTextPostEntityDetection as f$, type AutoChaptersResult as f0, type AutoChaptersResultError as f1, AutoChaptersResultErrorType as f2, type Chapter as f3, type JobInfo as f4, type LanguageIdentificationResult as f5, type LanguageIdentificationResultAlternative as f6, LanguageIdentificationResultError as f7, type LanguageIdentificationResultItem as f8, type LanguagePackInfo as f9, type SummarizationResult as fA, type TopicDetectionError as fB, TopicDetectionErrorType as fC, type TopicDetectionResult as fD, type TopicDetectionSegment as fE, type TopicDetectionSegmentTopic as fF, type TopicDetectionSummary as fG, type TopicDetectionSummaryOverall as fH, type TrackingDataDetails as fI, type TranscriptionConfigAdditionalVocabItem as fJ, TranscriptionConfigDiarization as fK, TranscriptionConfigMaxDelayMode as fL, type TranscriptionConfigPunctuationOverrides as fM, type TranscriptionConfigSpeakerDiarizationConfig as fN, type TranscriptionConfigTranscriptFilteringConfig as fO, type TranscriptionConfigTranscriptFilteringConfigReplacementsItem as fP, type TranslationError as fQ, TranslationErrorType as fR, type TranslationSentence as fS, type WrittenFormRecognitionResult as fT, WrittenFormRecognitionResultType as fU, type SpeechToTextChunkResponseModel as fV, type AdditionalFormatResponseModel as fW, type AdditionalFormats as fX, type BodySpeechToTextV1SpeechToTextPost as fY, type BodySpeechToTextV1SpeechToTextPostCloudStorageUrl as fZ, type BodySpeechToTextV1SpeechToTextPostDiarizationThreshold as f_, LanguagePackInfoWritingDirection as fa, type OutputConfigSrtOverrides as fb, type RecognitionAlternative as fc, type RecognitionDisplay as fd, RecognitionDisplayDirection as fe, type RecognitionMetadata as ff, type RecognitionResult as fg, RecognitionResultAttachesTo as fh, RecognitionResultType as fi, type RetrieveTranscriptResponse as fj, type RetrieveTranscriptResponseAudioEventSummary as fk, type RetrieveTranscriptResponseAudioEventSummaryChannels as fl, type RetrieveTranscriptResponseTranslations as fm, type SentimentAnalysisError as fn, SentimentAnalysisErrorType as fo, type SentimentAnalysisResult as fp, type SentimentAnalysisResultSentimentAnalysis as fq, type SentimentChannelSummary as fr, type SentimentSegment as fs, type SentimentSpeakerSummary as ft, type SentimentSummary as fu, type SentimentSummaryDetail as fv, type SpokenFormRecognitionResult as fw, SpokenFormRecognitionResultType as fx, type SummarizationError as fy, SummarizationErrorType as fz, StreamingSupportedModels as g, type GladiaExtendedData as g$, type BodySpeechToTextV1SpeechToTextPostFile as g0, BodySpeechToTextV1SpeechToTextPostFileFormat as g1, type BodySpeechToTextV1SpeechToTextPostLanguageCode as g2, BodySpeechToTextV1SpeechToTextPostModelId as g3, type BodySpeechToTextV1SpeechToTextPostNumSpeakers as g4, type BodySpeechToTextV1SpeechToTextPostSeed as g5, type BodySpeechToTextV1SpeechToTextPostTemperature as g6, BodySpeechToTextV1SpeechToTextPostTimestampsGranularity as g7, type BodySpeechToTextV1SpeechToTextPostWebhookId as g8, type BodySpeechToTextV1SpeechToTextPostWebhookMetadata as g9, type SpeechToTextChunkResponseModelAdditionalFormats as gA, type SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem as gB, type SpeechToTextChunkResponseModelChannelIndex as gC, type SpeechToTextChunkResponseModelEntities as gD, type SpeechToTextChunkResponseModelTranscriptionId as gE, type SpeechToTextWordResponseModel as gF, type SpeechToTextWordResponseModelCharacters as gG, type SpeechToTextWordResponseModelEnd as gH, type SpeechToTextWordResponseModelSpeakerId as gI, type SpeechToTextWordResponseModelStart as gJ, SpeechToTextWordResponseModelType as gK, type SrtExportOptions as gL, SrtExportOptionsFormat as gM, type SrtExportOptionsMaxCharactersPerLine as gN, type SrtExportOptionsMaxSegmentChars as gO, type SrtExportOptionsMaxSegmentDurationS as gP, type SrtExportOptionsSegmentOnSilenceLongerThanS as gQ, type TxtExportOptions as gR, TxtExportOptionsFormat as gS, type TxtExportOptionsMaxCharactersPerLine as gT, type TxtExportOptionsMaxSegmentChars as gU, type TxtExportOptionsMaxSegmentDurationS as gV, type TxtExportOptionsSegmentOnSilenceLongerThanS as gW, type SpeechmaticsOperatingPoint as gX, type TranscriptionModel as gY, type TranscriptionLanguage as gZ, type AssemblyAIExtendedData as g_, type BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf as ga, type DetectedEntity as gb, type DocxExportOptions as gc, DocxExportOptionsFormat as gd, type DocxExportOptionsMaxSegmentChars as ge, type DocxExportOptionsMaxSegmentDurationS as gf, type DocxExportOptionsSegmentOnSilenceLongerThanS as gg, type ExportOptions as gh, type HtmlExportOptions as gi, HtmlExportOptionsFormat as gj, type HtmlExportOptionsMaxSegmentChars as gk, type HtmlExportOptionsMaxSegmentDurationS as gl, type HtmlExportOptionsSegmentOnSilenceLongerThanS as gm, type PdfExportOptions as gn, PdfExportOptionsFormat as go, type PdfExportOptionsMaxSegmentChars as gp, type PdfExportOptionsMaxSegmentDurationS as gq, type PdfExportOptionsSegmentOnSilenceLongerThanS as gr, type SegmentedJsonExportOptions as gs, SegmentedJsonExportOptionsFormat as gt, type SegmentedJsonExportOptionsMaxSegmentChars as gu, type SegmentedJsonExportOptionsMaxSegmentDurationS as gv, type SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS as gw, type SpeechToTextCharacterResponseModel as gx, type SpeechToTextCharacterResponseModelEnd as gy, type SpeechToTextCharacterResponseModelStart as gz, type LanguageConfig as h, type DeepgramExtendedData as h0, type ElevenLabsExtendedData as h1, type ProviderExtendedDataMap as h2, type StreamingProvider as h3, type BatchOnlyProvider as h4, type SessionStatus as h5, type Speaker as h6, type Word as h7, type Utterance as h8, type TranscriptionStatus as h9, type StreamingEventMessage as hA, type StreamingWord as hB, type StreamingForceEndpoint as hC, type TranscriptMetadata as ha, type TranscriptData as hb, type ListTranscriptsResponse as hc, type ProviderRawResponseMap as hd, type StreamEventType as he, type SpeechEvent as hf, type TranslationEvent as hg, type SentimentEvent as hh, type EntityEvent as hi, type SummarizationEvent as hj, type ChapterizationEvent as hk, type AudioAckEvent as hl, type LifecycleEvent as hm, type AudioChunk as hn, type RawWebSocketMessage as ho, type AssemblyAIUpdateConfiguration as hp, type OpenAIStreamingOptions as hq, type SonioxStreamingOptions as hr, type ElevenLabsStreamingOptions as hs, type ProviderStreamingOptions as ht, type StreamingOptionsForProvider as hu, type TranscribeStreamParams as hv, type BeginEvent as hw, type TurnEvent as hx, type TerminationEvent as hy, type ErrorEvent as hz, type PostProcessingConfig as i, type TranscriptionMetadataDTO as j, type TranscriptionDTO as k, type TranslationDTO as l, type SummarizationDTO as m, type SentimentAnalysisDTO as n, type ChapterizationDTO as o, type PreRecordedResponse as p, type UtteranceDTO as q, TranscriptionLanguageCodeEnum as r, TranslationLanguageCodeEnum as s, StreamingSupportedRegions as t, type AddonErrorDTO as u, type AudioToLlmDTO as v, type AudioToLlmDTOError as w, type AudioToLlmDTOResults as x, type AudioToLlmListConfigDTO as y, type AudioToLlmListDTO as z };