voice-router-dev 0.8.3 → 0.8.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2501,29 +2501,609 @@ type CreateTranscription200One = CreateTranscriptionResponseJson | CreateTranscr
2501
2501
  * Generated by orval v7.9.0 🍺
2502
2502
  * Do not edit manually.
2503
2503
  * AssemblyAI API
2504
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2505
+ * OpenAPI spec version: 1.0.0
2506
+ */
2507
+ /**
2508
+ * The duration of this transcript object's media file, in seconds
2509
+ */
2510
+ type TranscriptAudioDuration = number | null;
2511
+
2512
+ /**
2513
+ * Generated by orval v7.9.0 🍺
2514
+ * Do not edit manually.
2515
+ * AssemblyAI API
2516
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2517
+ * OpenAPI spec version: 1.0.0
2518
+ */
2519
+ /**
2520
+ * The point in time, in milliseconds, in the file at which the transcription was terminated. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details.
2521
+ */
2522
+ type TranscriptAudioEndAt = number | null;
2523
+
2524
+ /**
2525
+ * Generated by orval v7.9.0 🍺
2526
+ * Do not edit manually.
2527
+ * AssemblyAI API
2528
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2529
+ * OpenAPI spec version: 1.0.0
2530
+ */
2531
+ /**
2532
+ * The point in time, in milliseconds, in the file at which the transcription was started. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details.
2533
+ */
2534
+ type TranscriptAudioStartFrom = number | null;
2535
+
2536
+ /**
2537
+ * Generated by orval v7.9.0 🍺
2538
+ * Do not edit manually.
2539
+ * AssemblyAI API
2540
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2541
+ * OpenAPI spec version: 1.0.0
2542
+ */
2543
+ /**
2544
+ * Whether [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) is enabled, can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible chapter summaries. See the [updated Auto Chapters page](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for details.
2545
+
2546
+ Note: This parameter is only supported for the Universal-2 model.
2547
+
2548
+ */
2549
+ type TranscriptAutoChapters = boolean | null;
2550
+
2551
+ /**
2552
+ * Generated by orval v7.9.0 🍺
2553
+ * Do not edit manually.
2554
+ * AssemblyAI API
2555
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2556
+ * OpenAPI spec version: 1.0.0
2557
+ */
2558
+ /**
2559
+ * Either success, or unavailable in the rare case that the model failed
2560
+ */
2561
+ type AudioIntelligenceModelStatus = (typeof AudioIntelligenceModelStatus)[keyof typeof AudioIntelligenceModelStatus];
2562
+ declare const AudioIntelligenceModelStatus: {
2563
+ readonly success: "success";
2564
+ readonly unavailable: "unavailable";
2565
+ };
2566
+
2567
+ /**
2568
+ * Generated by orval v7.9.0 🍺
2569
+ * Do not edit manually.
2570
+ * AssemblyAI API
2571
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2572
+ * OpenAPI spec version: 1.0.0
2573
+ */
2574
+ /**
2575
+ * Timestamp containing a start and end property in milliseconds
2576
+ */
2577
+ interface Timestamp {
2578
+ /** The start time in milliseconds */
2579
+ start: number;
2580
+ /** The end time in milliseconds */
2581
+ end: number;
2582
+ }
2583
+
2584
+ /**
2585
+ * Generated by orval v7.9.0 🍺
2586
+ * Do not edit manually.
2587
+ * AssemblyAI API
2588
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2589
+ * OpenAPI spec version: 1.0.0
2590
+ */
2591
+
2592
+ interface AutoHighlightResult {
2593
+ /** The total number of times the key phrase appears in the audio file */
2594
+ count: number;
2595
+ /** The total relevancy to the overall audio file of this key phrase - a greater number means more relevant */
2596
+ rank: number;
2597
+ /** The text itself of the key phrase */
2598
+ text: string;
2599
+ /** The timestamp of the of the key phrase */
2600
+ timestamps: Timestamp[];
2601
+ }
2602
+
2603
+ /**
2604
+ * Generated by orval v7.9.0 🍺
2605
+ * Do not edit manually.
2606
+ * AssemblyAI API
2607
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2608
+ * OpenAPI spec version: 1.0.0
2609
+ */
2610
+
2611
+ /**
2612
+ * An array of results for the Key Phrases model, if it is enabled.
2613
+ See [Key phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) for more information.
2614
+
2615
+ */
2616
+ interface AutoHighlightsResult {
2617
+ /** The status of the Key Phrases model. Either success, or unavailable in the rare case that the model failed. */
2618
+ status: AudioIntelligenceModelStatus;
2619
+ /** A temporally-sequential array of Key Phrases */
2620
+ results: AutoHighlightResult[];
2621
+ }
2622
+
2623
+ /**
2624
+ * Generated by orval v7.9.0 🍺
2625
+ * Do not edit manually.
2626
+ * AssemblyAI API
2627
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2628
+ * OpenAPI spec version: 1.0.0
2629
+ */
2630
+
2631
+ /**
2632
+ * An array of results for the Key Phrases model, if it is enabled.
2633
+ See [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) for more information.
2634
+
2635
+ */
2636
+ type TranscriptAutoHighlightsResult = AutoHighlightsResult | null;
2637
+
2638
+ /**
2639
+ * Generated by orval v7.9.0 🍺
2640
+ * Do not edit manually.
2641
+ * AssemblyAI API
2642
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2643
+ * OpenAPI spec version: 1.0.0
2644
+ */
2645
+ /**
2646
+ * Chapter of the audio file
2647
+ */
2648
+ interface Chapter$1 {
2649
+ /** An ultra-short summary (just a few words) of the content spoken in the chapter */
2650
+ gist: string;
2651
+ /** A single sentence summary of the content spoken during the chapter */
2652
+ headline: string;
2653
+ /** A one paragraph summary of the content spoken during the chapter */
2654
+ summary: string;
2655
+ /** The starting time, in milliseconds, for the chapter */
2656
+ start: number;
2657
+ /** The starting time, in milliseconds, for the chapter */
2658
+ end: number;
2659
+ }
2660
+
2661
+ /**
2662
+ * Generated by orval v7.9.0 🍺
2663
+ * Do not edit manually.
2664
+ * AssemblyAI API
2665
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2666
+ * OpenAPI spec version: 1.0.0
2667
+ */
2668
+
2669
+ /**
2670
+ * An array of temporally sequential chapters for the audio file. See [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for more information.
2671
+ */
2672
+ type TranscriptChapters = Chapter$1[] | null;
2673
+
2674
+ /**
2675
+ * Generated by orval v7.9.0 🍺
2676
+ * Do not edit manually.
2677
+ * AssemblyAI API
2678
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2679
+ * OpenAPI spec version: 1.0.0
2680
+ */
2681
+ /**
2682
+ * The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence)
2683
+ */
2684
+ type TranscriptConfidence = number | null;
2685
+
2686
+ /**
2687
+ * Generated by orval v7.9.0 🍺
2688
+ * Do not edit manually.
2689
+ * AssemblyAI API
2690
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2691
+ * OpenAPI spec version: 1.0.0
2692
+ */
2693
+ /**
2694
+ * Whether [Content Moderation](https://www.assemblyai.com/docs/content-moderation) is enabled, can be true or false
2695
+ */
2696
+ type TranscriptContentSafety = boolean | null;
2697
+
2698
+ /**
2699
+ * Generated by orval v7.9.0 🍺
2700
+ * Do not edit manually.
2701
+ * AssemblyAI API
2702
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2703
+ * OpenAPI spec version: 1.0.0
2704
+ */
2705
+ interface ContentSafetyLabel {
2706
+ /** The label of the sensitive topic */
2707
+ label: string;
2708
+ /** The confidence score for the topic being discussed, from 0 to 1 */
2709
+ confidence: number;
2710
+ /** How severely the topic is discussed in the section, from 0 to 1 */
2711
+ severity: number;
2712
+ }
2713
+
2714
+ /**
2715
+ * Generated by orval v7.9.0 🍺
2716
+ * Do not edit manually.
2717
+ * AssemblyAI API
2718
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2719
+ * OpenAPI spec version: 1.0.0
2720
+ */
2721
+
2722
+ interface ContentSafetyLabelResult {
2723
+ /** The transcript of the section flagged by the Content Moderation model */
2724
+ text: string;
2725
+ /** An array of safety labels, one per sensitive topic that was detected in the section */
2726
+ labels: ContentSafetyLabel[];
2727
+ /** The sentence index at which the section begins */
2728
+ sentences_idx_start: number;
2729
+ /** The sentence index at which the section ends */
2730
+ sentences_idx_end: number;
2731
+ /** Timestamp information for the section */
2732
+ timestamp: Timestamp;
2733
+ }
2734
+
2735
+ /**
2736
+ * Generated by orval v7.9.0 🍺
2737
+ * Do not edit manually.
2738
+ * AssemblyAI API
2739
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2740
+ * OpenAPI spec version: 1.0.0
2741
+ */
2742
+ /**
2743
+ * A summary of the Content Moderation confidence results for the entire audio file
2744
+ */
2745
+ type ContentSafetyLabelsResultSummary = {
2746
+ [key: string]: number;
2747
+ };
2748
+
2749
+ /**
2750
+ * Generated by orval v7.9.0 🍺
2751
+ * Do not edit manually.
2752
+ * AssemblyAI API
2753
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2754
+ * OpenAPI spec version: 1.0.0
2755
+ */
2756
+ interface SeverityScoreSummary {
2757
+ low: number;
2758
+ medium: number;
2759
+ high: number;
2760
+ }
2761
+
2762
+ /**
2763
+ * Generated by orval v7.9.0 🍺
2764
+ * Do not edit manually.
2765
+ * AssemblyAI API
2766
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2767
+ * OpenAPI spec version: 1.0.0
2768
+ */
2769
+
2770
+ /**
2771
+ * A summary of the Content Moderation severity results for the entire audio file
2772
+ */
2773
+ type ContentSafetyLabelsResultSeverityScoreSummary = {
2774
+ [key: string]: SeverityScoreSummary;
2775
+ };
2776
+
2777
+ /**
2778
+ * Generated by orval v7.9.0 🍺
2779
+ * Do not edit manually.
2780
+ * AssemblyAI API
2781
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2782
+ * OpenAPI spec version: 1.0.0
2783
+ */
2784
+
2785
+ /**
2786
+ * An array of results for the Content Moderation model, if it is enabled.
2787
+ See [Content moderation](https://www.assemblyai.com/docs/content-moderation) for more information.
2788
+
2789
+ */
2790
+ interface ContentSafetyLabelsResult {
2791
+ /** The status of the Content Moderation model. Either success, or unavailable in the rare case that the model failed. */
2792
+ status: AudioIntelligenceModelStatus;
2793
+ /** An array of results for the Content Moderation model */
2794
+ results: ContentSafetyLabelResult[];
2795
+ /** A summary of the Content Moderation confidence results for the entire audio file */
2796
+ summary: ContentSafetyLabelsResultSummary;
2797
+ /** A summary of the Content Moderation severity results for the entire audio file */
2798
+ severity_score_summary: ContentSafetyLabelsResultSeverityScoreSummary;
2799
+ }
2800
+
2801
+ /**
2802
+ * Generated by orval v7.9.0 🍺
2803
+ * Do not edit manually.
2804
+ * AssemblyAI API
2805
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2806
+ * OpenAPI spec version: 1.0.0
2807
+ */
2808
+
2809
+ /**
2810
+ * An array of results for the Content Moderation model, if it is enabled.
2811
+ See [Content moderation](https://www.assemblyai.com/docs/content-moderation) for more information.
2812
+
2813
+ */
2814
+ type TranscriptContentSafetyLabels = ContentSafetyLabelsResult | null;
2815
+
2816
+ /**
2817
+ * Generated by orval v7.9.0 🍺
2818
+ * Do not edit manually.
2819
+ * AssemblyAI API
2820
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2821
+ * OpenAPI spec version: 1.0.0
2822
+ */
2823
+ /**
2824
+ * Object containing words or phrases to replace, and the word or phrase to replace with
2825
+ */
2826
+ interface TranscriptCustomSpelling {
2827
+ /** Words or phrases to replace */
2828
+ from: string[];
2829
+ /** Word to replace with */
2830
+ to: string;
2831
+ }
2832
+
2833
+ /**
2834
+ * Generated by orval v7.9.0 🍺
2835
+ * Do not edit manually.
2836
+ * AssemblyAI API
2837
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2838
+ * OpenAPI spec version: 1.0.0
2839
+ */
2840
+
2841
+ /**
2842
+ * Customize how words are spelled and formatted using to and from values. See [Custom Spelling](https://www.assemblyai.com/docs/pre-recorded-audio/custom-spelling) for more details.
2843
+ */
2844
+ type TranscriptCustomSpellingProperty = TranscriptCustomSpelling[] | null;
2845
+
2846
+ /**
2847
+ * Generated by orval v7.9.0 🍺
2848
+ * Do not edit manually.
2849
+ * AssemblyAI API
2850
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2851
+ * OpenAPI spec version: 1.0.0
2852
+ */
2853
+ /**
2854
+ * Transcribe [Filler Words](https://www.assemblyai.com/docs/pre-recorded-audio/filler-words), like "umm", in your media file; can be true or false
2855
+ */
2856
+ type TranscriptDisfluencies = boolean | null;
2857
+
2858
+ /**
2859
+ * Generated by orval v7.9.0 🍺
2860
+ * Do not edit manually.
2861
+ * AssemblyAI API
2862
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2863
+ * OpenAPI spec version: 1.0.0
2864
+ */
2865
+ /**
2866
+ * The domain-specific model applied to the transcript. When set to `"medical-v1"`, [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) was used to improve accuracy for medical terminology.
2867
+
2868
+ */
2869
+ type TranscriptDomain = string | null;
2870
+
2871
+ /**
2872
+ * Generated by orval v7.9.0 🍺
2873
+ * Do not edit manually.
2874
+ * AssemblyAI API
2875
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2876
+ * OpenAPI spec version: 1.0.0
2877
+ */
2878
+ /**
2879
+ * The type of entity for the detected entity
2880
+ */
2881
+ type EntityType = (typeof EntityType)[keyof typeof EntityType];
2882
+ declare const EntityType: {
2883
+ readonly account_number: "account_number";
2884
+ readonly banking_information: "banking_information";
2885
+ readonly blood_type: "blood_type";
2886
+ readonly credit_card_cvv: "credit_card_cvv";
2887
+ readonly credit_card_expiration: "credit_card_expiration";
2888
+ readonly credit_card_number: "credit_card_number";
2889
+ readonly date: "date";
2890
+ readonly date_interval: "date_interval";
2891
+ readonly date_of_birth: "date_of_birth";
2892
+ readonly drivers_license: "drivers_license";
2893
+ readonly drug: "drug";
2894
+ readonly duration: "duration";
2895
+ readonly email_address: "email_address";
2896
+ readonly event: "event";
2897
+ readonly filename: "filename";
2898
+ readonly gender_sexuality: "gender_sexuality";
2899
+ readonly healthcare_number: "healthcare_number";
2900
+ readonly injury: "injury";
2901
+ readonly ip_address: "ip_address";
2902
+ readonly language: "language";
2903
+ readonly location: "location";
2904
+ readonly marital_status: "marital_status";
2905
+ readonly medical_condition: "medical_condition";
2906
+ readonly medical_process: "medical_process";
2907
+ readonly money_amount: "money_amount";
2908
+ readonly nationality: "nationality";
2909
+ readonly number_sequence: "number_sequence";
2910
+ readonly occupation: "occupation";
2911
+ readonly organization: "organization";
2912
+ readonly passport_number: "passport_number";
2913
+ readonly password: "password";
2914
+ readonly person_age: "person_age";
2915
+ readonly person_name: "person_name";
2916
+ readonly phone_number: "phone_number";
2917
+ readonly physical_attribute: "physical_attribute";
2918
+ readonly political_affiliation: "political_affiliation";
2919
+ readonly religion: "religion";
2920
+ readonly statistics: "statistics";
2921
+ readonly time: "time";
2922
+ readonly url: "url";
2923
+ readonly us_social_security_number: "us_social_security_number";
2924
+ readonly username: "username";
2925
+ readonly vehicle_id: "vehicle_id";
2926
+ readonly zodiac_sign: "zodiac_sign";
2927
+ };
2928
+
2929
+ /**
2930
+ * Generated by orval v7.9.0 🍺
2931
+ * Do not edit manually.
2932
+ * AssemblyAI API
2933
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2934
+ * OpenAPI spec version: 1.0.0
2935
+ */
2936
+
2937
+ /**
2938
+ * A detected entity
2939
+ */
2940
+ interface Entity {
2941
+ /** The type of entity for the detected entity */
2942
+ entity_type: EntityType;
2943
+ /** The text for the detected entity */
2944
+ text: string;
2945
+ /** The starting time, in milliseconds, at which the detected entity appears in the audio file */
2946
+ start: number;
2947
+ /** The ending time, in milliseconds, for the detected entity in the audio file */
2948
+ end: number;
2949
+ }
2950
+
2951
+ /**
2952
+ * Generated by orval v7.9.0 🍺
2953
+ * Do not edit manually.
2954
+ * AssemblyAI API
2955
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2956
+ * OpenAPI spec version: 1.0.0
2957
+ */
2958
+
2959
+ /**
2960
+ * An array of results for the Entity Detection model, if it is enabled.
2961
+ See [Entity detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) for more information.
2962
+
2963
+ */
2964
+ type TranscriptEntities = Entity[] | null;
2965
+
2966
+ /**
2967
+ * Generated by orval v7.9.0 🍺
2968
+ * Do not edit manually.
2969
+ * AssemblyAI API
2970
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2971
+ * OpenAPI spec version: 1.0.0
2972
+ */
2973
+ /**
2974
+ * Whether [Entity Detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) is enabled, can be true or false
2975
+ */
2976
+ type TranscriptEntityDetection = boolean | null;
2977
+
2978
+ /**
2979
+ * Generated by orval v7.9.0 🍺
2980
+ * Do not edit manually.
2981
+ * AssemblyAI API
2982
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2983
+ * OpenAPI spec version: 1.0.0
2984
+ */
2985
+ /**
2986
+ * Whether [Profanity Filtering](https://www.assemblyai.com/docs/profanity-filtering) is enabled, either true or false
2987
+ */
2988
+ type TranscriptFilterProfanity = boolean | null;
2989
+
2990
+ /**
2991
+ * Generated by orval v7.9.0 🍺
2992
+ * Do not edit manually.
2504
2993
  * AssemblyAI API
2505
- * OpenAPI spec version: 1.3.4
2994
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
2995
+ * OpenAPI spec version: 1.0.0
2506
2996
  */
2507
2997
  /**
2508
- * The status of your transcript. Possible values are queued, processing, completed, or error.
2998
+ * Whether [Text Formatting](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false
2509
2999
  */
2510
- type TranscriptStatus = (typeof TranscriptStatus)[keyof typeof TranscriptStatus];
2511
- declare const TranscriptStatus: {
2512
- readonly queued: "queued";
2513
- readonly processing: "processing";
2514
- readonly completed: "completed";
2515
- readonly error: "error";
3000
+ type TranscriptFormatText = boolean | null;
3001
+
3002
+ /**
3003
+ * Generated by orval v7.9.0 🍺
3004
+ * Do not edit manually.
3005
+ * AssemblyAI API
3006
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3007
+ * OpenAPI spec version: 1.0.0
3008
+ */
3009
+ /**
3010
+ * Whether [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) is enabled, can be true or false
3011
+ */
3012
+ type TranscriptIabCategories = boolean | null;
3013
+
3014
+ /**
3015
+ * Generated by orval v7.9.0 🍺
3016
+ * Do not edit manually.
3017
+ * AssemblyAI API
3018
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3019
+ * OpenAPI spec version: 1.0.0
3020
+ */
3021
+ interface TopicDetectionResultLabelsItems {
3022
+ /** How relevant the detected topic is of a detected topic */
3023
+ relevance: number;
3024
+ /** The IAB taxonomical label for the label of the detected topic, where > denotes supertopic/subtopic relationship */
3025
+ label: string;
3026
+ }
3027
+
3028
+ /**
3029
+ * Generated by orval v7.9.0 🍺
3030
+ * Do not edit manually.
3031
+ * AssemblyAI API
3032
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3033
+ * OpenAPI spec version: 1.0.0
3034
+ */
3035
+
3036
+ /**
3037
+ * The result of the topic detection model
3038
+ */
3039
+ interface TopicDetectionResult$1 {
3040
+ /** The text in the transcript in which a detected topic occurs */
3041
+ text: string;
3042
+ /** An array of detected topics in the text */
3043
+ labels?: TopicDetectionResultLabelsItems[];
3044
+ timestamp?: Timestamp;
3045
+ }
3046
+
3047
+ /**
3048
+ * Generated by orval v7.9.0 🍺
3049
+ * Do not edit manually.
3050
+ * AssemblyAI API
3051
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3052
+ * OpenAPI spec version: 1.0.0
3053
+ */
3054
+ /**
3055
+ * The overall relevance of topic to the entire audio file
3056
+ */
3057
+ type TopicDetectionModelResultSummary = {
3058
+ [key: string]: number;
2516
3059
  };
2517
3060
 
2518
3061
  /**
2519
3062
  * Generated by orval v7.9.0 🍺
2520
3063
  * Do not edit manually.
2521
3064
  * AssemblyAI API
3065
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3066
+ * OpenAPI spec version: 1.0.0
3067
+ */
3068
+
3069
+ /**
3070
+ * The result of the Topic Detection model, if it is enabled.
3071
+ See [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) for more information.
3072
+
3073
+ */
3074
+ interface TopicDetectionModelResult {
3075
+ /** The status of the Topic Detection model. Either success, or unavailable in the rare case that the model failed. */
3076
+ status: AudioIntelligenceModelStatus;
3077
+ /** An array of results for the Topic Detection model */
3078
+ results: TopicDetectionResult$1[];
3079
+ /** The overall relevance of topic to the entire audio file */
3080
+ summary: TopicDetectionModelResultSummary;
3081
+ }
3082
+
3083
+ /**
3084
+ * Generated by orval v7.9.0 🍺
3085
+ * Do not edit manually.
3086
+ * AssemblyAI API
3087
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3088
+ * OpenAPI spec version: 1.0.0
3089
+ */
3090
+
3091
+ /**
3092
+ * The result of the Topic Detection model, if it is enabled.
3093
+ See [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) for more information.
3094
+
3095
+ */
3096
+ type TranscriptIabCategoriesResult = TopicDetectionModelResult | null;
3097
+
3098
+ /**
3099
+ * Generated by orval v7.9.0 🍺
3100
+ * Do not edit manually.
2522
3101
  * AssemblyAI API
2523
- * OpenAPI spec version: 1.3.4
3102
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3103
+ * OpenAPI spec version: 1.0.0
2524
3104
  */
2525
3105
  /**
2526
- * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
3106
+ * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
2527
3107
  The default value is 'en_us'.
2528
3108
 
2529
3109
  */
@@ -2637,43 +3217,41 @@ declare const TranscriptLanguageCode: {
2637
3217
  * Generated by orval v7.9.0 🍺
2638
3218
  * Do not edit manually.
2639
3219
  * AssemblyAI API
2640
- * AssemblyAI API
2641
- * OpenAPI spec version: 1.3.4
3220
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3221
+ * OpenAPI spec version: 1.0.0
2642
3222
  */
2643
3223
 
2644
3224
  /**
2645
- * The language of your audio file.
2646
- Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
2647
- The default value is 'en_us'.
3225
+ * The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
3226
+ One of the values specified must be `en`.
2648
3227
 
2649
3228
  */
2650
- type TranscriptLanguageCodeProperty = TranscriptLanguageCode | string;
3229
+ type TranscriptLanguageCodes = TranscriptLanguageCode[] | null;
2651
3230
 
2652
3231
  /**
2653
3232
  * Generated by orval v7.9.0 🍺
2654
3233
  * Do not edit manually.
2655
3234
  * AssemblyAI API
2656
- * AssemblyAI API
2657
- * OpenAPI spec version: 1.3.4
3235
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3236
+ * OpenAPI spec version: 1.0.0
2658
3237
  */
2659
3238
  /**
2660
- * Whether [Automatic language detection](https://www.assemblyai.com/docs/models/speech-recognition#automatic-language-detection) is enabled, either true or false
3239
+ * The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence). See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
2661
3240
  */
2662
- type TranscriptLanguageDetection = boolean | null;
3241
+ type TranscriptLanguageConfidence = number | null;
2663
3242
 
2664
3243
  /**
2665
3244
  * Generated by orval v7.9.0 🍺
2666
3245
  * Do not edit manually.
2667
3246
  * AssemblyAI API
2668
- * AssemblyAI API
2669
- * OpenAPI spec version: 1.3.4
3247
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3248
+ * OpenAPI spec version: 1.0.0
2670
3249
  */
2671
3250
  /**
2672
3251
  * The confidence threshold for the automatically detected language.
2673
3252
  An error will be returned if the language confidence is below this threshold.
3253
+ See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
2674
3254
 
2675
- * @minimum 0
2676
- * @maximum 1
2677
3255
  */
2678
3256
  type TranscriptLanguageConfidenceThreshold = number | null;
2679
3257
 
@@ -2681,1415 +3259,1202 @@ type TranscriptLanguageConfidenceThreshold = number | null;
2681
3259
  * Generated by orval v7.9.0 🍺
2682
3260
  * Do not edit manually.
2683
3261
  * AssemblyAI API
2684
- * AssemblyAI API
2685
- * OpenAPI spec version: 1.3.4
2686
- */
2687
- /**
2688
- * The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence)
2689
- * @minimum 0
2690
- * @maximum 1
2691
- */
2692
- type TranscriptLanguageConfidence = number | null;
2693
-
2694
- /**
2695
- * Generated by orval v7.9.0 🍺
2696
- * Do not edit manually.
2697
- * AssemblyAI API
2698
- * AssemblyAI API
2699
- * OpenAPI spec version: 1.3.4
2700
- */
2701
- /**
2702
- * The speech model to use for the transcription.
2703
- */
2704
- type SpeechModel = (typeof SpeechModel)[keyof typeof SpeechModel];
2705
- declare const SpeechModel: {
2706
- readonly best: "best";
2707
- readonly "slam-1": "slam-1";
2708
- readonly universal: "universal";
2709
- };
2710
-
2711
- /**
2712
- * Generated by orval v7.9.0 🍺
2713
- * Do not edit manually.
2714
- * AssemblyAI API
2715
- * AssemblyAI API
2716
- * OpenAPI spec version: 1.3.4
2717
- */
2718
-
2719
- /**
2720
- * The speech model used for the transcription. When `null`, the default model is used.
2721
- */
2722
- type TranscriptSpeechModel = SpeechModel | null;
2723
-
2724
- /**
2725
- * Generated by orval v7.9.0 🍺
2726
- * Do not edit manually.
2727
- * AssemblyAI API
2728
- * AssemblyAI API
2729
- * OpenAPI spec version: 1.3.4
2730
- */
2731
- /**
2732
- * The textual transcript of your media file
2733
- */
2734
- type TranscriptText = string | null;
2735
-
2736
- /**
2737
- * Generated by orval v7.9.0 🍺
2738
- * Do not edit manually.
2739
- * AssemblyAI API
2740
- * AssemblyAI API
2741
- * OpenAPI spec version: 1.3.4
3262
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3263
+ * OpenAPI spec version: 1.0.0
2742
3264
  */
2743
3265
  /**
2744
- * The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3266
+ * Whether [Automatic language detection](/docs/pre-recorded-audio/automatic-language-detection) is enabled, either true or false
2745
3267
  */
2746
- type TranscriptWordChannel = string | null;
3268
+ type TranscriptLanguageDetection = boolean | null;
2747
3269
 
2748
3270
  /**
2749
3271
  * Generated by orval v7.9.0 🍺
2750
3272
  * Do not edit manually.
2751
3273
  * AssemblyAI API
2752
- * AssemblyAI API
2753
- * OpenAPI spec version: 1.3.4
2754
- */
2755
- /**
2756
- * The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null
3274
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3275
+ * OpenAPI spec version: 1.0.0
2757
3276
  */
2758
- type TranscriptWordSpeaker = string | null;
2759
-
2760
3277
  /**
2761
- * Generated by orval v7.9.0 🍺
2762
- * Do not edit manually.
2763
- * AssemblyAI API
2764
- * AssemblyAI API
2765
- * OpenAPI spec version: 1.3.4
3278
+ * Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection).
2766
3279
  */
2767
-
2768
- interface TranscriptWord {
2769
- /**
2770
- * The confidence score for the transcript of this word
2771
- * @minimum 0
2772
- * @maximum 1
3280
+ interface TranscriptLanguageDetectionOptions {
3281
+ /** List of languages expected in the audio file. Defaults to `["all"]` when unspecified. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details. */
3282
+ expected_languages?: string[];
3283
+ /** If the detected language of the audio file is not in the list of expected languages, the `fallback_language` is used. Specify `["auto"]` to let our model choose the fallback language from `expected_languages` with the highest confidence score. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
2773
3284
  */
2774
- confidence: number;
2775
- /** The starting time, in milliseconds, for the word */
2776
- start: number;
2777
- /** The ending time, in milliseconds, for the word */
2778
- end: number;
2779
- /** The text of the word */
2780
- text: string;
2781
- /** The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
2782
- channel?: TranscriptWordChannel;
2783
- /** The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null */
2784
- speaker: TranscriptWordSpeaker;
3285
+ fallback_language?: string;
3286
+ /** Whether [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) should be detected.
3287
+ */
3288
+ code_switching?: boolean;
3289
+ /** The confidence threshold for [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) detection. If the code switching confidence is below this threshold, the transcript will be processed in the language with the highest `language_detection_confidence` score.
3290
+ */
3291
+ code_switching_confidence_threshold?: number;
2785
3292
  }
2786
3293
 
2787
3294
  /**
2788
3295
  * Generated by orval v7.9.0 🍺
2789
3296
  * Do not edit manually.
2790
3297
  * AssemblyAI API
2791
- * AssemblyAI API
2792
- * OpenAPI spec version: 1.3.4
3298
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3299
+ * OpenAPI spec version: 1.0.0
2793
3300
  */
2794
-
2795
3301
  /**
2796
- * An array of temporally-sequential word objects, one for each word in the transcript.
2797
- See [Speech recognition](https://www.assemblyai.com/docs/models/speech-recognition) for more information.
2798
-
3302
+ * Whether [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) was enabled in the transcription request, either true or false
2799
3303
  */
2800
- type TranscriptWords = TranscriptWord[] | null;
3304
+ type TranscriptMultichannel = boolean | null;
2801
3305
 
2802
3306
  /**
2803
3307
  * Generated by orval v7.9.0 🍺
2804
3308
  * Do not edit manually.
2805
3309
  * AssemblyAI API
2806
- * AssemblyAI API
2807
- * OpenAPI spec version: 1.3.4
2808
- */
2809
- /**
2810
- * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3310
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3311
+ * OpenAPI spec version: 1.0.0
2811
3312
  */
2812
- type TranscriptUtteranceChannel = string | null;
2813
-
2814
3313
  /**
2815
- * Generated by orval v7.9.0 🍺
2816
- * Do not edit manually.
2817
- * AssemblyAI API
2818
- * AssemblyAI API
2819
- * OpenAPI spec version: 1.3.4
3314
+ * Whether [Automatic Punctuation](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false
2820
3315
  */
2821
-
2822
- interface TranscriptUtterance {
2823
- /**
2824
- * The confidence score for the transcript of this utterance
2825
- * @minimum 0
2826
- * @maximum 1
2827
- */
2828
- confidence: number;
2829
- /** The starting time, in milliseconds, of the utterance in the audio file */
2830
- start: number;
2831
- /** The ending time, in milliseconds, of the utterance in the audio file */
2832
- end: number;
2833
- /** The text for this utterance */
2834
- text: string;
2835
- /** The words in the utterance. */
2836
- words: TranscriptWord[];
2837
- /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
2838
- channel?: TranscriptUtteranceChannel;
2839
- /** The speaker of this utterance, where each speaker is assigned a sequential capital letter - e.g. "A" for Speaker A, "B" for Speaker B, etc. */
2840
- speaker: string;
2841
- }
3316
+ type TranscriptPunctuate = boolean | null;
2842
3317
 
2843
3318
  /**
2844
3319
  * Generated by orval v7.9.0 🍺
2845
3320
  * Do not edit manually.
2846
3321
  * AssemblyAI API
2847
- * AssemblyAI API
2848
- * OpenAPI spec version: 1.3.4
3322
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3323
+ * OpenAPI spec version: 1.0.0
2849
3324
  */
2850
-
2851
3325
  /**
2852
- * When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
2853
- See [Speaker diarization](https://www.assemblyai.com/docs/speech-to-text/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/speech-to-text/speech-recognition#multichannel-transcription) for more information.
3326
+ * Whether a redacted version of the audio file was generated,
3327
+ either true or false. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
2854
3328
 
2855
3329
  */
2856
- type TranscriptUtterances = TranscriptUtterance[] | null;
3330
+ type TranscriptRedactPiiAudio = boolean | null;
2857
3331
 
2858
3332
  /**
2859
3333
  * Generated by orval v7.9.0 🍺
2860
3334
  * Do not edit manually.
2861
3335
  * AssemblyAI API
2862
- * AssemblyAI API
2863
- * OpenAPI spec version: 1.3.4
3336
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3337
+ * OpenAPI spec version: 1.0.0
2864
3338
  */
2865
3339
  /**
2866
- * The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence)
2867
- * @minimum 0
2868
- * @maximum 1
3340
+ * Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep.
2869
3341
  */
2870
- type TranscriptConfidence = number | null;
3342
+ type TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod = (typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod)[keyof typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod];
3343
+ declare const TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod: {
3344
+ readonly silence: "silence";
3345
+ };
2871
3346
 
2872
3347
  /**
2873
3348
  * Generated by orval v7.9.0 🍺
2874
3349
  * Do not edit manually.
2875
3350
  * AssemblyAI API
2876
- * AssemblyAI API
2877
- * OpenAPI spec version: 1.3.4
2878
- */
2879
- /**
2880
- * The duration of this transcript object's media file, in seconds
3351
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3352
+ * OpenAPI spec version: 1.0.0
2881
3353
  */
2882
- type TranscriptAudioDuration = number | null;
2883
3354
 
2884
3355
  /**
2885
- * Generated by orval v7.9.0 🍺
2886
- * Do not edit manually.
2887
- * AssemblyAI API
2888
- * AssemblyAI API
2889
- * OpenAPI spec version: 1.3.4
2890
- */
2891
- /**
2892
- * Whether Automatic Punctuation is enabled, either true or false
2893
- */
2894
- type TranscriptPunctuate = boolean | null;
3356
+ * The options for PII-redacted audio, if redact_pii_audio is enabled.
3357
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
2895
3358
 
2896
- /**
2897
- * Generated by orval v7.9.0 🍺
2898
- * Do not edit manually.
2899
- * AssemblyAI API
2900
- * AssemblyAI API
2901
- * OpenAPI spec version: 1.3.4
2902
- */
2903
- /**
2904
- * Whether Text Formatting is enabled, either true or false
2905
3359
  */
2906
- type TranscriptFormatText = boolean | null;
3360
+ interface TranscriptRedactPiiAudioOptions {
3361
+ /** By default, audio redaction provides redacted audio URLs only when speech is detected. However, if your use-case specifically requires redacted audio files even for silent audio files without any dialogue, you can opt to receive these URLs by setting this parameter to `true`. */
3362
+ return_redacted_no_speech_audio?: boolean;
3363
+ /** Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep. */
3364
+ override_audio_redaction_method?: TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod;
3365
+ }
2907
3366
 
2908
3367
  /**
2909
3368
  * Generated by orval v7.9.0 🍺
2910
3369
  * Do not edit manually.
2911
3370
  * AssemblyAI API
2912
- * AssemblyAI API
2913
- * OpenAPI spec version: 1.3.4
3371
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3372
+ * OpenAPI spec version: 1.0.0
2914
3373
  */
2915
3374
  /**
2916
- * Transcribe Filler Words, like "umm", in your media file; can be true or false
3375
+ * Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more details.
2917
3376
  */
2918
- type TranscriptDisfluencies = boolean | null;
3377
+ type RedactPiiAudioQuality = (typeof RedactPiiAudioQuality)[keyof typeof RedactPiiAudioQuality];
3378
+ declare const RedactPiiAudioQuality: {
3379
+ readonly mp3: "mp3";
3380
+ readonly wav: "wav";
3381
+ };
2919
3382
 
2920
3383
  /**
2921
3384
  * Generated by orval v7.9.0 🍺
2922
3385
  * Do not edit manually.
2923
3386
  * AssemblyAI API
2924
- * AssemblyAI API
2925
- * OpenAPI spec version: 1.3.4
3387
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3388
+ * OpenAPI spec version: 1.0.0
2926
3389
  */
3390
+
2927
3391
  /**
2928
- * Whether [Multichannel transcription](https://www.assemblyai.com/docs/models/speech-recognition#multichannel-transcription) was enabled in the transcription request, either true or false
3392
+ * The audio quality of the PII-redacted audio file, if redact_pii_audio is enabled.
3393
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
3394
+
2929
3395
  */
2930
- type TranscriptMultichannel = boolean | null;
3396
+ type TranscriptRedactPiiAudioQuality = RedactPiiAudioQuality | null;
2931
3397
 
2932
3398
  /**
2933
3399
  * Generated by orval v7.9.0 🍺
2934
3400
  * Do not edit manually.
2935
3401
  * AssemblyAI API
2936
- * AssemblyAI API
2937
- * OpenAPI spec version: 1.3.4
3402
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3403
+ * OpenAPI spec version: 1.0.0
2938
3404
  */
2939
3405
  /**
2940
- * The URL to which we send webhook requests.
2941
- We sends two different types of webhook requests.
2942
- One request when a transcript is completed or failed, and one request when the redacted audio is ready if redact_pii_audio is enabled.
2943
-
3406
+ * The type of PII to redact
2944
3407
  */
2945
- type TranscriptWebhookUrl = string | null;
3408
+ type PiiPolicy = (typeof PiiPolicy)[keyof typeof PiiPolicy];
3409
+ declare const PiiPolicy: {
3410
+ readonly account_number: "account_number";
3411
+ readonly banking_information: "banking_information";
3412
+ readonly blood_type: "blood_type";
3413
+ readonly credit_card_cvv: "credit_card_cvv";
3414
+ readonly credit_card_expiration: "credit_card_expiration";
3415
+ readonly credit_card_number: "credit_card_number";
3416
+ readonly date: "date";
3417
+ readonly date_interval: "date_interval";
3418
+ readonly date_of_birth: "date_of_birth";
3419
+ readonly drivers_license: "drivers_license";
3420
+ readonly drug: "drug";
3421
+ readonly duration: "duration";
3422
+ readonly email_address: "email_address";
3423
+ readonly event: "event";
3424
+ readonly filename: "filename";
3425
+ readonly gender_sexuality: "gender_sexuality";
3426
+ readonly healthcare_number: "healthcare_number";
3427
+ readonly injury: "injury";
3428
+ readonly ip_address: "ip_address";
3429
+ readonly language: "language";
3430
+ readonly location: "location";
3431
+ readonly marital_status: "marital_status";
3432
+ readonly medical_condition: "medical_condition";
3433
+ readonly medical_process: "medical_process";
3434
+ readonly money_amount: "money_amount";
3435
+ readonly nationality: "nationality";
3436
+ readonly number_sequence: "number_sequence";
3437
+ readonly occupation: "occupation";
3438
+ readonly organization: "organization";
3439
+ readonly passport_number: "passport_number";
3440
+ readonly password: "password";
3441
+ readonly person_age: "person_age";
3442
+ readonly person_name: "person_name";
3443
+ readonly phone_number: "phone_number";
3444
+ readonly physical_attribute: "physical_attribute";
3445
+ readonly political_affiliation: "political_affiliation";
3446
+ readonly religion: "religion";
3447
+ readonly statistics: "statistics";
3448
+ readonly time: "time";
3449
+ readonly url: "url";
3450
+ readonly us_social_security_number: "us_social_security_number";
3451
+ readonly username: "username";
3452
+ readonly vehicle_id: "vehicle_id";
3453
+ readonly zodiac_sign: "zodiac_sign";
3454
+ };
2946
3455
 
2947
3456
  /**
2948
3457
  * Generated by orval v7.9.0 🍺
2949
3458
  * Do not edit manually.
2950
3459
  * AssemblyAI API
2951
- * AssemblyAI API
2952
- * OpenAPI spec version: 1.3.4
3460
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3461
+ * OpenAPI spec version: 1.0.0
2953
3462
  */
3463
+
2954
3464
  /**
2955
- * The status code we received from your server when delivering the transcript completed or failed webhook request, if a webhook URL was provided
3465
+ * The list of PII Redaction policies that were enabled, if PII Redaction is enabled.
3466
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more information.
3467
+
2956
3468
  */
2957
- type TranscriptWebhookStatusCode = number | null;
3469
+ type TranscriptRedactPiiPolicies = PiiPolicy[] | null;
2958
3470
 
2959
3471
  /**
2960
3472
  * Generated by orval v7.9.0 🍺
2961
3473
  * Do not edit manually.
2962
3474
  * AssemblyAI API
2963
- * AssemblyAI API
2964
- * OpenAPI spec version: 1.3.4
3475
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3476
+ * OpenAPI spec version: 1.0.0
2965
3477
  */
2966
3478
  /**
2967
- * The header name to be sent with the transcript completed or failed webhook requests
3479
+ * The replacement logic for detected PII, can be `entity_name` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details.
2968
3480
  */
2969
- type TranscriptWebhookAuthHeaderName = string | null;
3481
+ type SubstitutionPolicy = (typeof SubstitutionPolicy)[keyof typeof SubstitutionPolicy];
3482
+ declare const SubstitutionPolicy: {
3483
+ readonly entity_name: "entity_name";
3484
+ readonly hash: "hash";
3485
+ };
2970
3486
 
2971
3487
  /**
2972
3488
  * Generated by orval v7.9.0 🍺
2973
3489
  * Do not edit manually.
2974
3490
  * AssemblyAI API
2975
- * AssemblyAI API
2976
- * OpenAPI spec version: 1.3.4
3491
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3492
+ * OpenAPI spec version: 1.0.0
2977
3493
  */
2978
3494
  /**
2979
- * Whether speed boost is enabled
2980
- * @deprecated
3495
+ * Whether [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) is enabled, can be true or false
2981
3496
  */
2982
- type TranscriptSpeedBoost = boolean | null;
3497
+ type TranscriptSentimentAnalysis = boolean | null;
2983
3498
 
2984
3499
  /**
2985
3500
  * Generated by orval v7.9.0 🍺
2986
3501
  * Do not edit manually.
2987
3502
  * AssemblyAI API
2988
- * AssemblyAI API
2989
- * OpenAPI spec version: 1.3.4
2990
- */
2991
- /**
2992
- * Either success, or unavailable in the rare case that the model failed
3503
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3504
+ * OpenAPI spec version: 1.0.0
2993
3505
  */
2994
- type AudioIntelligenceModelStatus = (typeof AudioIntelligenceModelStatus)[keyof typeof AudioIntelligenceModelStatus];
2995
- declare const AudioIntelligenceModelStatus: {
2996
- readonly success: "success";
2997
- readonly unavailable: "unavailable";
3506
+ type Sentiment = (typeof Sentiment)[keyof typeof Sentiment];
3507
+ declare const Sentiment: {
3508
+ readonly POSITIVE: "POSITIVE";
3509
+ readonly NEUTRAL: "NEUTRAL";
3510
+ readonly NEGATIVE: "NEGATIVE";
2998
3511
  };
2999
3512
 
3000
3513
  /**
3001
3514
  * Generated by orval v7.9.0 🍺
3002
3515
  * Do not edit manually.
3003
3516
  * AssemblyAI API
3004
- * AssemblyAI API
3005
- * OpenAPI spec version: 1.3.4
3517
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3518
+ * OpenAPI spec version: 1.0.0
3006
3519
  */
3007
3520
  /**
3008
- * Timestamp containing a start and end property in milliseconds
3521
+ * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3009
3522
  */
3010
- interface Timestamp {
3011
- /** The start time in milliseconds */
3012
- start: number;
3013
- /** The end time in milliseconds */
3014
- end: number;
3015
- }
3523
+ type SentimentAnalysisResultChannel = string | null;
3016
3524
 
3017
3525
  /**
3018
3526
  * Generated by orval v7.9.0 🍺
3019
3527
  * Do not edit manually.
3020
3528
  * AssemblyAI API
3021
- * AssemblyAI API
3022
- * OpenAPI spec version: 1.3.4
3529
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3530
+ * OpenAPI spec version: 1.0.0
3023
3531
  */
3024
-
3025
- interface AutoHighlightResult {
3026
- /** The total number of times the key phrase appears in the audio file */
3027
- count: number;
3028
- /**
3029
- * The total relevancy to the overall audio file of this key phrase - a greater number means more relevant
3030
- * @minimum 0
3031
- * @maximum 1
3032
- */
3033
- rank: number;
3034
- /** The text itself of the key phrase */
3035
- text: string;
3036
- /** The timestamp of the of the key phrase */
3037
- timestamps: Timestamp[];
3038
- }
3532
+ /**
3533
+ * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
3534
+ */
3535
+ type SentimentAnalysisResultSpeaker = string | null;
3039
3536
 
3040
3537
  /**
3041
3538
  * Generated by orval v7.9.0 🍺
3042
3539
  * Do not edit manually.
3043
3540
  * AssemblyAI API
3044
- * AssemblyAI API
3045
- * OpenAPI spec version: 1.3.4
3541
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3542
+ * OpenAPI spec version: 1.0.0
3046
3543
  */
3047
3544
 
3048
3545
  /**
3049
- * An array of results for the Key Phrases model, if it is enabled.
3050
- See [Key phrases](https://www.assemblyai.com/docs/models/key-phrases) for more information.
3051
-
3546
+ * The result of the Sentiment Analysis model
3052
3547
  */
3053
- interface AutoHighlightsResult {
3054
- /** The status of the Key Phrases model. Either success, or unavailable in the rare case that the model failed. */
3055
- status: AudioIntelligenceModelStatus;
3056
- /** A temporally-sequential array of Key Phrases */
3057
- results: AutoHighlightResult[];
3548
+ interface SentimentAnalysisResult$1 {
3549
+ /** The transcript of the sentence */
3550
+ text: string;
3551
+ /** The starting time, in milliseconds, of the sentence */
3552
+ start: number;
3553
+ /** The ending time, in milliseconds, of the sentence */
3554
+ end: number;
3555
+ /** The detected sentiment for the sentence, one of POSITIVE, NEUTRAL, NEGATIVE */
3556
+ sentiment: Sentiment;
3557
+ /** The confidence score for the detected sentiment of the sentence, from 0 to 1 */
3558
+ confidence: number;
3559
+ /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
3560
+ channel?: SentimentAnalysisResultChannel;
3561
+ /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
3562
+ speaker: SentimentAnalysisResultSpeaker;
3058
3563
  }
3059
3564
 
3060
3565
  /**
3061
3566
  * Generated by orval v7.9.0 🍺
3062
3567
  * Do not edit manually.
3063
3568
  * AssemblyAI API
3064
- * AssemblyAI API
3065
- * OpenAPI spec version: 1.3.4
3569
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3570
+ * OpenAPI spec version: 1.0.0
3066
3571
  */
3067
3572
 
3068
3573
  /**
3069
- * An array of results for the Key Phrases model, if it is enabled.
3070
- See [Key Phrases](https://www.assemblyai.com/docs/models/key-phrases) for more information.
3574
+ * An array of results for the Sentiment Analysis model, if it is enabled.
3575
+ See [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) for more information.
3071
3576
 
3072
3577
  */
3073
- type TranscriptAutoHighlightsResult = AutoHighlightsResult | null;
3578
+ type TranscriptSentimentAnalysisResults = SentimentAnalysisResult$1[] | null;
3074
3579
 
3075
3580
  /**
3076
3581
  * Generated by orval v7.9.0 🍺
3077
3582
  * Do not edit manually.
3078
3583
  * AssemblyAI API
3079
- * AssemblyAI API
3080
- * OpenAPI spec version: 1.3.4
3584
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3585
+ * OpenAPI spec version: 1.0.0
3081
3586
  */
3082
3587
  /**
3083
- * The point in time, in milliseconds, in the file at which the transcription was started
3588
+ * Whether [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, can be true or false
3084
3589
  */
3085
- type TranscriptAudioStartFrom = number | null;
3590
+ type TranscriptSpeakerLabels = boolean | null;
3086
3591
 
3087
3592
  /**
3088
3593
  * Generated by orval v7.9.0 🍺
3089
3594
  * Do not edit manually.
3090
3595
  * AssemblyAI API
3091
- * AssemblyAI API
3092
- * OpenAPI spec version: 1.3.4
3596
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3597
+ * OpenAPI spec version: 1.0.0
3093
3598
  */
3094
3599
  /**
3095
- * The point in time, in milliseconds, in the file at which the transcription was terminated
3600
+ * Tell the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details.
3096
3601
  */
3097
- type TranscriptAudioEndAt = number | null;
3602
+ type TranscriptSpeakersExpected = number | null;
3098
3603
 
3099
3604
  /**
3100
3605
  * Generated by orval v7.9.0 🍺
3101
3606
  * Do not edit manually.
3102
3607
  * AssemblyAI API
3103
- * AssemblyAI API
3104
- * OpenAPI spec version: 1.3.4
3608
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3609
+ * OpenAPI spec version: 1.0.0
3105
3610
  */
3106
3611
  /**
3107
- * The word boost parameter value
3612
+ * The speech model to use for the transcription. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models.
3108
3613
  */
3109
- type TranscriptBoostParamProperty = string | null;
3614
+ type SpeechModel = string;
3110
3615
 
3111
3616
  /**
3112
3617
  * Generated by orval v7.9.0 🍺
3113
3618
  * Do not edit manually.
3114
3619
  * AssemblyAI API
3115
- * AssemblyAI API
3116
- * OpenAPI spec version: 1.3.4
3620
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3621
+ * OpenAPI spec version: 1.0.0
3117
3622
  */
3623
+
3118
3624
  /**
3119
- * Whether [Profanity Filtering](https://www.assemblyai.com/docs/models/speech-recognition#profanity-filtering) is enabled, either true or false
3625
+ * List multiple speech models in priority order, allowing our system to automatically route your audio to the best available option. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models and routing behavior.
3626
+
3120
3627
  */
3121
- type TranscriptFilterProfanity = boolean | null;
3628
+ type TranscriptSpeechModels = SpeechModel[] | null;
3122
3629
 
3123
3630
  /**
3124
3631
  * Generated by orval v7.9.0 🍺
3125
3632
  * Do not edit manually.
3126
3633
  * AssemblyAI API
3127
- * AssemblyAI API
3128
- * OpenAPI spec version: 1.3.4
3634
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3635
+ * OpenAPI spec version: 1.0.0
3129
3636
  */
3130
3637
  /**
3131
- * Whether a redacted version of the audio file was generated,
3132
- either true or false. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
3638
+ * Defaults to null. Reject audio files that contain less than this fraction of speech.
3639
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
3133
3640
 
3134
3641
  */
3135
- type TranscriptRedactPiiAudio = boolean | null;
3642
+ type TranscriptSpeechThreshold = number | null;
3136
3643
 
3137
3644
  /**
3138
3645
  * Generated by orval v7.9.0 🍺
3139
3646
  * Do not edit manually.
3140
3647
  * AssemblyAI API
3141
- * AssemblyAI API
3142
- * OpenAPI spec version: 1.3.4
3143
- */
3144
- /**
3145
- * Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details.
3648
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3649
+ * OpenAPI spec version: 1.0.0
3146
3650
  */
3147
- type RedactPiiAudioQuality = (typeof RedactPiiAudioQuality)[keyof typeof RedactPiiAudioQuality];
3148
- declare const RedactPiiAudioQuality: {
3149
- readonly mp3: "mp3";
3150
- readonly wav: "wav";
3151
- };
3651
+ interface TranslationRequestBodyTranslation {
3652
+ /** List of target language codes (e.g., `["es", "de"]`). See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for supported languages. */
3653
+ target_languages: string[];
3654
+ /** Use formal language style. See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for more details. */
3655
+ formal?: boolean;
3656
+ /** When enabled with Speaker Labels, returns translated text in the utterances array. Each utterance will include a `translated_texts` key containing translations for each target language. */
3657
+ match_original_utterance?: boolean;
3658
+ }
3152
3659
 
3153
3660
  /**
3154
3661
  * Generated by orval v7.9.0 🍺
3155
3662
  * Do not edit manually.
3156
3663
  * AssemblyAI API
3157
- * AssemblyAI API
3158
- * OpenAPI spec version: 1.3.4
3664
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3665
+ * OpenAPI spec version: 1.0.0
3159
3666
  */
3160
3667
 
3161
3668
  /**
3162
- * The audio quality of the PII-redacted audio file, if redact_pii_audio is enabled.
3163
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
3164
-
3669
+ * Request body for [Translation](https://www.assemblyai.com/docs/speech-understanding/translation).
3165
3670
  */
3166
- type TranscriptRedactPiiAudioQuality = RedactPiiAudioQuality | null;
3671
+ interface TranslationRequestBody {
3672
+ translation: TranslationRequestBodyTranslation;
3673
+ }
3167
3674
 
3168
3675
  /**
3169
3676
  * Generated by orval v7.9.0 🍺
3170
3677
  * Do not edit manually.
3171
3678
  * AssemblyAI API
3172
- * AssemblyAI API
3173
- * OpenAPI spec version: 1.3.4
3679
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3680
+ * OpenAPI spec version: 1.0.0
3174
3681
  */
3175
3682
  /**
3176
- * The type of PII to redact
3683
+ * Type of speaker identification. See [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification) for details on each type.
3177
3684
  */
3178
- type PiiPolicy = (typeof PiiPolicy)[keyof typeof PiiPolicy];
3179
- declare const PiiPolicy: {
3180
- readonly account_number: "account_number";
3181
- readonly banking_information: "banking_information";
3182
- readonly blood_type: "blood_type";
3183
- readonly credit_card_cvv: "credit_card_cvv";
3184
- readonly credit_card_expiration: "credit_card_expiration";
3185
- readonly credit_card_number: "credit_card_number";
3186
- readonly date: "date";
3187
- readonly date_interval: "date_interval";
3188
- readonly date_of_birth: "date_of_birth";
3189
- readonly drivers_license: "drivers_license";
3190
- readonly drug: "drug";
3191
- readonly duration: "duration";
3192
- readonly email_address: "email_address";
3193
- readonly event: "event";
3194
- readonly filename: "filename";
3195
- readonly gender_sexuality: "gender_sexuality";
3196
- readonly healthcare_number: "healthcare_number";
3197
- readonly injury: "injury";
3198
- readonly ip_address: "ip_address";
3199
- readonly language: "language";
3200
- readonly location: "location";
3201
- readonly marital_status: "marital_status";
3202
- readonly medical_condition: "medical_condition";
3203
- readonly medical_process: "medical_process";
3204
- readonly money_amount: "money_amount";
3205
- readonly nationality: "nationality";
3206
- readonly number_sequence: "number_sequence";
3207
- readonly occupation: "occupation";
3208
- readonly organization: "organization";
3209
- readonly passport_number: "passport_number";
3210
- readonly password: "password";
3211
- readonly person_age: "person_age";
3212
- readonly person_name: "person_name";
3213
- readonly phone_number: "phone_number";
3214
- readonly physical_attribute: "physical_attribute";
3215
- readonly political_affiliation: "political_affiliation";
3216
- readonly religion: "religion";
3217
- readonly statistics: "statistics";
3218
- readonly time: "time";
3219
- readonly url: "url";
3220
- readonly us_social_security_number: "us_social_security_number";
3221
- readonly username: "username";
3222
- readonly vehicle_id: "vehicle_id";
3223
- readonly zodiac_sign: "zodiac_sign";
3685
+ type SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType = (typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType)[keyof typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType];
3686
+ declare const SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType: {
3687
+ readonly role: "role";
3688
+ readonly name: "name";
3224
3689
  };
3225
3690
 
3226
3691
  /**
3227
3692
  * Generated by orval v7.9.0 🍺
3228
3693
  * Do not edit manually.
3229
3694
  * AssemblyAI API
3230
- * AssemblyAI API
3231
- * OpenAPI spec version: 1.3.4
3695
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3696
+ * OpenAPI spec version: 1.0.0
3232
3697
  */
3698
+ interface SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems {
3699
+ /** The role of the speaker. Required when `speaker_type` is "role". */
3700
+ role?: string;
3701
+ /** The name of the speaker. Required when `speaker_type` is "name". */
3702
+ name?: string;
3703
+ /** A description of the speaker to help the model identify them based on conversational context. */
3704
+ description?: string;
3705
+ }
3233
3706
 
3234
3707
  /**
3235
- * The list of PII Redaction policies that were enabled, if PII Redaction is enabled.
3236
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
3237
-
3708
+ * Generated by orval v7.9.0 🍺
3709
+ * Do not edit manually.
3710
+ * AssemblyAI API
3711
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3712
+ * OpenAPI spec version: 1.0.0
3238
3713
  */
3239
- type TranscriptRedactPiiPolicies = PiiPolicy[] | null;
3714
+
3715
+ interface SpeakerIdentificationRequestBodySpeakerIdentification {
3716
+ /** Type of speaker identification. See [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification) for details on each type. */
3717
+ speaker_type: SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType;
3718
+ /** Required if speaker_type is "role". Each value must be 35 characters or less. */
3719
+ known_values?: string[];
3720
+ /** An array of speaker objects with metadata to improve identification accuracy. Each object should include a `role` or `name` (depending on `speaker_type`) and an optional `description` to help the model identify the speaker. You can also include any additional custom properties (e.g., `company`, `title`) to provide more context. Use this as an alternative to `known_values` when you want to provide additional context about each speaker. */
3721
+ speakers?: SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems[];
3722
+ }
3240
3723
 
3241
3724
  /**
3242
3725
  * Generated by orval v7.9.0 🍺
3243
3726
  * Do not edit manually.
3244
3727
  * AssemblyAI API
3245
- * AssemblyAI API
3246
- * OpenAPI spec version: 1.3.4
3728
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3729
+ * OpenAPI spec version: 1.0.0
3247
3730
  */
3731
+
3248
3732
  /**
3249
- * The replacement logic for detected PII, can be "entity_name" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details.
3733
+ * Request body for [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification).
3250
3734
  */
3251
- type SubstitutionPolicy = (typeof SubstitutionPolicy)[keyof typeof SubstitutionPolicy];
3252
- declare const SubstitutionPolicy: {
3253
- readonly entity_name: "entity_name";
3254
- readonly hash: "hash";
3255
- };
3735
+ interface SpeakerIdentificationRequestBody {
3736
+ speaker_identification: SpeakerIdentificationRequestBodySpeakerIdentification;
3737
+ }
3256
3738
 
3257
3739
  /**
3258
3740
  * Generated by orval v7.9.0 🍺
3259
3741
  * Do not edit manually.
3260
3742
  * AssemblyAI API
3261
- * AssemblyAI API
3262
- * OpenAPI spec version: 1.3.4
3263
- */
3264
- /**
3265
- * Whether [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, can be true or false
3743
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3744
+ * OpenAPI spec version: 1.0.0
3266
3745
  */
3267
- type TranscriptSpeakerLabels = boolean | null;
3746
+ interface CustomFormattingRequestBodyCustomFormatting {
3747
+ /** Date format pattern (e.g., `"mm/dd/yyyy"`). See [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting) for more details. */
3748
+ date?: string;
3749
+ /** Phone number format pattern (e.g., `"(xxx)xxx-xxxx"`). See [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting) for more details. */
3750
+ phone_number?: string;
3751
+ /** Email format pattern (e.g., `"username@domain.com"`). See [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting) for more details. */
3752
+ email?: string;
3753
+ }
3268
3754
 
3269
3755
  /**
3270
3756
  * Generated by orval v7.9.0 🍺
3271
3757
  * Do not edit manually.
3272
3758
  * AssemblyAI API
3273
- * AssemblyAI API
3274
- * OpenAPI spec version: 1.3.4
3759
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3760
+ * OpenAPI spec version: 1.0.0
3275
3761
  */
3762
+
3276
3763
  /**
3277
- * Tell the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details.
3764
+ * Request body for [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting).
3278
3765
  */
3279
- type TranscriptSpeakersExpected = number | null;
3766
+ interface CustomFormattingRequestBody {
3767
+ custom_formatting: CustomFormattingRequestBodyCustomFormatting;
3768
+ }
3280
3769
 
3281
3770
  /**
3282
3771
  * Generated by orval v7.9.0 🍺
3283
3772
  * Do not edit manually.
3284
3773
  * AssemblyAI API
3285
- * AssemblyAI API
3286
- * OpenAPI spec version: 1.3.4
3287
- */
3288
- /**
3289
- * Whether [Content Moderation](https://www.assemblyai.com/docs/models/content-moderation) is enabled, can be true or false
3774
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3775
+ * OpenAPI spec version: 1.0.0
3290
3776
  */
3291
- type TranscriptContentSafety = boolean | null;
3777
+
3778
+ type TranscriptSpeechUnderstandingRequest = TranslationRequestBody | SpeakerIdentificationRequestBody | CustomFormattingRequestBody;
3292
3779
 
3293
3780
  /**
3294
3781
  * Generated by orval v7.9.0 🍺
3295
3782
  * Do not edit manually.
3296
3783
  * AssemblyAI API
3297
- * AssemblyAI API
3298
- * OpenAPI spec version: 1.3.4
3784
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3785
+ * OpenAPI spec version: 1.0.0
3299
3786
  */
3300
- interface ContentSafetyLabel {
3301
- /** The label of the sensitive topic */
3302
- label: string;
3303
- /**
3304
- * The confidence score for the topic being discussed, from 0 to 1
3305
- * @minimum 0
3306
- * @maximum 1
3307
- */
3308
- confidence: number;
3309
- /**
3310
- * How severely the topic is discussed in the section, from 0 to 1
3311
- * @minimum 0
3312
- * @maximum 1
3313
- */
3314
- severity: number;
3787
+ interface TranslationResponseTranslation {
3788
+ status?: string;
3315
3789
  }
3316
3790
 
3317
3791
  /**
3318
3792
  * Generated by orval v7.9.0 🍺
3319
3793
  * Do not edit manually.
3320
3794
  * AssemblyAI API
3321
- * AssemblyAI API
3322
- * OpenAPI spec version: 1.3.4
3795
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3796
+ * OpenAPI spec version: 1.0.0
3323
3797
  */
3324
3798
 
3325
- interface ContentSafetyLabelResult {
3326
- /** The transcript of the section flagged by the Content Moderation model */
3327
- text: string;
3328
- /** An array of safety labels, one per sensitive topic that was detected in the section */
3329
- labels: ContentSafetyLabel[];
3330
- /** The sentence index at which the section begins */
3331
- sentences_idx_start: number;
3332
- /** The sentence index at which the section ends */
3333
- sentences_idx_end: number;
3334
- /** Timestamp information for the section */
3335
- timestamp: Timestamp;
3799
+ interface TranslationResponse {
3800
+ translation?: TranslationResponseTranslation;
3336
3801
  }
3337
3802
 
3338
3803
  /**
3339
3804
  * Generated by orval v7.9.0 🍺
3340
3805
  * Do not edit manually.
3341
3806
  * AssemblyAI API
3342
- * AssemblyAI API
3343
- * OpenAPI spec version: 1.3.4
3807
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3808
+ * OpenAPI spec version: 1.0.0
3344
3809
  */
3345
3810
  /**
3346
- * A summary of the Content Moderation confidence results for the entire audio file
3811
+ * A mapping of the original generic speaker labels (e.g., "A", "B") to the identified speaker names or roles.
3347
3812
  */
3348
- type ContentSafetyLabelsResultSummary = {
3349
- [key: string]: number;
3813
+ type SpeakerIdentificationResponseSpeakerIdentificationMapping = {
3814
+ [key: string]: string;
3350
3815
  };
3351
3816
 
3352
3817
  /**
3353
3818
  * Generated by orval v7.9.0 🍺
3354
3819
  * Do not edit manually.
3355
3820
  * AssemblyAI API
3356
- * AssemblyAI API
3357
- * OpenAPI spec version: 1.3.4
3821
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3822
+ * OpenAPI spec version: 1.0.0
3358
3823
  */
3359
- interface SeverityScoreSummary {
3360
- /**
3361
- * @minimum 0
3362
- * @maximum 1
3363
- */
3364
- low: number;
3365
- /**
3366
- * @minimum 0
3367
- * @maximum 1
3368
- */
3369
- medium: number;
3370
- /**
3371
- * @minimum 0
3372
- * @maximum 1
3373
- */
3374
- high: number;
3824
+
3825
+ interface SpeakerIdentificationResponseSpeakerIdentification {
3826
+ /** A mapping of the original generic speaker labels (e.g., "A", "B") to the identified speaker names or roles. */
3827
+ mapping?: SpeakerIdentificationResponseSpeakerIdentificationMapping;
3828
+ status?: string;
3375
3829
  }
3376
3830
 
3377
3831
  /**
3378
3832
  * Generated by orval v7.9.0 🍺
3379
3833
  * Do not edit manually.
3380
3834
  * AssemblyAI API
3381
- * AssemblyAI API
3382
- * OpenAPI spec version: 1.3.4
3835
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3836
+ * OpenAPI spec version: 1.0.0
3383
3837
  */
3384
3838
 
3385
- /**
3386
- * A summary of the Content Moderation severity results for the entire audio file
3387
- */
3388
- type ContentSafetyLabelsResultSeverityScoreSummary = {
3389
- [key: string]: SeverityScoreSummary;
3390
- };
3839
+ interface SpeakerIdentificationResponse {
3840
+ speaker_identification?: SpeakerIdentificationResponseSpeakerIdentification;
3841
+ }
3391
3842
 
3392
3843
  /**
3393
3844
  * Generated by orval v7.9.0 🍺
3394
3845
  * Do not edit manually.
3395
3846
  * AssemblyAI API
3396
- * AssemblyAI API
3397
- * OpenAPI spec version: 1.3.4
3398
- */
3399
-
3400
- /**
3401
- * An array of results for the Content Moderation model, if it is enabled.
3402
- See [Content moderation](https://www.assemblyai.com/docs/models/content-moderation) for more information.
3403
-
3847
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3848
+ * OpenAPI spec version: 1.0.0
3404
3849
  */
3405
- interface ContentSafetyLabelsResult {
3406
- /** The status of the Content Moderation model. Either success, or unavailable in the rare case that the model failed. */
3407
- status: AudioIntelligenceModelStatus;
3408
- /** An array of results for the Content Moderation model */
3409
- results: ContentSafetyLabelResult[];
3410
- /** A summary of the Content Moderation confidence results for the entire audio file */
3411
- summary: ContentSafetyLabelsResultSummary;
3412
- /** A summary of the Content Moderation severity results for the entire audio file */
3413
- severity_score_summary: ContentSafetyLabelsResultSeverityScoreSummary;
3414
- }
3850
+ type CustomFormattingResponseCustomFormattingMapping = {
3851
+ [key: string]: string;
3852
+ };
3415
3853
 
3416
3854
  /**
3417
3855
  * Generated by orval v7.9.0 🍺
3418
3856
  * Do not edit manually.
3419
3857
  * AssemblyAI API
3420
- * AssemblyAI API
3421
- * OpenAPI spec version: 1.3.4
3858
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3859
+ * OpenAPI spec version: 1.0.0
3422
3860
  */
3423
3861
 
3424
- /**
3425
- * An array of results for the Content Moderation model, if it is enabled.
3426
- See [Content moderation](https://www.assemblyai.com/docs/models/content-moderation) for more information.
3427
-
3428
- */
3429
- type TranscriptContentSafetyLabels = ContentSafetyLabelsResult | null;
3862
+ interface CustomFormattingResponseCustomFormatting {
3863
+ mapping?: CustomFormattingResponseCustomFormattingMapping;
3864
+ formatted_text?: string;
3865
+ }
3430
3866
 
3431
3867
  /**
3432
3868
  * Generated by orval v7.9.0 🍺
3433
3869
  * Do not edit manually.
3434
3870
  * AssemblyAI API
3435
- * AssemblyAI API
3436
- * OpenAPI spec version: 1.3.4
3437
- */
3438
- /**
3439
- * Whether [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) is enabled, can be true or false
3871
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3872
+ * OpenAPI spec version: 1.0.0
3440
3873
  */
3441
- type TranscriptIabCategories = boolean | null;
3874
+
3875
+ interface CustomFormattingResponse {
3876
+ custom_formatting?: CustomFormattingResponseCustomFormatting;
3877
+ }
3442
3878
 
3443
3879
  /**
3444
3880
  * Generated by orval v7.9.0 🍺
3445
3881
  * Do not edit manually.
3446
3882
  * AssemblyAI API
3447
- * AssemblyAI API
3448
- * OpenAPI spec version: 1.3.4
3883
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3884
+ * OpenAPI spec version: 1.0.0
3449
3885
  */
3450
- type TopicDetectionResultLabelsItem = {
3451
- /**
3452
- * How relevant the detected topic is of a detected topic
3453
- * @minimum 0
3454
- * @maximum 1
3455
- */
3456
- relevance: number;
3457
- /** The IAB taxonomical label for the label of the detected topic, where > denotes supertopic/subtopic relationship */
3458
- label: string;
3459
- };
3886
+
3887
+ type TranscriptSpeechUnderstandingResponse = TranslationResponse | SpeakerIdentificationResponse | CustomFormattingResponse;
3460
3888
 
3461
3889
  /**
3462
3890
  * Generated by orval v7.9.0 🍺
3463
3891
  * Do not edit manually.
3464
3892
  * AssemblyAI API
3465
- * AssemblyAI API
3466
- * OpenAPI spec version: 1.3.4
3893
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3894
+ * OpenAPI spec version: 1.0.0
3467
3895
  */
3468
3896
 
3469
3897
  /**
3470
- * The result of the topic detection model
3898
+ * Speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
3899
+
3471
3900
  */
3472
- interface TopicDetectionResult$1 {
3473
- /** The text in the transcript in which a detected topic occurs */
3474
- text: string;
3475
- /** An array of detected topics in the text */
3476
- labels?: TopicDetectionResultLabelsItem[];
3477
- timestamp?: Timestamp;
3901
+ interface TranscriptSpeechUnderstanding {
3902
+ request?: TranscriptSpeechUnderstandingRequest;
3903
+ response?: TranscriptSpeechUnderstandingResponse;
3478
3904
  }
3479
3905
 
3480
3906
  /**
3481
3907
  * Generated by orval v7.9.0 🍺
3482
3908
  * Do not edit manually.
3483
3909
  * AssemblyAI API
3484
- * AssemblyAI API
3485
- * OpenAPI spec version: 1.3.4
3910
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3911
+ * OpenAPI spec version: 1.0.0
3486
3912
  */
3487
3913
  /**
3488
- * The overall relevance of topic to the entire audio file
3914
+ * The status of your transcript. Possible values are queued, processing, completed, or error.
3489
3915
  */
3490
- type TopicDetectionModelResultSummary = {
3491
- [key: string]: number;
3916
+ type TranscriptStatus = (typeof TranscriptStatus)[keyof typeof TranscriptStatus];
3917
+ declare const TranscriptStatus: {
3918
+ readonly queued: "queued";
3919
+ readonly processing: "processing";
3920
+ readonly completed: "completed";
3921
+ readonly error: "error";
3492
3922
  };
3493
3923
 
3494
3924
  /**
3495
3925
  * Generated by orval v7.9.0 🍺
3496
3926
  * Do not edit manually.
3497
3927
  * AssemblyAI API
3498
- * AssemblyAI API
3499
- * OpenAPI spec version: 1.3.4
3928
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3929
+ * OpenAPI spec version: 1.0.0
3500
3930
  */
3501
-
3502
3931
  /**
3503
- * The result of the Topic Detection model, if it is enabled.
3504
- See [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) for more information.
3505
-
3932
+ * The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
3506
3933
  */
3507
- interface TopicDetectionModelResult {
3508
- /** The status of the Topic Detection model. Either success, or unavailable in the rare case that the model failed. */
3509
- status: AudioIntelligenceModelStatus;
3510
- /** An array of results for the Topic Detection model */
3511
- results: TopicDetectionResult$1[];
3512
- /** The overall relevance of topic to the entire audio file */
3513
- summary: TopicDetectionModelResultSummary;
3514
- }
3934
+ type TranscriptSummary = string | null;
3515
3935
 
3516
3936
  /**
3517
3937
  * Generated by orval v7.9.0 🍺
3518
3938
  * Do not edit manually.
3519
3939
  * AssemblyAI API
3520
- * AssemblyAI API
3521
- * OpenAPI spec version: 1.3.4
3940
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3941
+ * OpenAPI spec version: 1.0.0
3522
3942
  */
3523
-
3524
3943
  /**
3525
- * The result of the Topic Detection model, if it is enabled.
3526
- See [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) for more information.
3944
+ * The Summarization model used to generate the summary,
3945
+ if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-models) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
3527
3946
 
3528
3947
  */
3529
- type TranscriptIabCategoriesResult = TopicDetectionModelResult | null;
3948
+ type TranscriptSummaryModel = string | null;
3530
3949
 
3531
3950
  /**
3532
3951
  * Generated by orval v7.9.0 🍺
3533
3952
  * Do not edit manually.
3534
3953
  * AssemblyAI API
3535
- * AssemblyAI API
3536
- * OpenAPI spec version: 1.3.4
3954
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3955
+ * OpenAPI spec version: 1.0.0
3537
3956
  */
3538
3957
  /**
3539
- * Object containing words or phrases to replace, and the word or phrase to replace with
3958
+ * The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-types) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
3540
3959
  */
3541
- interface TranscriptCustomSpelling {
3542
- /** Words or phrases to replace */
3543
- from: string[];
3544
- /** Word to replace with */
3545
- to: string;
3546
- }
3960
+ type TranscriptSummaryType = string | null;
3547
3961
 
3548
3962
  /**
3549
3963
  * Generated by orval v7.9.0 🍺
3550
3964
  * Do not edit manually.
3551
3965
  * AssemblyAI API
3552
- * AssemblyAI API
3553
- * OpenAPI spec version: 1.3.4
3966
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3967
+ * OpenAPI spec version: 1.0.0
3554
3968
  */
3555
-
3556
3969
  /**
3557
- * Customize how words are spelled and formatted using to and from values
3970
+ * Whether [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) were removed from the transcript text.
3971
+
3972
+ Note: This parameter is only supported for the Universal-3 Pro model.
3973
+
3558
3974
  */
3559
- type TranscriptCustomSpellingProperty = TranscriptCustomSpelling[] | null;
3975
+ type TranscriptRemoveAudioTags = (typeof TranscriptRemoveAudioTags)[keyof typeof TranscriptRemoveAudioTags];
3976
+ declare const TranscriptRemoveAudioTags: {
3977
+ readonly all: "all";
3978
+ };
3560
3979
 
3561
3980
  /**
3562
3981
  * Generated by orval v7.9.0 🍺
3563
3982
  * Do not edit manually.
3564
3983
  * AssemblyAI API
3565
- * AssemblyAI API
3566
- * OpenAPI spec version: 1.3.4
3984
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
3985
+ * OpenAPI spec version: 1.0.0
3567
3986
  */
3987
+
3568
3988
  /**
3569
- * Whether [Auto Chapters](https://www.assemblyai.com/docs/models/auto-chapters) is enabled, can be true or false
3989
+ * Whether [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) were removed from the transcript text.
3990
+
3991
+ Note: This parameter is only supported for the Universal-3 Pro model.
3992
+
3570
3993
  */
3571
- type TranscriptAutoChapters = boolean | null;
3994
+ type TranscriptRemoveAudioTagsProperty = TranscriptRemoveAudioTags | null;
3572
3995
 
3573
3996
  /**
3574
3997
  * Generated by orval v7.9.0 🍺
3575
3998
  * Do not edit manually.
3576
3999
  * AssemblyAI API
3577
- * AssemblyAI API
3578
- * OpenAPI spec version: 1.3.4
4000
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4001
+ * OpenAPI spec version: 1.0.0
3579
4002
  */
3580
4003
  /**
3581
- * Chapter of the audio file
4004
+ * The temperature that was used for the model's response. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for more details.
4005
+
4006
+ Note: This parameter can only be used with the Universal-3 Pro model.
4007
+
3582
4008
  */
3583
- interface Chapter$1 {
3584
- /** An ultra-short summary (just a few words) of the content spoken in the chapter */
3585
- gist: string;
3586
- /** A single sentence summary of the content spoken during the chapter */
3587
- headline: string;
3588
- /** A one paragraph summary of the content spoken during the chapter */
3589
- summary: string;
3590
- /** The starting time, in milliseconds, for the chapter */
3591
- start: number;
3592
- /** The starting time, in milliseconds, for the chapter */
3593
- end: number;
3594
- }
4009
+ type TranscriptTemperature = number | null;
3595
4010
 
3596
4011
  /**
3597
4012
  * Generated by orval v7.9.0 🍺
3598
4013
  * Do not edit manually.
3599
4014
  * AssemblyAI API
3600
- * AssemblyAI API
3601
- * OpenAPI spec version: 1.3.4
4015
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4016
+ * OpenAPI spec version: 1.0.0
3602
4017
  */
3603
-
3604
4018
  /**
3605
- * An array of temporally sequential chapters for the audio file
4019
+ * The textual transcript of your media file
3606
4020
  */
3607
- type TranscriptChapters = Chapter$1[] | null;
4021
+ type TranscriptText = string | null;
3608
4022
 
3609
4023
  /**
3610
4024
  * Generated by orval v7.9.0 🍺
3611
4025
  * Do not edit manually.
3612
4026
  * AssemblyAI API
3613
- * AssemblyAI API
3614
- * OpenAPI spec version: 1.3.4
4027
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4028
+ * OpenAPI spec version: 1.0.0
3615
4029
  */
3616
4030
  /**
3617
- * The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
4031
+ * True while a request is throttled and false when a request is no longer throttled
3618
4032
  */
3619
- type TranscriptSummaryType = string | null;
4033
+ type TranscriptThrottled = boolean | null;
3620
4034
 
3621
4035
  /**
3622
4036
  * Generated by orval v7.9.0 🍺
3623
4037
  * Do not edit manually.
3624
4038
  * AssemblyAI API
3625
- * AssemblyAI API
3626
- * OpenAPI spec version: 1.3.4
4039
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4040
+ * OpenAPI spec version: 1.0.0
3627
4041
  */
3628
4042
  /**
3629
- * The Summarization model used to generate the summary,
3630
- if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
3631
-
4043
+ * The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3632
4044
  */
3633
- type TranscriptSummaryModel = string | null;
4045
+ type TranscriptWordChannel = string | null;
3634
4046
 
3635
4047
  /**
3636
4048
  * Generated by orval v7.9.0 🍺
3637
4049
  * Do not edit manually.
3638
4050
  * AssemblyAI API
3639
- * AssemblyAI API
3640
- * OpenAPI spec version: 1.3.4
4051
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4052
+ * OpenAPI spec version: 1.0.0
3641
4053
  */
3642
4054
  /**
3643
- * The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
4055
+ * The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
3644
4056
  */
3645
- type TranscriptSummary = string | null;
4057
+ type TranscriptWordSpeaker = string | null;
3646
4058
 
3647
4059
  /**
3648
4060
  * Generated by orval v7.9.0 🍺
3649
4061
  * Do not edit manually.
3650
4062
  * AssemblyAI API
3651
- * AssemblyAI API
3652
- * OpenAPI spec version: 1.3.4
3653
- */
3654
- /**
3655
- * Whether custom topics is enabled, either true or false
3656
- * @deprecated
4063
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4064
+ * OpenAPI spec version: 1.0.0
3657
4065
  */
3658
- type TranscriptCustomTopics = boolean | null;
4066
+
4067
+ interface TranscriptWord {
4068
+ /** The confidence score for the transcript of this word */
4069
+ confidence: number;
4070
+ /** The starting time, in milliseconds, for the word */
4071
+ start: number;
4072
+ /** The ending time, in milliseconds, for the word */
4073
+ end: number;
4074
+ /** The text of the word */
4075
+ text: string;
4076
+ /** The channel of the word. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
4077
+ channel?: TranscriptWordChannel;
4078
+ /** The speaker of the word if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
4079
+ speaker: TranscriptWordSpeaker;
4080
+ }
3659
4081
 
3660
4082
  /**
3661
4083
  * Generated by orval v7.9.0 🍺
3662
4084
  * Do not edit manually.
3663
4085
  * AssemblyAI API
3664
- * AssemblyAI API
3665
- * OpenAPI spec version: 1.3.4
4086
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4087
+ * OpenAPI spec version: 1.0.0
3666
4088
  */
3667
4089
  /**
3668
- * Whether [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) is enabled, can be true or false
4090
+ * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
3669
4091
  */
3670
- type TranscriptSentimentAnalysis = boolean | null;
4092
+ type TranscriptUtteranceChannel = string | null;
3671
4093
 
3672
4094
  /**
3673
4095
  * Generated by orval v7.9.0 🍺
3674
4096
  * Do not edit manually.
3675
4097
  * AssemblyAI API
3676
- * AssemblyAI API
3677
- * OpenAPI spec version: 1.3.4
4098
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4099
+ * OpenAPI spec version: 1.0.0
3678
4100
  */
3679
- type Sentiment = (typeof Sentiment)[keyof typeof Sentiment];
3680
- declare const Sentiment: {
3681
- readonly POSITIVE: "POSITIVE";
3682
- readonly NEUTRAL: "NEUTRAL";
3683
- readonly NEGATIVE: "NEGATIVE";
4101
+ /**
4102
+ * Translations keyed by language code (e.g., `{"es": "Texto traducido", "de": "Übersetzter Text"}`). Only present when `match_original_utterance` is enabled with translation.
4103
+ */
4104
+ type TranscriptUtteranceTranslatedTexts = {
4105
+ [key: string]: string;
3684
4106
  };
3685
4107
 
3686
4108
  /**
3687
4109
  * Generated by orval v7.9.0 🍺
3688
4110
  * Do not edit manually.
3689
4111
  * AssemblyAI API
3690
- * AssemblyAI API
3691
- * OpenAPI spec version: 1.3.4
3692
- */
3693
- /**
3694
- * The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
4112
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4113
+ * OpenAPI spec version: 1.0.0
3695
4114
  */
3696
- type SentimentAnalysisResultChannel = string | null;
4115
+
4116
+ interface TranscriptUtterance {
4117
+ /** The confidence score for the transcript of this utterance */
4118
+ confidence: number;
4119
+ /** The starting time, in milliseconds, of the utterance in the audio file */
4120
+ start: number;
4121
+ /** The ending time, in milliseconds, of the utterance in the audio file */
4122
+ end: number;
4123
+ /** The text for this utterance */
4124
+ text: string;
4125
+ /** The words in the utterance. */
4126
+ words: TranscriptWord[];
4127
+ /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
4128
+ channel?: TranscriptUtteranceChannel;
4129
+ /** The speaker of this utterance, where each speaker is assigned a sequential capital letter - e.g. "A" for Speaker A, "B" for Speaker B, etc. */
4130
+ speaker: string;
4131
+ /** Translations keyed by language code (e.g., `{"es": "Texto traducido", "de": "Übersetzter Text"}`). Only present when `match_original_utterance` is enabled with translation. */
4132
+ translated_texts?: TranscriptUtteranceTranslatedTexts;
4133
+ }
3697
4134
 
3698
4135
  /**
3699
4136
  * Generated by orval v7.9.0 🍺
3700
4137
  * Do not edit manually.
3701
4138
  * AssemblyAI API
3702
- * AssemblyAI API
3703
- * OpenAPI spec version: 1.3.4
4139
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4140
+ * OpenAPI spec version: 1.0.0
3704
4141
  */
4142
+
3705
4143
  /**
3706
- * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null
4144
+ * When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
4145
+ See [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) for more information.
4146
+
3707
4147
  */
3708
- type SentimentAnalysisResultSpeaker = string | null;
4148
+ type TranscriptUtterances = TranscriptUtterance[] | null;
3709
4149
 
3710
4150
  /**
3711
4151
  * Generated by orval v7.9.0 🍺
3712
4152
  * Do not edit manually.
3713
4153
  * AssemblyAI API
3714
- * AssemblyAI API
3715
- * OpenAPI spec version: 1.3.4
4154
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4155
+ * OpenAPI spec version: 1.0.0
3716
4156
  */
3717
-
3718
4157
  /**
3719
- * The result of the Sentiment Analysis model
4158
+ * The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests
3720
4159
  */
3721
- interface SentimentAnalysisResult$1 {
3722
- /** The transcript of the sentence */
3723
- text: string;
3724
- /** The starting time, in milliseconds, of the sentence */
3725
- start: number;
3726
- /** The ending time, in milliseconds, of the sentence */
3727
- end: number;
3728
- /** The detected sentiment for the sentence, one of POSITIVE, NEUTRAL, NEGATIVE */
3729
- sentiment: Sentiment;
3730
- /**
3731
- * The confidence score for the detected sentiment of the sentence, from 0 to 1
3732
- * @minimum 0
3733
- * @maximum 1
3734
- */
3735
- confidence: number;
3736
- /** The channel of this utterance. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
3737
- channel?: SentimentAnalysisResultChannel;
3738
- /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null */
3739
- speaker: SentimentAnalysisResultSpeaker;
3740
- }
4160
+ type TranscriptWebhookAuthHeaderName = string | null;
3741
4161
 
3742
4162
  /**
3743
4163
  * Generated by orval v7.9.0 🍺
3744
4164
  * Do not edit manually.
3745
4165
  * AssemblyAI API
3746
- * AssemblyAI API
3747
- * OpenAPI spec version: 1.3.4
4166
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4167
+ * OpenAPI spec version: 1.0.0
3748
4168
  */
3749
-
3750
4169
  /**
3751
- * An array of results for the Sentiment Analysis model, if it is enabled.
3752
- See [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) for more information.
3753
-
4170
+ * The status code we received from your server when delivering the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) request, if a webhook URL was provided
3754
4171
  */
3755
- type TranscriptSentimentAnalysisResults = SentimentAnalysisResult$1[] | null;
4172
+ type TranscriptWebhookStatusCode = number | null;
3756
4173
 
3757
4174
  /**
3758
4175
  * Generated by orval v7.9.0 🍺
3759
4176
  * Do not edit manually.
3760
4177
  * AssemblyAI API
3761
- * AssemblyAI API
3762
- * OpenAPI spec version: 1.3.4
4178
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4179
+ * OpenAPI spec version: 1.0.0
3763
4180
  */
3764
4181
  /**
3765
- * Whether [Entity Detection](https://www.assemblyai.com/docs/models/entity-detection) is enabled, can be true or false
4182
+ * The URL to which we send [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests.
4183
+
3766
4184
  */
3767
- type TranscriptEntityDetection = boolean | null;
4185
+ type TranscriptWebhookUrl = string | null;
3768
4186
 
3769
4187
  /**
3770
4188
  * Generated by orval v7.9.0 🍺
3771
4189
  * Do not edit manually.
3772
4190
  * AssemblyAI API
3773
- * AssemblyAI API
3774
- * OpenAPI spec version: 1.3.4
4191
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4192
+ * OpenAPI spec version: 1.0.0
3775
4193
  */
4194
+
3776
4195
  /**
3777
- * The type of entity for the detected entity
4196
+ * An array of temporally-sequential word objects, one for each word in the transcript.
4197
+
3778
4198
  */
3779
- type EntityType = (typeof EntityType)[keyof typeof EntityType];
3780
- declare const EntityType: {
3781
- readonly account_number: "account_number";
3782
- readonly banking_information: "banking_information";
3783
- readonly blood_type: "blood_type";
3784
- readonly credit_card_cvv: "credit_card_cvv";
3785
- readonly credit_card_expiration: "credit_card_expiration";
3786
- readonly credit_card_number: "credit_card_number";
3787
- readonly date: "date";
3788
- readonly date_interval: "date_interval";
3789
- readonly date_of_birth: "date_of_birth";
3790
- readonly drivers_license: "drivers_license";
3791
- readonly drug: "drug";
3792
- readonly duration: "duration";
3793
- readonly email_address: "email_address";
3794
- readonly event: "event";
3795
- readonly filename: "filename";
3796
- readonly gender_sexuality: "gender_sexuality";
3797
- readonly healthcare_number: "healthcare_number";
3798
- readonly injury: "injury";
3799
- readonly ip_address: "ip_address";
3800
- readonly language: "language";
3801
- readonly location: "location";
3802
- readonly marital_status: "marital_status";
3803
- readonly medical_condition: "medical_condition";
3804
- readonly medical_process: "medical_process";
3805
- readonly money_amount: "money_amount";
3806
- readonly nationality: "nationality";
3807
- readonly number_sequence: "number_sequence";
3808
- readonly occupation: "occupation";
3809
- readonly organization: "organization";
3810
- readonly passport_number: "passport_number";
3811
- readonly password: "password";
3812
- readonly person_age: "person_age";
3813
- readonly person_name: "person_name";
3814
- readonly phone_number: "phone_number";
3815
- readonly physical_attribute: "physical_attribute";
3816
- readonly political_affiliation: "political_affiliation";
3817
- readonly religion: "religion";
3818
- readonly statistics: "statistics";
3819
- readonly time: "time";
3820
- readonly url: "url";
3821
- readonly us_social_security_number: "us_social_security_number";
3822
- readonly username: "username";
3823
- readonly vehicle_id: "vehicle_id";
3824
- readonly zodiac_sign: "zodiac_sign";
3825
- };
4199
+ type TranscriptWords = TranscriptWord[] | null;
3826
4200
 
3827
4201
  /**
3828
4202
  * Generated by orval v7.9.0 🍺
3829
4203
  * Do not edit manually.
3830
4204
  * AssemblyAI API
3831
- * AssemblyAI API
3832
- * OpenAPI spec version: 1.3.4
4205
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4206
+ * OpenAPI spec version: 1.0.0
3833
4207
  */
3834
-
3835
4208
  /**
3836
- * A detected entity
4209
+ * This parameter does not currently have any functionality attached to it.
3837
4210
  */
3838
- interface Entity {
3839
- /** The type of entity for the detected entity */
3840
- entity_type: EntityType;
3841
- /** The text for the detected entity */
3842
- text: string;
3843
- /** The starting time, in milliseconds, at which the detected entity appears in the audio file */
3844
- start: number;
3845
- /** The ending time, in milliseconds, for the detected entity in the audio file */
3846
- end: number;
3847
- }
4211
+ type TranscriptCustomTopics = boolean | null;
3848
4212
 
3849
4213
  /**
3850
4214
  * Generated by orval v7.9.0 🍺
3851
4215
  * Do not edit manually.
3852
4216
  * AssemblyAI API
3853
- * AssemblyAI API
3854
- * OpenAPI spec version: 1.3.4
4217
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4218
+ * OpenAPI spec version: 1.0.0
3855
4219
  */
3856
4220
 
3857
4221
  /**
3858
- * An array of results for the Entity Detection model, if it is enabled.
3859
- See [Entity detection](https://www.assemblyai.com/docs/models/entity-detection) for more information.
4222
+ * This parameter has been replaced with the `speech_models` parameter, learn more about the `speech_models` parameter [here](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model).
3860
4223
 
3861
4224
  */
3862
- type TranscriptEntities = Entity[] | null;
4225
+ type TranscriptSpeechModel = SpeechModel | null;
3863
4226
 
3864
4227
  /**
3865
4228
  * Generated by orval v7.9.0 🍺
3866
4229
  * Do not edit manually.
3867
4230
  * AssemblyAI API
3868
- * AssemblyAI API
3869
- * OpenAPI spec version: 1.3.4
4231
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4232
+ * OpenAPI spec version: 1.0.0
3870
4233
  */
3871
4234
  /**
3872
- * Defaults to null. Reject audio files that contain less than this fraction of speech.
3873
- Valid values are in the range [0, 1] inclusive.
3874
-
3875
- * @minimum 0
3876
- * @maximum 1
4235
+ * This parameter does not currently have any functionality attached to it.
3877
4236
  */
3878
- type TranscriptSpeechThreshold = number | null;
4237
+ type TranscriptSpeedBoost = boolean | null;
3879
4238
 
3880
4239
  /**
3881
4240
  * Generated by orval v7.9.0 🍺
3882
4241
  * Do not edit manually.
3883
4242
  * AssemblyAI API
3884
- * AssemblyAI API
3885
- * OpenAPI spec version: 1.3.4
4243
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4244
+ * OpenAPI spec version: 1.0.0
3886
4245
  */
3887
4246
  /**
3888
- * True while a request is throttled and false when a request is no longer throttled
4247
+ * Translated text keyed by language code. See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for more details.
3889
4248
  */
3890
- type TranscriptThrottled = boolean | null;
4249
+ interface TranscriptTranslatedTexts {
4250
+ /** Translated text for this language code */
4251
+ language_code?: string;
4252
+ }
3891
4253
 
3892
4254
  /**
3893
4255
  * Generated by orval v7.9.0 🍺
3894
4256
  * Do not edit manually.
3895
4257
  * AssemblyAI API
3896
- * AssemblyAI API
3897
- * OpenAPI spec version: 1.3.4
4258
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4259
+ * OpenAPI spec version: 1.0.0
3898
4260
  */
3899
4261
 
3900
4262
  /**
3901
4263
  * A transcript object
3902
4264
  */
3903
4265
  interface Transcript {
3904
- /** The unique identifier of your transcript */
3905
- id: string;
4266
+ /** The number of audio channels in the audio file. This is only present when [multichannel](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) is enabled. */
4267
+ audio_channels?: number;
4268
+ /** The duration of this transcript object's media file, in seconds */
4269
+ audio_duration?: TranscriptAudioDuration;
4270
+ /** The point in time, in milliseconds, in the file at which the transcription was terminated. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
4271
+ audio_end_at?: TranscriptAudioEndAt;
4272
+ /** The point in time, in milliseconds, in the file at which the transcription was started. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
4273
+ audio_start_from?: TranscriptAudioStartFrom;
3906
4274
  /** The URL of the media that was transcribed */
3907
4275
  audio_url: string;
3908
- /** The status of your transcript. Possible values are queued, processing, completed, or error. */
3909
- status: TranscriptStatus;
3910
- /** The language of your audio file.
3911
- Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
3912
- The default value is 'en_us'.
3913
- */
3914
- language_code?: TranscriptLanguageCodeProperty;
3915
- /** Whether [Automatic language detection](https://www.assemblyai.com/docs/models/speech-recognition#automatic-language-detection) is enabled, either true or false */
3916
- language_detection?: TranscriptLanguageDetection;
3917
- /**
3918
- * The confidence threshold for the automatically detected language.
3919
- An error will be returned if the language confidence is below this threshold.
4276
+ /** Whether [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) is enabled, can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible chapter summaries. See the [updated Auto Chapters page](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for details.
3920
4277
 
3921
- * @minimum 0
3922
- * @maximum 1
3923
- */
3924
- language_confidence_threshold: TranscriptLanguageConfidenceThreshold;
3925
- /**
3926
- * The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence)
3927
- * @minimum 0
3928
- * @maximum 1
3929
- */
3930
- language_confidence: TranscriptLanguageConfidence;
3931
- /** The speech model used for the transcription. When `null`, the default model is used. */
3932
- speech_model: TranscriptSpeechModel;
3933
- /** The textual transcript of your media file */
3934
- text?: TranscriptText;
3935
- /** An array of temporally-sequential word objects, one for each word in the transcript.
3936
- See [Speech recognition](https://www.assemblyai.com/docs/models/speech-recognition) for more information.
3937
- */
3938
- words?: TranscriptWords;
3939
- /** When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
3940
- See [Speaker diarization](https://www.assemblyai.com/docs/speech-to-text/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/speech-to-text/speech-recognition#multichannel-transcription) for more information.
3941
- */
3942
- utterances?: TranscriptUtterances;
3943
- /**
3944
- * The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence)
3945
- * @minimum 0
3946
- * @maximum 1
3947
- */
3948
- confidence?: TranscriptConfidence;
3949
- /** The duration of this transcript object's media file, in seconds */
3950
- audio_duration?: TranscriptAudioDuration;
3951
- /** Whether Automatic Punctuation is enabled, either true or false */
3952
- punctuate?: TranscriptPunctuate;
3953
- /** Whether Text Formatting is enabled, either true or false */
3954
- format_text?: TranscriptFormatText;
3955
- /** Transcribe Filler Words, like "umm", in your media file; can be true or false */
3956
- disfluencies?: TranscriptDisfluencies;
3957
- /** Whether [Multichannel transcription](https://www.assemblyai.com/docs/models/speech-recognition#multichannel-transcription) was enabled in the transcription request, either true or false */
3958
- multichannel?: TranscriptMultichannel;
3959
- /** The number of audio channels in the audio file. This is only present when multichannel is enabled. */
3960
- audio_channels?: number;
3961
- /** The URL to which we send webhook requests.
3962
- We sends two different types of webhook requests.
3963
- One request when a transcript is completed or failed, and one request when the redacted audio is ready if redact_pii_audio is enabled.
4278
+ Note: This parameter is only supported for the Universal-2 model.
3964
4279
  */
3965
- webhook_url?: TranscriptWebhookUrl;
3966
- /** The status code we received from your server when delivering the transcript completed or failed webhook request, if a webhook URL was provided */
3967
- webhook_status_code?: TranscriptWebhookStatusCode;
3968
- /** Whether webhook authentication details were provided */
3969
- webhook_auth: boolean;
3970
- /** The header name to be sent with the transcript completed or failed webhook requests */
3971
- webhook_auth_header_name?: TranscriptWebhookAuthHeaderName;
3972
- /**
3973
- * Whether speed boost is enabled
3974
- * @deprecated
3975
- */
3976
- speed_boost?: TranscriptSpeedBoost;
3977
- /** Whether Key Phrases is enabled, either true or false */
4280
+ auto_chapters?: TranscriptAutoChapters;
4281
+ /** Whether [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) is enabled, either true or false */
3978
4282
  auto_highlights: boolean;
3979
4283
  /** An array of results for the Key Phrases model, if it is enabled.
3980
- See [Key Phrases](https://www.assemblyai.com/docs/models/key-phrases) for more information.
4284
+ See [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases) for more information.
3981
4285
  */
3982
4286
  auto_highlights_result?: TranscriptAutoHighlightsResult;
3983
- /** The point in time, in milliseconds, in the file at which the transcription was started */
3984
- audio_start_from?: TranscriptAudioStartFrom;
3985
- /** The point in time, in milliseconds, in the file at which the transcription was terminated */
3986
- audio_end_at?: TranscriptAudioEndAt;
3987
- /**
3988
- * The list of custom vocabulary to boost transcription probability for
3989
- * @deprecated
4287
+ /** An array of temporally sequential chapters for the audio file. See [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for more information. */
4288
+ chapters?: TranscriptChapters;
4289
+ /** The confidence score for the transcript, between 0.0 (low confidence) and 1.0 (high confidence) */
4290
+ confidence?: TranscriptConfidence;
4291
+ /** Whether [Content Moderation](https://www.assemblyai.com/docs/content-moderation) is enabled, can be true or false */
4292
+ content_safety?: TranscriptContentSafety;
4293
+ /** An array of results for the Content Moderation model, if it is enabled.
4294
+ See [Content moderation](https://www.assemblyai.com/docs/content-moderation) for more information.
4295
+ */
4296
+ content_safety_labels?: TranscriptContentSafetyLabels;
4297
+ /** Customize how words are spelled and formatted using to and from values. See [Custom Spelling](https://www.assemblyai.com/docs/pre-recorded-audio/custom-spelling) for more details. */
4298
+ custom_spelling?: TranscriptCustomSpellingProperty;
4299
+ /** Transcribe [Filler Words](https://www.assemblyai.com/docs/pre-recorded-audio/filler-words), like "umm", in your media file; can be true or false */
4300
+ disfluencies?: TranscriptDisfluencies;
4301
+ /** The domain-specific model applied to the transcript. When set to `"medical-v1"`, [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) was used to improve accuracy for medical terminology.
3990
4302
  */
3991
- word_boost?: string[];
3992
- /** The word boost parameter value */
3993
- boost_param?: TranscriptBoostParamProperty;
3994
- /** Whether [Profanity Filtering](https://www.assemblyai.com/docs/models/speech-recognition#profanity-filtering) is enabled, either true or false */
4303
+ domain?: TranscriptDomain;
4304
+ /** An array of results for the Entity Detection model, if it is enabled.
4305
+ See [Entity detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) for more information.
4306
+ */
4307
+ entities?: TranscriptEntities;
4308
+ /** Whether [Entity Detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection) is enabled, can be true or false */
4309
+ entity_detection?: TranscriptEntityDetection;
4310
+ /** Error message of why the transcript failed */
4311
+ error?: string;
4312
+ /** Whether [Profanity Filtering](https://www.assemblyai.com/docs/profanity-filtering) is enabled, either true or false */
3995
4313
  filter_profanity?: TranscriptFilterProfanity;
3996
- /** Whether [PII Redaction](https://www.assemblyai.com/docs/models/pii-redaction) is enabled, either true or false */
4314
+ /** Whether [Text Formatting](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false */
4315
+ format_text?: TranscriptFormatText;
4316
+ /** Whether [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) is enabled, can be true or false */
4317
+ iab_categories?: TranscriptIabCategories;
4318
+ /** The result of the Topic Detection model, if it is enabled.
4319
+ See [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection) for more information.
4320
+ */
4321
+ iab_categories_result?: TranscriptIabCategoriesResult;
4322
+ /** The unique identifier of your transcript */
4323
+ id: string;
4324
+ /** Improve accuracy with up to 200 (for Universal-2) or 1000 (for Universal-3 Pro) domain-specific words or phrases (maximum 6 words per phrase). See [Keyterms Prompting](https://www.assemblyai.com/docs/pre-recorded-audio/keyterms-prompting) for more details.
4325
+ */
4326
+ keyterms_prompt?: string[];
4327
+ /** The language of your audio file.
4328
+ Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
4329
+ The default value is 'en_us'.
4330
+ */
4331
+ language_code?: TranscriptLanguageCode;
4332
+ /** The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
4333
+ One of the values specified must be `en`.
4334
+ */
4335
+ language_codes?: TranscriptLanguageCodes;
4336
+ /** The confidence score for the detected language, between 0.0 (low confidence) and 1.0 (high confidence). See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details. */
4337
+ language_confidence: TranscriptLanguageConfidence;
4338
+ /** The confidence threshold for the automatically detected language.
4339
+ An error will be returned if the language confidence is below this threshold.
4340
+ See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
4341
+ */
4342
+ language_confidence_threshold: TranscriptLanguageConfidenceThreshold;
4343
+ /** Whether [Automatic language detection](/docs/pre-recorded-audio/automatic-language-detection) is enabled, either true or false */
4344
+ language_detection?: TranscriptLanguageDetection;
4345
+ /** Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection). */
4346
+ language_detection_options?: TranscriptLanguageDetectionOptions;
4347
+ /** Whether [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) was enabled in the transcription request, either true or false */
4348
+ multichannel?: TranscriptMultichannel;
4349
+ /** Provide natural language prompting of up to 1,500 words of contextual information to the model. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for best practices.
4350
+
4351
+ Note: This parameter is only supported for the Universal-3 Pro model.
4352
+ */
4353
+ prompt?: string;
4354
+ /** Whether [Automatic Punctuation](https://www.assemblyai.com/docs/pre-recorded-audio) is enabled, either true or false */
4355
+ punctuate?: TranscriptPunctuate;
4356
+ /** Whether [PII Redaction](https://www.assemblyai.com/docs/pii-redaction) is enabled, either true or false */
3997
4357
  redact_pii: boolean;
3998
4358
  /** Whether a redacted version of the audio file was generated,
3999
- either true or false. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
4359
+ either true or false. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
4000
4360
  */
4001
4361
  redact_pii_audio?: TranscriptRedactPiiAudio;
4362
+ /** The options for PII-redacted audio, if redact_pii_audio is enabled.
4363
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
4364
+ */
4365
+ redact_pii_audio_options?: TranscriptRedactPiiAudioOptions;
4002
4366
  /** The audio quality of the PII-redacted audio file, if redact_pii_audio is enabled.
4003
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
4367
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more information.
4004
4368
  */
4005
4369
  redact_pii_audio_quality?: TranscriptRedactPiiAudioQuality;
4006
4370
  /** The list of PII Redaction policies that were enabled, if PII Redaction is enabled.
4007
- See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more information.
4371
+ See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more information.
4008
4372
  */
4009
4373
  redact_pii_policies?: TranscriptRedactPiiPolicies;
4010
- /** The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
4374
+ /** The replacement logic for detected PII, can be `entity_type` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
4011
4375
  redact_pii_sub?: SubstitutionPolicy;
4012
- /** Whether [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, can be true or false */
4376
+ /** Whether [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) is enabled, can be true or false */
4377
+ sentiment_analysis?: TranscriptSentimentAnalysis;
4378
+ /** An array of results for the Sentiment Analysis model, if it is enabled.
4379
+ See [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis) for more information.
4380
+ */
4381
+ sentiment_analysis_results?: TranscriptSentimentAnalysisResults;
4382
+ /** Whether [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, can be true or false */
4013
4383
  speaker_labels?: TranscriptSpeakerLabels;
4014
- /** Tell the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details. */
4384
+ /** Tell the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details. */
4015
4385
  speakers_expected?: TranscriptSpeakersExpected;
4016
- /** Whether [Content Moderation](https://www.assemblyai.com/docs/models/content-moderation) is enabled, can be true or false */
4017
- content_safety?: TranscriptContentSafety;
4018
- /** An array of results for the Content Moderation model, if it is enabled.
4019
- See [Content moderation](https://www.assemblyai.com/docs/models/content-moderation) for more information.
4020
- */
4021
- content_safety_labels?: TranscriptContentSafetyLabels;
4022
- /** Whether [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) is enabled, can be true or false */
4023
- iab_categories?: TranscriptIabCategories;
4024
- /** The result of the Topic Detection model, if it is enabled.
4025
- See [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection) for more information.
4026
- */
4027
- iab_categories_result?: TranscriptIabCategoriesResult;
4028
- /** Customize how words are spelled and formatted using to and from values */
4029
- custom_spelling?: TranscriptCustomSpellingProperty;
4030
- /** Improve accuracy with up to 1000 domain-specific words or phrases (maximum 6 words per phrase).
4386
+ /** The speech model that was actually used for the transcription. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models. */
4387
+ speech_model_used?: SpeechModel;
4388
+ /** List multiple speech models in priority order, allowing our system to automatically route your audio to the best available option. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models and routing behavior.
4031
4389
  */
4032
- keyterms_prompt?: string[];
4033
- /**
4034
- * This parameter does not currently have any functionality attached to it.
4035
- * @deprecated
4390
+ speech_models?: TranscriptSpeechModels;
4391
+ /** Defaults to null. Reject audio files that contain less than this fraction of speech.
4392
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
4393
+ */
4394
+ speech_threshold?: TranscriptSpeechThreshold;
4395
+ /** Speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
4036
4396
  */
4037
- prompt?: string;
4038
- /** Whether [Auto Chapters](https://www.assemblyai.com/docs/models/auto-chapters) is enabled, can be true or false */
4039
- auto_chapters?: TranscriptAutoChapters;
4040
- /** An array of temporally sequential chapters for the audio file */
4041
- chapters?: TranscriptChapters;
4042
- /** Whether [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled, either true or false */
4397
+ speech_understanding?: TranscriptSpeechUnderstanding;
4398
+ /** The status of your transcript. Possible values are queued, processing, completed, or error. */
4399
+ status: TranscriptStatus;
4400
+ /** Whether [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization) is enabled, either true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
4401
+
4402
+ Note: This parameter is only supported for the Universal-2 model.
4403
+ */
4043
4404
  summarization: boolean;
4044
- /** The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled */
4045
- summary_type?: TranscriptSummaryType;
4405
+ /** The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
4406
+ summary?: TranscriptSummary;
4046
4407
  /** The Summarization model used to generate the summary,
4047
- if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled
4408
+ if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-models) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
4048
4409
  */
4049
4410
  summary_model?: TranscriptSummaryModel;
4050
- /** The generated summary of the media file, if [Summarization](https://www.assemblyai.com/docs/models/summarization) is enabled */
4051
- summary?: TranscriptSummary;
4052
- /**
4053
- * Whether custom topics is enabled, either true or false
4054
- * @deprecated
4055
- */
4056
- custom_topics?: TranscriptCustomTopics;
4057
- /** The list of custom topics provided if custom topics is enabled */
4058
- topics?: string[];
4059
- /** Whether [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) is enabled, can be true or false */
4060
- sentiment_analysis?: TranscriptSentimentAnalysis;
4061
- /** An array of results for the Sentiment Analysis model, if it is enabled.
4062
- See [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) for more information.
4063
- */
4064
- sentiment_analysis_results?: TranscriptSentimentAnalysisResults;
4065
- /** Whether [Entity Detection](https://www.assemblyai.com/docs/models/entity-detection) is enabled, can be true or false */
4066
- entity_detection?: TranscriptEntityDetection;
4067
- /** An array of results for the Entity Detection model, if it is enabled.
4068
- See [Entity detection](https://www.assemblyai.com/docs/models/entity-detection) for more information.
4411
+ /** The type of summary generated, if [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization#summary-types) is enabled. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
4412
+ summary_type?: TranscriptSummaryType;
4413
+ /** Whether [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) were removed from the transcript text.
4414
+
4415
+ Note: This parameter is only supported for the Universal-3 Pro model.
4069
4416
  */
4070
- entities?: TranscriptEntities;
4071
- /**
4072
- * Defaults to null. Reject audio files that contain less than this fraction of speech.
4073
- Valid values are in the range [0, 1] inclusive.
4417
+ remove_audio_tags?: TranscriptRemoveAudioTagsProperty;
4418
+ /** The temperature that was used for the model's response. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for more details.
4074
4419
 
4075
- * @minimum 0
4076
- * @maximum 1
4077
- */
4078
- speech_threshold?: TranscriptSpeechThreshold;
4420
+ Note: This parameter can only be used with the Universal-3 Pro model.
4421
+ */
4422
+ temperature?: TranscriptTemperature;
4423
+ /** The textual transcript of your media file */
4424
+ text?: TranscriptText;
4079
4425
  /** True while a request is throttled and false when a request is no longer throttled */
4080
4426
  throttled?: TranscriptThrottled;
4081
- /** Error message of why the transcript failed */
4082
- error?: string;
4083
- /**
4084
- * The language model that was used for the transcript
4085
- * @deprecated
4427
+ /** When multichannel or speaker_labels is enabled, a list of turn-by-turn utterance objects.
4428
+ See [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) and [Multichannel transcription](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) for more information.
4429
+ */
4430
+ utterances?: TranscriptUtterances;
4431
+ /** Whether [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) authentication details were provided */
4432
+ webhook_auth: boolean;
4433
+ /** The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests */
4434
+ webhook_auth_header_name?: TranscriptWebhookAuthHeaderName;
4435
+ /** The status code we received from your server when delivering the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) request, if a webhook URL was provided */
4436
+ webhook_status_code?: TranscriptWebhookStatusCode;
4437
+ /** The URL to which we send [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests.
4086
4438
  */
4087
- language_model: string;
4088
- /**
4089
- * The acoustic model that was used for the transcript
4090
- * @deprecated
4439
+ webhook_url?: TranscriptWebhookUrl;
4440
+ /** An array of temporally-sequential word objects, one for each word in the transcript.
4091
4441
  */
4442
+ words?: TranscriptWords;
4443
+ /** This parameter does not currently have any functionality attached to it. */
4092
4444
  acoustic_model: string;
4445
+ /** This parameter does not currently have any functionality attached to it. */
4446
+ custom_topics?: TranscriptCustomTopics;
4447
+ /** This parameter does not currently have any functionality attached to it. */
4448
+ language_model: string;
4449
+ /** This parameter has been replaced with the `speech_models` parameter, learn more about the `speech_models` parameter [here](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model).
4450
+ */
4451
+ speech_model: TranscriptSpeechModel;
4452
+ /** This parameter does not currently have any functionality attached to it. */
4453
+ speed_boost?: TranscriptSpeedBoost;
4454
+ /** This parameter does not currently have any functionality attached to it. */
4455
+ topics?: string[];
4456
+ /** Translated text keyed by language code. See [Translation](https://www.assemblyai.com/docs/speech-understanding/translation) for more details. */
4457
+ translated_texts?: TranscriptTranslatedTexts;
4093
4458
  }
4094
4459
 
4095
4460
  /**
@@ -4446,6 +4811,24 @@ type SharedCallbackParameter = string;
4446
4811
  /**
4447
4812
  * SharedCallbackMethodParameter type definition
4448
4813
  */
4814
+ /**
4815
+ * SharedCallbackMethodParameter type definition
4816
+ */
4817
+ /**
4818
+ * SharedCallbackMethodParameter type definition
4819
+ */
4820
+ /**
4821
+ * SharedCallbackMethodParameter type definition
4822
+ */
4823
+ /**
4824
+ * SharedCallbackMethodParameter type definition
4825
+ */
4826
+ /**
4827
+ * SharedCallbackMethodParameter type definition
4828
+ */
4829
+ /**
4830
+ * SharedCallbackMethodParameter type definition
4831
+ */
4449
4832
  type SharedCallbackMethodParameter = typeof SharedCallbackMethodParameter[keyof typeof SharedCallbackMethodParameter];
4450
4833
  declare const SharedCallbackMethodParameter: {
4451
4834
  readonly POST: "POST";
@@ -4550,6 +4933,24 @@ type SharedCustomTopicParameter = string | string[];
4550
4933
  /**
4551
4934
  * SharedCustomTopicModeParameter type definition
4552
4935
  */
4936
+ /**
4937
+ * SharedCustomTopicModeParameter type definition
4938
+ */
4939
+ /**
4940
+ * SharedCustomTopicModeParameter type definition
4941
+ */
4942
+ /**
4943
+ * SharedCustomTopicModeParameter type definition
4944
+ */
4945
+ /**
4946
+ * SharedCustomTopicModeParameter type definition
4947
+ */
4948
+ /**
4949
+ * SharedCustomTopicModeParameter type definition
4950
+ */
4951
+ /**
4952
+ * SharedCustomTopicModeParameter type definition
4953
+ */
4553
4954
  type SharedCustomTopicModeParameter = typeof SharedCustomTopicModeParameter[keyof typeof SharedCustomTopicModeParameter];
4554
4955
  declare const SharedCustomTopicModeParameter: {
4555
4956
  readonly extended: "extended";
@@ -4611,6 +5012,24 @@ type SharedCustomIntentParameter = string | string[];
4611
5012
  /**
4612
5013
  * SharedCustomIntentModeParameter type definition
4613
5014
  */
5015
+ /**
5016
+ * SharedCustomIntentModeParameter type definition
5017
+ */
5018
+ /**
5019
+ * SharedCustomIntentModeParameter type definition
5020
+ */
5021
+ /**
5022
+ * SharedCustomIntentModeParameter type definition
5023
+ */
5024
+ /**
5025
+ * SharedCustomIntentModeParameter type definition
5026
+ */
5027
+ /**
5028
+ * SharedCustomIntentModeParameter type definition
5029
+ */
5030
+ /**
5031
+ * SharedCustomIntentModeParameter type definition
5032
+ */
4614
5033
  type SharedCustomIntentModeParameter = typeof SharedCustomIntentModeParameter[keyof typeof SharedCustomIntentModeParameter];
4615
5034
  declare const SharedCustomIntentModeParameter: {
4616
5035
  readonly extended: "extended";
@@ -4698,6 +5117,24 @@ type ListenV1DictationParameter = boolean;
4698
5117
  /**
4699
5118
  * ListenV1EncodingParameter type definition
4700
5119
  */
5120
+ /**
5121
+ * ListenV1EncodingParameter type definition
5122
+ */
5123
+ /**
5124
+ * ListenV1EncodingParameter type definition
5125
+ */
5126
+ /**
5127
+ * ListenV1EncodingParameter type definition
5128
+ */
5129
+ /**
5130
+ * ListenV1EncodingParameter type definition
5131
+ */
5132
+ /**
5133
+ * ListenV1EncodingParameter type definition
5134
+ */
5135
+ /**
5136
+ * ListenV1EncodingParameter type definition
5137
+ */
4701
5138
  type ListenV1EncodingParameter = typeof ListenV1EncodingParameter[keyof typeof ListenV1EncodingParameter];
4702
5139
  declare const ListenV1EncodingParameter: {
4703
5140
  readonly linear16: "linear16";
@@ -5113,128 +5550,194 @@ type ListenV1MediaTranscribeParams = {
5113
5550
  * Generated by orval v7.9.0 🍺
5114
5551
  * Do not edit manually.
5115
5552
  * AssemblyAI API
5116
- * AssemblyAI API
5117
- * OpenAPI spec version: 1.3.4
5553
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5554
+ * OpenAPI spec version: 1.0.0
5118
5555
  */
5556
+ /**
5557
+ * Enable domain-specific transcription models to improve accuracy for specialized terminology. Set to `"medical-v1"` to enable [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) for improved accuracy of medical terms such as medications, procedures, conditions, and dosages.
5119
5558
 
5120
- type TranscriptOptionalParamsLanguageCodeOneOf = TranscriptLanguageCode | string;
5559
+ Supported languages: English (`en`), Spanish (`es`), German (`de`), French (`fr`). If used with an unsupported language, the parameter is ignored and a warning is returned.
5560
+
5561
+ */
5562
+ type TranscriptParamsDomain = string | null;
5121
5563
 
5122
5564
  /**
5123
5565
  * Generated by orval v7.9.0 🍺
5124
5566
  * Do not edit manually.
5125
5567
  * AssemblyAI API
5126
- * AssemblyAI API
5127
- * OpenAPI spec version: 1.3.4
5568
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5569
+ * OpenAPI spec version: 1.0.0
5128
5570
  */
5129
5571
 
5130
5572
  /**
5131
- * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
5573
+ * The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
5132
5574
  The default value is 'en_us'.
5133
5575
 
5134
5576
  */
5135
- type TranscriptOptionalParamsLanguageCode = TranscriptOptionalParamsLanguageCodeOneOf | null;
5577
+ type TranscriptParamsLanguageCode = TranscriptLanguageCode | null;
5136
5578
 
5137
5579
  /**
5138
5580
  * Generated by orval v7.9.0 🍺
5139
5581
  * Do not edit manually.
5140
5582
  * AssemblyAI API
5141
- * AssemblyAI API
5142
- * OpenAPI spec version: 1.3.4
5583
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5584
+ * OpenAPI spec version: 1.0.0
5143
5585
  */
5144
5586
 
5145
5587
  /**
5146
- * The speech model to use for the transcription. When `null`, the "best" model is used.
5588
+ * The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
5589
+ One of the values specified must be `en`.
5590
+
5147
5591
  */
5148
- type TranscriptOptionalParamsSpeechModel = SpeechModel | null;
5592
+ type TranscriptParamsLanguageCodes = TranscriptLanguageCode[] | null;
5149
5593
 
5150
5594
  /**
5151
5595
  * Generated by orval v7.9.0 🍺
5152
5596
  * Do not edit manually.
5153
5597
  * AssemblyAI API
5154
- * AssemblyAI API
5155
- * OpenAPI spec version: 1.3.4
5598
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5599
+ * OpenAPI spec version: 1.0.0
5156
5600
  */
5157
5601
  /**
5158
- * The header name to be sent with the transcript completed or failed webhook requests
5602
+ * Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection).
5159
5603
  */
5160
- type TranscriptOptionalParamsWebhookAuthHeaderName = string | null;
5604
+ interface TranscriptOptionalParamsLanguageDetectionOptions {
5605
+ /** List of languages expected in the audio file. Defaults to `["all"]` when unspecified. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details. */
5606
+ expected_languages?: string[];
5607
+ /** If the detected language of the audio file is not in the list of expected languages, the `fallback_language` is used. Specify `["auto"]` to let our model choose the fallback language from `expected_languages` with the highest confidence score. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
5608
+ */
5609
+ fallback_language?: string;
5610
+ /** Whether [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) should be detected.
5611
+ */
5612
+ code_switching?: boolean;
5613
+ /** The confidence threshold for [code switching](/docs/speech-to-text/pre-recorded-audio/code-switching) detection. If the code switching confidence is below this threshold, the transcript will be processed in the language with the highest `language_detection_confidence` score.
5614
+ */
5615
+ code_switching_confidence_threshold?: number;
5616
+ }
5161
5617
 
5162
5618
  /**
5163
5619
  * Generated by orval v7.9.0 🍺
5164
5620
  * Do not edit manually.
5165
5621
  * AssemblyAI API
5166
- * AssemblyAI API
5167
- * OpenAPI spec version: 1.3.4
5622
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5623
+ * OpenAPI spec version: 1.0.0
5168
5624
  */
5169
5625
  /**
5170
- * The header value to send back with the transcript completed or failed webhook requests for added security
5626
+ * Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep.
5171
5627
  */
5172
- type TranscriptOptionalParamsWebhookAuthHeaderValue = string | null;
5628
+ type TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod = (typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod)[keyof typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod];
5629
+ declare const TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod: {
5630
+ readonly silence: "silence";
5631
+ };
5173
5632
 
5174
5633
  /**
5175
5634
  * Generated by orval v7.9.0 🍺
5176
5635
  * Do not edit manually.
5177
5636
  * AssemblyAI API
5178
- * AssemblyAI API
5179
- * OpenAPI spec version: 1.3.4
5637
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5638
+ * OpenAPI spec version: 1.0.0
5180
5639
  */
5640
+
5181
5641
  /**
5182
- * How much to boost specified words
5642
+ * Specify options for [PII redacted audio](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) files.
5183
5643
  */
5184
- type TranscriptBoostParam = (typeof TranscriptBoostParam)[keyof typeof TranscriptBoostParam];
5185
- declare const TranscriptBoostParam: {
5186
- readonly low: "low";
5187
- readonly default: "default";
5188
- readonly high: "high";
5189
- };
5644
+ interface TranscriptOptionalParamsRedactPiiAudioOptions {
5645
+ /** By default, audio redaction provides redacted audio URLs only when speech is detected. However, if your use-case specifically requires redacted audio files even for silent audio files without any dialogue, you can opt to receive these URLs by setting this parameter to `true`. */
5646
+ return_redacted_no_speech_audio?: boolean;
5647
+ /** Specify the method used to redact audio. By default, redacted audio uses a beep sound. Set to `silence` to replace PII with silence instead of a beep. */
5648
+ override_audio_redaction_method?: TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod;
5649
+ }
5190
5650
 
5191
5651
  /**
5192
5652
  * Generated by orval v7.9.0 🍺
5193
5653
  * Do not edit manually.
5194
5654
  * AssemblyAI API
5195
- * AssemblyAI API
5196
- * OpenAPI spec version: 1.3.4
5655
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5656
+ * OpenAPI spec version: 1.0.0
5197
5657
  */
5198
5658
 
5199
5659
  /**
5200
- * The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details.
5660
+ * The replacement logic for detected PII, can be `entity_type` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details.
5201
5661
  */
5202
- type TranscriptOptionalParamsRedactPiiSub = SubstitutionPolicy | null;
5662
+ type TranscriptParamsRedactPiiSub = SubstitutionPolicy | null;
5203
5663
 
5204
5664
  /**
5205
5665
  * Generated by orval v7.9.0 🍺
5206
5666
  * Do not edit manually.
5207
5667
  * AssemblyAI API
5208
- * AssemblyAI API
5209
- * OpenAPI spec version: 1.3.4
5668
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5669
+ * OpenAPI spec version: 1.0.0
5210
5670
  */
5211
5671
  /**
5212
- * Tells the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details.
5672
+ * Specify options for [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers). Use this to set a range of possible speakers.
5213
5673
  */
5214
- type TranscriptOptionalParamsSpeakersExpected = number | null;
5674
+ interface TranscriptOptionalParamsSpeakerOptions {
5675
+ /** The minimum number of speakers expected in the audio file. See [Set a range of possible speakers](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers) for more details. */
5676
+ min_speakers_expected?: number;
5677
+ /** <Warning>Setting this parameter too high may hurt model accuracy</Warning>
5678
+ The maximum number of speakers expected in the audio file. The default depends on audio duration: no limit for 0-2 minutes, 10 for 2-10 minutes, and 30 for 10+ minutes. See [Set a range of possible speakers](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers) for more details.
5679
+ */
5680
+ max_speakers_expected?: number;
5681
+ }
5215
5682
 
5216
5683
  /**
5217
5684
  * Generated by orval v7.9.0 🍺
5218
5685
  * Do not edit manually.
5219
5686
  * AssemblyAI API
5687
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5688
+ * OpenAPI spec version: 1.0.0
5689
+ */
5690
+ /**
5691
+ * Tells the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details.
5692
+ */
5693
+ type TranscriptParamsSpeakersExpected = number | null;
5694
+
5695
+ /**
5696
+ * Generated by orval v7.9.0 🍺
5697
+ * Do not edit manually.
5220
5698
  * AssemblyAI API
5221
- * OpenAPI spec version: 1.3.4
5699
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5700
+ * OpenAPI spec version: 1.0.0
5222
5701
  */
5223
5702
  /**
5224
5703
  * Reject audio files that contain less than this fraction of speech.
5225
- Valid values are in the range [0, 1] inclusive.
5704
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
5705
+
5706
+ */
5707
+ type TranscriptParamsSpeechThreshold = number | null;
5226
5708
 
5227
- * @minimum 0
5228
- * @maximum 1
5709
+ /**
5710
+ * Generated by orval v7.9.0 🍺
5711
+ * Do not edit manually.
5712
+ * AssemblyAI API
5713
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5714
+ * OpenAPI spec version: 1.0.0
5229
5715
  */
5230
- type TranscriptOptionalParamsSpeechThreshold = number | null;
5716
+
5717
+ type TranscriptOptionalParamsSpeechUnderstandingRequest = TranslationRequestBody | SpeakerIdentificationRequestBody | CustomFormattingRequestBody;
5231
5718
 
5232
5719
  /**
5233
5720
  * Generated by orval v7.9.0 🍺
5234
5721
  * Do not edit manually.
5235
5722
  * AssemblyAI API
5723
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5724
+ * OpenAPI spec version: 1.0.0
5725
+ */
5726
+
5727
+ /**
5728
+ * Enable speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
5729
+
5730
+ */
5731
+ interface TranscriptOptionalParamsSpeechUnderstanding {
5732
+ request: TranscriptOptionalParamsSpeechUnderstandingRequest;
5733
+ }
5734
+
5735
+ /**
5736
+ * Generated by orval v7.9.0 🍺
5737
+ * Do not edit manually.
5236
5738
  * AssemblyAI API
5237
- * OpenAPI spec version: 1.3.4
5739
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5740
+ * OpenAPI spec version: 1.0.0
5238
5741
  */
5239
5742
  /**
5240
5743
  * The model to summarize the transcript
@@ -5250,8 +5753,8 @@ declare const SummaryModel: {
5250
5753
  * Generated by orval v7.9.0 🍺
5251
5754
  * Do not edit manually.
5252
5755
  * AssemblyAI API
5253
- * AssemblyAI API
5254
- * OpenAPI spec version: 1.3.4
5756
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5757
+ * OpenAPI spec version: 1.0.0
5255
5758
  */
5256
5759
  /**
5257
5760
  * The type of summary
@@ -5269,125 +5772,195 @@ declare const SummaryType: {
5269
5772
  * Generated by orval v7.9.0 🍺
5270
5773
  * Do not edit manually.
5271
5774
  * AssemblyAI API
5272
- * AssemblyAI API
5273
- * OpenAPI spec version: 1.3.4
5775
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5776
+ * OpenAPI spec version: 1.0.0
5777
+ */
5778
+ /**
5779
+ * Remove [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) from the transcript text. Set to `"all"` to remove all audio tags.
5780
+
5781
+ Note: This parameter is only supported for the Universal-3 Pro model.
5782
+
5783
+ */
5784
+ type TranscriptOptionalParamsRemoveAudioTags = (typeof TranscriptOptionalParamsRemoveAudioTags)[keyof typeof TranscriptOptionalParamsRemoveAudioTags];
5785
+ declare const TranscriptOptionalParamsRemoveAudioTags: {
5786
+ readonly all: "all";
5787
+ };
5788
+
5789
+ /**
5790
+ * Generated by orval v7.9.0 🍺
5791
+ * Do not edit manually.
5792
+ * AssemblyAI API
5793
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5794
+ * OpenAPI spec version: 1.0.0
5795
+ */
5796
+
5797
+ /**
5798
+ * Remove [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) from the transcript text. Set to `"all"` to remove all audio tags.
5799
+
5800
+ Note: This parameter is only supported for the Universal-3 Pro model.
5801
+
5802
+ */
5803
+ type TranscriptParamsRemoveAudioTags = TranscriptOptionalParamsRemoveAudioTags | null;
5804
+
5805
+ /**
5806
+ * Generated by orval v7.9.0 🍺
5807
+ * Do not edit manually.
5808
+ * AssemblyAI API
5809
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5810
+ * OpenAPI spec version: 1.0.0
5811
+ */
5812
+ /**
5813
+ * The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests
5814
+ */
5815
+ type TranscriptParamsWebhookAuthHeaderName = string | null;
5816
+
5817
+ /**
5818
+ * Generated by orval v7.9.0 🍺
5819
+ * Do not edit manually.
5820
+ * AssemblyAI API
5821
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5822
+ * OpenAPI spec version: 1.0.0
5823
+ */
5824
+ /**
5825
+ * The header value to send back with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests for added security
5826
+ */
5827
+ type TranscriptParamsWebhookAuthHeaderValue = string | null;
5828
+
5829
+ /**
5830
+ * Generated by orval v7.9.0 🍺
5831
+ * Do not edit manually.
5832
+ * AssemblyAI API
5833
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
5834
+ * OpenAPI spec version: 1.0.0
5274
5835
  */
5275
5836
 
5276
5837
  /**
5277
5838
  * The parameters for creating a transcript
5278
5839
  */
5279
- interface TranscriptOptionalParams {
5280
- /** The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/concepts/supported-languages).
5281
- The default value is 'en_us'.
5840
+ interface TranscriptParams {
5841
+ /** The point in time, in milliseconds, to stop transcribing in your media file. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
5842
+ audio_end_at?: number;
5843
+ /** The point in time, in milliseconds, to begin transcribing in your media file. See [Set the start and end of the transcript](https://www.assemblyai.com/docs/pre-recorded-audio/set-the-start-and-end-of-the-transcript) for more details. */
5844
+ audio_start_from?: number;
5845
+ /** Enable [Auto Chapters](https://www.assemblyai.com/docs/speech-understanding/auto-chapters), can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible chapter summaries. See the [updated Auto Chapters page](https://www.assemblyai.com/docs/speech-understanding/auto-chapters) for details.
5846
+
5847
+ Note: This parameter is only supported for the Universal-2 model.
5282
5848
  */
5283
- language_code?: TranscriptOptionalParamsLanguageCode;
5284
- /** Enable [Automatic language detection](https://www.assemblyai.com/docs/models/speech-recognition#automatic-language-detection), either true or false. */
5285
- language_detection?: boolean;
5286
- /**
5287
- * The confidence threshold for the automatically detected language.
5288
- An error will be returned if the language confidence is below this threshold.
5289
- Defaults to 0.
5849
+ auto_chapters?: boolean;
5850
+ /** Enable [Key Phrases](https://www.assemblyai.com/docs/speech-understanding/key-phrases), either true or false */
5851
+ auto_highlights?: boolean;
5852
+ /** Enable [Content Moderation](https://www.assemblyai.com/docs/content-moderation), can be true or false */
5853
+ content_safety?: boolean;
5854
+ /** The confidence threshold for the [Content Moderation](https://www.assemblyai.com/docs/content-moderation) model. Values must be between 25 and 100. */
5855
+ content_safety_confidence?: number;
5856
+ /** Customize how words are spelled and formatted using to and from values. See [Custom Spelling](https://www.assemblyai.com/docs/pre-recorded-audio/custom-spelling) for more details. */
5857
+ custom_spelling?: TranscriptCustomSpelling[];
5858
+ /** Transcribe [Filler Words](https://www.assemblyai.com/docs/pre-recorded-audio/filler-words), like "umm", in your media file; can be true or false */
5859
+ disfluencies?: boolean;
5860
+ /** Enable domain-specific transcription models to improve accuracy for specialized terminology. Set to `"medical-v1"` to enable [Medical Mode](https://www.assemblyai.com/docs/pre-recorded-audio/medical-mode) for improved accuracy of medical terms such as medications, procedures, conditions, and dosages.
5290
5861
 
5291
- * @minimum 0
5292
- * @maximum 1
5862
+ Supported languages: English (`en`), Spanish (`es`), German (`de`), French (`fr`). If used with an unsupported language, the parameter is ignored and a warning is returned.
5863
+ */
5864
+ domain?: TranscriptParamsDomain;
5865
+ /** Enable [Entity Detection](https://www.assemblyai.com/docs/speech-understanding/entity-detection), can be true or false */
5866
+ entity_detection?: boolean;
5867
+ /** Filter profanity from the transcribed text, can be true or false. See [Profanity Filtering](https://www.assemblyai.com/docs/profanity-filtering) for more details. */
5868
+ filter_profanity?: boolean;
5869
+ /** Enable [Text Formatting](https://www.assemblyai.com/docs/pre-recorded-audio), can be true or false */
5870
+ format_text?: boolean;
5871
+ /** Enable [Topic Detection](https://www.assemblyai.com/docs/speech-understanding/topic-detection), can be true or false */
5872
+ iab_categories?: boolean;
5873
+ /** Improve accuracy with up to 200 (for Universal-2) or 1000 (for Universal-3 Pro) domain-specific words or phrases (maximum 6 words per phrase). See [Keyterms Prompting](https://www.assemblyai.com/docs/pre-recorded-audio/keyterms-prompting) for more details.
5293
5874
  */
5875
+ keyterms_prompt?: string[];
5876
+ /** The language of your audio file. Possible values are found in [Supported Languages](https://www.assemblyai.com/docs/pre-recorded-audio/supported-languages).
5877
+ The default value is 'en_us'.
5878
+ */
5879
+ language_code?: TranscriptParamsLanguageCode;
5880
+ /** The language codes of your audio file. Used for [Code switching](/docs/speech-to-text/pre-recorded-audio/code-switching)
5881
+ One of the values specified must be `en`.
5882
+ */
5883
+ language_codes?: TranscriptParamsLanguageCodes;
5884
+ /** The confidence threshold for the automatically detected language.
5885
+ An error will be returned if the language confidence is below this threshold.
5886
+ Defaults to 0. See [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection) for more details.
5887
+ */
5294
5888
  language_confidence_threshold?: number;
5295
- /** The speech model to use for the transcription. When `null`, the "best" model is used. */
5296
- speech_model?: TranscriptOptionalParamsSpeechModel;
5297
- /** Enable Automatic Punctuation, can be true or false */
5298
- punctuate?: boolean;
5299
- /** Enable Text Formatting, can be true or false */
5300
- format_text?: boolean;
5301
- /** Transcribe Filler Words, like "umm", in your media file; can be true or false */
5302
- disfluencies?: boolean;
5303
- /** Enable [Multichannel](https://www.assemblyai.com/docs/models/speech-recognition#multichannel-transcription) transcription, can be true or false. */
5889
+ /** Enable [Automatic language detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection), either true or false. */
5890
+ language_detection?: boolean;
5891
+ /** Specify options for [Automatic Language Detection](https://www.assemblyai.com/docs/pre-recorded-audio/language-detection). */
5892
+ language_detection_options?: TranscriptOptionalParamsLanguageDetectionOptions;
5893
+ /** Enable [Multichannel](https://www.assemblyai.com/docs/pre-recorded-audio/multichannel) transcription, can be true or false. */
5304
5894
  multichannel?: boolean;
5305
- /** The URL to which we send webhook requests.
5306
- We sends two different types of webhook requests.
5307
- One request when a transcript is completed or failed, and one request when the redacted audio is ready if redact_pii_audio is enabled.
5895
+ /** Provide natural language prompting of up to 1,500 words of contextual information to the model. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for best practices.
5896
+
5897
+ Note: This parameter is only supported for the Universal-3 Pro model.
5308
5898
  */
5309
- webhook_url?: string;
5310
- /** The header name to be sent with the transcript completed or failed webhook requests */
5311
- webhook_auth_header_name?: TranscriptOptionalParamsWebhookAuthHeaderName;
5312
- /** The header value to send back with the transcript completed or failed webhook requests for added security */
5313
- webhook_auth_header_value?: TranscriptOptionalParamsWebhookAuthHeaderValue;
5314
- /** Enable Key Phrases, either true or false */
5315
- auto_highlights?: boolean;
5316
- /** The point in time, in milliseconds, to begin transcribing in your media file */
5317
- audio_start_from?: number;
5318
- /** The point in time, in milliseconds, to stop transcribing in your media file */
5319
- audio_end_at?: number;
5320
- /**
5321
- * The list of custom vocabulary to boost transcription probability for
5322
- * @deprecated
5323
- */
5324
- word_boost?: string[];
5325
- /** How much to boost specified words */
5326
- boost_param?: TranscriptBoostParam;
5327
- /** Filter profanity from the transcribed text, can be true or false */
5328
- filter_profanity?: boolean;
5329
- /** Redact PII from the transcribed text using the Redact PII model, can be true or false */
5899
+ prompt?: string;
5900
+ /** Enable [Automatic Punctuation](https://www.assemblyai.com/docs/pre-recorded-audio), can be true or false */
5901
+ punctuate?: boolean;
5902
+ /** Redact PII from the transcribed text using the Redact PII model, can be true or false. See [PII Redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
5330
5903
  redact_pii?: boolean;
5331
- /** Generate a copy of the original media file with spoken PII "beeped" out, can be true or false. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5904
+ /** Generate a copy of the original media file with spoken PII "beeped" out, can be true or false. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more details. */
5332
5905
  redact_pii_audio?: boolean;
5333
- /** Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5906
+ /** Specify options for [PII redacted audio](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) files. */
5907
+ redact_pii_audio_options?: TranscriptOptionalParamsRedactPiiAudioOptions;
5908
+ /** Controls the filetype of the audio created by redact_pii_audio. Currently supports mp3 (default) and wav. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction#request-for-redacted-audio) for more details. */
5334
5909
  redact_pii_audio_quality?: RedactPiiAudioQuality;
5335
- /** The list of PII Redaction policies to enable. See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5910
+ /** The list of PII Redaction policies to enable. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
5336
5911
  redact_pii_policies?: PiiPolicy[];
5337
- /** The replacement logic for detected PII, can be "entity_type" or "hash". See [PII redaction](https://www.assemblyai.com/docs/models/pii-redaction) for more details. */
5338
- redact_pii_sub?: TranscriptOptionalParamsRedactPiiSub;
5339
- /** Enable [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization), can be true or false */
5912
+ /** The replacement logic for detected PII, can be `entity_type` or `hash`. See [PII redaction](https://www.assemblyai.com/docs/pii-redaction) for more details. */
5913
+ redact_pii_sub?: TranscriptParamsRedactPiiSub;
5914
+ /** Enable [Sentiment Analysis](https://www.assemblyai.com/docs/speech-understanding/sentiment-analysis), can be true or false */
5915
+ sentiment_analysis?: boolean;
5916
+ /** Enable [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization), can be true or false */
5340
5917
  speaker_labels?: boolean;
5341
- /** Tells the speaker label model how many speakers it should attempt to identify. See [Speaker diarization](https://www.assemblyai.com/docs/models/speaker-diarization) for more details. */
5342
- speakers_expected?: TranscriptOptionalParamsSpeakersExpected;
5343
- /** Enable [Content Moderation](https://www.assemblyai.com/docs/models/content-moderation), can be true or false */
5344
- content_safety?: boolean;
5345
- /**
5346
- * The confidence threshold for the Content Moderation model. Values must be between 25 and 100.
5347
- * @minimum 25
5348
- * @maximum 100
5349
- */
5350
- content_safety_confidence?: number;
5351
- /** Enable [Topic Detection](https://www.assemblyai.com/docs/models/topic-detection), can be true or false */
5352
- iab_categories?: boolean;
5353
- /** Customize how words are spelled and formatted using to and from values */
5354
- custom_spelling?: TranscriptCustomSpelling[];
5355
- /** <Warning>`keyterms_prompt` is only supported when the `speech_model` is specified as `slam-1`</Warning>
5356
- Improve accuracy with up to 1000 domain-specific words or phrases (maximum 6 words per phrase).
5918
+ /** Specify options for [Speaker diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers). Use this to set a range of possible speakers. */
5919
+ speaker_options?: TranscriptOptionalParamsSpeakerOptions;
5920
+ /** Tells the speaker label model how many speakers it should attempt to identify. See [Set number of speakers expected](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization#set-number-of-speakers-expected) for more details. */
5921
+ speakers_expected?: TranscriptParamsSpeakersExpected;
5922
+ /** List multiple speech models in priority order, allowing our system to automatically route your audio to the best available option. See [Model Selection](https://www.assemblyai.com/docs/pre-recorded-audio/select-the-speech-model) for available models and routing behavior.
5923
+ */
5924
+ speech_models: SpeechModel[];
5925
+ /** Reject audio files that contain less than this fraction of speech.
5926
+ Valid values are in the range [0, 1] inclusive. See [Speech Threshold](https://www.assemblyai.com/docs/speech-threshold) for more details.
5357
5927
  */
5358
- keyterms_prompt?: string[];
5359
- /**
5360
- * This parameter does not currently have any functionality attached to it.
5361
- * @deprecated
5928
+ speech_threshold?: TranscriptParamsSpeechThreshold;
5929
+ /** Enable speech understanding tasks like [Translation](https://www.assemblyai.com/docs/speech-understanding/translation), [Speaker Identification](https://www.assemblyai.com/docs/speech-understanding/speaker-identification), and [Custom Formatting](https://www.assemblyai.com/docs/speech-understanding/custom-formatting). See the task-specific docs for available options and configuration.
5362
5930
  */
5363
- prompt?: string;
5364
- /** Enable [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis), can be true or false */
5365
- sentiment_analysis?: boolean;
5366
- /** Enable [Auto Chapters](https://www.assemblyai.com/docs/models/auto-chapters), can be true or false */
5367
- auto_chapters?: boolean;
5368
- /** Enable [Entity Detection](https://www.assemblyai.com/docs/models/entity-detection), can be true or false */
5369
- entity_detection?: boolean;
5370
- /**
5371
- * Reject audio files that contain less than this fraction of speech.
5372
- Valid values are in the range [0, 1] inclusive.
5931
+ speech_understanding?: TranscriptOptionalParamsSpeechUnderstanding;
5932
+ /** Enable [Summarization](https://www.assemblyai.com/docs/speech-understanding/summarization), can be true or false. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details.
5373
5933
 
5374
- * @minimum 0
5375
- * @maximum 1
5376
- */
5377
- speech_threshold?: TranscriptOptionalParamsSpeechThreshold;
5378
- /** Enable [Summarization](https://www.assemblyai.com/docs/models/summarization), can be true or false */
5934
+ Note: This parameter is only supported for the Universal-2 model.
5935
+ */
5379
5936
  summarization?: boolean;
5380
- /** The model to summarize the transcript */
5937
+ /** The model to summarize the transcript. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
5381
5938
  summary_model?: SummaryModel;
5382
- /** The type of summary */
5939
+ /** The type of summary. Deprecated - use [LLM Gateway](https://www.assemblyai.com/docs/llm-gateway/overview) instead for more flexible summaries. See the [updated Summarization page](https://www.assemblyai.com/docs/speech-understanding/summarization) for details. */
5383
5940
  summary_type?: SummaryType;
5384
- /**
5385
- * Enable custom topics, either true or false
5386
- * @deprecated
5941
+ /** Remove [audio event tags](https://www.assemblyai.com/docs/pre-recorded-audio/universal-3-pro#audio-event-tags) from the transcript text. Set to `"all"` to remove all audio tags.
5942
+
5943
+ Note: This parameter is only supported for the Universal-3 Pro model.
5944
+ */
5945
+ remove_audio_tags?: TranscriptParamsRemoveAudioTags;
5946
+ /** Control the amount of randomness injected into the model's response. See the [Prompting Guide](https://www.assemblyai.com/docs/pre-recorded-audio/prompting) for more details.
5947
+
5948
+ Note: This parameter can only be used with the Universal-3 Pro model.
5949
+ */
5950
+ temperature?: number;
5951
+ /** The header name to be sent with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests */
5952
+ webhook_auth_header_name?: TranscriptParamsWebhookAuthHeaderName;
5953
+ /** The header value to send back with the transcript completed or failed [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests for added security */
5954
+ webhook_auth_header_value?: TranscriptParamsWebhookAuthHeaderValue;
5955
+ /** The URL to which we send [webhook](https://www.assemblyai.com/docs/deployment/webhooks-for-pre-recorded-audio) requests.
5387
5956
  */
5957
+ webhook_url?: string;
5958
+ /** This parameter does not currently have any functionality attached to it. */
5388
5959
  custom_topics?: boolean;
5389
- /** The list of custom topics */
5960
+ /** This parameter does not currently have any functionality attached to it. */
5390
5961
  topics?: string[];
5962
+ /** The URL of the audio or video file to transcribe. */
5963
+ audio_url: string;
5391
5964
  }
5392
5965
 
5393
5966
  /**
@@ -7204,75 +7777,15 @@ interface TranscribeStreamParams<P extends StreamingProvider> {
7204
7777
  * Generated by orval v7.9.0 🍺
7205
7778
  * Do not edit manually.
7206
7779
  * AssemblyAI API
7207
- * AssemblyAI API
7208
- * OpenAPI spec version: 1.3.4
7209
- */
7210
- /**
7211
- * Maximum amount of transcripts to retrieve
7212
- */
7213
- type Limit = number;
7214
-
7215
- /**
7216
- * Generated by orval v7.9.0 🍺
7217
- * Do not edit manually.
7218
- * AssemblyAI API
7219
- * AssemblyAI API
7220
- * OpenAPI spec version: 1.3.4
7221
- */
7222
- /**
7223
- * Only get transcripts created on this date
7224
- */
7225
- type CreatedOn = string;
7226
-
7227
- /**
7228
- * Generated by orval v7.9.0 🍺
7229
- * Do not edit manually.
7230
- * AssemblyAI API
7231
- * AssemblyAI API
7232
- * OpenAPI spec version: 1.3.4
7233
- */
7234
- /**
7235
- * Get transcripts that were created before this transcript ID
7236
- */
7237
- type BeforeId = string;
7238
-
7239
- /**
7240
- * Generated by orval v7.9.0 🍺
7241
- * Do not edit manually.
7242
- * AssemblyAI API
7243
- * AssemblyAI API
7244
- * OpenAPI spec version: 1.3.4
7245
- */
7246
- /**
7247
- * Get transcripts that were created after this transcript ID
7248
- */
7249
- type AfterId = string;
7250
-
7251
- /**
7252
- * Generated by orval v7.9.0 🍺
7253
- * Do not edit manually.
7254
- * AssemblyAI API
7255
- * AssemblyAI API
7256
- * OpenAPI spec version: 1.3.4
7257
- */
7258
- /**
7259
- * Only get throttled transcripts, overrides the status filter
7260
- */
7261
- type ThrottledOnly = boolean;
7262
-
7263
- /**
7264
- * Generated by orval v7.9.0 🍺
7265
- * Do not edit manually.
7266
- * AssemblyAI API
7267
- * AssemblyAI API
7268
- * OpenAPI spec version: 1.3.4
7780
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
7781
+ * OpenAPI spec version: 1.0.0
7269
7782
  */
7270
7783
 
7271
7784
  type ListTranscriptsParams = {
7272
7785
  /**
7273
7786
  * Maximum amount of transcripts to retrieve
7274
7787
  */
7275
- limit?: Limit;
7788
+ limit?: number;
7276
7789
  /**
7277
7790
  * Filter by transcript status
7278
7791
  */
@@ -7280,19 +7793,19 @@ type ListTranscriptsParams = {
7280
7793
  /**
7281
7794
  * Only get transcripts created on this date
7282
7795
  */
7283
- created_on?: CreatedOn;
7796
+ created_on?: string;
7284
7797
  /**
7285
7798
  * Get transcripts that were created before this transcript ID
7286
7799
  */
7287
- before_id?: BeforeId;
7800
+ before_id?: string;
7288
7801
  /**
7289
7802
  * Get transcripts that were created after this transcript ID
7290
7803
  */
7291
- after_id?: AfterId;
7804
+ after_id?: string;
7292
7805
  /**
7293
7806
  * Only get throttled transcripts, overrides the status filter
7294
7807
  */
7295
- throttled_only?: ThrottledOnly;
7808
+ throttled_only?: boolean;
7296
7809
  };
7297
7810
 
7298
7811
  /**
@@ -7475,6 +7988,24 @@ type ManageV1FilterRequestIdParameter = string;
7475
7988
  /**
7476
7989
  * ManageV1FilterDeploymentParameter type definition
7477
7990
  */
7991
+ /**
7992
+ * ManageV1FilterDeploymentParameter type definition
7993
+ */
7994
+ /**
7995
+ * ManageV1FilterDeploymentParameter type definition
7996
+ */
7997
+ /**
7998
+ * ManageV1FilterDeploymentParameter type definition
7999
+ */
8000
+ /**
8001
+ * ManageV1FilterDeploymentParameter type definition
8002
+ */
8003
+ /**
8004
+ * ManageV1FilterDeploymentParameter type definition
8005
+ */
8006
+ /**
8007
+ * ManageV1FilterDeploymentParameter type definition
8008
+ */
7478
8009
  type ManageV1FilterDeploymentParameter = typeof ManageV1FilterDeploymentParameter[keyof typeof ManageV1FilterDeploymentParameter];
7479
8010
  declare const ManageV1FilterDeploymentParameter: {
7480
8011
  readonly hosted: "hosted";
@@ -7513,6 +8044,24 @@ declare const ManageV1FilterDeploymentParameter: {
7513
8044
  /**
7514
8045
  * ManageV1FilterEndpointParameter type definition
7515
8046
  */
8047
+ /**
8048
+ * ManageV1FilterEndpointParameter type definition
8049
+ */
8050
+ /**
8051
+ * ManageV1FilterEndpointParameter type definition
8052
+ */
8053
+ /**
8054
+ * ManageV1FilterEndpointParameter type definition
8055
+ */
8056
+ /**
8057
+ * ManageV1FilterEndpointParameter type definition
8058
+ */
8059
+ /**
8060
+ * ManageV1FilterEndpointParameter type definition
8061
+ */
8062
+ /**
8063
+ * ManageV1FilterEndpointParameter type definition
8064
+ */
7516
8065
  type ManageV1FilterEndpointParameter = typeof ManageV1FilterEndpointParameter[keyof typeof ManageV1FilterEndpointParameter];
7517
8066
  declare const ManageV1FilterEndpointParameter: {
7518
8067
  readonly listen: "listen";
@@ -7556,6 +8105,24 @@ declare const ManageV1FilterEndpointParameter: {
7556
8105
  /**
7557
8106
  * ManageV1FilterMethodParameter type definition
7558
8107
  */
8108
+ /**
8109
+ * ManageV1FilterMethodParameter type definition
8110
+ */
8111
+ /**
8112
+ * ManageV1FilterMethodParameter type definition
8113
+ */
8114
+ /**
8115
+ * ManageV1FilterMethodParameter type definition
8116
+ */
8117
+ /**
8118
+ * ManageV1FilterMethodParameter type definition
8119
+ */
8120
+ /**
8121
+ * ManageV1FilterMethodParameter type definition
8122
+ */
8123
+ /**
8124
+ * ManageV1FilterMethodParameter type definition
8125
+ */
7559
8126
  type ManageV1FilterMethodParameter = typeof ManageV1FilterMethodParameter[keyof typeof ManageV1FilterMethodParameter];
7560
8127
  declare const ManageV1FilterMethodParameter: {
7561
8128
  readonly sync: "sync";
@@ -7595,6 +8162,24 @@ declare const ManageV1FilterMethodParameter: {
7595
8162
  /**
7596
8163
  * ManageV1FilterStatusParameter type definition
7597
8164
  */
8165
+ /**
8166
+ * ManageV1FilterStatusParameter type definition
8167
+ */
8168
+ /**
8169
+ * ManageV1FilterStatusParameter type definition
8170
+ */
8171
+ /**
8172
+ * ManageV1FilterStatusParameter type definition
8173
+ */
8174
+ /**
8175
+ * ManageV1FilterStatusParameter type definition
8176
+ */
8177
+ /**
8178
+ * ManageV1FilterStatusParameter type definition
8179
+ */
8180
+ /**
8181
+ * ManageV1FilterStatusParameter type definition
8182
+ */
7598
8183
  type ManageV1FilterStatusParameter = typeof ManageV1FilterStatusParameter[keyof typeof ManageV1FilterStatusParameter];
7599
8184
  declare const ManageV1FilterStatusParameter: {
7600
8185
  readonly succeeded: "succeeded";
@@ -7939,7 +8524,7 @@ interface TranscribeOptions {
7939
8524
  * AssemblyAI-specific options (passed directly to API)
7940
8525
  * @see https://www.assemblyai.com/docs/api-reference/transcripts/submit
7941
8526
  */
7942
- assemblyai?: Partial<TranscriptOptionalParams>;
8527
+ assemblyai?: Partial<TranscriptParams>;
7943
8528
  /**
7944
8529
  * Gladia-specific options (passed directly to API)
7945
8530
  * @see https://docs.gladia.io/
@@ -8884,96 +9469,6 @@ interface CallbackTranscriptionSuccessPayload {
8884
9469
  custom_metadata?: CallbackTranscriptionSuccessPayloadCustomMetadata;
8885
9470
  }
8886
9471
 
8887
- /**
8888
- * Generated by orval v7.9.0 🍺
8889
- * Do not edit manually.
8890
- * AssemblyAI API
8891
- * AssemblyAI API
8892
- * OpenAPI spec version: 1.3.4
8893
- */
8894
- /**
8895
- * The status of the redacted audio
8896
- */
8897
- type RedactedAudioStatus = (typeof RedactedAudioStatus)[keyof typeof RedactedAudioStatus];
8898
- declare const RedactedAudioStatus: {
8899
- readonly redacted_audio_ready: "redacted_audio_ready";
8900
- };
8901
-
8902
- /**
8903
- * Generated by orval v7.9.0 🍺
8904
- * Do not edit manually.
8905
- * AssemblyAI API
8906
- * AssemblyAI API
8907
- * OpenAPI spec version: 1.3.4
8908
- */
8909
-
8910
- interface RedactedAudioResponse {
8911
- /** The status of the redacted audio */
8912
- status: RedactedAudioStatus;
8913
- /** The URL of the redacted audio file */
8914
- redacted_audio_url: string;
8915
- }
8916
-
8917
- /**
8918
- * Generated by orval v7.9.0 🍺
8919
- * Do not edit manually.
8920
- * AssemblyAI API
8921
- * AssemblyAI API
8922
- * OpenAPI spec version: 1.3.4
8923
- */
8924
-
8925
- /**
8926
- * The notification when the redacted audio is ready.
8927
- */
8928
- type RedactedAudioNotification = RedactedAudioResponse;
8929
-
8930
- /**
8931
- * Generated by orval v7.9.0 🍺
8932
- * Do not edit manually.
8933
- * AssemblyAI API
8934
- * AssemblyAI API
8935
- * OpenAPI spec version: 1.3.4
8936
- */
8937
- /**
8938
- * The status of the transcript. Either completed or error.
8939
- */
8940
- type TranscriptReadyStatus = (typeof TranscriptReadyStatus)[keyof typeof TranscriptReadyStatus];
8941
- declare const TranscriptReadyStatus: {
8942
- readonly completed: "completed";
8943
- readonly error: "error";
8944
- };
8945
-
8946
- /**
8947
- * Generated by orval v7.9.0 🍺
8948
- * Do not edit manually.
8949
- * AssemblyAI API
8950
- * AssemblyAI API
8951
- * OpenAPI spec version: 1.3.4
8952
- */
8953
-
8954
- /**
8955
- * The notification when the transcript status is completed or error.
8956
- */
8957
- interface TranscriptReadyNotification {
8958
- /** The ID of the transcript */
8959
- transcript_id: string;
8960
- /** The status of the transcript. Either completed or error. */
8961
- status: TranscriptReadyStatus;
8962
- }
8963
-
8964
- /**
8965
- * Generated by orval v7.9.0 🍺
8966
- * Do not edit manually.
8967
- * AssemblyAI API
8968
- * AssemblyAI API
8969
- * OpenAPI spec version: 1.3.4
8970
- */
8971
-
8972
- /**
8973
- * The notifications sent to the webhook URL.
8974
- */
8975
- type TranscriptWebhookNotification = TranscriptReadyNotification | RedactedAudioNotification;
8976
-
8977
9472
  /**
8978
9473
  * Generated by orval v7.9.0 🍺
8979
9474
  * Do not edit manually.
@@ -10350,4 +10845,4 @@ interface SpeechToTextChunkResponseModel {
10350
10845
  entities?: SpeechToTextChunkResponseModelEntities;
10351
10846
  }
10352
10847
 
10353
- export { type CustomSpellingConfigDTO as $, type AssemblyAIStreamingOptions as A, type AudioToLlmListDTOError as B, type CallbackConfig as C, type DeepgramStreamingOptions as D, type AudioToLlmResultDTO as E, type FileResponse as F, type GladiaStreamingOptions as G, type CallbackConfigDto as H, CallbackMethodEnum as I, type CallbackTranscriptionErrorPayload as J, type CallbackTranscriptionErrorPayloadCustomMetadata as K, type ListTranscriptsOptions as L, type MessagesConfig as M, type NamedEntityRecognitionDTO as N, CallbackTranscriptionErrorPayloadEvent as O, type PreProcessingConfig as P, type CallbackTranscriptionSuccessPayload as Q, type RealtimeProcessingConfig as R, type StreamingOptions as S, type TranscribeOptions as T, type UnifiedTranscriptResponse as U, type CallbackTranscriptionSuccessPayloadCustomMetadata as V, type WordDTO as W, CallbackTranscriptionSuccessPayloadEvent as X, type ChapterizationDTOError as Y, type ChapterizationDTOResults as Z, type CodeSwitchingConfigDTO as _, type StreamingCallbacks as a, type Entity as a$, type CustomSpellingConfigDTOSpellingDictionary as a0, type CustomVocabularyConfigDTO as a1, type CustomVocabularyConfigDTOVocabularyItem as a2, type CustomVocabularyEntryDTO as a3, type DiarizationConfigDTO as a4, type DiarizationDTO as a5, type DiarizationDTOError as a6, type DisplayModeDTO as a7, type DisplayModeDTOError as a8, type ErrorDTO as a9, type SubtitlesConfigDTO as aA, SubtitlesFormatEnum as aB, SubtitlesStyleEnum as aC, type SummarizationConfigDTO as aD, type SummarizationDTOError as aE, SummaryTypesEnum as aF, TranscriptionControllerListV2KindItem as aG, type TranscriptionControllerListV2Params as aH, TranscriptionControllerListV2StatusItem as aI, type TranscriptionResultDTO as aJ, type TranslationConfigDTO as aK, type TranslationDTOError as aL, TranslationModelEnum as aM, type TranslationResultDTO as aN, type TranslationResultDTOError as aO, type TranscriptOptionalParams as aP, TranscriptStatus as aQ, type TranscriptWord as aR, AudioIntelligenceModelStatus as aS, type AutoHighlightResult as aT, type AutoHighlightsResult as aU, type Chapter$1 as aV, type ContentSafetyLabel as aW, type ContentSafetyLabelResult as aX, type ContentSafetyLabelsResult as aY, type ContentSafetyLabelsResultSeverityScoreSummary as aZ, type ContentSafetyLabelsResultSummary as a_, type InitTranscriptionRequest as aa, type InitTranscriptionRequestCustomMetadata as ab, type ModerationDTO as ac, type ModerationDTOError as ad, type NamedEntityRecognitionDTOError as ae, type NamesConsistencyDTO as af, type NamesConsistencyDTOError as ag, type PreRecordedRequestParamsResponse as ah, type PreRecordedResponseCustomMetadata as ai, type PreRecordedResponseFile as aj, PreRecordedResponseKind as ak, type PreRecordedResponsePostSessionMetadata as al, type PreRecordedResponseRequestParams as am, type PreRecordedResponseResult as an, PreRecordedResponseStatus as ao, type SentencesDTO as ap, type SentencesDTOError as aq, type SentimentAnalysisDTOError as ar, type SpeakerReidentificationDTO as as, type SpeakerReidentificationDTOError as at, type StreamingRequest as au, type StreamingRequestCustomMetadata as av, type StructuredDataExtractionConfigDTO as aw, type StructuredDataExtractionDTO as ax, type StructuredDataExtractionDTOError as ay, type SubtitleDTO as az, type StreamingSession as b, type TranscriptSentimentAnalysis as b$, EntityType as b0, type ListTranscriptsParams as b1, PiiPolicy as b2, RedactPiiAudioQuality as b3, type RedactedAudioNotification as b4, type RedactedAudioResponse as b5, RedactedAudioStatus as b6, Sentiment as b7, type SentimentAnalysisResult$1 as b8, type SentimentAnalysisResultChannel as b9, type TranscriptDisfluencies as bA, type TranscriptEntities as bB, type TranscriptEntityDetection as bC, type TranscriptFilterProfanity as bD, type TranscriptFormatText as bE, type TranscriptIabCategories as bF, type TranscriptIabCategoriesResult as bG, TranscriptLanguageCode as bH, type TranscriptLanguageCodeProperty as bI, type TranscriptLanguageConfidence as bJ, type TranscriptLanguageConfidenceThreshold as bK, type TranscriptLanguageDetection as bL, type TranscriptMultichannel as bM, type TranscriptOptionalParamsLanguageCode as bN, type TranscriptOptionalParamsLanguageCodeOneOf as bO, type TranscriptOptionalParamsRedactPiiSub as bP, type TranscriptOptionalParamsSpeakersExpected as bQ, type TranscriptOptionalParamsSpeechModel as bR, type TranscriptOptionalParamsSpeechThreshold as bS, type TranscriptOptionalParamsWebhookAuthHeaderName as bT, type TranscriptOptionalParamsWebhookAuthHeaderValue as bU, type TranscriptPunctuate as bV, type TranscriptReadyNotification as bW, TranscriptReadyStatus as bX, type TranscriptRedactPiiAudio as bY, type TranscriptRedactPiiAudioQuality as bZ, type TranscriptRedactPiiPolicies as b_, type SentimentAnalysisResultSpeaker as ba, type SeverityScoreSummary as bb, SpeechModel as bc, SubstitutionPolicy as bd, SummaryModel as be, SummaryType as bf, type Timestamp as bg, type TopicDetectionModelResult as bh, type TopicDetectionModelResultSummary as bi, type TopicDetectionResult$1 as bj, type TopicDetectionResultLabelsItem as bk, type Transcript as bl, type TranscriptAudioDuration as bm, type TranscriptAudioEndAt as bn, type TranscriptAudioStartFrom as bo, type TranscriptAutoChapters as bp, type TranscriptAutoHighlightsResult as bq, TranscriptBoostParam as br, type TranscriptBoostParamProperty as bs, type TranscriptChapters as bt, type TranscriptConfidence as bu, type TranscriptContentSafety as bv, type TranscriptContentSafetyLabels as bw, type TranscriptCustomSpelling as bx, type TranscriptCustomSpellingProperty as by, type TranscriptCustomTopics as bz, type StreamEvent as c, type TranscriptionWord as c$, type TranscriptSentimentAnalysisResults as c0, type TranscriptSpeakerLabels as c1, type TranscriptSpeakersExpected as c2, type TranscriptSpeechModel as c3, type TranscriptSpeechThreshold as c4, type TranscriptSpeedBoost as c5, type TranscriptSummary as c6, type TranscriptSummaryModel as c7, type TranscriptSummaryType as c8, type TranscriptText as c9, type TranscriptTextUsageTokens as cA, type TranscriptionSegment as cB, type RealtimeSessionCreateRequestGAModel as cC, RealtimeTranscriptionSessionCreateRequestTurnDetectionType as cD, RealtimeTranscriptionSessionCreateRequestInputAudioFormat as cE, AudioResponseFormat as cF, type CreateTranscription200One as cG, type CreateTranscriptionRequest as cH, type CreateTranscriptionRequestModel as cI, type CreateTranscriptionRequestStream as cJ, CreateTranscriptionRequestTimestampGranularitiesItem as cK, type CreateTranscriptionResponseDiarizedJson as cL, CreateTranscriptionResponseDiarizedJsonTask as cM, type CreateTranscriptionResponseDiarizedJsonUsage as cN, type CreateTranscriptionResponseJson as cO, type CreateTranscriptionResponseJsonLogprobsItem as cP, type CreateTranscriptionResponseJsonUsage as cQ, type CreateTranscriptionResponseVerboseJson as cR, type TranscriptTextUsageDuration as cS, TranscriptTextUsageDurationType as cT, type TranscriptTextUsageTokensInputTokenDetails as cU, TranscriptTextUsageTokensType as cV, type TranscriptionChunkingStrategy as cW, type TranscriptionChunkingStrategyAnyOf as cX, type TranscriptionDiarizedSegment as cY, TranscriptionDiarizedSegmentType as cZ, TranscriptionInclude as c_, type TranscriptThrottled as ca, type TranscriptUtterance as cb, type TranscriptUtteranceChannel as cc, type TranscriptUtterances as cd, type TranscriptWebhookAuthHeaderName as ce, type TranscriptWebhookNotification as cf, type TranscriptWebhookStatusCode as cg, type TranscriptWebhookUrl as ch, type TranscriptWordChannel as ci, type TranscriptWordSpeaker as cj, type TranscriptWords as ck, type StreamingUpdateConfiguration as cl, type Transcription as cm, type EntityError as cn, Status as co, type EntityReference as cp, type DiarizationProperties as cq, type DiarizationSpeakersProperties as cr, LanguageIdentificationMode as cs, type LanguageIdentificationProperties as ct, type LanguageIdentificationPropertiesSpeechModelMapping as cu, ProfanityFilterMode as cv, PunctuationMode as cw, type TranscriptionCustomProperties as cx, type TranscriptionLinks as cy, type TranscriptionProperties as cz, StreamingSupportedEncodingEnum as d, type ListenV1ResponseResultsChannelsItemSearchItemHitsItem as d$, type VadConfig as d0, VadConfigType as d1, type ListenV1Response as d2, type ManageV1FilterAccessorParameter as d3, ManageV1FilterDeploymentParameter as d4, type ManageV1LimitParameter as d5, type ManageV1PageParameter as d6, ManageV1FilterEndpointParameter as d7, ManageV1FilterMethodParameter as d8, type SharedTopics as d9, type ListenV1ModelParameter as dA, type ListenV1MultichannelParameter as dB, type ListenV1NumeralsParameter as dC, type ListenV1ParagraphsParameter as dD, type ListenV1ProfanityFilterParameter as dE, type ListenV1PunctuateParameter as dF, type ListenV1RedactParameter as dG, ListenV1RedactParameterOneOfItem as dH, type ListenV1ReplaceParameter as dI, type ListenV1ResponseMetadata as dJ, type ListenV1ResponseMetadataIntentsInfo as dK, type ListenV1ResponseMetadataModelInfo as dL, type ListenV1ResponseMetadataSentimentInfo as dM, type ListenV1ResponseMetadataSummaryInfo as dN, type ListenV1ResponseMetadataTopicsInfo as dO, type ListenV1ResponseResults as dP, type ListenV1ResponseResultsChannels as dQ, type ListenV1ResponseResultsChannelsItem as dR, type ListenV1ResponseResultsChannelsItemAlternativesItem as dS, type ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem as dT, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs as dU, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem as dV, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem as dW, type ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem as dX, type ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem as dY, type ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem as dZ, type ListenV1ResponseResultsChannelsItemSearchItem as d_, type SharedIntents as da, type SharedSentiments as db, type SharedCallbackParameter as dc, SharedCallbackMethodParameter as dd, type SharedSentimentParameter as de, type SharedSummarizeParameter as df, type SharedTagParameter as dg, type SharedTopicsParameter as dh, type SharedCustomTopicParameter as di, SharedCustomTopicModeParameter as dj, type SharedIntentsParameter as dk, type SharedCustomIntentParameter as dl, SharedCustomIntentModeParameter as dm, type SharedMipOptOutParameter as dn, type ListenV1DetectEntitiesParameter as dp, type ListenV1DetectLanguageParameter as dq, type ListenV1DiarizeParameter as dr, type ListenV1DictationParameter as ds, ListenV1EncodingParameter as dt, type ListenV1FillerWordsParameter as du, type ListenV1KeytermParameter as dv, type ListenV1KeywordsParameter as dw, type ListenV1LanguageParameter as dx, type ListenV1MeasurementsParameter as dy, type ListenV1MediaTranscribeParams as dz, StreamingSupportedBitDepthEnum as e, type SentimentSpeakerSummary as e$, type ListenV1ResponseResultsSummary as e0, type ListenV1ResponseResultsUtterances as e1, type ListenV1ResponseResultsUtterancesItem as e2, type ListenV1ResponseResultsUtterancesItemWordsItem as e3, type ListenV1SearchParameter as e4, type ListenV1SmartFormatParameter as e5, type ListenV1UttSplitParameter as e6, type ListenV1UtterancesParameter as e7, type ListenV1VersionParameter as e8, type ManageV1EndDateTimeParameter as e9, AutoChaptersResultErrorType as eA, type Chapter as eB, type JobInfo as eC, type LanguageIdentificationResult as eD, type LanguageIdentificationResultAlternative as eE, LanguageIdentificationResultError as eF, type LanguageIdentificationResultItem as eG, type LanguagePackInfo as eH, LanguagePackInfoWritingDirection as eI, type OutputConfigSrtOverrides as eJ, type RecognitionAlternative as eK, type RecognitionDisplay as eL, RecognitionDisplayDirection as eM, type RecognitionMetadata as eN, type RecognitionResult as eO, RecognitionResultAttachesTo as eP, RecognitionResultType as eQ, type RetrieveTranscriptResponse as eR, type RetrieveTranscriptResponseAudioEventSummary as eS, type RetrieveTranscriptResponseAudioEventSummaryChannels as eT, type RetrieveTranscriptResponseTranslations as eU, type SentimentAnalysisError as eV, SentimentAnalysisErrorType as eW, type SentimentAnalysisResult as eX, type SentimentAnalysisResultSentimentAnalysis as eY, type SentimentChannelSummary as eZ, type SentimentSegment as e_, type ManageV1FilterRequestIdParameter as ea, ManageV1FilterStatusParameter as eb, type ManageV1ProjectsRequestsListParams as ec, type ManageV1StartDateTimeParameter as ed, type SharedExtraParameter as ee, type SharedIntentsResults as ef, type SharedIntentsResultsIntents as eg, type SharedIntentsResultsIntentsSegmentsItem as eh, type SharedIntentsResultsIntentsSegmentsItemIntentsItem as ei, type SharedSentimentsAverage as ej, type SharedSentimentsSegmentsItem as ek, type SharedTopicsResults as el, type SharedTopicsResultsTopics as em, type SharedTopicsResultsTopicsSegmentsItem as en, type SharedTopicsResultsTopicsSegmentsItemTopicsItem as eo, JobType as ep, type AlignmentConfig as eq, type TranscriptionConfig as er, type TrackingData as es, type OutputConfig as et, OperatingPoint as eu, type AudioEventItem as ev, type AudioEventSummary as ew, type AudioEventSummaryItem as ex, type AutoChaptersResult as ey, type AutoChaptersResultError as ez, StreamingSupportedSampleRateEnum as f, SegmentedJsonExportOptionsFormat as f$, type SentimentSummary as f0, type SentimentSummaryDetail as f1, type SpokenFormRecognitionResult as f2, SpokenFormRecognitionResultType as f3, type SummarizationError as f4, SummarizationErrorType as f5, type SummarizationResult as f6, type TopicDetectionError as f7, TopicDetectionErrorType as f8, type TopicDetectionResult as f9, type BodySpeechToTextV1SpeechToTextPostLanguageCode as fA, BodySpeechToTextV1SpeechToTextPostModelId as fB, type BodySpeechToTextV1SpeechToTextPostNumSpeakers as fC, type BodySpeechToTextV1SpeechToTextPostSeed as fD, type BodySpeechToTextV1SpeechToTextPostTemperature as fE, BodySpeechToTextV1SpeechToTextPostTimestampsGranularity as fF, type BodySpeechToTextV1SpeechToTextPostWebhookId as fG, type BodySpeechToTextV1SpeechToTextPostWebhookMetadata as fH, type BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf as fI, type DetectedEntity as fJ, type DocxExportOptions as fK, DocxExportOptionsFormat as fL, type DocxExportOptionsMaxSegmentChars as fM, type DocxExportOptionsMaxSegmentDurationS as fN, type DocxExportOptionsSegmentOnSilenceLongerThanS as fO, type ExportOptions as fP, type HtmlExportOptions as fQ, HtmlExportOptionsFormat as fR, type HtmlExportOptionsMaxSegmentChars as fS, type HtmlExportOptionsMaxSegmentDurationS as fT, type HtmlExportOptionsSegmentOnSilenceLongerThanS as fU, type PdfExportOptions as fV, PdfExportOptionsFormat as fW, type PdfExportOptionsMaxSegmentChars as fX, type PdfExportOptionsMaxSegmentDurationS as fY, type PdfExportOptionsSegmentOnSilenceLongerThanS as fZ, type SegmentedJsonExportOptions as f_, type TopicDetectionSegment as fa, type TopicDetectionSegmentTopic as fb, type TopicDetectionSummary as fc, type TopicDetectionSummaryOverall as fd, type TrackingDataDetails as fe, type TranscriptionConfigAdditionalVocabItem as ff, TranscriptionConfigDiarization as fg, TranscriptionConfigMaxDelayMode as fh, type TranscriptionConfigPunctuationOverrides as fi, type TranscriptionConfigSpeakerDiarizationConfig as fj, type TranscriptionConfigTranscriptFilteringConfig as fk, type TranscriptionConfigTranscriptFilteringConfigReplacementsItem as fl, type TranslationError as fm, TranslationErrorType as fn, type TranslationSentence as fo, type WrittenFormRecognitionResult as fp, WrittenFormRecognitionResultType as fq, type SpeechToTextChunkResponseModel as fr, type AdditionalFormatResponseModel as fs, type AdditionalFormats as ft, type BodySpeechToTextV1SpeechToTextPost as fu, type BodySpeechToTextV1SpeechToTextPostCloudStorageUrl as fv, type BodySpeechToTextV1SpeechToTextPostDiarizationThreshold as fw, type BodySpeechToTextV1SpeechToTextPostEntityDetection as fx, type BodySpeechToTextV1SpeechToTextPostFile as fy, BodySpeechToTextV1SpeechToTextPostFileFormat as fz, StreamingSupportedModels as g, type ProviderStreamingOptions as g$, type SegmentedJsonExportOptionsMaxSegmentChars as g0, type SegmentedJsonExportOptionsMaxSegmentDurationS as g1, type SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS as g2, type SpeechToTextCharacterResponseModel as g3, type SpeechToTextCharacterResponseModelEnd as g4, type SpeechToTextCharacterResponseModelStart as g5, type SpeechToTextChunkResponseModelAdditionalFormats as g6, type SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem as g7, type SpeechToTextChunkResponseModelChannelIndex as g8, type SpeechToTextChunkResponseModelEntities as g9, type ProviderExtendedDataMap as gA, type StreamingProvider as gB, type BatchOnlyProvider as gC, type SessionStatus as gD, type Speaker as gE, type Word as gF, type Utterance as gG, type TranscriptionStatus as gH, type TranscriptMetadata as gI, type TranscriptData as gJ, type ListTranscriptsResponse as gK, type ProviderRawResponseMap as gL, type StreamEventType as gM, type SpeechEvent as gN, type TranslationEvent as gO, type SentimentEvent as gP, type EntityEvent as gQ, type SummarizationEvent as gR, type ChapterizationEvent as gS, type AudioAckEvent as gT, type LifecycleEvent as gU, type AudioChunk as gV, type RawWebSocketMessage as gW, type AssemblyAIUpdateConfiguration as gX, type OpenAIStreamingOptions as gY, type SonioxStreamingOptions as gZ, type ElevenLabsStreamingOptions as g_, type SpeechToTextChunkResponseModelTranscriptionId as ga, type SpeechToTextWordResponseModel as gb, type SpeechToTextWordResponseModelCharacters as gc, type SpeechToTextWordResponseModelEnd as gd, type SpeechToTextWordResponseModelSpeakerId as ge, type SpeechToTextWordResponseModelStart as gf, SpeechToTextWordResponseModelType as gg, type SrtExportOptions as gh, SrtExportOptionsFormat as gi, type SrtExportOptionsMaxCharactersPerLine as gj, type SrtExportOptionsMaxSegmentChars as gk, type SrtExportOptionsMaxSegmentDurationS as gl, type SrtExportOptionsSegmentOnSilenceLongerThanS as gm, type TxtExportOptions as gn, TxtExportOptionsFormat as go, type TxtExportOptionsMaxCharactersPerLine as gp, type TxtExportOptionsMaxSegmentChars as gq, type TxtExportOptionsMaxSegmentDurationS as gr, type TxtExportOptionsSegmentOnSilenceLongerThanS as gs, type SpeechmaticsOperatingPoint as gt, type TranscriptionModel as gu, type TranscriptionLanguage as gv, type AssemblyAIExtendedData as gw, type GladiaExtendedData as gx, type DeepgramExtendedData as gy, type ElevenLabsExtendedData as gz, type LanguageConfig as h, type StreamingOptionsForProvider as h0, type TranscribeStreamParams as h1, type BeginEvent as h2, type TurnEvent as h3, type TerminationEvent as h4, type ErrorEvent as h5, type StreamingEventMessage as h6, type StreamingWord as h7, type StreamingForceEndpoint as h8, type PostProcessingConfig as i, type TranscriptionMetadataDTO as j, type TranscriptionDTO as k, type TranslationDTO as l, type SummarizationDTO as m, type SentimentAnalysisDTO as n, type ChapterizationDTO as o, type PreRecordedResponse as p, type UtteranceDTO as q, TranscriptionLanguageCodeEnum as r, TranslationLanguageCodeEnum as s, StreamingSupportedRegions as t, type AddonErrorDTO as u, type AudioToLlmDTO as v, type AudioToLlmDTOError as w, type AudioToLlmDTOResults as x, type AudioToLlmListConfigDTO as y, type AudioToLlmListDTO as z };
10848
+ export { type CustomSpellingConfigDTO as $, type AssemblyAIStreamingOptions as A, type AudioToLlmListDTOError as B, type CallbackConfig as C, type DeepgramStreamingOptions as D, type AudioToLlmResultDTO as E, type FileResponse as F, type GladiaStreamingOptions as G, type CallbackConfigDto as H, CallbackMethodEnum as I, type CallbackTranscriptionErrorPayload as J, type CallbackTranscriptionErrorPayloadCustomMetadata as K, type ListTranscriptsOptions as L, type MessagesConfig as M, type NamedEntityRecognitionDTO as N, CallbackTranscriptionErrorPayloadEvent as O, type PreProcessingConfig as P, type CallbackTranscriptionSuccessPayload as Q, type RealtimeProcessingConfig as R, type StreamingOptions as S, type TranscribeOptions as T, type UnifiedTranscriptResponse as U, type CallbackTranscriptionSuccessPayloadCustomMetadata as V, type WordDTO as W, CallbackTranscriptionSuccessPayloadEvent as X, type ChapterizationDTOError as Y, type ChapterizationDTOResults as Z, type CodeSwitchingConfigDTO as _, type StreamingCallbacks as a, type CustomFormattingRequestBodyCustomFormatting as a$, type CustomSpellingConfigDTOSpellingDictionary as a0, type CustomVocabularyConfigDTO as a1, type CustomVocabularyConfigDTOVocabularyItem as a2, type CustomVocabularyEntryDTO as a3, type DiarizationConfigDTO as a4, type DiarizationDTO as a5, type DiarizationDTOError as a6, type DisplayModeDTO as a7, type DisplayModeDTOError as a8, type ErrorDTO as a9, type SubtitlesConfigDTO as aA, SubtitlesFormatEnum as aB, SubtitlesStyleEnum as aC, type SummarizationConfigDTO as aD, type SummarizationDTOError as aE, SummaryTypesEnum as aF, TranscriptionControllerListV2KindItem as aG, type TranscriptionControllerListV2Params as aH, TranscriptionControllerListV2StatusItem as aI, type TranscriptionResultDTO as aJ, type TranslationConfigDTO as aK, type TranslationDTOError as aL, TranslationModelEnum as aM, type TranslationResultDTO as aN, type TranslationResultDTOError as aO, TranscriptStatus as aP, type TranscriptWord as aQ, AudioIntelligenceModelStatus as aR, type AutoHighlightResult as aS, type AutoHighlightsResult as aT, type Chapter$1 as aU, type ContentSafetyLabel as aV, type ContentSafetyLabelResult as aW, type ContentSafetyLabelsResult as aX, type ContentSafetyLabelsResultSeverityScoreSummary as aY, type ContentSafetyLabelsResultSummary as aZ, type CustomFormattingRequestBody as a_, type InitTranscriptionRequest as aa, type InitTranscriptionRequestCustomMetadata as ab, type ModerationDTO as ac, type ModerationDTOError as ad, type NamedEntityRecognitionDTOError as ae, type NamesConsistencyDTO as af, type NamesConsistencyDTOError as ag, type PreRecordedRequestParamsResponse as ah, type PreRecordedResponseCustomMetadata as ai, type PreRecordedResponseFile as aj, PreRecordedResponseKind as ak, type PreRecordedResponsePostSessionMetadata as al, type PreRecordedResponseRequestParams as am, type PreRecordedResponseResult as an, PreRecordedResponseStatus as ao, type SentencesDTO as ap, type SentencesDTOError as aq, type SentimentAnalysisDTOError as ar, type SpeakerReidentificationDTO as as, type SpeakerReidentificationDTOError as at, type StreamingRequest as au, type StreamingRequestCustomMetadata as av, type StructuredDataExtractionConfigDTO as aw, type StructuredDataExtractionDTO as ax, type StructuredDataExtractionDTOError as ay, type SubtitleDTO as az, type StreamingSession as b, type TranscriptOptionalParamsSpeechUnderstandingRequest as b$, type CustomFormattingResponse as b0, type CustomFormattingResponseCustomFormatting as b1, type CustomFormattingResponseCustomFormattingMapping as b2, type Entity as b3, EntityType as b4, type ListTranscriptsParams as b5, PiiPolicy as b6, RedactPiiAudioQuality as b7, Sentiment as b8, type SentimentAnalysisResult$1 as b9, type TranscriptConfidence as bA, type TranscriptContentSafety as bB, type TranscriptContentSafetyLabels as bC, type TranscriptCustomSpelling as bD, type TranscriptCustomSpellingProperty as bE, type TranscriptCustomTopics as bF, type TranscriptDisfluencies as bG, type TranscriptDomain as bH, type TranscriptEntities as bI, type TranscriptEntityDetection as bJ, type TranscriptFilterProfanity as bK, type TranscriptFormatText as bL, type TranscriptIabCategories as bM, type TranscriptIabCategoriesResult as bN, TranscriptLanguageCode as bO, type TranscriptLanguageCodes as bP, type TranscriptLanguageConfidence as bQ, type TranscriptLanguageConfidenceThreshold as bR, type TranscriptLanguageDetection as bS, type TranscriptLanguageDetectionOptions as bT, type TranscriptMultichannel as bU, type TranscriptOptionalParamsLanguageDetectionOptions as bV, type TranscriptOptionalParamsRedactPiiAudioOptions as bW, TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod as bX, TranscriptOptionalParamsRemoveAudioTags as bY, type TranscriptOptionalParamsSpeakerOptions as bZ, type TranscriptOptionalParamsSpeechUnderstanding as b_, type SentimentAnalysisResultChannel as ba, type SentimentAnalysisResultSpeaker as bb, type SeverityScoreSummary as bc, type SpeakerIdentificationRequestBody as bd, type SpeakerIdentificationRequestBodySpeakerIdentification as be, SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType as bf, type SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems as bg, type SpeakerIdentificationResponse as bh, type SpeakerIdentificationResponseSpeakerIdentification as bi, type SpeakerIdentificationResponseSpeakerIdentificationMapping as bj, type SpeechModel as bk, SubstitutionPolicy as bl, SummaryModel as bm, SummaryType as bn, type Timestamp as bo, type TopicDetectionModelResult as bp, type TopicDetectionModelResultSummary as bq, type TopicDetectionResult$1 as br, type TopicDetectionResultLabelsItems as bs, type Transcript as bt, type TranscriptAudioDuration as bu, type TranscriptAudioEndAt as bv, type TranscriptAudioStartFrom as bw, type TranscriptAutoChapters as bx, type TranscriptAutoHighlightsResult as by, type TranscriptChapters as bz, type StreamEvent as c, type TranscriptionLinks as c$, type TranscriptParams as c0, type TranscriptParamsDomain as c1, type TranscriptParamsLanguageCode as c2, type TranscriptParamsLanguageCodes as c3, type TranscriptParamsRedactPiiSub as c4, type TranscriptParamsRemoveAudioTags as c5, type TranscriptParamsSpeakersExpected as c6, type TranscriptParamsSpeechThreshold as c7, type TranscriptParamsWebhookAuthHeaderName as c8, type TranscriptParamsWebhookAuthHeaderValue as c9, type TranscriptUtterance as cA, type TranscriptUtteranceChannel as cB, type TranscriptUtteranceTranslatedTexts as cC, type TranscriptUtterances as cD, type TranscriptWebhookAuthHeaderName as cE, type TranscriptWebhookStatusCode as cF, type TranscriptWebhookUrl as cG, type TranscriptWordChannel as cH, type TranscriptWordSpeaker as cI, type TranscriptWords as cJ, type TranslationRequestBody as cK, type TranslationRequestBodyTranslation as cL, type TranslationResponse as cM, type TranslationResponseTranslation as cN, type StreamingUpdateConfiguration as cO, type Transcription as cP, type EntityError as cQ, Status as cR, type EntityReference as cS, type DiarizationProperties as cT, type DiarizationSpeakersProperties as cU, LanguageIdentificationMode as cV, type LanguageIdentificationProperties as cW, type LanguageIdentificationPropertiesSpeechModelMapping as cX, ProfanityFilterMode as cY, PunctuationMode as cZ, type TranscriptionCustomProperties as c_, type TranscriptPunctuate as ca, type TranscriptRedactPiiAudio as cb, type TranscriptRedactPiiAudioOptions as cc, TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod as cd, type TranscriptRedactPiiAudioQuality as ce, type TranscriptRedactPiiPolicies as cf, TranscriptRemoveAudioTags as cg, type TranscriptRemoveAudioTagsProperty as ch, type TranscriptSentimentAnalysis as ci, type TranscriptSentimentAnalysisResults as cj, type TranscriptSpeakerLabels as ck, type TranscriptSpeakersExpected as cl, type TranscriptSpeechModel as cm, type TranscriptSpeechModels as cn, type TranscriptSpeechThreshold as co, type TranscriptSpeechUnderstanding as cp, type TranscriptSpeechUnderstandingRequest as cq, type TranscriptSpeechUnderstandingResponse as cr, type TranscriptSpeedBoost as cs, type TranscriptSummary as ct, type TranscriptSummaryModel as cu, type TranscriptSummaryType as cv, type TranscriptTemperature as cw, type TranscriptText as cx, type TranscriptThrottled as cy, type TranscriptTranslatedTexts as cz, StreamingSupportedEncodingEnum as d, type ListenV1MeasurementsParameter as d$, type TranscriptionProperties as d0, type TranscriptTextUsageTokens as d1, type TranscriptionSegment as d2, type RealtimeSessionCreateRequestGAModel as d3, RealtimeTranscriptionSessionCreateRequestTurnDetectionType as d4, RealtimeTranscriptionSessionCreateRequestInputAudioFormat as d5, AudioResponseFormat as d6, type CreateTranscription200One as d7, type CreateTranscriptionRequest as d8, type CreateTranscriptionRequestModel as d9, type ManageV1PageParameter as dA, ManageV1FilterEndpointParameter as dB, ManageV1FilterMethodParameter as dC, type SharedTopics as dD, type SharedIntents as dE, type SharedSentiments as dF, type SharedCallbackParameter as dG, SharedCallbackMethodParameter as dH, type SharedSentimentParameter as dI, type SharedSummarizeParameter as dJ, type SharedTagParameter as dK, type SharedTopicsParameter as dL, type SharedCustomTopicParameter as dM, SharedCustomTopicModeParameter as dN, type SharedIntentsParameter as dO, type SharedCustomIntentParameter as dP, SharedCustomIntentModeParameter as dQ, type SharedMipOptOutParameter as dR, type ListenV1DetectEntitiesParameter as dS, type ListenV1DetectLanguageParameter as dT, type ListenV1DiarizeParameter as dU, type ListenV1DictationParameter as dV, ListenV1EncodingParameter as dW, type ListenV1FillerWordsParameter as dX, type ListenV1KeytermParameter as dY, type ListenV1KeywordsParameter as dZ, type ListenV1LanguageParameter as d_, type CreateTranscriptionRequestStream as da, CreateTranscriptionRequestTimestampGranularitiesItem as db, type CreateTranscriptionResponseDiarizedJson as dc, CreateTranscriptionResponseDiarizedJsonTask as dd, type CreateTranscriptionResponseDiarizedJsonUsage as de, type CreateTranscriptionResponseJson as df, type CreateTranscriptionResponseJsonLogprobsItem as dg, type CreateTranscriptionResponseJsonUsage as dh, type CreateTranscriptionResponseVerboseJson as di, type TranscriptTextUsageDuration as dj, TranscriptTextUsageDurationType as dk, type TranscriptTextUsageTokensInputTokenDetails as dl, TranscriptTextUsageTokensType as dm, type TranscriptionChunkingStrategy as dn, type TranscriptionChunkingStrategyAnyOf as dp, type TranscriptionDiarizedSegment as dq, TranscriptionDiarizedSegmentType as dr, TranscriptionInclude as ds, type TranscriptionWord as dt, type VadConfig as du, VadConfigType as dv, type ListenV1Response as dw, type ManageV1FilterAccessorParameter as dx, ManageV1FilterDeploymentParameter as dy, type ManageV1LimitParameter as dz, StreamingSupportedBitDepthEnum as e, type AutoChaptersResult as e$, type ListenV1MediaTranscribeParams as e0, type ListenV1ModelParameter as e1, type ListenV1MultichannelParameter as e2, type ListenV1NumeralsParameter as e3, type ListenV1ParagraphsParameter as e4, type ListenV1ProfanityFilterParameter as e5, type ListenV1PunctuateParameter as e6, type ListenV1RedactParameter as e7, ListenV1RedactParameterOneOfItem as e8, type ListenV1ReplaceParameter as e9, type ListenV1UtterancesParameter as eA, type ListenV1VersionParameter as eB, type ManageV1EndDateTimeParameter as eC, type ManageV1FilterRequestIdParameter as eD, ManageV1FilterStatusParameter as eE, type ManageV1ProjectsRequestsListParams as eF, type ManageV1StartDateTimeParameter as eG, type SharedExtraParameter as eH, type SharedIntentsResults as eI, type SharedIntentsResultsIntents as eJ, type SharedIntentsResultsIntentsSegmentsItem as eK, type SharedIntentsResultsIntentsSegmentsItemIntentsItem as eL, type SharedSentimentsAverage as eM, type SharedSentimentsSegmentsItem as eN, type SharedTopicsResults as eO, type SharedTopicsResultsTopics as eP, type SharedTopicsResultsTopicsSegmentsItem as eQ, type SharedTopicsResultsTopicsSegmentsItemTopicsItem as eR, JobType as eS, type AlignmentConfig as eT, type TranscriptionConfig as eU, type TrackingData as eV, type OutputConfig as eW, OperatingPoint as eX, type AudioEventItem as eY, type AudioEventSummary as eZ, type AudioEventSummaryItem as e_, type ListenV1ResponseMetadata as ea, type ListenV1ResponseMetadataIntentsInfo as eb, type ListenV1ResponseMetadataModelInfo as ec, type ListenV1ResponseMetadataSentimentInfo as ed, type ListenV1ResponseMetadataSummaryInfo as ee, type ListenV1ResponseMetadataTopicsInfo as ef, type ListenV1ResponseResults as eg, type ListenV1ResponseResultsChannels as eh, type ListenV1ResponseResultsChannelsItem as ei, type ListenV1ResponseResultsChannelsItemAlternativesItem as ej, type ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem as ek, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs as el, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem as em, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem as en, type ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem as eo, type ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem as ep, type ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem as eq, type ListenV1ResponseResultsChannelsItemSearchItem as er, type ListenV1ResponseResultsChannelsItemSearchItemHitsItem as es, type ListenV1ResponseResultsSummary as et, type ListenV1ResponseResultsUtterances as eu, type ListenV1ResponseResultsUtterancesItem as ev, type ListenV1ResponseResultsUtterancesItemWordsItem as ew, type ListenV1SearchParameter as ex, type ListenV1SmartFormatParameter as ey, type ListenV1UttSplitParameter as ez, StreamingSupportedSampleRateEnum as f, type BodySpeechToTextV1SpeechToTextPostFile as f$, type AutoChaptersResultError as f0, AutoChaptersResultErrorType as f1, type Chapter as f2, type JobInfo as f3, type LanguageIdentificationResult as f4, type LanguageIdentificationResultAlternative as f5, LanguageIdentificationResultError as f6, type LanguageIdentificationResultItem as f7, type LanguagePackInfo as f8, LanguagePackInfoWritingDirection as f9, type TopicDetectionError as fA, TopicDetectionErrorType as fB, type TopicDetectionResult as fC, type TopicDetectionSegment as fD, type TopicDetectionSegmentTopic as fE, type TopicDetectionSummary as fF, type TopicDetectionSummaryOverall as fG, type TrackingDataDetails as fH, type TranscriptionConfigAdditionalVocabItem as fI, TranscriptionConfigDiarization as fJ, TranscriptionConfigMaxDelayMode as fK, type TranscriptionConfigPunctuationOverrides as fL, type TranscriptionConfigSpeakerDiarizationConfig as fM, type TranscriptionConfigTranscriptFilteringConfig as fN, type TranscriptionConfigTranscriptFilteringConfigReplacementsItem as fO, type TranslationError as fP, TranslationErrorType as fQ, type TranslationSentence as fR, type WrittenFormRecognitionResult as fS, WrittenFormRecognitionResultType as fT, type SpeechToTextChunkResponseModel as fU, type AdditionalFormatResponseModel as fV, type AdditionalFormats as fW, type BodySpeechToTextV1SpeechToTextPost as fX, type BodySpeechToTextV1SpeechToTextPostCloudStorageUrl as fY, type BodySpeechToTextV1SpeechToTextPostDiarizationThreshold as fZ, type BodySpeechToTextV1SpeechToTextPostEntityDetection as f_, type OutputConfigSrtOverrides as fa, type RecognitionAlternative as fb, type RecognitionDisplay as fc, RecognitionDisplayDirection as fd, type RecognitionMetadata as fe, type RecognitionResult as ff, RecognitionResultAttachesTo as fg, RecognitionResultType as fh, type RetrieveTranscriptResponse as fi, type RetrieveTranscriptResponseAudioEventSummary as fj, type RetrieveTranscriptResponseAudioEventSummaryChannels as fk, type RetrieveTranscriptResponseTranslations as fl, type SentimentAnalysisError as fm, SentimentAnalysisErrorType as fn, type SentimentAnalysisResult as fo, type SentimentAnalysisResultSentimentAnalysis as fp, type SentimentChannelSummary as fq, type SentimentSegment as fr, type SentimentSpeakerSummary as fs, type SentimentSummary as ft, type SentimentSummaryDetail as fu, type SpokenFormRecognitionResult as fv, SpokenFormRecognitionResultType as fw, type SummarizationError as fx, SummarizationErrorType as fy, type SummarizationResult as fz, StreamingSupportedModels as g, type DeepgramExtendedData as g$, BodySpeechToTextV1SpeechToTextPostFileFormat as g0, type BodySpeechToTextV1SpeechToTextPostLanguageCode as g1, BodySpeechToTextV1SpeechToTextPostModelId as g2, type BodySpeechToTextV1SpeechToTextPostNumSpeakers as g3, type BodySpeechToTextV1SpeechToTextPostSeed as g4, type BodySpeechToTextV1SpeechToTextPostTemperature as g5, BodySpeechToTextV1SpeechToTextPostTimestampsGranularity as g6, type BodySpeechToTextV1SpeechToTextPostWebhookId as g7, type BodySpeechToTextV1SpeechToTextPostWebhookMetadata as g8, type BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf as g9, type SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem as gA, type SpeechToTextChunkResponseModelChannelIndex as gB, type SpeechToTextChunkResponseModelEntities as gC, type SpeechToTextChunkResponseModelTranscriptionId as gD, type SpeechToTextWordResponseModel as gE, type SpeechToTextWordResponseModelCharacters as gF, type SpeechToTextWordResponseModelEnd as gG, type SpeechToTextWordResponseModelSpeakerId as gH, type SpeechToTextWordResponseModelStart as gI, SpeechToTextWordResponseModelType as gJ, type SrtExportOptions as gK, SrtExportOptionsFormat as gL, type SrtExportOptionsMaxCharactersPerLine as gM, type SrtExportOptionsMaxSegmentChars as gN, type SrtExportOptionsMaxSegmentDurationS as gO, type SrtExportOptionsSegmentOnSilenceLongerThanS as gP, type TxtExportOptions as gQ, TxtExportOptionsFormat as gR, type TxtExportOptionsMaxCharactersPerLine as gS, type TxtExportOptionsMaxSegmentChars as gT, type TxtExportOptionsMaxSegmentDurationS as gU, type TxtExportOptionsSegmentOnSilenceLongerThanS as gV, type SpeechmaticsOperatingPoint as gW, type TranscriptionModel as gX, type TranscriptionLanguage as gY, type AssemblyAIExtendedData as gZ, type GladiaExtendedData as g_, type DetectedEntity as ga, type DocxExportOptions as gb, DocxExportOptionsFormat as gc, type DocxExportOptionsMaxSegmentChars as gd, type DocxExportOptionsMaxSegmentDurationS as ge, type DocxExportOptionsSegmentOnSilenceLongerThanS as gf, type ExportOptions as gg, type HtmlExportOptions as gh, HtmlExportOptionsFormat as gi, type HtmlExportOptionsMaxSegmentChars as gj, type HtmlExportOptionsMaxSegmentDurationS as gk, type HtmlExportOptionsSegmentOnSilenceLongerThanS as gl, type PdfExportOptions as gm, PdfExportOptionsFormat as gn, type PdfExportOptionsMaxSegmentChars as go, type PdfExportOptionsMaxSegmentDurationS as gp, type PdfExportOptionsSegmentOnSilenceLongerThanS as gq, type SegmentedJsonExportOptions as gr, SegmentedJsonExportOptionsFormat as gs, type SegmentedJsonExportOptionsMaxSegmentChars as gt, type SegmentedJsonExportOptionsMaxSegmentDurationS as gu, type SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS as gv, type SpeechToTextCharacterResponseModel as gw, type SpeechToTextCharacterResponseModelEnd as gx, type SpeechToTextCharacterResponseModelStart as gy, type SpeechToTextChunkResponseModelAdditionalFormats as gz, type LanguageConfig as h, type ElevenLabsExtendedData as h0, type ProviderExtendedDataMap as h1, type StreamingProvider as h2, type BatchOnlyProvider as h3, type SessionStatus as h4, type Speaker as h5, type Word as h6, type Utterance as h7, type TranscriptionStatus as h8, type TranscriptMetadata as h9, type StreamingWord as hA, type StreamingForceEndpoint as hB, type TranscriptData as ha, type ListTranscriptsResponse as hb, type ProviderRawResponseMap as hc, type StreamEventType as hd, type SpeechEvent as he, type TranslationEvent as hf, type SentimentEvent as hg, type EntityEvent as hh, type SummarizationEvent as hi, type ChapterizationEvent as hj, type AudioAckEvent as hk, type LifecycleEvent as hl, type AudioChunk as hm, type RawWebSocketMessage as hn, type AssemblyAIUpdateConfiguration as ho, type OpenAIStreamingOptions as hp, type SonioxStreamingOptions as hq, type ElevenLabsStreamingOptions as hr, type ProviderStreamingOptions as hs, type StreamingOptionsForProvider as ht, type TranscribeStreamParams as hu, type BeginEvent as hv, type TurnEvent as hw, type TerminationEvent as hx, type ErrorEvent as hy, type StreamingEventMessage as hz, type PostProcessingConfig as i, type TranscriptionMetadataDTO as j, type TranscriptionDTO as k, type TranslationDTO as l, type SummarizationDTO as m, type SentimentAnalysisDTO as n, type ChapterizationDTO as o, type PreRecordedResponse as p, type UtteranceDTO as q, TranscriptionLanguageCodeEnum as r, TranslationLanguageCodeEnum as s, StreamingSupportedRegions as t, type AddonErrorDTO as u, type AudioToLlmDTO as v, type AudioToLlmDTOError as w, type AudioToLlmDTOResults as x, type AudioToLlmListConfigDTO as y, type AudioToLlmListDTO as z };