voice-router-dev 0.8.2 → 0.8.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +211 -0
- package/dist/constants.d.mts +33 -3
- package/dist/constants.d.ts +33 -3
- package/dist/constants.js +14 -1
- package/dist/constants.mjs +13 -1
- package/dist/{field-configs-DN2_WrYr.d.mts → field-configs-t_lVCkE5.d.mts} +1194 -1194
- package/dist/{field-configs-DN2_WrYr.d.ts → field-configs-t_lVCkE5.d.ts} +1194 -1194
- package/dist/field-configs.d.mts +1 -1
- package/dist/field-configs.d.ts +1 -1
- package/dist/field-configs.js +8 -8
- package/dist/field-configs.mjs +8 -8
- package/dist/index.d.mts +651 -1628
- package/dist/index.d.ts +651 -1628
- package/dist/index.js +203 -51
- package/dist/index.mjs +200 -51
- package/dist/{provider-metadata-BnkedpXm.d.mts → provider-metadata-MDUUEuqF.d.mts} +4 -4
- package/dist/{provider-metadata-DbsSGAO7.d.ts → provider-metadata-_gUWlRXS.d.ts} +4 -4
- package/dist/provider-metadata.d.mts +1 -1
- package/dist/provider-metadata.d.ts +1 -1
- package/dist/{speechToTextChunkResponseModel-DExUFZT3.d.ts → speechToTextChunkResponseModel-DjL2ncnf.d.ts} +1147 -10
- package/dist/{speechToTextChunkResponseModel-3IUnJXKx.d.mts → speechToTextChunkResponseModel-DvIT4xai.d.mts} +1147 -10
- package/dist/webhooks.d.mts +234 -215
- package/dist/webhooks.d.ts +234 -215
- package/package.json +1 -1
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { DeepgramModelCode, DeepgramTopicModeType, DeepgramRedactType, AssemblyAISampleRateType, AssemblyAIEncodingType, AssemblyAISpeechModelType, SonioxRealtimeModelCode, SonioxLanguageCode, ElevenLabsRealtimeModelCode, ElevenLabsAudioFormatType, SonioxModelCode, ElevenLabsModelCode, DeepgramLanguageCode, ElevenLabsLanguageCode, SpeechmaticsLanguageCode, AzureLocaleCode } from './constants.js';
|
|
2
|
-
import { e as StreamingProviderType, B as BatchOnlyProviderType, T as TranscriptionProvider } from './provider-metadata-
|
|
2
|
+
import { e as StreamingProviderType, B as BatchOnlyProviderType, T as TranscriptionProvider } from './provider-metadata-_gUWlRXS.js';
|
|
3
3
|
|
|
4
4
|
/**
|
|
5
5
|
* Unified audio encoding types for Voice Router SDK
|
|
@@ -3469,7 +3469,7 @@ type TopicDetectionResultLabelsItem = {
|
|
|
3469
3469
|
/**
|
|
3470
3470
|
* The result of the topic detection model
|
|
3471
3471
|
*/
|
|
3472
|
-
interface TopicDetectionResult {
|
|
3472
|
+
interface TopicDetectionResult$1 {
|
|
3473
3473
|
/** The text in the transcript in which a detected topic occurs */
|
|
3474
3474
|
text: string;
|
|
3475
3475
|
/** An array of detected topics in the text */
|
|
@@ -3508,7 +3508,7 @@ interface TopicDetectionModelResult {
|
|
|
3508
3508
|
/** The status of the Topic Detection model. Either success, or unavailable in the rare case that the model failed. */
|
|
3509
3509
|
status: AudioIntelligenceModelStatus;
|
|
3510
3510
|
/** An array of results for the Topic Detection model */
|
|
3511
|
-
results: TopicDetectionResult[];
|
|
3511
|
+
results: TopicDetectionResult$1[];
|
|
3512
3512
|
/** The overall relevance of topic to the entire audio file */
|
|
3513
3513
|
summary: TopicDetectionModelResultSummary;
|
|
3514
3514
|
}
|
|
@@ -3580,7 +3580,7 @@ type TranscriptAutoChapters = boolean | null;
|
|
|
3580
3580
|
/**
|
|
3581
3581
|
* Chapter of the audio file
|
|
3582
3582
|
*/
|
|
3583
|
-
interface Chapter {
|
|
3583
|
+
interface Chapter$1 {
|
|
3584
3584
|
/** An ultra-short summary (just a few words) of the content spoken in the chapter */
|
|
3585
3585
|
gist: string;
|
|
3586
3586
|
/** A single sentence summary of the content spoken during the chapter */
|
|
@@ -3604,7 +3604,7 @@ interface Chapter {
|
|
|
3604
3604
|
/**
|
|
3605
3605
|
* An array of temporally sequential chapters for the audio file
|
|
3606
3606
|
*/
|
|
3607
|
-
type TranscriptChapters = Chapter[] | null;
|
|
3607
|
+
type TranscriptChapters = Chapter$1[] | null;
|
|
3608
3608
|
|
|
3609
3609
|
/**
|
|
3610
3610
|
* Generated by orval v7.9.0 🍺
|
|
@@ -3718,7 +3718,7 @@ type SentimentAnalysisResultSpeaker = string | null;
|
|
|
3718
3718
|
/**
|
|
3719
3719
|
* The result of the Sentiment Analysis model
|
|
3720
3720
|
*/
|
|
3721
|
-
interface SentimentAnalysisResult {
|
|
3721
|
+
interface SentimentAnalysisResult$1 {
|
|
3722
3722
|
/** The transcript of the sentence */
|
|
3723
3723
|
text: string;
|
|
3724
3724
|
/** The starting time, in milliseconds, of the sentence */
|
|
@@ -3752,7 +3752,7 @@ interface SentimentAnalysisResult {
|
|
|
3752
3752
|
See [Sentiment Analysis](https://www.assemblyai.com/docs/models/sentiment-analysis) for more information.
|
|
3753
3753
|
|
|
3754
3754
|
*/
|
|
3755
|
-
type TranscriptSentimentAnalysisResults = SentimentAnalysisResult[] | null;
|
|
3755
|
+
type TranscriptSentimentAnalysisResults = SentimentAnalysisResult$1[] | null;
|
|
3756
3756
|
|
|
3757
3757
|
/**
|
|
3758
3758
|
* Generated by orval v7.9.0 🍺
|
|
@@ -4443,6 +4443,9 @@ type SharedCallbackParameter = string;
|
|
|
4443
4443
|
/**
|
|
4444
4444
|
* SharedCallbackMethodParameter type definition
|
|
4445
4445
|
*/
|
|
4446
|
+
/**
|
|
4447
|
+
* SharedCallbackMethodParameter type definition
|
|
4448
|
+
*/
|
|
4446
4449
|
type SharedCallbackMethodParameter = typeof SharedCallbackMethodParameter[keyof typeof SharedCallbackMethodParameter];
|
|
4447
4450
|
declare const SharedCallbackMethodParameter: {
|
|
4448
4451
|
readonly POST: "POST";
|
|
@@ -4544,6 +4547,9 @@ type SharedCustomTopicParameter = string | string[];
|
|
|
4544
4547
|
/**
|
|
4545
4548
|
* SharedCustomTopicModeParameter type definition
|
|
4546
4549
|
*/
|
|
4550
|
+
/**
|
|
4551
|
+
* SharedCustomTopicModeParameter type definition
|
|
4552
|
+
*/
|
|
4547
4553
|
type SharedCustomTopicModeParameter = typeof SharedCustomTopicModeParameter[keyof typeof SharedCustomTopicModeParameter];
|
|
4548
4554
|
declare const SharedCustomTopicModeParameter: {
|
|
4549
4555
|
readonly extended: "extended";
|
|
@@ -4602,6 +4608,9 @@ type SharedCustomIntentParameter = string | string[];
|
|
|
4602
4608
|
/**
|
|
4603
4609
|
* SharedCustomIntentModeParameter type definition
|
|
4604
4610
|
*/
|
|
4611
|
+
/**
|
|
4612
|
+
* SharedCustomIntentModeParameter type definition
|
|
4613
|
+
*/
|
|
4605
4614
|
type SharedCustomIntentModeParameter = typeof SharedCustomIntentModeParameter[keyof typeof SharedCustomIntentModeParameter];
|
|
4606
4615
|
declare const SharedCustomIntentModeParameter: {
|
|
4607
4616
|
readonly extended: "extended";
|
|
@@ -4686,6 +4695,9 @@ type ListenV1DictationParameter = boolean;
|
|
|
4686
4695
|
/**
|
|
4687
4696
|
* ListenV1EncodingParameter type definition
|
|
4688
4697
|
*/
|
|
4698
|
+
/**
|
|
4699
|
+
* ListenV1EncodingParameter type definition
|
|
4700
|
+
*/
|
|
4689
4701
|
type ListenV1EncodingParameter = typeof ListenV1EncodingParameter[keyof typeof ListenV1EncodingParameter];
|
|
4690
4702
|
declare const ListenV1EncodingParameter: {
|
|
4691
4703
|
readonly linear16: "linear16";
|
|
@@ -7460,6 +7472,9 @@ type ManageV1FilterRequestIdParameter = string;
|
|
|
7460
7472
|
/**
|
|
7461
7473
|
* ManageV1FilterDeploymentParameter type definition
|
|
7462
7474
|
*/
|
|
7475
|
+
/**
|
|
7476
|
+
* ManageV1FilterDeploymentParameter type definition
|
|
7477
|
+
*/
|
|
7463
7478
|
type ManageV1FilterDeploymentParameter = typeof ManageV1FilterDeploymentParameter[keyof typeof ManageV1FilterDeploymentParameter];
|
|
7464
7479
|
declare const ManageV1FilterDeploymentParameter: {
|
|
7465
7480
|
readonly hosted: "hosted";
|
|
@@ -7495,6 +7510,9 @@ declare const ManageV1FilterDeploymentParameter: {
|
|
|
7495
7510
|
/**
|
|
7496
7511
|
* ManageV1FilterEndpointParameter type definition
|
|
7497
7512
|
*/
|
|
7513
|
+
/**
|
|
7514
|
+
* ManageV1FilterEndpointParameter type definition
|
|
7515
|
+
*/
|
|
7498
7516
|
type ManageV1FilterEndpointParameter = typeof ManageV1FilterEndpointParameter[keyof typeof ManageV1FilterEndpointParameter];
|
|
7499
7517
|
declare const ManageV1FilterEndpointParameter: {
|
|
7500
7518
|
readonly listen: "listen";
|
|
@@ -7535,6 +7553,9 @@ declare const ManageV1FilterEndpointParameter: {
|
|
|
7535
7553
|
/**
|
|
7536
7554
|
* ManageV1FilterMethodParameter type definition
|
|
7537
7555
|
*/
|
|
7556
|
+
/**
|
|
7557
|
+
* ManageV1FilterMethodParameter type definition
|
|
7558
|
+
*/
|
|
7538
7559
|
type ManageV1FilterMethodParameter = typeof ManageV1FilterMethodParameter[keyof typeof ManageV1FilterMethodParameter];
|
|
7539
7560
|
declare const ManageV1FilterMethodParameter: {
|
|
7540
7561
|
readonly sync: "sync";
|
|
@@ -7571,6 +7592,9 @@ declare const ManageV1FilterMethodParameter: {
|
|
|
7571
7592
|
/**
|
|
7572
7593
|
* ManageV1FilterStatusParameter type definition
|
|
7573
7594
|
*/
|
|
7595
|
+
/**
|
|
7596
|
+
* ManageV1FilterStatusParameter type definition
|
|
7597
|
+
*/
|
|
7574
7598
|
type ManageV1FilterStatusParameter = typeof ManageV1FilterStatusParameter[keyof typeof ManageV1FilterStatusParameter];
|
|
7575
7599
|
declare const ManageV1FilterStatusParameter: {
|
|
7576
7600
|
readonly succeeded: "succeeded";
|
|
@@ -7688,11 +7712,11 @@ type TranscriptionLanguage = TranscriptLanguageCode | TranscriptionLanguageCodeE
|
|
|
7688
7712
|
*/
|
|
7689
7713
|
interface AssemblyAIExtendedData {
|
|
7690
7714
|
/** Auto-generated chapters with summaries */
|
|
7691
|
-
chapters?: Chapter[];
|
|
7715
|
+
chapters?: Chapter$1[];
|
|
7692
7716
|
/** Detected named entities (people, organizations, locations) */
|
|
7693
7717
|
entities?: Entity[];
|
|
7694
7718
|
/** Per-utterance sentiment analysis results */
|
|
7695
|
-
sentimentResults?: SentimentAnalysisResult[];
|
|
7719
|
+
sentimentResults?: SentimentAnalysisResult$1[];
|
|
7696
7720
|
/** Key phrases and highlights */
|
|
7697
7721
|
highlights?: AutoHighlightsResult;
|
|
7698
7722
|
/** Content safety/moderation labels */
|
|
@@ -8950,6 +8974,1119 @@ interface TranscriptReadyNotification {
|
|
|
8950
8974
|
*/
|
|
8951
8975
|
type TranscriptWebhookNotification = TranscriptReadyNotification | RedactedAudioNotification;
|
|
8952
8976
|
|
|
8977
|
+
/**
|
|
8978
|
+
* Generated by orval v7.9.0 🍺
|
|
8979
|
+
* Do not edit manually.
|
|
8980
|
+
* Speechmatics ASR REST API
|
|
8981
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
8982
|
+
* OpenAPI spec version: 2.0.0
|
|
8983
|
+
*/
|
|
8984
|
+
interface AlignmentConfig {
|
|
8985
|
+
language: string;
|
|
8986
|
+
}
|
|
8987
|
+
|
|
8988
|
+
/**
|
|
8989
|
+
* Generated by orval v7.9.0 🍺
|
|
8990
|
+
* Do not edit manually.
|
|
8991
|
+
* Speechmatics ASR REST API
|
|
8992
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
8993
|
+
* OpenAPI spec version: 2.0.0
|
|
8994
|
+
*/
|
|
8995
|
+
interface AudioEventItem {
|
|
8996
|
+
/** Kind of audio event. E.g. music */
|
|
8997
|
+
type?: string;
|
|
8998
|
+
/** Time (in seconds) at which the audio event starts */
|
|
8999
|
+
start_time?: number;
|
|
9000
|
+
/** Time (in seconds) at which the audio event ends */
|
|
9001
|
+
end_time?: number;
|
|
9002
|
+
/** Prediction confidence associated with this event */
|
|
9003
|
+
confidence?: number;
|
|
9004
|
+
/** Input channel this event occurred on */
|
|
9005
|
+
channel?: string;
|
|
9006
|
+
}
|
|
9007
|
+
|
|
9008
|
+
/**
|
|
9009
|
+
* Generated by orval v7.9.0 🍺
|
|
9010
|
+
* Do not edit manually.
|
|
9011
|
+
* Speechmatics ASR REST API
|
|
9012
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9013
|
+
* OpenAPI spec version: 2.0.0
|
|
9014
|
+
*/
|
|
9015
|
+
/**
|
|
9016
|
+
* Summary statistics for this audio event type
|
|
9017
|
+
*/
|
|
9018
|
+
interface AudioEventSummaryItem {
|
|
9019
|
+
/** Total duration (in seconds) of all audio events of this type */
|
|
9020
|
+
total_duration?: number;
|
|
9021
|
+
/** Number of events of this type */
|
|
9022
|
+
count?: number;
|
|
9023
|
+
}
|
|
9024
|
+
|
|
9025
|
+
/**
|
|
9026
|
+
* Generated by orval v7.9.0 🍺
|
|
9027
|
+
* Do not edit manually.
|
|
9028
|
+
* Speechmatics ASR REST API
|
|
9029
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9030
|
+
* OpenAPI spec version: 2.0.0
|
|
9031
|
+
*/
|
|
9032
|
+
|
|
9033
|
+
interface AudioEventSummary {
|
|
9034
|
+
[key: string]: AudioEventSummaryItem;
|
|
9035
|
+
}
|
|
9036
|
+
|
|
9037
|
+
/**
|
|
9038
|
+
* Generated by orval v7.9.0 🍺
|
|
9039
|
+
* Do not edit manually.
|
|
9040
|
+
* Speechmatics ASR REST API
|
|
9041
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9042
|
+
* OpenAPI spec version: 2.0.0
|
|
9043
|
+
*/
|
|
9044
|
+
interface Chapter {
|
|
9045
|
+
title?: string;
|
|
9046
|
+
summary?: string;
|
|
9047
|
+
start_time?: number;
|
|
9048
|
+
end_time?: number;
|
|
9049
|
+
}
|
|
9050
|
+
|
|
9051
|
+
/**
|
|
9052
|
+
* Generated by orval v7.9.0 🍺
|
|
9053
|
+
* Do not edit manually.
|
|
9054
|
+
* Speechmatics ASR REST API
|
|
9055
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9056
|
+
* OpenAPI spec version: 2.0.0
|
|
9057
|
+
*/
|
|
9058
|
+
|
|
9059
|
+
/**
|
|
9060
|
+
* An array of objects that represent summarized chapters of the transcript
|
|
9061
|
+
*/
|
|
9062
|
+
type AutoChaptersResult = Chapter[];
|
|
9063
|
+
|
|
9064
|
+
/**
|
|
9065
|
+
* Generated by orval v7.9.0 🍺
|
|
9066
|
+
* Do not edit manually.
|
|
9067
|
+
* Speechmatics ASR REST API
|
|
9068
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9069
|
+
* OpenAPI spec version: 2.0.0
|
|
9070
|
+
*/
|
|
9071
|
+
type AutoChaptersResultErrorType = (typeof AutoChaptersResultErrorType)[keyof typeof AutoChaptersResultErrorType];
|
|
9072
|
+
declare const AutoChaptersResultErrorType: {
|
|
9073
|
+
readonly auto_chapters_failed: "auto_chapters_failed";
|
|
9074
|
+
readonly unsupported_language: "unsupported_language";
|
|
9075
|
+
};
|
|
9076
|
+
|
|
9077
|
+
/**
|
|
9078
|
+
* Generated by orval v7.9.0 🍺
|
|
9079
|
+
* Do not edit manually.
|
|
9080
|
+
* Speechmatics ASR REST API
|
|
9081
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9082
|
+
* OpenAPI spec version: 2.0.0
|
|
9083
|
+
*/
|
|
9084
|
+
|
|
9085
|
+
interface AutoChaptersResultError {
|
|
9086
|
+
type?: AutoChaptersResultErrorType;
|
|
9087
|
+
/** Human readable error message */
|
|
9088
|
+
message?: string;
|
|
9089
|
+
}
|
|
9090
|
+
|
|
9091
|
+
/**
|
|
9092
|
+
* Generated by orval v7.9.0 🍺
|
|
9093
|
+
* Do not edit manually.
|
|
9094
|
+
* Speechmatics ASR REST API
|
|
9095
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9096
|
+
* OpenAPI spec version: 2.0.0
|
|
9097
|
+
*/
|
|
9098
|
+
type JobType = (typeof JobType)[keyof typeof JobType];
|
|
9099
|
+
declare const JobType: {
|
|
9100
|
+
readonly alignment: "alignment";
|
|
9101
|
+
readonly transcription: "transcription";
|
|
9102
|
+
};
|
|
9103
|
+
|
|
9104
|
+
/**
|
|
9105
|
+
* Generated by orval v7.9.0 🍺
|
|
9106
|
+
* Do not edit manually.
|
|
9107
|
+
* Speechmatics ASR REST API
|
|
9108
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9109
|
+
* OpenAPI spec version: 2.0.0
|
|
9110
|
+
*/
|
|
9111
|
+
type OperatingPoint = (typeof OperatingPoint)[keyof typeof OperatingPoint];
|
|
9112
|
+
declare const OperatingPoint: {
|
|
9113
|
+
readonly standard: "standard";
|
|
9114
|
+
readonly enhanced: "enhanced";
|
|
9115
|
+
};
|
|
9116
|
+
|
|
9117
|
+
/**
|
|
9118
|
+
* Generated by orval v7.9.0 🍺
|
|
9119
|
+
* Do not edit manually.
|
|
9120
|
+
* Speechmatics ASR REST API
|
|
9121
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9122
|
+
* OpenAPI spec version: 2.0.0
|
|
9123
|
+
*/
|
|
9124
|
+
type TranscriptionConfigAdditionalVocabItem = {
|
|
9125
|
+
content: string;
|
|
9126
|
+
sounds_like?: string[];
|
|
9127
|
+
};
|
|
9128
|
+
|
|
9129
|
+
/**
|
|
9130
|
+
* Generated by orval v7.9.0 🍺
|
|
9131
|
+
* Do not edit manually.
|
|
9132
|
+
* Speechmatics ASR REST API
|
|
9133
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9134
|
+
* OpenAPI spec version: 2.0.0
|
|
9135
|
+
*/
|
|
9136
|
+
/**
|
|
9137
|
+
* Control punctuation settings.
|
|
9138
|
+
*/
|
|
9139
|
+
type TranscriptionConfigPunctuationOverrides = {
|
|
9140
|
+
/**
|
|
9141
|
+
* Ranges between zero and one. Higher values will produce more punctuation. The default is 0.5.
|
|
9142
|
+
* @minimum 0
|
|
9143
|
+
* @maximum 1
|
|
9144
|
+
*/
|
|
9145
|
+
sensitivity?: number;
|
|
9146
|
+
/** The punctuation marks which the client is prepared to accept in transcription output, or the special value 'all' (the default). Unsupported marks are ignored. This value is used to guide the transcription process. */
|
|
9147
|
+
permitted_marks?: string[];
|
|
9148
|
+
};
|
|
9149
|
+
|
|
9150
|
+
/**
|
|
9151
|
+
* Generated by orval v7.9.0 🍺
|
|
9152
|
+
* Do not edit manually.
|
|
9153
|
+
* Speechmatics ASR REST API
|
|
9154
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9155
|
+
* OpenAPI spec version: 2.0.0
|
|
9156
|
+
*/
|
|
9157
|
+
/**
|
|
9158
|
+
* Specify whether speaker or channel labels are added to the transcript.
|
|
9159
|
+
The default is `none`.
|
|
9160
|
+
- **none**: no speaker or channel labels are added.
|
|
9161
|
+
- **speaker**: speaker attribution is performed based on acoustic matching;
|
|
9162
|
+
all input channels are mixed into a single stream for processing.
|
|
9163
|
+
- **channel**: multiple input channels are processed individually and collated
|
|
9164
|
+
into a single transcript.
|
|
9165
|
+
*/
|
|
9166
|
+
type TranscriptionConfigDiarization = (typeof TranscriptionConfigDiarization)[keyof typeof TranscriptionConfigDiarization];
|
|
9167
|
+
declare const TranscriptionConfigDiarization: {
|
|
9168
|
+
readonly none: "none";
|
|
9169
|
+
readonly speaker: "speaker";
|
|
9170
|
+
readonly channel: "channel";
|
|
9171
|
+
};
|
|
9172
|
+
|
|
9173
|
+
/**
|
|
9174
|
+
* Generated by orval v7.9.0 🍺
|
|
9175
|
+
* Do not edit manually.
|
|
9176
|
+
* Speechmatics ASR REST API
|
|
9177
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9178
|
+
* OpenAPI spec version: 2.0.0
|
|
9179
|
+
*/
|
|
9180
|
+
/**
|
|
9181
|
+
* Whether or not to enable flexible endpointing and allow the entity to continue to be spoken.
|
|
9182
|
+
*/
|
|
9183
|
+
type TranscriptionConfigMaxDelayMode = (typeof TranscriptionConfigMaxDelayMode)[keyof typeof TranscriptionConfigMaxDelayMode];
|
|
9184
|
+
declare const TranscriptionConfigMaxDelayMode: {
|
|
9185
|
+
readonly fixed: "fixed";
|
|
9186
|
+
readonly flexible: "flexible";
|
|
9187
|
+
};
|
|
9188
|
+
|
|
9189
|
+
/**
|
|
9190
|
+
* Generated by orval v7.9.0 🍺
|
|
9191
|
+
* Do not edit manually.
|
|
9192
|
+
* Speechmatics ASR REST API
|
|
9193
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9194
|
+
* OpenAPI spec version: 2.0.0
|
|
9195
|
+
*/
|
|
9196
|
+
type TranscriptionConfigTranscriptFilteringConfigReplacementsItem = {
|
|
9197
|
+
from: string;
|
|
9198
|
+
to: string;
|
|
9199
|
+
};
|
|
9200
|
+
|
|
9201
|
+
/**
|
|
9202
|
+
* Generated by orval v7.9.0 🍺
|
|
9203
|
+
* Do not edit manually.
|
|
9204
|
+
* Speechmatics ASR REST API
|
|
9205
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9206
|
+
* OpenAPI spec version: 2.0.0
|
|
9207
|
+
*/
|
|
9208
|
+
|
|
9209
|
+
/**
|
|
9210
|
+
* Configuration for applying filtering to the transcription
|
|
9211
|
+
*/
|
|
9212
|
+
type TranscriptionConfigTranscriptFilteringConfig = {
|
|
9213
|
+
/** If true, words that are identified as disfluencies will be removed from the transcript. If false (default), they are tagged in the transcript as 'disfluency'. */
|
|
9214
|
+
remove_disfluencies?: boolean;
|
|
9215
|
+
/** A list of replacements to apply to the transcript. Each replacement is a pair of strings, where the first string is the pattern to be replaced and the second string is the replacement text. */
|
|
9216
|
+
replacements?: TranscriptionConfigTranscriptFilteringConfigReplacementsItem[];
|
|
9217
|
+
};
|
|
9218
|
+
|
|
9219
|
+
/**
|
|
9220
|
+
* Generated by orval v7.9.0 🍺
|
|
9221
|
+
* Do not edit manually.
|
|
9222
|
+
* Speechmatics ASR REST API
|
|
9223
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9224
|
+
* OpenAPI spec version: 2.0.0
|
|
9225
|
+
*/
|
|
9226
|
+
/**
|
|
9227
|
+
* Configuration for speaker diarization
|
|
9228
|
+
*/
|
|
9229
|
+
type TranscriptionConfigSpeakerDiarizationConfig = {
|
|
9230
|
+
/** If true, the algorithm will prefer to stay with the current active speaker if it is a close enough match, even if other speakers may be closer. This is useful for cases where we can flip incorrectly between similar speakers during a single speaker section." */
|
|
9231
|
+
prefer_current_speaker?: boolean;
|
|
9232
|
+
/**
|
|
9233
|
+
* Controls how sensitive the algorithm is in terms of keeping similar speakers separate, as opposed to combining them into a single speaker. Higher values will typically lead to more speakers, as the degree of difference between speakers in order to allow them to remain distinct will be lower. A lower value for this parameter will conversely guide the algorithm towards being less sensitive in terms of retaining similar speakers, and as such may lead to fewer speakers overall. The default is 0.5.
|
|
9234
|
+
* @minimum 0
|
|
9235
|
+
* @maximum 1
|
|
9236
|
+
*/
|
|
9237
|
+
speaker_sensitivity?: number;
|
|
9238
|
+
};
|
|
9239
|
+
|
|
9240
|
+
/**
|
|
9241
|
+
* Generated by orval v7.9.0 🍺
|
|
9242
|
+
* Do not edit manually.
|
|
9243
|
+
* Speechmatics ASR REST API
|
|
9244
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9245
|
+
* OpenAPI spec version: 2.0.0
|
|
9246
|
+
*/
|
|
9247
|
+
|
|
9248
|
+
interface TranscriptionConfig {
|
|
9249
|
+
/** Language model to process the audio input, normally specified as an ISO language code */
|
|
9250
|
+
language: string;
|
|
9251
|
+
/** Request a specialized model based on 'language' but optimized for a particular field, e.g. "finance" or "medical". */
|
|
9252
|
+
domain?: string;
|
|
9253
|
+
/** Language locale to be used when generating the transcription output, normally specified as an ISO language code */
|
|
9254
|
+
output_locale?: string;
|
|
9255
|
+
operating_point?: OperatingPoint;
|
|
9256
|
+
/** List of custom words or phrases that should be recognized. Alternative pronunciations can be specified to aid recognition. */
|
|
9257
|
+
additional_vocab?: TranscriptionConfigAdditionalVocabItem[];
|
|
9258
|
+
/** Control punctuation settings. */
|
|
9259
|
+
punctuation_overrides?: TranscriptionConfigPunctuationOverrides;
|
|
9260
|
+
/** Specify whether speaker or channel labels are added to the transcript.
|
|
9261
|
+
The default is `none`.
|
|
9262
|
+
- **none**: no speaker or channel labels are added.
|
|
9263
|
+
- **speaker**: speaker attribution is performed based on acoustic matching;
|
|
9264
|
+
all input channels are mixed into a single stream for processing.
|
|
9265
|
+
- **channel**: multiple input channels are processed individually and collated
|
|
9266
|
+
into a single transcript. */
|
|
9267
|
+
diarization?: TranscriptionConfigDiarization;
|
|
9268
|
+
/** Transcript labels to use when using collating separate input channels. */
|
|
9269
|
+
channel_diarization_labels?: string[];
|
|
9270
|
+
/** Include additional 'entity' objects in the transcription results (e.g. dates, numbers) and their original spoken form. These entities are interleaved with other types of results. The concatenation of these words is represented as a single entity with the concatenated written form present in the 'content' field. The entities contain a 'spoken_form' field, which can be used in place of the corresponding 'word' type results, in case a spoken form is preferred to a written form. They also contain a 'written_form', which can be used instead of the entity, if you want a breakdown of the words without spaces. They can still contain non-breaking spaces and other special whitespace characters, as they are considered part of the word for the formatting output. In case of a written_form, the individual word times are estimated and might not be accurate if the order of the words in the written form does not correspond to the order they were actually spoken (such as 'one hundred million dollars' and '$100 million'). */
|
|
9271
|
+
enable_entities?: boolean;
|
|
9272
|
+
/** Whether or not to enable flexible endpointing and allow the entity to continue to be spoken. */
|
|
9273
|
+
max_delay_mode?: TranscriptionConfigMaxDelayMode;
|
|
9274
|
+
/** Configuration for applying filtering to the transcription */
|
|
9275
|
+
transcript_filtering_config?: TranscriptionConfigTranscriptFilteringConfig;
|
|
9276
|
+
/** Configuration for speaker diarization */
|
|
9277
|
+
speaker_diarization_config?: TranscriptionConfigSpeakerDiarizationConfig;
|
|
9278
|
+
}
|
|
9279
|
+
|
|
9280
|
+
/**
|
|
9281
|
+
* Generated by orval v7.9.0 🍺
|
|
9282
|
+
* Do not edit manually.
|
|
9283
|
+
* Speechmatics ASR REST API
|
|
9284
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9285
|
+
* OpenAPI spec version: 2.0.0
|
|
9286
|
+
*/
|
|
9287
|
+
/**
|
|
9288
|
+
* Customer-defined JSON structure.
|
|
9289
|
+
*/
|
|
9290
|
+
type TrackingDataDetails = {
|
|
9291
|
+
[key: string]: unknown;
|
|
9292
|
+
};
|
|
9293
|
+
|
|
9294
|
+
/**
|
|
9295
|
+
* Generated by orval v7.9.0 🍺
|
|
9296
|
+
* Do not edit manually.
|
|
9297
|
+
* Speechmatics ASR REST API
|
|
9298
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9299
|
+
* OpenAPI spec version: 2.0.0
|
|
9300
|
+
*/
|
|
9301
|
+
|
|
9302
|
+
interface TrackingData {
|
|
9303
|
+
/** The title of the job. */
|
|
9304
|
+
title?: string;
|
|
9305
|
+
/** External system reference. */
|
|
9306
|
+
reference?: string;
|
|
9307
|
+
tags?: string[];
|
|
9308
|
+
/** Customer-defined JSON structure. */
|
|
9309
|
+
details?: TrackingDataDetails;
|
|
9310
|
+
}
|
|
9311
|
+
|
|
9312
|
+
/**
|
|
9313
|
+
* Generated by orval v7.9.0 🍺
|
|
9314
|
+
* Do not edit manually.
|
|
9315
|
+
* Speechmatics ASR REST API
|
|
9316
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9317
|
+
* OpenAPI spec version: 2.0.0
|
|
9318
|
+
*/
|
|
9319
|
+
/**
|
|
9320
|
+
* Parameters that override default values of srt conversion. max_line_length: sets maximum count of characters per subtitle line including white space. max_lines: sets maximum count of lines in a subtitle section.
|
|
9321
|
+
*/
|
|
9322
|
+
type OutputConfigSrtOverrides = {
|
|
9323
|
+
max_line_length?: number;
|
|
9324
|
+
max_lines?: number;
|
|
9325
|
+
};
|
|
9326
|
+
|
|
9327
|
+
/**
|
|
9328
|
+
* Generated by orval v7.9.0 🍺
|
|
9329
|
+
* Do not edit manually.
|
|
9330
|
+
* Speechmatics ASR REST API
|
|
9331
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9332
|
+
* OpenAPI spec version: 2.0.0
|
|
9333
|
+
*/
|
|
9334
|
+
|
|
9335
|
+
interface OutputConfig {
|
|
9336
|
+
/** Parameters that override default values of srt conversion. max_line_length: sets maximum count of characters per subtitle line including white space. max_lines: sets maximum count of lines in a subtitle section. */
|
|
9337
|
+
srt_overrides?: OutputConfigSrtOverrides;
|
|
9338
|
+
}
|
|
9339
|
+
|
|
9340
|
+
/**
|
|
9341
|
+
* Generated by orval v7.9.0 🍺
|
|
9342
|
+
* Do not edit manually.
|
|
9343
|
+
* Speechmatics ASR REST API
|
|
9344
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9345
|
+
* OpenAPI spec version: 2.0.0
|
|
9346
|
+
*/
|
|
9347
|
+
|
|
9348
|
+
/**
|
|
9349
|
+
* Summary information about an ASR job, to support identification and tracking.
|
|
9350
|
+
*/
|
|
9351
|
+
interface JobInfo {
|
|
9352
|
+
/** The UTC date time the job was created. */
|
|
9353
|
+
created_at: string;
|
|
9354
|
+
/** Name of data file submitted for job. */
|
|
9355
|
+
data_name: string;
|
|
9356
|
+
/**
|
|
9357
|
+
* The data file audio duration (in seconds).
|
|
9358
|
+
* @minimum 0
|
|
9359
|
+
*/
|
|
9360
|
+
duration: number;
|
|
9361
|
+
/** The unique id assigned to the job. */
|
|
9362
|
+
id: string;
|
|
9363
|
+
/** Name of the text file submitted to be aligned to audio. */
|
|
9364
|
+
text_name?: string;
|
|
9365
|
+
tracking?: TrackingData;
|
|
9366
|
+
}
|
|
9367
|
+
|
|
9368
|
+
/**
|
|
9369
|
+
* Generated by orval v7.9.0 🍺
|
|
9370
|
+
* Do not edit manually.
|
|
9371
|
+
* Speechmatics ASR REST API
|
|
9372
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9373
|
+
* OpenAPI spec version: 2.0.0
|
|
9374
|
+
*/
|
|
9375
|
+
interface LanguageIdentificationResultAlternative {
|
|
9376
|
+
language?: string;
|
|
9377
|
+
confidence?: number;
|
|
9378
|
+
}
|
|
9379
|
+
|
|
9380
|
+
/**
|
|
9381
|
+
* Generated by orval v7.9.0 🍺
|
|
9382
|
+
* Do not edit manually.
|
|
9383
|
+
* Speechmatics ASR REST API
|
|
9384
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9385
|
+
* OpenAPI spec version: 2.0.0
|
|
9386
|
+
*/
|
|
9387
|
+
|
|
9388
|
+
interface LanguageIdentificationResultItem {
|
|
9389
|
+
alternatives?: LanguageIdentificationResultAlternative[];
|
|
9390
|
+
start_time?: number;
|
|
9391
|
+
end_time?: number;
|
|
9392
|
+
}
|
|
9393
|
+
|
|
9394
|
+
/**
|
|
9395
|
+
* Generated by orval v7.9.0 🍺
|
|
9396
|
+
* Do not edit manually.
|
|
9397
|
+
* Speechmatics ASR REST API
|
|
9398
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9399
|
+
* OpenAPI spec version: 2.0.0
|
|
9400
|
+
*/
|
|
9401
|
+
type LanguageIdentificationResultError = (typeof LanguageIdentificationResultError)[keyof typeof LanguageIdentificationResultError];
|
|
9402
|
+
declare const LanguageIdentificationResultError: {
|
|
9403
|
+
readonly LOW_CONFIDENCE: "LOW_CONFIDENCE";
|
|
9404
|
+
readonly UNEXPECTED_LANGUAGE: "UNEXPECTED_LANGUAGE";
|
|
9405
|
+
readonly NO_SPEECH: "NO_SPEECH";
|
|
9406
|
+
readonly FILE_UNREADABLE: "FILE_UNREADABLE";
|
|
9407
|
+
readonly OTHER: "OTHER";
|
|
9408
|
+
};
|
|
9409
|
+
|
|
9410
|
+
/**
|
|
9411
|
+
* Generated by orval v7.9.0 🍺
|
|
9412
|
+
* Do not edit manually.
|
|
9413
|
+
* Speechmatics ASR REST API
|
|
9414
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9415
|
+
* OpenAPI spec version: 2.0.0
|
|
9416
|
+
*/
|
|
9417
|
+
|
|
9418
|
+
interface LanguageIdentificationResult {
|
|
9419
|
+
results?: LanguageIdentificationResultItem[];
|
|
9420
|
+
error?: LanguageIdentificationResultError;
|
|
9421
|
+
message?: string;
|
|
9422
|
+
}
|
|
9423
|
+
|
|
9424
|
+
/**
|
|
9425
|
+
* Generated by orval v7.9.0 🍺
|
|
9426
|
+
* Do not edit manually.
|
|
9427
|
+
* Speechmatics ASR REST API
|
|
9428
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9429
|
+
* OpenAPI spec version: 2.0.0
|
|
9430
|
+
*/
|
|
9431
|
+
/**
|
|
9432
|
+
* The direction that words in the language should be written and read in.
|
|
9433
|
+
*/
|
|
9434
|
+
type LanguagePackInfoWritingDirection = (typeof LanguagePackInfoWritingDirection)[keyof typeof LanguagePackInfoWritingDirection];
|
|
9435
|
+
declare const LanguagePackInfoWritingDirection: {
|
|
9436
|
+
readonly "left-to-right": "left-to-right";
|
|
9437
|
+
readonly "right-to-left": "right-to-left";
|
|
9438
|
+
};
|
|
9439
|
+
|
|
9440
|
+
/**
|
|
9441
|
+
* Generated by orval v7.9.0 🍺
|
|
9442
|
+
* Do not edit manually.
|
|
9443
|
+
* Speechmatics ASR REST API
|
|
9444
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9445
|
+
* OpenAPI spec version: 2.0.0
|
|
9446
|
+
*/
|
|
9447
|
+
|
|
9448
|
+
/**
|
|
9449
|
+
* Properties of the language pack.
|
|
9450
|
+
*/
|
|
9451
|
+
interface LanguagePackInfo {
|
|
9452
|
+
/** Full descriptive name of the language, e.g. 'Japanese'. */
|
|
9453
|
+
language_description?: string;
|
|
9454
|
+
/** The character to use to separate words. */
|
|
9455
|
+
word_delimiter: string;
|
|
9456
|
+
/** The direction that words in the language should be written and read in. */
|
|
9457
|
+
writing_direction?: LanguagePackInfoWritingDirection;
|
|
9458
|
+
/** Whether or not ITN (inverse text normalization) is available for the language pack. */
|
|
9459
|
+
itn?: boolean;
|
|
9460
|
+
/** Whether or not language model adaptation has been applied to the language pack. */
|
|
9461
|
+
adapted?: boolean;
|
|
9462
|
+
}
|
|
9463
|
+
|
|
9464
|
+
/**
|
|
9465
|
+
* Generated by orval v7.9.0 🍺
|
|
9466
|
+
* Do not edit manually.
|
|
9467
|
+
* Speechmatics ASR REST API
|
|
9468
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9469
|
+
* OpenAPI spec version: 2.0.0
|
|
9470
|
+
*/
|
|
9471
|
+
type RecognitionDisplayDirection = (typeof RecognitionDisplayDirection)[keyof typeof RecognitionDisplayDirection];
|
|
9472
|
+
declare const RecognitionDisplayDirection: {
|
|
9473
|
+
readonly ltr: "ltr";
|
|
9474
|
+
readonly rtl: "rtl";
|
|
9475
|
+
};
|
|
9476
|
+
|
|
9477
|
+
/**
|
|
9478
|
+
* Generated by orval v7.9.0 🍺
|
|
9479
|
+
* Do not edit manually.
|
|
9480
|
+
* Speechmatics ASR REST API
|
|
9481
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9482
|
+
* OpenAPI spec version: 2.0.0
|
|
9483
|
+
*/
|
|
9484
|
+
|
|
9485
|
+
interface RecognitionDisplay {
|
|
9486
|
+
direction: RecognitionDisplayDirection;
|
|
9487
|
+
}
|
|
9488
|
+
|
|
9489
|
+
/**
|
|
9490
|
+
* Generated by orval v7.9.0 🍺
|
|
9491
|
+
* Do not edit manually.
|
|
9492
|
+
* Speechmatics ASR REST API
|
|
9493
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9494
|
+
* OpenAPI spec version: 2.0.0
|
|
9495
|
+
*/
|
|
9496
|
+
|
|
9497
|
+
/**
|
|
9498
|
+
* List of possible job output item values, ordered by likelihood.
|
|
9499
|
+
*/
|
|
9500
|
+
interface RecognitionAlternative {
|
|
9501
|
+
content: string;
|
|
9502
|
+
confidence: number;
|
|
9503
|
+
language: string;
|
|
9504
|
+
display?: RecognitionDisplay;
|
|
9505
|
+
speaker?: string;
|
|
9506
|
+
tags?: string[];
|
|
9507
|
+
}
|
|
9508
|
+
|
|
9509
|
+
/**
|
|
9510
|
+
* Generated by orval v7.9.0 🍺
|
|
9511
|
+
* Do not edit manually.
|
|
9512
|
+
* Speechmatics ASR REST API
|
|
9513
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9514
|
+
* OpenAPI spec version: 2.0.0
|
|
9515
|
+
*/
|
|
9516
|
+
type TranslationErrorType = (typeof TranslationErrorType)[keyof typeof TranslationErrorType];
|
|
9517
|
+
declare const TranslationErrorType: {
|
|
9518
|
+
readonly translation_failed: "translation_failed";
|
|
9519
|
+
readonly unsupported_translation_pair: "unsupported_translation_pair";
|
|
9520
|
+
};
|
|
9521
|
+
|
|
9522
|
+
/**
|
|
9523
|
+
* Generated by orval v7.9.0 🍺
|
|
9524
|
+
* Do not edit manually.
|
|
9525
|
+
* Speechmatics ASR REST API
|
|
9526
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9527
|
+
* OpenAPI spec version: 2.0.0
|
|
9528
|
+
*/
|
|
9529
|
+
|
|
9530
|
+
interface TranslationError {
|
|
9531
|
+
type?: TranslationErrorType;
|
|
9532
|
+
/** Human readable error message */
|
|
9533
|
+
message?: string;
|
|
9534
|
+
}
|
|
9535
|
+
|
|
9536
|
+
/**
|
|
9537
|
+
* Generated by orval v7.9.0 🍺
|
|
9538
|
+
* Do not edit manually.
|
|
9539
|
+
* Speechmatics ASR REST API
|
|
9540
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9541
|
+
* OpenAPI spec version: 2.0.0
|
|
9542
|
+
*/
|
|
9543
|
+
type SummarizationErrorType = (typeof SummarizationErrorType)[keyof typeof SummarizationErrorType];
|
|
9544
|
+
declare const SummarizationErrorType: {
|
|
9545
|
+
readonly summarization_failed: "summarization_failed";
|
|
9546
|
+
readonly unsupported_language: "unsupported_language";
|
|
9547
|
+
};
|
|
9548
|
+
|
|
9549
|
+
/**
|
|
9550
|
+
* Generated by orval v7.9.0 🍺
|
|
9551
|
+
* Do not edit manually.
|
|
9552
|
+
* Speechmatics ASR REST API
|
|
9553
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9554
|
+
* OpenAPI spec version: 2.0.0
|
|
9555
|
+
*/
|
|
9556
|
+
|
|
9557
|
+
interface SummarizationError {
|
|
9558
|
+
type?: SummarizationErrorType;
|
|
9559
|
+
/** Human readable error message */
|
|
9560
|
+
message?: string;
|
|
9561
|
+
}
|
|
9562
|
+
|
|
9563
|
+
/**
|
|
9564
|
+
* Generated by orval v7.9.0 🍺
|
|
9565
|
+
* Do not edit manually.
|
|
9566
|
+
* Speechmatics ASR REST API
|
|
9567
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9568
|
+
* OpenAPI spec version: 2.0.0
|
|
9569
|
+
*/
|
|
9570
|
+
type SentimentAnalysisErrorType = (typeof SentimentAnalysisErrorType)[keyof typeof SentimentAnalysisErrorType];
|
|
9571
|
+
declare const SentimentAnalysisErrorType: {
|
|
9572
|
+
readonly sentiment_analysis_failed: "sentiment_analysis_failed";
|
|
9573
|
+
readonly unsupported_language: "unsupported_language";
|
|
9574
|
+
};
|
|
9575
|
+
|
|
9576
|
+
/**
|
|
9577
|
+
* Generated by orval v7.9.0 🍺
|
|
9578
|
+
* Do not edit manually.
|
|
9579
|
+
* Speechmatics ASR REST API
|
|
9580
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9581
|
+
* OpenAPI spec version: 2.0.0
|
|
9582
|
+
*/
|
|
9583
|
+
|
|
9584
|
+
interface SentimentAnalysisError {
|
|
9585
|
+
type?: SentimentAnalysisErrorType;
|
|
9586
|
+
/** Human readable error message */
|
|
9587
|
+
message?: string;
|
|
9588
|
+
}
|
|
9589
|
+
|
|
9590
|
+
/**
|
|
9591
|
+
* Generated by orval v7.9.0 🍺
|
|
9592
|
+
* Do not edit manually.
|
|
9593
|
+
* Speechmatics ASR REST API
|
|
9594
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9595
|
+
* OpenAPI spec version: 2.0.0
|
|
9596
|
+
*/
|
|
9597
|
+
type TopicDetectionErrorType = (typeof TopicDetectionErrorType)[keyof typeof TopicDetectionErrorType];
|
|
9598
|
+
declare const TopicDetectionErrorType: {
|
|
9599
|
+
readonly topic_detection_failed: "topic_detection_failed";
|
|
9600
|
+
readonly unsupported_list_of_topics: "unsupported_list_of_topics";
|
|
9601
|
+
readonly unsupported_language: "unsupported_language";
|
|
9602
|
+
};
|
|
9603
|
+
|
|
9604
|
+
/**
|
|
9605
|
+
* Generated by orval v7.9.0 🍺
|
|
9606
|
+
* Do not edit manually.
|
|
9607
|
+
* Speechmatics ASR REST API
|
|
9608
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9609
|
+
* OpenAPI spec version: 2.0.0
|
|
9610
|
+
*/
|
|
9611
|
+
|
|
9612
|
+
interface TopicDetectionError {
|
|
9613
|
+
type?: TopicDetectionErrorType;
|
|
9614
|
+
/** Human readable error message */
|
|
9615
|
+
message?: string;
|
|
9616
|
+
}
|
|
9617
|
+
|
|
9618
|
+
/**
|
|
9619
|
+
* Generated by orval v7.9.0 🍺
|
|
9620
|
+
* Do not edit manually.
|
|
9621
|
+
* Speechmatics ASR REST API
|
|
9622
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9623
|
+
* OpenAPI spec version: 2.0.0
|
|
9624
|
+
*/
|
|
9625
|
+
|
|
9626
|
+
/**
|
|
9627
|
+
* Summary information about the output from an ASR job, comprising the job type and configuration parameters used when generating the output.
|
|
9628
|
+
*/
|
|
9629
|
+
interface RecognitionMetadata {
|
|
9630
|
+
/** The UTC date time the transcription output was created. */
|
|
9631
|
+
created_at: string;
|
|
9632
|
+
type: JobType;
|
|
9633
|
+
transcription_config?: TranscriptionConfig;
|
|
9634
|
+
/** List of errors that occurred in the translation stage. */
|
|
9635
|
+
translation_errors?: TranslationError[];
|
|
9636
|
+
/** List of errors that occurred in the summarization stage. */
|
|
9637
|
+
summarization_errors?: SummarizationError[];
|
|
9638
|
+
/** List of errors that occurred in the sentiment analysis stage. */
|
|
9639
|
+
sentiment_analysis_errors?: SentimentAnalysisError[];
|
|
9640
|
+
/** List of errors that occurred in the topic detection stage. */
|
|
9641
|
+
topic_detection_errors?: TopicDetectionError[];
|
|
9642
|
+
/** List of errors that occurred in the auto chapters stage. */
|
|
9643
|
+
auto_chapters_errors?: AutoChaptersResultError[];
|
|
9644
|
+
alignment_config?: AlignmentConfig;
|
|
9645
|
+
output_config?: OutputConfig;
|
|
9646
|
+
language_pack_info?: LanguagePackInfo;
|
|
9647
|
+
language_identification?: LanguageIdentificationResult;
|
|
9648
|
+
/** Orchestrator version in PEP 440 Format or set to 'version_not_found' as default. */
|
|
9649
|
+
orchestrator_version?: string;
|
|
9650
|
+
}
|
|
9651
|
+
|
|
9652
|
+
/**
|
|
9653
|
+
* Generated by orval v7.9.0 🍺
|
|
9654
|
+
* Do not edit manually.
|
|
9655
|
+
* Speechmatics ASR REST API
|
|
9656
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9657
|
+
* OpenAPI spec version: 2.0.0
|
|
9658
|
+
*/
|
|
9659
|
+
/**
|
|
9660
|
+
* New types of items may appear without being requested; unrecognized item types can be ignored.
|
|
9661
|
+
*/
|
|
9662
|
+
type RecognitionResultType = (typeof RecognitionResultType)[keyof typeof RecognitionResultType];
|
|
9663
|
+
declare const RecognitionResultType: {
|
|
9664
|
+
readonly word: "word";
|
|
9665
|
+
readonly punctuation: "punctuation";
|
|
9666
|
+
readonly entity: "entity";
|
|
9667
|
+
};
|
|
9668
|
+
|
|
9669
|
+
/**
|
|
9670
|
+
* Generated by orval v7.9.0 🍺
|
|
9671
|
+
* Do not edit manually.
|
|
9672
|
+
* Speechmatics ASR REST API
|
|
9673
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9674
|
+
* OpenAPI spec version: 2.0.0
|
|
9675
|
+
*/
|
|
9676
|
+
/**
|
|
9677
|
+
* What kind of object this is. See #/Definitions/RecognitionResult for definitions of the enums.
|
|
9678
|
+
*/
|
|
9679
|
+
type WrittenFormRecognitionResultType = (typeof WrittenFormRecognitionResultType)[keyof typeof WrittenFormRecognitionResultType];
|
|
9680
|
+
declare const WrittenFormRecognitionResultType: {
|
|
9681
|
+
readonly word: "word";
|
|
9682
|
+
};
|
|
9683
|
+
|
|
9684
|
+
/**
|
|
9685
|
+
* Generated by orval v7.9.0 🍺
|
|
9686
|
+
* Do not edit manually.
|
|
9687
|
+
* Speechmatics ASR REST API
|
|
9688
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9689
|
+
* OpenAPI spec version: 2.0.0
|
|
9690
|
+
*/
|
|
9691
|
+
|
|
9692
|
+
/**
|
|
9693
|
+
* A WrittenFormRecognitionResult describes a simple object which consists solely of 'word' type entries with a start and end time. It can occur only inside the written_form property of a full RecognitionResult"
|
|
9694
|
+
*/
|
|
9695
|
+
interface WrittenFormRecognitionResult {
|
|
9696
|
+
alternatives: RecognitionAlternative[];
|
|
9697
|
+
end_time: number;
|
|
9698
|
+
start_time: number;
|
|
9699
|
+
/** What kind of object this is. See #/Definitions/RecognitionResult for definitions of the enums. */
|
|
9700
|
+
type: WrittenFormRecognitionResultType;
|
|
9701
|
+
}
|
|
9702
|
+
|
|
9703
|
+
/**
|
|
9704
|
+
* Generated by orval v7.9.0 🍺
|
|
9705
|
+
* Do not edit manually.
|
|
9706
|
+
* Speechmatics ASR REST API
|
|
9707
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9708
|
+
* OpenAPI spec version: 2.0.0
|
|
9709
|
+
*/
|
|
9710
|
+
/**
|
|
9711
|
+
* What kind of object this is. See #/Definitions/RecognitionResult for definitions of the enums.
|
|
9712
|
+
*/
|
|
9713
|
+
type SpokenFormRecognitionResultType = (typeof SpokenFormRecognitionResultType)[keyof typeof SpokenFormRecognitionResultType];
|
|
9714
|
+
declare const SpokenFormRecognitionResultType: {
|
|
9715
|
+
readonly word: "word";
|
|
9716
|
+
readonly punctuation: "punctuation";
|
|
9717
|
+
};
|
|
9718
|
+
|
|
9719
|
+
/**
|
|
9720
|
+
* Generated by orval v7.9.0 🍺
|
|
9721
|
+
* Do not edit manually.
|
|
9722
|
+
* Speechmatics ASR REST API
|
|
9723
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9724
|
+
* OpenAPI spec version: 2.0.0
|
|
9725
|
+
*/
|
|
9726
|
+
|
|
9727
|
+
/**
|
|
9728
|
+
* A SpokenFormRecognitionResult describes a simple object which consists solely of 'word' or 'punctuation' type entries with a start and end time. It can occur only inside the spoken_form property of a full "RecognitionResult"
|
|
9729
|
+
*/
|
|
9730
|
+
interface SpokenFormRecognitionResult {
|
|
9731
|
+
alternatives: RecognitionAlternative[];
|
|
9732
|
+
end_time: number;
|
|
9733
|
+
start_time: number;
|
|
9734
|
+
/** What kind of object this is. See #/Definitions/RecognitionResult for definitions of the enums. */
|
|
9735
|
+
type: SpokenFormRecognitionResultType;
|
|
9736
|
+
}
|
|
9737
|
+
|
|
9738
|
+
/**
|
|
9739
|
+
* Generated by orval v7.9.0 🍺
|
|
9740
|
+
* Do not edit manually.
|
|
9741
|
+
* Speechmatics ASR REST API
|
|
9742
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9743
|
+
* OpenAPI spec version: 2.0.0
|
|
9744
|
+
*/
|
|
9745
|
+
/**
|
|
9746
|
+
* Attachment direction of the punctuation mark. This only applies to punctuation marks. This information can be used to produce a well-formed text representation by placing the `word_delimiter` from `language_pack_info` on the correct side of the punctuation mark.
|
|
9747
|
+
*/
|
|
9748
|
+
type RecognitionResultAttachesTo = (typeof RecognitionResultAttachesTo)[keyof typeof RecognitionResultAttachesTo];
|
|
9749
|
+
declare const RecognitionResultAttachesTo: {
|
|
9750
|
+
readonly previous: "previous";
|
|
9751
|
+
readonly next: "next";
|
|
9752
|
+
readonly both: "both";
|
|
9753
|
+
readonly none: "none";
|
|
9754
|
+
};
|
|
9755
|
+
|
|
9756
|
+
/**
|
|
9757
|
+
* Generated by orval v7.9.0 🍺
|
|
9758
|
+
* Do not edit manually.
|
|
9759
|
+
* Speechmatics ASR REST API
|
|
9760
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9761
|
+
* OpenAPI spec version: 2.0.0
|
|
9762
|
+
*/
|
|
9763
|
+
|
|
9764
|
+
/**
|
|
9765
|
+
* An ASR job output item. The primary item types are `word` and `punctuation`. Other item types may be present, for example to provide semantic information of different forms.
|
|
9766
|
+
*/
|
|
9767
|
+
interface RecognitionResult {
|
|
9768
|
+
channel?: string;
|
|
9769
|
+
start_time: number;
|
|
9770
|
+
end_time: number;
|
|
9771
|
+
/**
|
|
9772
|
+
* An indication of the volume of audio across the time period the word was spoken.
|
|
9773
|
+
* @minimum 0
|
|
9774
|
+
* @maximum 100
|
|
9775
|
+
*/
|
|
9776
|
+
volume?: number;
|
|
9777
|
+
/** Whether the punctuation mark is an end of sentence character. Only applies to punctuation marks. */
|
|
9778
|
+
is_eos?: boolean;
|
|
9779
|
+
/** New types of items may appear without being requested; unrecognized item types can be ignored. */
|
|
9780
|
+
type: RecognitionResultType;
|
|
9781
|
+
written_form?: WrittenFormRecognitionResult[];
|
|
9782
|
+
spoken_form?: SpokenFormRecognitionResult[];
|
|
9783
|
+
alternatives?: RecognitionAlternative[];
|
|
9784
|
+
/** Attachment direction of the punctuation mark. This only applies to punctuation marks. This information can be used to produce a well-formed text representation by placing the `word_delimiter` from `language_pack_info` on the correct side of the punctuation mark. */
|
|
9785
|
+
attaches_to?: RecognitionResultAttachesTo;
|
|
9786
|
+
}
|
|
9787
|
+
|
|
9788
|
+
/**
|
|
9789
|
+
* Generated by orval v7.9.0 🍺
|
|
9790
|
+
* Do not edit manually.
|
|
9791
|
+
* Speechmatics ASR REST API
|
|
9792
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9793
|
+
* OpenAPI spec version: 2.0.0
|
|
9794
|
+
*/
|
|
9795
|
+
interface TranslationSentence {
|
|
9796
|
+
start_time?: number;
|
|
9797
|
+
end_time?: number;
|
|
9798
|
+
content?: string;
|
|
9799
|
+
speaker?: string;
|
|
9800
|
+
channel?: string;
|
|
9801
|
+
}
|
|
9802
|
+
|
|
9803
|
+
/**
|
|
9804
|
+
* Generated by orval v7.9.0 🍺
|
|
9805
|
+
* Do not edit manually.
|
|
9806
|
+
* Speechmatics ASR REST API
|
|
9807
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9808
|
+
* OpenAPI spec version: 2.0.0
|
|
9809
|
+
*/
|
|
9810
|
+
|
|
9811
|
+
/**
|
|
9812
|
+
* Translations of the transcript into other languages. It is a map of ISO language codes to arrays of translated sentences. Configured using `translation_config`.
|
|
9813
|
+
*/
|
|
9814
|
+
type RetrieveTranscriptResponseTranslations = {
|
|
9815
|
+
[key: string]: TranslationSentence[];
|
|
9816
|
+
};
|
|
9817
|
+
|
|
9818
|
+
/**
|
|
9819
|
+
* Generated by orval v7.9.0 🍺
|
|
9820
|
+
* Do not edit manually.
|
|
9821
|
+
* Speechmatics ASR REST API
|
|
9822
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9823
|
+
* OpenAPI spec version: 2.0.0
|
|
9824
|
+
*/
|
|
9825
|
+
/**
|
|
9826
|
+
* Summary of the transcript, configured using `summarization_config`.
|
|
9827
|
+
*/
|
|
9828
|
+
interface SummarizationResult {
|
|
9829
|
+
content?: string;
|
|
9830
|
+
}
|
|
9831
|
+
|
|
9832
|
+
/**
|
|
9833
|
+
* Generated by orval v7.9.0 🍺
|
|
9834
|
+
* Do not edit manually.
|
|
9835
|
+
* Speechmatics ASR REST API
|
|
9836
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9837
|
+
* OpenAPI spec version: 2.0.0
|
|
9838
|
+
*/
|
|
9839
|
+
/**
|
|
9840
|
+
* Represents a segment of text and its associated sentiment.
|
|
9841
|
+
*/
|
|
9842
|
+
interface SentimentSegment {
|
|
9843
|
+
text?: string;
|
|
9844
|
+
start_time?: number;
|
|
9845
|
+
end_time?: number;
|
|
9846
|
+
sentiment?: string;
|
|
9847
|
+
speaker?: string;
|
|
9848
|
+
channel?: string;
|
|
9849
|
+
confidence?: number;
|
|
9850
|
+
}
|
|
9851
|
+
|
|
9852
|
+
/**
|
|
9853
|
+
* Generated by orval v7.9.0 🍺
|
|
9854
|
+
* Do not edit manually.
|
|
9855
|
+
* Speechmatics ASR REST API
|
|
9856
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9857
|
+
* OpenAPI spec version: 2.0.0
|
|
9858
|
+
*/
|
|
9859
|
+
/**
|
|
9860
|
+
* Holds the count of sentiment information grouped by positive, neutral and negative.
|
|
9861
|
+
*/
|
|
9862
|
+
interface SentimentSummaryDetail {
|
|
9863
|
+
positive_count?: number;
|
|
9864
|
+
negative_count?: number;
|
|
9865
|
+
neutral_count?: number;
|
|
9866
|
+
}
|
|
9867
|
+
|
|
9868
|
+
/**
|
|
9869
|
+
* Generated by orval v7.9.0 🍺
|
|
9870
|
+
* Do not edit manually.
|
|
9871
|
+
* Speechmatics ASR REST API
|
|
9872
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9873
|
+
* OpenAPI spec version: 2.0.0
|
|
9874
|
+
*/
|
|
9875
|
+
/**
|
|
9876
|
+
* Holds sentiment information for a specific speaker.
|
|
9877
|
+
*/
|
|
9878
|
+
interface SentimentSpeakerSummary {
|
|
9879
|
+
speaker?: string;
|
|
9880
|
+
positive_count?: number;
|
|
9881
|
+
negative_count?: number;
|
|
9882
|
+
neutral_count?: number;
|
|
9883
|
+
}
|
|
9884
|
+
|
|
9885
|
+
/**
|
|
9886
|
+
* Generated by orval v7.9.0 🍺
|
|
9887
|
+
* Do not edit manually.
|
|
9888
|
+
* Speechmatics ASR REST API
|
|
9889
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9890
|
+
* OpenAPI spec version: 2.0.0
|
|
9891
|
+
*/
|
|
9892
|
+
/**
|
|
9893
|
+
* Holds sentiment information for a specific channel.
|
|
9894
|
+
*/
|
|
9895
|
+
interface SentimentChannelSummary {
|
|
9896
|
+
channel?: string;
|
|
9897
|
+
positive_count?: number;
|
|
9898
|
+
negative_count?: number;
|
|
9899
|
+
neutral_count?: number;
|
|
9900
|
+
}
|
|
9901
|
+
|
|
9902
|
+
/**
|
|
9903
|
+
* Generated by orval v7.9.0 🍺
|
|
9904
|
+
* Do not edit manually.
|
|
9905
|
+
* Speechmatics ASR REST API
|
|
9906
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9907
|
+
* OpenAPI spec version: 2.0.0
|
|
9908
|
+
*/
|
|
9909
|
+
|
|
9910
|
+
/**
|
|
9911
|
+
* Holds overall sentiment information, as well as detailed per-speaker and per-channel sentiment data.
|
|
9912
|
+
*/
|
|
9913
|
+
interface SentimentSummary {
|
|
9914
|
+
overall?: SentimentSummaryDetail;
|
|
9915
|
+
/** An array of objects that represent sentiment data for a specific speaker. */
|
|
9916
|
+
speakers?: SentimentSpeakerSummary[];
|
|
9917
|
+
/** An array of objects that represent sentiment data for a specific channel. */
|
|
9918
|
+
channels?: SentimentChannelSummary[];
|
|
9919
|
+
}
|
|
9920
|
+
|
|
9921
|
+
/**
|
|
9922
|
+
* Generated by orval v7.9.0 🍺
|
|
9923
|
+
* Do not edit manually.
|
|
9924
|
+
* Speechmatics ASR REST API
|
|
9925
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9926
|
+
* OpenAPI spec version: 2.0.0
|
|
9927
|
+
*/
|
|
9928
|
+
|
|
9929
|
+
/**
|
|
9930
|
+
* Holds the detailed sentiment analysis information.
|
|
9931
|
+
*/
|
|
9932
|
+
type SentimentAnalysisResultSentimentAnalysis = {
|
|
9933
|
+
/** An array of objects that represent a segment of text and its associated sentiment. */
|
|
9934
|
+
segments?: SentimentSegment[];
|
|
9935
|
+
summary?: SentimentSummary;
|
|
9936
|
+
};
|
|
9937
|
+
|
|
9938
|
+
/**
|
|
9939
|
+
* Generated by orval v7.9.0 🍺
|
|
9940
|
+
* Do not edit manually.
|
|
9941
|
+
* Speechmatics ASR REST API
|
|
9942
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9943
|
+
* OpenAPI spec version: 2.0.0
|
|
9944
|
+
*/
|
|
9945
|
+
|
|
9946
|
+
/**
|
|
9947
|
+
* The main object that holds sentiment analysis data.
|
|
9948
|
+
*/
|
|
9949
|
+
interface SentimentAnalysisResult {
|
|
9950
|
+
/** Holds the detailed sentiment analysis information. */
|
|
9951
|
+
sentiment_analysis?: SentimentAnalysisResultSentimentAnalysis;
|
|
9952
|
+
}
|
|
9953
|
+
|
|
9954
|
+
/**
|
|
9955
|
+
* Generated by orval v7.9.0 🍺
|
|
9956
|
+
* Do not edit manually.
|
|
9957
|
+
* Speechmatics ASR REST API
|
|
9958
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9959
|
+
* OpenAPI spec version: 2.0.0
|
|
9960
|
+
*/
|
|
9961
|
+
/**
|
|
9962
|
+
* Represents a topic and its associated information.
|
|
9963
|
+
*/
|
|
9964
|
+
interface TopicDetectionSegmentTopic {
|
|
9965
|
+
topic?: string;
|
|
9966
|
+
}
|
|
9967
|
+
|
|
9968
|
+
/**
|
|
9969
|
+
* Generated by orval v7.9.0 🍺
|
|
9970
|
+
* Do not edit manually.
|
|
9971
|
+
* Speechmatics ASR REST API
|
|
9972
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9973
|
+
* OpenAPI spec version: 2.0.0
|
|
9974
|
+
*/
|
|
9975
|
+
|
|
9976
|
+
/**
|
|
9977
|
+
* Represents a segment of text and its associated topic information.
|
|
9978
|
+
*/
|
|
9979
|
+
interface TopicDetectionSegment {
|
|
9980
|
+
text?: string;
|
|
9981
|
+
start_time?: number;
|
|
9982
|
+
end_time?: number;
|
|
9983
|
+
topics?: TopicDetectionSegmentTopic[];
|
|
9984
|
+
}
|
|
9985
|
+
|
|
9986
|
+
/**
|
|
9987
|
+
* Generated by orval v7.9.0 🍺
|
|
9988
|
+
* Do not edit manually.
|
|
9989
|
+
* Speechmatics ASR REST API
|
|
9990
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
9991
|
+
* OpenAPI spec version: 2.0.0
|
|
9992
|
+
*/
|
|
9993
|
+
/**
|
|
9994
|
+
* Holds the count of topics detected.
|
|
9995
|
+
*/
|
|
9996
|
+
interface TopicDetectionSummaryOverall {
|
|
9997
|
+
[key: string]: number;
|
|
9998
|
+
}
|
|
9999
|
+
|
|
10000
|
+
/**
|
|
10001
|
+
* Generated by orval v7.9.0 🍺
|
|
10002
|
+
* Do not edit manually.
|
|
10003
|
+
* Speechmatics ASR REST API
|
|
10004
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
10005
|
+
* OpenAPI spec version: 2.0.0
|
|
10006
|
+
*/
|
|
10007
|
+
|
|
10008
|
+
/**
|
|
10009
|
+
* Holds overall information on the topics detected.
|
|
10010
|
+
*/
|
|
10011
|
+
interface TopicDetectionSummary {
|
|
10012
|
+
overall?: TopicDetectionSummaryOverall;
|
|
10013
|
+
}
|
|
10014
|
+
|
|
10015
|
+
/**
|
|
10016
|
+
* Generated by orval v7.9.0 🍺
|
|
10017
|
+
* Do not edit manually.
|
|
10018
|
+
* Speechmatics ASR REST API
|
|
10019
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
10020
|
+
* OpenAPI spec version: 2.0.0
|
|
10021
|
+
*/
|
|
10022
|
+
|
|
10023
|
+
/**
|
|
10024
|
+
* Main object that holds topic detection results.
|
|
10025
|
+
*/
|
|
10026
|
+
interface TopicDetectionResult {
|
|
10027
|
+
/** An array of objects that represent a segment of text and its associated topic information. */
|
|
10028
|
+
segments?: TopicDetectionSegment[];
|
|
10029
|
+
summary?: TopicDetectionSummary;
|
|
10030
|
+
}
|
|
10031
|
+
|
|
10032
|
+
/**
|
|
10033
|
+
* Generated by orval v7.9.0 🍺
|
|
10034
|
+
* Do not edit manually.
|
|
10035
|
+
* Speechmatics ASR REST API
|
|
10036
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
10037
|
+
* OpenAPI spec version: 2.0.0
|
|
10038
|
+
*/
|
|
10039
|
+
|
|
10040
|
+
/**
|
|
10041
|
+
* Summary keyed by channel, only set if channel diarization is enabled
|
|
10042
|
+
*/
|
|
10043
|
+
type RetrieveTranscriptResponseAudioEventSummaryChannels = {
|
|
10044
|
+
[key: string]: AudioEventSummary;
|
|
10045
|
+
};
|
|
10046
|
+
|
|
10047
|
+
/**
|
|
10048
|
+
* Generated by orval v7.9.0 🍺
|
|
10049
|
+
* Do not edit manually.
|
|
10050
|
+
* Speechmatics ASR REST API
|
|
10051
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
10052
|
+
* OpenAPI spec version: 2.0.0
|
|
10053
|
+
*/
|
|
10054
|
+
|
|
10055
|
+
/**
|
|
10056
|
+
* Summary statistics per event type, keyed by `type`, e.g. music
|
|
10057
|
+
*/
|
|
10058
|
+
type RetrieveTranscriptResponseAudioEventSummary = {
|
|
10059
|
+
overall?: AudioEventSummary;
|
|
10060
|
+
/** Summary keyed by channel, only set if channel diarization is enabled */
|
|
10061
|
+
channels?: RetrieveTranscriptResponseAudioEventSummaryChannels;
|
|
10062
|
+
};
|
|
10063
|
+
|
|
10064
|
+
/**
|
|
10065
|
+
* Generated by orval v7.9.0 🍺
|
|
10066
|
+
* Do not edit manually.
|
|
10067
|
+
* Speechmatics ASR REST API
|
|
10068
|
+
* The Speechmatics Automatic Speech Recognition REST API is used to submit ASR jobs and receive the results. The supported job type is transcription of audio files.
|
|
10069
|
+
* OpenAPI spec version: 2.0.0
|
|
10070
|
+
*/
|
|
10071
|
+
|
|
10072
|
+
interface RetrieveTranscriptResponse {
|
|
10073
|
+
/** Speechmatics JSON transcript format version number. */
|
|
10074
|
+
format: string;
|
|
10075
|
+
job: JobInfo;
|
|
10076
|
+
metadata: RecognitionMetadata;
|
|
10077
|
+
results: RecognitionResult[];
|
|
10078
|
+
/** Translations of the transcript into other languages. It is a map of ISO language codes to arrays of translated sentences. Configured using `translation_config`. */
|
|
10079
|
+
translations?: RetrieveTranscriptResponseTranslations;
|
|
10080
|
+
summary?: SummarizationResult;
|
|
10081
|
+
sentiment_analysis?: SentimentAnalysisResult;
|
|
10082
|
+
topics?: TopicDetectionResult;
|
|
10083
|
+
chapters?: AutoChaptersResult;
|
|
10084
|
+
/** Timestamped audio events, only set if `audio_events_config` is in the config */
|
|
10085
|
+
audio_events?: AudioEventItem[];
|
|
10086
|
+
/** Summary statistics per event type, keyed by `type`, e.g. music */
|
|
10087
|
+
audio_event_summary?: RetrieveTranscriptResponseAudioEventSummary;
|
|
10088
|
+
}
|
|
10089
|
+
|
|
8953
10090
|
/**
|
|
8954
10091
|
* Generated by orval v7.9.0 🍺
|
|
8955
10092
|
* Do not edit manually.
|
|
@@ -9213,4 +10350,4 @@ interface SpeechToTextChunkResponseModel {
|
|
|
9213
10350
|
entities?: SpeechToTextChunkResponseModelEntities;
|
|
9214
10351
|
}
|
|
9215
10352
|
|
|
9216
|
-
export { type CustomSpellingConfigDTO as $, type AssemblyAIStreamingOptions as A, type AudioToLlmListDTOError as B, type CallbackConfig as C, type DeepgramStreamingOptions as D, type AudioToLlmResultDTO as E, type FileResponse as F, type GladiaStreamingOptions as G, type CallbackConfigDto as H, CallbackMethodEnum as I, type CallbackTranscriptionErrorPayload as J, type CallbackTranscriptionErrorPayloadCustomMetadata as K, type ListTranscriptsOptions as L, type MessagesConfig as M, type NamedEntityRecognitionDTO as N, CallbackTranscriptionErrorPayloadEvent as O, type PreProcessingConfig as P, type CallbackTranscriptionSuccessPayload as Q, type RealtimeProcessingConfig as R, type StreamingOptions as S, type TranscribeOptions as T, type UnifiedTranscriptResponse as U, type CallbackTranscriptionSuccessPayloadCustomMetadata as V, type WordDTO as W, CallbackTranscriptionSuccessPayloadEvent as X, type ChapterizationDTOError as Y, type ChapterizationDTOResults as Z, type CodeSwitchingConfigDTO as _, type StreamingCallbacks as a, type Entity as a$, type CustomSpellingConfigDTOSpellingDictionary as a0, type CustomVocabularyConfigDTO as a1, type CustomVocabularyConfigDTOVocabularyItem as a2, type CustomVocabularyEntryDTO as a3, type DiarizationConfigDTO as a4, type DiarizationDTO as a5, type DiarizationDTOError as a6, type DisplayModeDTO as a7, type DisplayModeDTOError as a8, type ErrorDTO as a9, type SubtitlesConfigDTO as aA, SubtitlesFormatEnum as aB, SubtitlesStyleEnum as aC, type SummarizationConfigDTO as aD, type SummarizationDTOError as aE, SummaryTypesEnum as aF, TranscriptionControllerListV2KindItem as aG, type TranscriptionControllerListV2Params as aH, TranscriptionControllerListV2StatusItem as aI, type TranscriptionResultDTO as aJ, type TranslationConfigDTO as aK, type TranslationDTOError as aL, TranslationModelEnum as aM, type TranslationResultDTO as aN, type TranslationResultDTOError as aO, type TranscriptOptionalParams as aP, TranscriptStatus as aQ, type TranscriptWord as aR, AudioIntelligenceModelStatus as aS, type AutoHighlightResult as aT, type AutoHighlightsResult as aU, type Chapter as aV, type ContentSafetyLabel as aW, type ContentSafetyLabelResult as aX, type ContentSafetyLabelsResult as aY, type ContentSafetyLabelsResultSeverityScoreSummary as aZ, type ContentSafetyLabelsResultSummary as a_, type InitTranscriptionRequest as aa, type InitTranscriptionRequestCustomMetadata as ab, type ModerationDTO as ac, type ModerationDTOError as ad, type NamedEntityRecognitionDTOError as ae, type NamesConsistencyDTO as af, type NamesConsistencyDTOError as ag, type PreRecordedRequestParamsResponse as ah, type PreRecordedResponseCustomMetadata as ai, type PreRecordedResponseFile as aj, PreRecordedResponseKind as ak, type PreRecordedResponsePostSessionMetadata as al, type PreRecordedResponseRequestParams as am, type PreRecordedResponseResult as an, PreRecordedResponseStatus as ao, type SentencesDTO as ap, type SentencesDTOError as aq, type SentimentAnalysisDTOError as ar, type SpeakerReidentificationDTO as as, type SpeakerReidentificationDTOError as at, type StreamingRequest as au, type StreamingRequestCustomMetadata as av, type StructuredDataExtractionConfigDTO as aw, type StructuredDataExtractionDTO as ax, type StructuredDataExtractionDTOError as ay, type SubtitleDTO as az, type StreamingSession as b, type TranscriptSentimentAnalysis as b$, EntityType as b0, type ListTranscriptsParams as b1, PiiPolicy as b2, RedactPiiAudioQuality as b3, type RedactedAudioNotification as b4, type RedactedAudioResponse as b5, RedactedAudioStatus as b6, Sentiment as b7, type SentimentAnalysisResult as b8, type SentimentAnalysisResultChannel as b9, type TranscriptDisfluencies as bA, type TranscriptEntities as bB, type TranscriptEntityDetection as bC, type TranscriptFilterProfanity as bD, type TranscriptFormatText as bE, type TranscriptIabCategories as bF, type TranscriptIabCategoriesResult as bG, TranscriptLanguageCode as bH, type TranscriptLanguageCodeProperty as bI, type TranscriptLanguageConfidence as bJ, type TranscriptLanguageConfidenceThreshold as bK, type TranscriptLanguageDetection as bL, type TranscriptMultichannel as bM, type TranscriptOptionalParamsLanguageCode as bN, type TranscriptOptionalParamsLanguageCodeOneOf as bO, type TranscriptOptionalParamsRedactPiiSub as bP, type TranscriptOptionalParamsSpeakersExpected as bQ, type TranscriptOptionalParamsSpeechModel as bR, type TranscriptOptionalParamsSpeechThreshold as bS, type TranscriptOptionalParamsWebhookAuthHeaderName as bT, type TranscriptOptionalParamsWebhookAuthHeaderValue as bU, type TranscriptPunctuate as bV, type TranscriptReadyNotification as bW, TranscriptReadyStatus as bX, type TranscriptRedactPiiAudio as bY, type TranscriptRedactPiiAudioQuality as bZ, type TranscriptRedactPiiPolicies as b_, type SentimentAnalysisResultSpeaker as ba, type SeverityScoreSummary as bb, SpeechModel as bc, SubstitutionPolicy as bd, SummaryModel as be, SummaryType as bf, type Timestamp as bg, type TopicDetectionModelResult as bh, type TopicDetectionModelResultSummary as bi, type TopicDetectionResult as bj, type TopicDetectionResultLabelsItem as bk, type Transcript as bl, type TranscriptAudioDuration as bm, type TranscriptAudioEndAt as bn, type TranscriptAudioStartFrom as bo, type TranscriptAutoChapters as bp, type TranscriptAutoHighlightsResult as bq, TranscriptBoostParam as br, type TranscriptBoostParamProperty as bs, type TranscriptChapters as bt, type TranscriptConfidence as bu, type TranscriptContentSafety as bv, type TranscriptContentSafetyLabels as bw, type TranscriptCustomSpelling as bx, type TranscriptCustomSpellingProperty as by, type TranscriptCustomTopics as bz, type StreamEvent as c, type TranscriptionWord as c$, type TranscriptSentimentAnalysisResults as c0, type TranscriptSpeakerLabels as c1, type TranscriptSpeakersExpected as c2, type TranscriptSpeechModel as c3, type TranscriptSpeechThreshold as c4, type TranscriptSpeedBoost as c5, type TranscriptSummary as c6, type TranscriptSummaryModel as c7, type TranscriptSummaryType as c8, type TranscriptText as c9, type TranscriptTextUsageTokens as cA, type TranscriptionSegment as cB, type RealtimeSessionCreateRequestGAModel as cC, RealtimeTranscriptionSessionCreateRequestTurnDetectionType as cD, RealtimeTranscriptionSessionCreateRequestInputAudioFormat as cE, AudioResponseFormat as cF, type CreateTranscription200One as cG, type CreateTranscriptionRequest as cH, type CreateTranscriptionRequestModel as cI, type CreateTranscriptionRequestStream as cJ, CreateTranscriptionRequestTimestampGranularitiesItem as cK, type CreateTranscriptionResponseDiarizedJson as cL, CreateTranscriptionResponseDiarizedJsonTask as cM, type CreateTranscriptionResponseDiarizedJsonUsage as cN, type CreateTranscriptionResponseJson as cO, type CreateTranscriptionResponseJsonLogprobsItem as cP, type CreateTranscriptionResponseJsonUsage as cQ, type CreateTranscriptionResponseVerboseJson as cR, type TranscriptTextUsageDuration as cS, TranscriptTextUsageDurationType as cT, type TranscriptTextUsageTokensInputTokenDetails as cU, TranscriptTextUsageTokensType as cV, type TranscriptionChunkingStrategy as cW, type TranscriptionChunkingStrategyAnyOf as cX, type TranscriptionDiarizedSegment as cY, TranscriptionDiarizedSegmentType as cZ, TranscriptionInclude as c_, type TranscriptThrottled as ca, type TranscriptUtterance as cb, type TranscriptUtteranceChannel as cc, type TranscriptUtterances as cd, type TranscriptWebhookAuthHeaderName as ce, type TranscriptWebhookNotification as cf, type TranscriptWebhookStatusCode as cg, type TranscriptWebhookUrl as ch, type TranscriptWordChannel as ci, type TranscriptWordSpeaker as cj, type TranscriptWords as ck, type StreamingUpdateConfiguration as cl, type Transcription as cm, Status as cn, type EntityError as co, type EntityReference as cp, type DiarizationProperties as cq, type DiarizationSpeakersProperties as cr, LanguageIdentificationMode as cs, type LanguageIdentificationProperties as ct, type LanguageIdentificationPropertiesSpeechModelMapping as cu, ProfanityFilterMode as cv, PunctuationMode as cw, type TranscriptionCustomProperties as cx, type TranscriptionLinks as cy, type TranscriptionProperties as cz, StreamingSupportedEncodingEnum as d, type ListenV1ResponseResultsChannelsItemSearchItemHitsItem as d$, type VadConfig as d0, VadConfigType as d1, type ListenV1Response as d2, type ManageV1FilterAccessorParameter as d3, ManageV1FilterDeploymentParameter as d4, type ManageV1LimitParameter as d5, type ManageV1PageParameter as d6, ManageV1FilterEndpointParameter as d7, ManageV1FilterMethodParameter as d8, type SharedTopics as d9, type ListenV1ModelParameter as dA, type ListenV1MultichannelParameter as dB, type ListenV1NumeralsParameter as dC, type ListenV1ParagraphsParameter as dD, type ListenV1ProfanityFilterParameter as dE, type ListenV1PunctuateParameter as dF, type ListenV1RedactParameter as dG, ListenV1RedactParameterOneOfItem as dH, type ListenV1ReplaceParameter as dI, type ListenV1ResponseMetadata as dJ, type ListenV1ResponseMetadataIntentsInfo as dK, type ListenV1ResponseMetadataModelInfo as dL, type ListenV1ResponseMetadataSentimentInfo as dM, type ListenV1ResponseMetadataSummaryInfo as dN, type ListenV1ResponseMetadataTopicsInfo as dO, type ListenV1ResponseResults as dP, type ListenV1ResponseResultsChannels as dQ, type ListenV1ResponseResultsChannelsItem as dR, type ListenV1ResponseResultsChannelsItemAlternativesItem as dS, type ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem as dT, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs as dU, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem as dV, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem as dW, type ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem as dX, type ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem as dY, type ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem as dZ, type ListenV1ResponseResultsChannelsItemSearchItem as d_, type SharedIntents as da, type SharedSentiments as db, type SharedCallbackParameter as dc, SharedCallbackMethodParameter as dd, type SharedSentimentParameter as de, type SharedSummarizeParameter as df, type SharedTagParameter as dg, type SharedTopicsParameter as dh, type SharedCustomTopicParameter as di, SharedCustomTopicModeParameter as dj, type SharedIntentsParameter as dk, type SharedCustomIntentParameter as dl, SharedCustomIntentModeParameter as dm, type SharedMipOptOutParameter as dn, type ListenV1DetectEntitiesParameter as dp, type ListenV1DetectLanguageParameter as dq, type ListenV1DiarizeParameter as dr, type ListenV1DictationParameter as ds, ListenV1EncodingParameter as dt, type ListenV1FillerWordsParameter as du, type ListenV1KeytermParameter as dv, type ListenV1KeywordsParameter as dw, type ListenV1LanguageParameter as dx, type ListenV1MeasurementsParameter as dy, type ListenV1MediaTranscribeParams as dz, StreamingSupportedBitDepthEnum as e, type SegmentedJsonExportOptionsMaxSegmentDurationS as e$, type ListenV1ResponseResultsSummary as e0, type ListenV1ResponseResultsUtterances as e1, type ListenV1ResponseResultsUtterancesItem as e2, type ListenV1ResponseResultsUtterancesItemWordsItem as e3, type ListenV1SearchParameter as e4, type ListenV1SmartFormatParameter as e5, type ListenV1UttSplitParameter as e6, type ListenV1UtterancesParameter as e7, type ListenV1VersionParameter as e8, type ManageV1EndDateTimeParameter as e9, type BodySpeechToTextV1SpeechToTextPostNumSpeakers as eA, type BodySpeechToTextV1SpeechToTextPostSeed as eB, type BodySpeechToTextV1SpeechToTextPostTemperature as eC, BodySpeechToTextV1SpeechToTextPostTimestampsGranularity as eD, type BodySpeechToTextV1SpeechToTextPostWebhookId as eE, type BodySpeechToTextV1SpeechToTextPostWebhookMetadata as eF, type BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf as eG, type DetectedEntity as eH, type DocxExportOptions as eI, DocxExportOptionsFormat as eJ, type DocxExportOptionsMaxSegmentChars as eK, type DocxExportOptionsMaxSegmentDurationS as eL, type DocxExportOptionsSegmentOnSilenceLongerThanS as eM, type ExportOptions as eN, type HtmlExportOptions as eO, HtmlExportOptionsFormat as eP, type HtmlExportOptionsMaxSegmentChars as eQ, type HtmlExportOptionsMaxSegmentDurationS as eR, type HtmlExportOptionsSegmentOnSilenceLongerThanS as eS, type PdfExportOptions as eT, PdfExportOptionsFormat as eU, type PdfExportOptionsMaxSegmentChars as eV, type PdfExportOptionsMaxSegmentDurationS as eW, type PdfExportOptionsSegmentOnSilenceLongerThanS as eX, type SegmentedJsonExportOptions as eY, SegmentedJsonExportOptionsFormat as eZ, type SegmentedJsonExportOptionsMaxSegmentChars as e_, type ManageV1FilterRequestIdParameter as ea, ManageV1FilterStatusParameter as eb, type ManageV1ProjectsRequestsListParams as ec, type ManageV1StartDateTimeParameter as ed, type SharedExtraParameter as ee, type SharedIntentsResults as ef, type SharedIntentsResultsIntents as eg, type SharedIntentsResultsIntentsSegmentsItem as eh, type SharedIntentsResultsIntentsSegmentsItemIntentsItem as ei, type SharedSentimentsAverage as ej, type SharedSentimentsSegmentsItem as ek, type SharedTopicsResults as el, type SharedTopicsResultsTopics as em, type SharedTopicsResultsTopicsSegmentsItem as en, type SharedTopicsResultsTopicsSegmentsItemTopicsItem as eo, type SpeechToTextChunkResponseModel as ep, type AdditionalFormatResponseModel as eq, type AdditionalFormats as er, type BodySpeechToTextV1SpeechToTextPost as es, type BodySpeechToTextV1SpeechToTextPostCloudStorageUrl as et, type BodySpeechToTextV1SpeechToTextPostDiarizationThreshold as eu, type BodySpeechToTextV1SpeechToTextPostEntityDetection as ev, type BodySpeechToTextV1SpeechToTextPostFile as ew, BodySpeechToTextV1SpeechToTextPostFileFormat as ex, type BodySpeechToTextV1SpeechToTextPostLanguageCode as ey, BodySpeechToTextV1SpeechToTextPostModelId as ez, StreamingSupportedSampleRateEnum as f, type TranscribeStreamParams as f$, type SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS as f0, type SpeechToTextCharacterResponseModel as f1, type SpeechToTextCharacterResponseModelEnd as f2, type SpeechToTextCharacterResponseModelStart as f3, type SpeechToTextChunkResponseModelAdditionalFormats as f4, type SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem as f5, type SpeechToTextChunkResponseModelChannelIndex as f6, type SpeechToTextChunkResponseModelEntities as f7, type SpeechToTextChunkResponseModelTranscriptionId as f8, type SpeechToTextWordResponseModel as f9, type BatchOnlyProvider as fA, type SessionStatus as fB, type Speaker as fC, type Word as fD, type Utterance as fE, type TranscriptionStatus as fF, type TranscriptMetadata as fG, type TranscriptData as fH, type ListTranscriptsResponse as fI, type ProviderRawResponseMap as fJ, type StreamEventType as fK, type SpeechEvent as fL, type TranslationEvent as fM, type SentimentEvent as fN, type EntityEvent as fO, type SummarizationEvent as fP, type ChapterizationEvent as fQ, type AudioAckEvent as fR, type LifecycleEvent as fS, type AudioChunk as fT, type RawWebSocketMessage as fU, type AssemblyAIUpdateConfiguration as fV, type OpenAIStreamingOptions as fW, type SonioxStreamingOptions as fX, type ElevenLabsStreamingOptions as fY, type ProviderStreamingOptions as fZ, type StreamingOptionsForProvider as f_, type SpeechToTextWordResponseModelCharacters as fa, type SpeechToTextWordResponseModelEnd as fb, type SpeechToTextWordResponseModelSpeakerId as fc, type SpeechToTextWordResponseModelStart as fd, SpeechToTextWordResponseModelType as fe, type SrtExportOptions as ff, SrtExportOptionsFormat as fg, type SrtExportOptionsMaxCharactersPerLine as fh, type SrtExportOptionsMaxSegmentChars as fi, type SrtExportOptionsMaxSegmentDurationS as fj, type SrtExportOptionsSegmentOnSilenceLongerThanS as fk, type TxtExportOptions as fl, TxtExportOptionsFormat as fm, type TxtExportOptionsMaxCharactersPerLine as fn, type TxtExportOptionsMaxSegmentChars as fo, type TxtExportOptionsMaxSegmentDurationS as fp, type TxtExportOptionsSegmentOnSilenceLongerThanS as fq, type SpeechmaticsOperatingPoint as fr, type TranscriptionModel as fs, type TranscriptionLanguage as ft, type AssemblyAIExtendedData as fu, type GladiaExtendedData as fv, type DeepgramExtendedData as fw, type ElevenLabsExtendedData as fx, type ProviderExtendedDataMap as fy, type StreamingProvider as fz, StreamingSupportedModels as g, type BeginEvent as g0, type TurnEvent as g1, type TerminationEvent as g2, type ErrorEvent as g3, type StreamingEventMessage as g4, type StreamingWord as g5, type StreamingForceEndpoint as g6, type LanguageConfig as h, type PostProcessingConfig as i, type TranscriptionMetadataDTO as j, type TranscriptionDTO as k, type TranslationDTO as l, type SummarizationDTO as m, type SentimentAnalysisDTO as n, type ChapterizationDTO as o, type PreRecordedResponse as p, type UtteranceDTO as q, TranscriptionLanguageCodeEnum as r, TranslationLanguageCodeEnum as s, StreamingSupportedRegions as t, type AddonErrorDTO as u, type AudioToLlmDTO as v, type AudioToLlmDTOError as w, type AudioToLlmDTOResults as x, type AudioToLlmListConfigDTO as y, type AudioToLlmListDTO as z };
|
|
10353
|
+
export { type CustomSpellingConfigDTO as $, type AssemblyAIStreamingOptions as A, type AudioToLlmListDTOError as B, type CallbackConfig as C, type DeepgramStreamingOptions as D, type AudioToLlmResultDTO as E, type FileResponse as F, type GladiaStreamingOptions as G, type CallbackConfigDto as H, CallbackMethodEnum as I, type CallbackTranscriptionErrorPayload as J, type CallbackTranscriptionErrorPayloadCustomMetadata as K, type ListTranscriptsOptions as L, type MessagesConfig as M, type NamedEntityRecognitionDTO as N, CallbackTranscriptionErrorPayloadEvent as O, type PreProcessingConfig as P, type CallbackTranscriptionSuccessPayload as Q, type RealtimeProcessingConfig as R, type StreamingOptions as S, type TranscribeOptions as T, type UnifiedTranscriptResponse as U, type CallbackTranscriptionSuccessPayloadCustomMetadata as V, type WordDTO as W, CallbackTranscriptionSuccessPayloadEvent as X, type ChapterizationDTOError as Y, type ChapterizationDTOResults as Z, type CodeSwitchingConfigDTO as _, type StreamingCallbacks as a, type Entity as a$, type CustomSpellingConfigDTOSpellingDictionary as a0, type CustomVocabularyConfigDTO as a1, type CustomVocabularyConfigDTOVocabularyItem as a2, type CustomVocabularyEntryDTO as a3, type DiarizationConfigDTO as a4, type DiarizationDTO as a5, type DiarizationDTOError as a6, type DisplayModeDTO as a7, type DisplayModeDTOError as a8, type ErrorDTO as a9, type SubtitlesConfigDTO as aA, SubtitlesFormatEnum as aB, SubtitlesStyleEnum as aC, type SummarizationConfigDTO as aD, type SummarizationDTOError as aE, SummaryTypesEnum as aF, TranscriptionControllerListV2KindItem as aG, type TranscriptionControllerListV2Params as aH, TranscriptionControllerListV2StatusItem as aI, type TranscriptionResultDTO as aJ, type TranslationConfigDTO as aK, type TranslationDTOError as aL, TranslationModelEnum as aM, type TranslationResultDTO as aN, type TranslationResultDTOError as aO, type TranscriptOptionalParams as aP, TranscriptStatus as aQ, type TranscriptWord as aR, AudioIntelligenceModelStatus as aS, type AutoHighlightResult as aT, type AutoHighlightsResult as aU, type Chapter$1 as aV, type ContentSafetyLabel as aW, type ContentSafetyLabelResult as aX, type ContentSafetyLabelsResult as aY, type ContentSafetyLabelsResultSeverityScoreSummary as aZ, type ContentSafetyLabelsResultSummary as a_, type InitTranscriptionRequest as aa, type InitTranscriptionRequestCustomMetadata as ab, type ModerationDTO as ac, type ModerationDTOError as ad, type NamedEntityRecognitionDTOError as ae, type NamesConsistencyDTO as af, type NamesConsistencyDTOError as ag, type PreRecordedRequestParamsResponse as ah, type PreRecordedResponseCustomMetadata as ai, type PreRecordedResponseFile as aj, PreRecordedResponseKind as ak, type PreRecordedResponsePostSessionMetadata as al, type PreRecordedResponseRequestParams as am, type PreRecordedResponseResult as an, PreRecordedResponseStatus as ao, type SentencesDTO as ap, type SentencesDTOError as aq, type SentimentAnalysisDTOError as ar, type SpeakerReidentificationDTO as as, type SpeakerReidentificationDTOError as at, type StreamingRequest as au, type StreamingRequestCustomMetadata as av, type StructuredDataExtractionConfigDTO as aw, type StructuredDataExtractionDTO as ax, type StructuredDataExtractionDTOError as ay, type SubtitleDTO as az, type StreamingSession as b, type TranscriptSentimentAnalysis as b$, EntityType as b0, type ListTranscriptsParams as b1, PiiPolicy as b2, RedactPiiAudioQuality as b3, type RedactedAudioNotification as b4, type RedactedAudioResponse as b5, RedactedAudioStatus as b6, Sentiment as b7, type SentimentAnalysisResult$1 as b8, type SentimentAnalysisResultChannel as b9, type TranscriptDisfluencies as bA, type TranscriptEntities as bB, type TranscriptEntityDetection as bC, type TranscriptFilterProfanity as bD, type TranscriptFormatText as bE, type TranscriptIabCategories as bF, type TranscriptIabCategoriesResult as bG, TranscriptLanguageCode as bH, type TranscriptLanguageCodeProperty as bI, type TranscriptLanguageConfidence as bJ, type TranscriptLanguageConfidenceThreshold as bK, type TranscriptLanguageDetection as bL, type TranscriptMultichannel as bM, type TranscriptOptionalParamsLanguageCode as bN, type TranscriptOptionalParamsLanguageCodeOneOf as bO, type TranscriptOptionalParamsRedactPiiSub as bP, type TranscriptOptionalParamsSpeakersExpected as bQ, type TranscriptOptionalParamsSpeechModel as bR, type TranscriptOptionalParamsSpeechThreshold as bS, type TranscriptOptionalParamsWebhookAuthHeaderName as bT, type TranscriptOptionalParamsWebhookAuthHeaderValue as bU, type TranscriptPunctuate as bV, type TranscriptReadyNotification as bW, TranscriptReadyStatus as bX, type TranscriptRedactPiiAudio as bY, type TranscriptRedactPiiAudioQuality as bZ, type TranscriptRedactPiiPolicies as b_, type SentimentAnalysisResultSpeaker as ba, type SeverityScoreSummary as bb, SpeechModel as bc, SubstitutionPolicy as bd, SummaryModel as be, SummaryType as bf, type Timestamp as bg, type TopicDetectionModelResult as bh, type TopicDetectionModelResultSummary as bi, type TopicDetectionResult$1 as bj, type TopicDetectionResultLabelsItem as bk, type Transcript as bl, type TranscriptAudioDuration as bm, type TranscriptAudioEndAt as bn, type TranscriptAudioStartFrom as bo, type TranscriptAutoChapters as bp, type TranscriptAutoHighlightsResult as bq, TranscriptBoostParam as br, type TranscriptBoostParamProperty as bs, type TranscriptChapters as bt, type TranscriptConfidence as bu, type TranscriptContentSafety as bv, type TranscriptContentSafetyLabels as bw, type TranscriptCustomSpelling as bx, type TranscriptCustomSpellingProperty as by, type TranscriptCustomTopics as bz, type StreamEvent as c, type TranscriptionWord as c$, type TranscriptSentimentAnalysisResults as c0, type TranscriptSpeakerLabels as c1, type TranscriptSpeakersExpected as c2, type TranscriptSpeechModel as c3, type TranscriptSpeechThreshold as c4, type TranscriptSpeedBoost as c5, type TranscriptSummary as c6, type TranscriptSummaryModel as c7, type TranscriptSummaryType as c8, type TranscriptText as c9, type TranscriptTextUsageTokens as cA, type TranscriptionSegment as cB, type RealtimeSessionCreateRequestGAModel as cC, RealtimeTranscriptionSessionCreateRequestTurnDetectionType as cD, RealtimeTranscriptionSessionCreateRequestInputAudioFormat as cE, AudioResponseFormat as cF, type CreateTranscription200One as cG, type CreateTranscriptionRequest as cH, type CreateTranscriptionRequestModel as cI, type CreateTranscriptionRequestStream as cJ, CreateTranscriptionRequestTimestampGranularitiesItem as cK, type CreateTranscriptionResponseDiarizedJson as cL, CreateTranscriptionResponseDiarizedJsonTask as cM, type CreateTranscriptionResponseDiarizedJsonUsage as cN, type CreateTranscriptionResponseJson as cO, type CreateTranscriptionResponseJsonLogprobsItem as cP, type CreateTranscriptionResponseJsonUsage as cQ, type CreateTranscriptionResponseVerboseJson as cR, type TranscriptTextUsageDuration as cS, TranscriptTextUsageDurationType as cT, type TranscriptTextUsageTokensInputTokenDetails as cU, TranscriptTextUsageTokensType as cV, type TranscriptionChunkingStrategy as cW, type TranscriptionChunkingStrategyAnyOf as cX, type TranscriptionDiarizedSegment as cY, TranscriptionDiarizedSegmentType as cZ, TranscriptionInclude as c_, type TranscriptThrottled as ca, type TranscriptUtterance as cb, type TranscriptUtteranceChannel as cc, type TranscriptUtterances as cd, type TranscriptWebhookAuthHeaderName as ce, type TranscriptWebhookNotification as cf, type TranscriptWebhookStatusCode as cg, type TranscriptWebhookUrl as ch, type TranscriptWordChannel as ci, type TranscriptWordSpeaker as cj, type TranscriptWords as ck, type StreamingUpdateConfiguration as cl, type Transcription as cm, type EntityError as cn, Status as co, type EntityReference as cp, type DiarizationProperties as cq, type DiarizationSpeakersProperties as cr, LanguageIdentificationMode as cs, type LanguageIdentificationProperties as ct, type LanguageIdentificationPropertiesSpeechModelMapping as cu, ProfanityFilterMode as cv, PunctuationMode as cw, type TranscriptionCustomProperties as cx, type TranscriptionLinks as cy, type TranscriptionProperties as cz, StreamingSupportedEncodingEnum as d, type ListenV1ResponseResultsChannelsItemSearchItemHitsItem as d$, type VadConfig as d0, VadConfigType as d1, type ListenV1Response as d2, type ManageV1FilterAccessorParameter as d3, ManageV1FilterDeploymentParameter as d4, type ManageV1LimitParameter as d5, type ManageV1PageParameter as d6, ManageV1FilterEndpointParameter as d7, ManageV1FilterMethodParameter as d8, type SharedTopics as d9, type ListenV1ModelParameter as dA, type ListenV1MultichannelParameter as dB, type ListenV1NumeralsParameter as dC, type ListenV1ParagraphsParameter as dD, type ListenV1ProfanityFilterParameter as dE, type ListenV1PunctuateParameter as dF, type ListenV1RedactParameter as dG, ListenV1RedactParameterOneOfItem as dH, type ListenV1ReplaceParameter as dI, type ListenV1ResponseMetadata as dJ, type ListenV1ResponseMetadataIntentsInfo as dK, type ListenV1ResponseMetadataModelInfo as dL, type ListenV1ResponseMetadataSentimentInfo as dM, type ListenV1ResponseMetadataSummaryInfo as dN, type ListenV1ResponseMetadataTopicsInfo as dO, type ListenV1ResponseResults as dP, type ListenV1ResponseResultsChannels as dQ, type ListenV1ResponseResultsChannelsItem as dR, type ListenV1ResponseResultsChannelsItemAlternativesItem as dS, type ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem as dT, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs as dU, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem as dV, type ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem as dW, type ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem as dX, type ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem as dY, type ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem as dZ, type ListenV1ResponseResultsChannelsItemSearchItem as d_, type SharedIntents as da, type SharedSentiments as db, type SharedCallbackParameter as dc, SharedCallbackMethodParameter as dd, type SharedSentimentParameter as de, type SharedSummarizeParameter as df, type SharedTagParameter as dg, type SharedTopicsParameter as dh, type SharedCustomTopicParameter as di, SharedCustomTopicModeParameter as dj, type SharedIntentsParameter as dk, type SharedCustomIntentParameter as dl, SharedCustomIntentModeParameter as dm, type SharedMipOptOutParameter as dn, type ListenV1DetectEntitiesParameter as dp, type ListenV1DetectLanguageParameter as dq, type ListenV1DiarizeParameter as dr, type ListenV1DictationParameter as ds, ListenV1EncodingParameter as dt, type ListenV1FillerWordsParameter as du, type ListenV1KeytermParameter as dv, type ListenV1KeywordsParameter as dw, type ListenV1LanguageParameter as dx, type ListenV1MeasurementsParameter as dy, type ListenV1MediaTranscribeParams as dz, StreamingSupportedBitDepthEnum as e, type SentimentSpeakerSummary as e$, type ListenV1ResponseResultsSummary as e0, type ListenV1ResponseResultsUtterances as e1, type ListenV1ResponseResultsUtterancesItem as e2, type ListenV1ResponseResultsUtterancesItemWordsItem as e3, type ListenV1SearchParameter as e4, type ListenV1SmartFormatParameter as e5, type ListenV1UttSplitParameter as e6, type ListenV1UtterancesParameter as e7, type ListenV1VersionParameter as e8, type ManageV1EndDateTimeParameter as e9, AutoChaptersResultErrorType as eA, type Chapter as eB, type JobInfo as eC, type LanguageIdentificationResult as eD, type LanguageIdentificationResultAlternative as eE, LanguageIdentificationResultError as eF, type LanguageIdentificationResultItem as eG, type LanguagePackInfo as eH, LanguagePackInfoWritingDirection as eI, type OutputConfigSrtOverrides as eJ, type RecognitionAlternative as eK, type RecognitionDisplay as eL, RecognitionDisplayDirection as eM, type RecognitionMetadata as eN, type RecognitionResult as eO, RecognitionResultAttachesTo as eP, RecognitionResultType as eQ, type RetrieveTranscriptResponse as eR, type RetrieveTranscriptResponseAudioEventSummary as eS, type RetrieveTranscriptResponseAudioEventSummaryChannels as eT, type RetrieveTranscriptResponseTranslations as eU, type SentimentAnalysisError as eV, SentimentAnalysisErrorType as eW, type SentimentAnalysisResult as eX, type SentimentAnalysisResultSentimentAnalysis as eY, type SentimentChannelSummary as eZ, type SentimentSegment as e_, type ManageV1FilterRequestIdParameter as ea, ManageV1FilterStatusParameter as eb, type ManageV1ProjectsRequestsListParams as ec, type ManageV1StartDateTimeParameter as ed, type SharedExtraParameter as ee, type SharedIntentsResults as ef, type SharedIntentsResultsIntents as eg, type SharedIntentsResultsIntentsSegmentsItem as eh, type SharedIntentsResultsIntentsSegmentsItemIntentsItem as ei, type SharedSentimentsAverage as ej, type SharedSentimentsSegmentsItem as ek, type SharedTopicsResults as el, type SharedTopicsResultsTopics as em, type SharedTopicsResultsTopicsSegmentsItem as en, type SharedTopicsResultsTopicsSegmentsItemTopicsItem as eo, JobType as ep, type AlignmentConfig as eq, type TranscriptionConfig as er, type TrackingData as es, type OutputConfig as et, OperatingPoint as eu, type AudioEventItem as ev, type AudioEventSummary as ew, type AudioEventSummaryItem as ex, type AutoChaptersResult as ey, type AutoChaptersResultError as ez, StreamingSupportedSampleRateEnum as f, SegmentedJsonExportOptionsFormat as f$, type SentimentSummary as f0, type SentimentSummaryDetail as f1, type SpokenFormRecognitionResult as f2, SpokenFormRecognitionResultType as f3, type SummarizationError as f4, SummarizationErrorType as f5, type SummarizationResult as f6, type TopicDetectionError as f7, TopicDetectionErrorType as f8, type TopicDetectionResult as f9, type BodySpeechToTextV1SpeechToTextPostLanguageCode as fA, BodySpeechToTextV1SpeechToTextPostModelId as fB, type BodySpeechToTextV1SpeechToTextPostNumSpeakers as fC, type BodySpeechToTextV1SpeechToTextPostSeed as fD, type BodySpeechToTextV1SpeechToTextPostTemperature as fE, BodySpeechToTextV1SpeechToTextPostTimestampsGranularity as fF, type BodySpeechToTextV1SpeechToTextPostWebhookId as fG, type BodySpeechToTextV1SpeechToTextPostWebhookMetadata as fH, type BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf as fI, type DetectedEntity as fJ, type DocxExportOptions as fK, DocxExportOptionsFormat as fL, type DocxExportOptionsMaxSegmentChars as fM, type DocxExportOptionsMaxSegmentDurationS as fN, type DocxExportOptionsSegmentOnSilenceLongerThanS as fO, type ExportOptions as fP, type HtmlExportOptions as fQ, HtmlExportOptionsFormat as fR, type HtmlExportOptionsMaxSegmentChars as fS, type HtmlExportOptionsMaxSegmentDurationS as fT, type HtmlExportOptionsSegmentOnSilenceLongerThanS as fU, type PdfExportOptions as fV, PdfExportOptionsFormat as fW, type PdfExportOptionsMaxSegmentChars as fX, type PdfExportOptionsMaxSegmentDurationS as fY, type PdfExportOptionsSegmentOnSilenceLongerThanS as fZ, type SegmentedJsonExportOptions as f_, type TopicDetectionSegment as fa, type TopicDetectionSegmentTopic as fb, type TopicDetectionSummary as fc, type TopicDetectionSummaryOverall as fd, type TrackingDataDetails as fe, type TranscriptionConfigAdditionalVocabItem as ff, TranscriptionConfigDiarization as fg, TranscriptionConfigMaxDelayMode as fh, type TranscriptionConfigPunctuationOverrides as fi, type TranscriptionConfigSpeakerDiarizationConfig as fj, type TranscriptionConfigTranscriptFilteringConfig as fk, type TranscriptionConfigTranscriptFilteringConfigReplacementsItem as fl, type TranslationError as fm, TranslationErrorType as fn, type TranslationSentence as fo, type WrittenFormRecognitionResult as fp, WrittenFormRecognitionResultType as fq, type SpeechToTextChunkResponseModel as fr, type AdditionalFormatResponseModel as fs, type AdditionalFormats as ft, type BodySpeechToTextV1SpeechToTextPost as fu, type BodySpeechToTextV1SpeechToTextPostCloudStorageUrl as fv, type BodySpeechToTextV1SpeechToTextPostDiarizationThreshold as fw, type BodySpeechToTextV1SpeechToTextPostEntityDetection as fx, type BodySpeechToTextV1SpeechToTextPostFile as fy, BodySpeechToTextV1SpeechToTextPostFileFormat as fz, StreamingSupportedModels as g, type ProviderStreamingOptions as g$, type SegmentedJsonExportOptionsMaxSegmentChars as g0, type SegmentedJsonExportOptionsMaxSegmentDurationS as g1, type SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS as g2, type SpeechToTextCharacterResponseModel as g3, type SpeechToTextCharacterResponseModelEnd as g4, type SpeechToTextCharacterResponseModelStart as g5, type SpeechToTextChunkResponseModelAdditionalFormats as g6, type SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem as g7, type SpeechToTextChunkResponseModelChannelIndex as g8, type SpeechToTextChunkResponseModelEntities as g9, type ProviderExtendedDataMap as gA, type StreamingProvider as gB, type BatchOnlyProvider as gC, type SessionStatus as gD, type Speaker as gE, type Word as gF, type Utterance as gG, type TranscriptionStatus as gH, type TranscriptMetadata as gI, type TranscriptData as gJ, type ListTranscriptsResponse as gK, type ProviderRawResponseMap as gL, type StreamEventType as gM, type SpeechEvent as gN, type TranslationEvent as gO, type SentimentEvent as gP, type EntityEvent as gQ, type SummarizationEvent as gR, type ChapterizationEvent as gS, type AudioAckEvent as gT, type LifecycleEvent as gU, type AudioChunk as gV, type RawWebSocketMessage as gW, type AssemblyAIUpdateConfiguration as gX, type OpenAIStreamingOptions as gY, type SonioxStreamingOptions as gZ, type ElevenLabsStreamingOptions as g_, type SpeechToTextChunkResponseModelTranscriptionId as ga, type SpeechToTextWordResponseModel as gb, type SpeechToTextWordResponseModelCharacters as gc, type SpeechToTextWordResponseModelEnd as gd, type SpeechToTextWordResponseModelSpeakerId as ge, type SpeechToTextWordResponseModelStart as gf, SpeechToTextWordResponseModelType as gg, type SrtExportOptions as gh, SrtExportOptionsFormat as gi, type SrtExportOptionsMaxCharactersPerLine as gj, type SrtExportOptionsMaxSegmentChars as gk, type SrtExportOptionsMaxSegmentDurationS as gl, type SrtExportOptionsSegmentOnSilenceLongerThanS as gm, type TxtExportOptions as gn, TxtExportOptionsFormat as go, type TxtExportOptionsMaxCharactersPerLine as gp, type TxtExportOptionsMaxSegmentChars as gq, type TxtExportOptionsMaxSegmentDurationS as gr, type TxtExportOptionsSegmentOnSilenceLongerThanS as gs, type SpeechmaticsOperatingPoint as gt, type TranscriptionModel as gu, type TranscriptionLanguage as gv, type AssemblyAIExtendedData as gw, type GladiaExtendedData as gx, type DeepgramExtendedData as gy, type ElevenLabsExtendedData as gz, type LanguageConfig as h, type StreamingOptionsForProvider as h0, type TranscribeStreamParams as h1, type BeginEvent as h2, type TurnEvent as h3, type TerminationEvent as h4, type ErrorEvent as h5, type StreamingEventMessage as h6, type StreamingWord as h7, type StreamingForceEndpoint as h8, type PostProcessingConfig as i, type TranscriptionMetadataDTO as j, type TranscriptionDTO as k, type TranslationDTO as l, type SummarizationDTO as m, type SentimentAnalysisDTO as n, type ChapterizationDTO as o, type PreRecordedResponse as p, type UtteranceDTO as q, TranscriptionLanguageCodeEnum as r, TranslationLanguageCodeEnum as s, StreamingSupportedRegions as t, type AddonErrorDTO as u, type AudioToLlmDTO as v, type AudioToLlmDTOError as w, type AudioToLlmDTOResults as x, type AudioToLlmListConfigDTO as y, type AudioToLlmListDTO as z };
|