voice-router-dev 0.8.3 → 0.8.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.mts CHANGED
@@ -1,10 +1,10 @@
1
- import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptOptionalParams, aQ as TranscriptStatus, aR as TranscriptWord, aS as AudioIntelligenceModelStatus, aT as AutoHighlightResult, aU as AutoHighlightsResult, aV as Chapter, aW as ContentSafetyLabel, aX as ContentSafetyLabelResult, aY as ContentSafetyLabelsResult, aZ as ContentSafetyLabelsResultSeverityScoreSummary, a_ as ContentSafetyLabelsResultSummary, a$ as Entity, b0 as EntityType, b1 as ListTranscriptsParams, b2 as PiiPolicy, b3 as RedactPiiAudioQuality, b4 as RedactedAudioNotification, b5 as RedactedAudioResponse, b6 as RedactedAudioStatus, b7 as Sentiment, b8 as SentimentAnalysisResult$1, b9 as SentimentAnalysisResultChannel, ba as SentimentAnalysisResultSpeaker, bb as SeverityScoreSummary, bc as SpeechModel, bd as SubstitutionPolicy, be as SummaryModel, bf as SummaryType, bg as Timestamp, bh as TopicDetectionModelResult, bi as TopicDetectionModelResultSummary, bj as TopicDetectionResult, bk as TopicDetectionResultLabelsItem, bl as Transcript, bm as TranscriptAudioDuration, bn as TranscriptAudioEndAt, bo as TranscriptAudioStartFrom, bp as TranscriptAutoChapters, bq as TranscriptAutoHighlightsResult, br as TranscriptBoostParam, bs as TranscriptBoostParamProperty, bt as TranscriptChapters, bu as TranscriptConfidence, bv as TranscriptContentSafety, bw as TranscriptContentSafetyLabels, bx as TranscriptCustomSpelling, by as TranscriptCustomSpellingProperty, bz as TranscriptCustomTopics, bA as TranscriptDisfluencies, bB as TranscriptEntities, bC as TranscriptEntityDetection, bD as TranscriptFilterProfanity, bE as TranscriptFormatText, bF as TranscriptIabCategories, bG as TranscriptIabCategoriesResult, bH as TranscriptLanguageCode, bI as TranscriptLanguageCodeProperty, bJ as TranscriptLanguageConfidence, bK as TranscriptLanguageConfidenceThreshold, bL as TranscriptLanguageDetection, bM as TranscriptMultichannel, bN as TranscriptOptionalParamsLanguageCode, bO as TranscriptOptionalParamsLanguageCodeOneOf, bP as TranscriptOptionalParamsRedactPiiSub, bQ as TranscriptOptionalParamsSpeakersExpected, bR as TranscriptOptionalParamsSpeechModel, bS as TranscriptOptionalParamsSpeechThreshold, bT as TranscriptOptionalParamsWebhookAuthHeaderName, bU as TranscriptOptionalParamsWebhookAuthHeaderValue, bV as TranscriptPunctuate, bW as TranscriptReadyNotification, bX as TranscriptReadyStatus, bY as TranscriptRedactPiiAudio, bZ as TranscriptRedactPiiAudioQuality, b_ as TranscriptRedactPiiPolicies, b$ as TranscriptSentimentAnalysis, c0 as TranscriptSentimentAnalysisResults, c1 as TranscriptSpeakerLabels, c2 as TranscriptSpeakersExpected, c3 as TranscriptSpeechModel, c4 as TranscriptSpeechThreshold, c5 as TranscriptSpeedBoost, c6 as TranscriptSummary, c7 as TranscriptSummaryModel, c8 as TranscriptSummaryType, c9 as TranscriptText, ca as TranscriptThrottled, cb as TranscriptUtterance, cc as TranscriptUtteranceChannel, cd as TranscriptUtterances, ce as TranscriptWebhookAuthHeaderName, cf as TranscriptWebhookNotification, cg as TranscriptWebhookStatusCode, ch as TranscriptWebhookUrl, ci as TranscriptWordChannel, cj as TranscriptWordSpeaker, ck as TranscriptWords, cl as StreamingUpdateConfiguration, cm as Transcription$1, cn as EntityError, co as Status, cp as EntityReference, cq as DiarizationProperties, cr as DiarizationSpeakersProperties, cs as LanguageIdentificationMode, ct as LanguageIdentificationProperties, cu as LanguageIdentificationPropertiesSpeechModelMapping, cv as ProfanityFilterMode, cw as PunctuationMode, cx as TranscriptionCustomProperties, cy as TranscriptionLinks, cz as TranscriptionProperties, cA as TranscriptTextUsageTokens, cB as TranscriptionSegment, cC as RealtimeSessionCreateRequestGAModel, cD as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, cE as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, cF as AudioResponseFormat, cG as CreateTranscription200One, cH as CreateTranscriptionRequest, cI as CreateTranscriptionRequestModel, cJ as CreateTranscriptionRequestStream, cK as CreateTranscriptionRequestTimestampGranularitiesItem, cL as CreateTranscriptionResponseDiarizedJson, cM as CreateTranscriptionResponseDiarizedJsonTask, cN as CreateTranscriptionResponseDiarizedJsonUsage, cO as CreateTranscriptionResponseJson, cP as CreateTranscriptionResponseJsonLogprobsItem, cQ as CreateTranscriptionResponseJsonUsage, cR as CreateTranscriptionResponseVerboseJson, cS as TranscriptTextUsageDuration, cT as TranscriptTextUsageDurationType, cU as TranscriptTextUsageTokensInputTokenDetails, cV as TranscriptTextUsageTokensType, cW as TranscriptionChunkingStrategy, cX as TranscriptionChunkingStrategyAnyOf, cY as TranscriptionDiarizedSegment, cZ as TranscriptionDiarizedSegmentType, c_ as TranscriptionInclude, c$ as TranscriptionWord, d0 as VadConfig, d1 as VadConfigType, d2 as ListenV1Response, d3 as ManageV1FilterAccessorParameter, d4 as ManageV1FilterDeploymentParameter, d5 as ManageV1LimitParameter, d6 as ManageV1PageParameter, d7 as ManageV1FilterEndpointParameter, d8 as ManageV1FilterMethodParameter, d9 as SharedTopics, da as SharedIntents, db as SharedSentiments, dc as SharedCallbackParameter, dd as SharedCallbackMethodParameter, de as SharedSentimentParameter, df as SharedSummarizeParameter, dg as SharedTagParameter, dh as SharedTopicsParameter, di as SharedCustomTopicParameter, dj as SharedCustomTopicModeParameter, dk as SharedIntentsParameter, dl as SharedCustomIntentParameter, dm as SharedCustomIntentModeParameter, dn as SharedMipOptOutParameter, dp as ListenV1DetectEntitiesParameter, dq as ListenV1DetectLanguageParameter, dr as ListenV1DiarizeParameter, ds as ListenV1DictationParameter, dt as ListenV1EncodingParameter, du as ListenV1FillerWordsParameter, dv as ListenV1KeytermParameter, dw as ListenV1KeywordsParameter, dx as ListenV1LanguageParameter, dy as ListenV1MeasurementsParameter, dz as ListenV1MediaTranscribeParams, dA as ListenV1ModelParameter, dB as ListenV1MultichannelParameter, dC as ListenV1NumeralsParameter, dD as ListenV1ParagraphsParameter, dE as ListenV1ProfanityFilterParameter, dF as ListenV1PunctuateParameter, dG as ListenV1RedactParameter, dH as ListenV1RedactParameterOneOfItem, dI as ListenV1ReplaceParameter, dJ as ListenV1ResponseMetadata, dK as ListenV1ResponseMetadataIntentsInfo, dL as ListenV1ResponseMetadataModelInfo, dM as ListenV1ResponseMetadataSentimentInfo, dN as ListenV1ResponseMetadataSummaryInfo, dO as ListenV1ResponseMetadataTopicsInfo, dP as ListenV1ResponseResults, dQ as ListenV1ResponseResultsChannels, dR as ListenV1ResponseResultsChannelsItem, dS as ListenV1ResponseResultsChannelsItemAlternativesItem, dT as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, dU as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, dV as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, dW as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, dX as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, dY as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, dZ as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, d_ as ListenV1ResponseResultsChannelsItemSearchItem, d$ as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, e0 as ListenV1ResponseResultsSummary, e1 as ListenV1ResponseResultsUtterances, e2 as ListenV1ResponseResultsUtterancesItem, e3 as ListenV1ResponseResultsUtterancesItemWordsItem, e4 as ListenV1SearchParameter, e5 as ListenV1SmartFormatParameter, e6 as ListenV1UttSplitParameter, e7 as ListenV1UtterancesParameter, e8 as ListenV1VersionParameter, e9 as ManageV1EndDateTimeParameter, ea as ManageV1FilterRequestIdParameter, eb as ManageV1FilterStatusParameter, ec as ManageV1ProjectsRequestsListParams, ed as ManageV1StartDateTimeParameter, ee as SharedExtraParameter, ef as SharedIntentsResults, eg as SharedIntentsResultsIntents, eh as SharedIntentsResultsIntentsSegmentsItem, ei as SharedIntentsResultsIntentsSegmentsItemIntentsItem, ej as SharedSentimentsAverage, ek as SharedSentimentsSegmentsItem, el as SharedTopicsResults, em as SharedTopicsResultsTopics, en as SharedTopicsResultsTopicsSegmentsItem, eo as SharedTopicsResultsTopicsSegmentsItemTopicsItem, ep as JobType, eq as AlignmentConfig, er as TranscriptionConfig, es as TrackingData, et as OutputConfig, eu as OperatingPoint, ev as AudioEventItem, ew as AudioEventSummary, ex as AudioEventSummaryItem, ey as AutoChaptersResult, ez as AutoChaptersResultError, eA as AutoChaptersResultErrorType, eB as Chapter$1, eC as JobInfo, eD as LanguageIdentificationResult, eE as LanguageIdentificationResultAlternative, eF as LanguageIdentificationResultError, eG as LanguageIdentificationResultItem, eH as LanguagePackInfo, eI as LanguagePackInfoWritingDirection, eJ as OutputConfigSrtOverrides, eK as RecognitionAlternative, eL as RecognitionDisplay, eM as RecognitionDisplayDirection, eN as RecognitionMetadata, eO as RecognitionResult, eP as RecognitionResultAttachesTo, eQ as RecognitionResultType, eR as RetrieveTranscriptResponse, eS as RetrieveTranscriptResponseAudioEventSummary, eT as RetrieveTranscriptResponseAudioEventSummaryChannels, eU as RetrieveTranscriptResponseTranslations, eV as SentimentAnalysisError, eW as SentimentAnalysisErrorType, eX as SentimentAnalysisResult$2, eY as SentimentAnalysisResultSentimentAnalysis, eZ as SentimentChannelSummary, e_ as SentimentSegment, e$ as SentimentSpeakerSummary, f0 as SentimentSummary, f1 as SentimentSummaryDetail, f2 as SpokenFormRecognitionResult, f3 as SpokenFormRecognitionResultType, f4 as SummarizationError, f5 as SummarizationErrorType, f6 as SummarizationResult, f7 as TopicDetectionError, f8 as TopicDetectionErrorType, f9 as TopicDetectionResult$1, fa as TopicDetectionSegment, fb as TopicDetectionSegmentTopic, fc as TopicDetectionSummary, fd as TopicDetectionSummaryOverall, fe as TrackingDataDetails, ff as TranscriptionConfigAdditionalVocabItem, fg as TranscriptionConfigDiarization, fh as TranscriptionConfigMaxDelayMode, fi as TranscriptionConfigPunctuationOverrides, fj as TranscriptionConfigSpeakerDiarizationConfig, fk as TranscriptionConfigTranscriptFilteringConfig, fl as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fm as TranslationError, fn as TranslationErrorType, fo as TranslationSentence, fp as WrittenFormRecognitionResult, fq as WrittenFormRecognitionResultType, fr as SpeechToTextChunkResponseModel, fs as AdditionalFormatResponseModel, ft as AdditionalFormats, fu as BodySpeechToTextV1SpeechToTextPost, fv as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fw as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, fx as BodySpeechToTextV1SpeechToTextPostEntityDetection, fy as BodySpeechToTextV1SpeechToTextPostFile, fz as BodySpeechToTextV1SpeechToTextPostFileFormat, fA as BodySpeechToTextV1SpeechToTextPostLanguageCode, fB as BodySpeechToTextV1SpeechToTextPostModelId, fC as BodySpeechToTextV1SpeechToTextPostNumSpeakers, fD as BodySpeechToTextV1SpeechToTextPostSeed, fE as BodySpeechToTextV1SpeechToTextPostTemperature, fF as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, fG as BodySpeechToTextV1SpeechToTextPostWebhookId, fH as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, fI as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, fJ as DetectedEntity, fK as DocxExportOptions, fL as DocxExportOptionsFormat, fM as DocxExportOptionsMaxSegmentChars, fN as DocxExportOptionsMaxSegmentDurationS, fO as DocxExportOptionsSegmentOnSilenceLongerThanS, fP as ExportOptions, fQ as HtmlExportOptions, fR as HtmlExportOptionsFormat, fS as HtmlExportOptionsMaxSegmentChars, fT as HtmlExportOptionsMaxSegmentDurationS, fU as HtmlExportOptionsSegmentOnSilenceLongerThanS, fV as PdfExportOptions, fW as PdfExportOptionsFormat, fX as PdfExportOptionsMaxSegmentChars, fY as PdfExportOptionsMaxSegmentDurationS, fZ as PdfExportOptionsSegmentOnSilenceLongerThanS, f_ as SegmentedJsonExportOptions, f$ as SegmentedJsonExportOptionsFormat, g0 as SegmentedJsonExportOptionsMaxSegmentChars, g1 as SegmentedJsonExportOptionsMaxSegmentDurationS, g2 as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, g3 as SpeechToTextCharacterResponseModel, g4 as SpeechToTextCharacterResponseModelEnd, g5 as SpeechToTextCharacterResponseModelStart, g6 as SpeechToTextChunkResponseModelAdditionalFormats, g7 as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, g8 as SpeechToTextChunkResponseModelChannelIndex, g9 as SpeechToTextChunkResponseModelEntities, ga as SpeechToTextChunkResponseModelTranscriptionId, gb as SpeechToTextWordResponseModel, gc as SpeechToTextWordResponseModelCharacters, gd as SpeechToTextWordResponseModelEnd, ge as SpeechToTextWordResponseModelSpeakerId, gf as SpeechToTextWordResponseModelStart, gg as SpeechToTextWordResponseModelType, gh as SrtExportOptions, gi as SrtExportOptionsFormat, gj as SrtExportOptionsMaxCharactersPerLine, gk as SrtExportOptionsMaxSegmentChars, gl as SrtExportOptionsMaxSegmentDurationS, gm as SrtExportOptionsSegmentOnSilenceLongerThanS, gn as TxtExportOptions, go as TxtExportOptionsFormat, gp as TxtExportOptionsMaxCharactersPerLine, gq as TxtExportOptionsMaxSegmentChars, gr as TxtExportOptionsMaxSegmentDurationS, gs as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-DvIT4xai.mjs';
2
- export { gw as AssemblyAIExtendedData, gX as AssemblyAIUpdateConfiguration, gT as AudioAckEvent, gV as AudioChunk, gC as BatchOnlyProvider, h2 as BeginEvent, gS as ChapterizationEvent, gy as DeepgramExtendedData, gz as ElevenLabsExtendedData, g_ as ElevenLabsStreamingOptions, gQ as EntityEvent, h5 as ErrorEvent, gx as GladiaExtendedData, gU as LifecycleEvent, gK as ListTranscriptsResponse, gY as OpenAIStreamingOptions, gA as ProviderExtendedDataMap, gL as ProviderRawResponseMap, g$ as ProviderStreamingOptions, gW as RawWebSocketMessage, gP as SentimentEvent, gD as SessionStatus, gZ as SonioxStreamingOptions, gE as Speaker, gN as SpeechEvent, gt as SpeechmaticsOperatingPoint, gM as StreamEventType, h6 as StreamingEventMessage, h8 as StreamingForceEndpoint, h0 as StreamingOptionsForProvider, gB as StreamingProvider, h7 as StreamingWord, gR as SummarizationEvent, h4 as TerminationEvent, h1 as TranscribeStreamParams, gJ as TranscriptData, gI as TranscriptMetadata, gv as TranscriptionLanguage, gu as TranscriptionModel, gH as TranscriptionStatus, gO as TranslationEvent, h3 as TurnEvent, gG as Utterance, gF as Word } from './speechToTextChunkResponseModel-DvIT4xai.mjs';
1
+ import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptStatus, aQ as TranscriptWord, aR as AudioIntelligenceModelStatus, aS as AutoHighlightResult, aT as AutoHighlightsResult, aU as Chapter, aV as ContentSafetyLabel, aW as ContentSafetyLabelResult, aX as ContentSafetyLabelsResult, aY as ContentSafetyLabelsResultSeverityScoreSummary, aZ as ContentSafetyLabelsResultSummary, a_ as CustomFormattingRequestBody, a$ as CustomFormattingRequestBodyCustomFormatting, b0 as CustomFormattingResponse, b1 as CustomFormattingResponseCustomFormatting, b2 as CustomFormattingResponseCustomFormattingMapping, b3 as Entity, b4 as EntityType, b5 as ListTranscriptsParams, b6 as PiiPolicy, b7 as RedactPiiAudioQuality, b8 as Sentiment, b9 as SentimentAnalysisResult$1, ba as SentimentAnalysisResultChannel, bb as SentimentAnalysisResultSpeaker, bc as SeverityScoreSummary, bd as SpeakerIdentificationRequestBody, be as SpeakerIdentificationRequestBodySpeakerIdentification, bf as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, bg as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, bh as SpeakerIdentificationResponse, bi as SpeakerIdentificationResponseSpeakerIdentification, bj as SpeakerIdentificationResponseSpeakerIdentificationMapping, bk as SpeechModel, bl as SubstitutionPolicy, bm as SummaryModel, bn as SummaryType, bo as Timestamp, bp as TopicDetectionModelResult, bq as TopicDetectionModelResultSummary, br as TopicDetectionResult, bs as TopicDetectionResultLabelsItems, bt as Transcript, bu as TranscriptAudioDuration, bv as TranscriptAudioEndAt, bw as TranscriptAudioStartFrom, bx as TranscriptAutoChapters, by as TranscriptAutoHighlightsResult, bz as TranscriptChapters, bA as TranscriptConfidence, bB as TranscriptContentSafety, bC as TranscriptContentSafetyLabels, bD as TranscriptCustomSpelling, bE as TranscriptCustomSpellingProperty, bF as TranscriptCustomTopics, bG as TranscriptDisfluencies, bH as TranscriptDomain, bI as TranscriptEntities, bJ as TranscriptEntityDetection, bK as TranscriptFilterProfanity, bL as TranscriptFormatText, bM as TranscriptIabCategories, bN as TranscriptIabCategoriesResult, bO as TranscriptLanguageCode, bP as TranscriptLanguageCodes, bQ as TranscriptLanguageConfidence, bR as TranscriptLanguageConfidenceThreshold, bS as TranscriptLanguageDetection, bT as TranscriptLanguageDetectionOptions, bU as TranscriptMultichannel, bV as TranscriptOptionalParamsLanguageDetectionOptions, bW as TranscriptOptionalParamsRedactPiiAudioOptions, bX as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, bY as TranscriptOptionalParamsRemoveAudioTags, bZ as TranscriptOptionalParamsSpeakerOptions, b_ as TranscriptOptionalParamsSpeechUnderstanding, b$ as TranscriptOptionalParamsSpeechUnderstandingRequest, c0 as TranscriptParams, c1 as TranscriptParamsDomain, c2 as TranscriptParamsLanguageCode, c3 as TranscriptParamsLanguageCodes, c4 as TranscriptParamsRedactPiiSub, c5 as TranscriptParamsRemoveAudioTags, c6 as TranscriptParamsSpeakersExpected, c7 as TranscriptParamsSpeechModel, c8 as TranscriptParamsSpeechThreshold, c9 as TranscriptParamsWebhookAuthHeaderName, ca as TranscriptParamsWebhookAuthHeaderValue, cb as TranscriptPunctuate, cc as TranscriptRedactPiiAudio, cd as TranscriptRedactPiiAudioOptions, ce as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, cf as TranscriptRedactPiiAudioQuality, cg as TranscriptRedactPiiPolicies, ch as TranscriptRemoveAudioTags, ci as TranscriptRemoveAudioTagsProperty, cj as TranscriptSentimentAnalysis, ck as TranscriptSentimentAnalysisResults, cl as TranscriptSpeakerLabels, cm as TranscriptSpeakersExpected, cn as TranscriptSpeechModel, co as TranscriptSpeechModels, cp as TranscriptSpeechThreshold, cq as TranscriptSpeechUnderstanding, cr as TranscriptSpeechUnderstandingRequest, cs as TranscriptSpeechUnderstandingResponse, ct as TranscriptSpeedBoost, cu as TranscriptSummary, cv as TranscriptSummaryModel, cw as TranscriptSummaryType, cx as TranscriptTemperature, cy as TranscriptText, cz as TranscriptThrottled, cA as TranscriptTranslatedTexts, cB as TranscriptUtterance, cC as TranscriptUtteranceChannel, cD as TranscriptUtteranceTranslatedTexts, cE as TranscriptUtterances, cF as TranscriptWebhookAuthHeaderName, cG as TranscriptWebhookStatusCode, cH as TranscriptWebhookUrl, cI as TranscriptWordChannel, cJ as TranscriptWordSpeaker, cK as TranscriptWords, cL as TranslationRequestBody, cM as TranslationRequestBodyTranslation, cN as TranslationResponse, cO as TranslationResponseTranslation, cP as StreamingUpdateConfiguration, cQ as Transcription$1, cR as EntityError, cS as Status, cT as EntityReference, cU as DiarizationProperties, cV as DiarizationSpeakersProperties, cW as LanguageIdentificationMode, cX as LanguageIdentificationProperties, cY as LanguageIdentificationPropertiesSpeechModelMapping, cZ as ProfanityFilterMode, c_ as PunctuationMode, c$ as TranscriptionCustomProperties, d0 as TranscriptionLinks, d1 as TranscriptionProperties, d2 as TranscriptTextUsageTokens, d3 as TranscriptionSegment, d4 as RealtimeSessionCreateRequestGAModel, d5 as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, d6 as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, d7 as AudioResponseFormat, d8 as CreateTranscription200One, d9 as CreateTranscriptionRequest, da as CreateTranscriptionRequestModel, db as CreateTranscriptionRequestStream, dc as CreateTranscriptionRequestTimestampGranularitiesItem, dd as CreateTranscriptionResponseDiarizedJson, de as CreateTranscriptionResponseDiarizedJsonTask, df as CreateTranscriptionResponseDiarizedJsonUsage, dg as CreateTranscriptionResponseJson, dh as CreateTranscriptionResponseJsonLogprobsItem, di as CreateTranscriptionResponseJsonUsage, dj as CreateTranscriptionResponseVerboseJson, dk as TranscriptTextUsageDuration, dl as TranscriptTextUsageDurationType, dm as TranscriptTextUsageTokensInputTokenDetails, dn as TranscriptTextUsageTokensType, dp as TranscriptionChunkingStrategy, dq as TranscriptionChunkingStrategyAnyOf, dr as TranscriptionDiarizedSegment, ds as TranscriptionDiarizedSegmentType, dt as TranscriptionInclude, du as TranscriptionWord, dv as VadConfig, dw as VadConfigType, dx as ListenV1Response, dy as ManageV1FilterAccessorParameter, dz as ManageV1FilterDeploymentParameter, dA as ManageV1LimitParameter, dB as ManageV1PageParameter, dC as ManageV1FilterEndpointParameter, dD as ManageV1FilterMethodParameter, dE as SharedTopics, dF as SharedIntents, dG as SharedSentiments, dH as SharedCallbackParameter, dI as SharedCallbackMethodParameter, dJ as SharedSentimentParameter, dK as SharedSummarizeParameter, dL as SharedTagParameter, dM as SharedTopicsParameter, dN as SharedCustomTopicParameter, dO as SharedCustomTopicModeParameter, dP as SharedIntentsParameter, dQ as SharedCustomIntentParameter, dR as SharedCustomIntentModeParameter, dS as SharedMipOptOutParameter, dT as ListenV1DetectEntitiesParameter, dU as ListenV1DetectLanguageParameter, dV as ListenV1DiarizeParameter, dW as ListenV1DictationParameter, dX as ListenV1EncodingParameter, dY as ListenV1FillerWordsParameter, dZ as ListenV1KeytermParameter, d_ as ListenV1KeywordsParameter, d$ as ListenV1LanguageParameter, e0 as ListenV1MeasurementsParameter, e1 as ListenV1MediaTranscribeParams, e2 as ListenV1ModelParameter, e3 as ListenV1MultichannelParameter, e4 as ListenV1NumeralsParameter, e5 as ListenV1ParagraphsParameter, e6 as ListenV1ProfanityFilterParameter, e7 as ListenV1PunctuateParameter, e8 as ListenV1RedactParameter, e9 as ListenV1RedactParameterOneOfItem, ea as ListenV1ReplaceParameter, eb as ListenV1ResponseMetadata, ec as ListenV1ResponseMetadataIntentsInfo, ed as ListenV1ResponseMetadataModelInfo, ee as ListenV1ResponseMetadataSentimentInfo, ef as ListenV1ResponseMetadataSummaryInfo, eg as ListenV1ResponseMetadataTopicsInfo, eh as ListenV1ResponseResults, ei as ListenV1ResponseResultsChannels, ej as ListenV1ResponseResultsChannelsItem, ek as ListenV1ResponseResultsChannelsItemAlternativesItem, el as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, em as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, en as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, eo as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, ep as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, eq as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, er as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, es as ListenV1ResponseResultsChannelsItemSearchItem, et as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, eu as ListenV1ResponseResultsSummary, ev as ListenV1ResponseResultsUtterances, ew as ListenV1ResponseResultsUtterancesItem, ex as ListenV1ResponseResultsUtterancesItemWordsItem, ey as ListenV1SearchParameter, ez as ListenV1SmartFormatParameter, eA as ListenV1UttSplitParameter, eB as ListenV1UtterancesParameter, eC as ListenV1VersionParameter, eD as ManageV1EndDateTimeParameter, eE as ManageV1FilterRequestIdParameter, eF as ManageV1FilterStatusParameter, eG as ManageV1ProjectsRequestsListParams, eH as ManageV1StartDateTimeParameter, eI as SharedExtraParameter, eJ as SharedIntentsResults, eK as SharedIntentsResultsIntents, eL as SharedIntentsResultsIntentsSegmentsItem, eM as SharedIntentsResultsIntentsSegmentsItemIntentsItem, eN as SharedSentimentsAverage, eO as SharedSentimentsSegmentsItem, eP as SharedTopicsResults, eQ as SharedTopicsResultsTopics, eR as SharedTopicsResultsTopicsSegmentsItem, eS as SharedTopicsResultsTopicsSegmentsItemTopicsItem, eT as JobType, eU as AlignmentConfig, eV as TranscriptionConfig, eW as TrackingData, eX as OutputConfig, eY as OperatingPoint, eZ as AudioEventItem, e_ as AudioEventSummary, e$ as AudioEventSummaryItem, f0 as AutoChaptersResult, f1 as AutoChaptersResultError, f2 as AutoChaptersResultErrorType, f3 as Chapter$1, f4 as JobInfo, f5 as LanguageIdentificationResult, f6 as LanguageIdentificationResultAlternative, f7 as LanguageIdentificationResultError, f8 as LanguageIdentificationResultItem, f9 as LanguagePackInfo, fa as LanguagePackInfoWritingDirection, fb as OutputConfigSrtOverrides, fc as RecognitionAlternative, fd as RecognitionDisplay, fe as RecognitionDisplayDirection, ff as RecognitionMetadata, fg as RecognitionResult, fh as RecognitionResultAttachesTo, fi as RecognitionResultType, fj as RetrieveTranscriptResponse, fk as RetrieveTranscriptResponseAudioEventSummary, fl as RetrieveTranscriptResponseAudioEventSummaryChannels, fm as RetrieveTranscriptResponseTranslations, fn as SentimentAnalysisError, fo as SentimentAnalysisErrorType, fp as SentimentAnalysisResult$2, fq as SentimentAnalysisResultSentimentAnalysis, fr as SentimentChannelSummary, fs as SentimentSegment, ft as SentimentSpeakerSummary, fu as SentimentSummary, fv as SentimentSummaryDetail, fw as SpokenFormRecognitionResult, fx as SpokenFormRecognitionResultType, fy as SummarizationError, fz as SummarizationErrorType, fA as SummarizationResult, fB as TopicDetectionError, fC as TopicDetectionErrorType, fD as TopicDetectionResult$1, fE as TopicDetectionSegment, fF as TopicDetectionSegmentTopic, fG as TopicDetectionSummary, fH as TopicDetectionSummaryOverall, fI as TrackingDataDetails, fJ as TranscriptionConfigAdditionalVocabItem, fK as TranscriptionConfigDiarization, fL as TranscriptionConfigMaxDelayMode, fM as TranscriptionConfigPunctuationOverrides, fN as TranscriptionConfigSpeakerDiarizationConfig, fO as TranscriptionConfigTranscriptFilteringConfig, fP as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fQ as TranslationError, fR as TranslationErrorType, fS as TranslationSentence, fT as WrittenFormRecognitionResult, fU as WrittenFormRecognitionResultType, fV as SpeechToTextChunkResponseModel, fW as AdditionalFormatResponseModel, fX as AdditionalFormats, fY as BodySpeechToTextV1SpeechToTextPost, fZ as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, f_ as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, f$ as BodySpeechToTextV1SpeechToTextPostEntityDetection, g0 as BodySpeechToTextV1SpeechToTextPostFile, g1 as BodySpeechToTextV1SpeechToTextPostFileFormat, g2 as BodySpeechToTextV1SpeechToTextPostLanguageCode, g3 as BodySpeechToTextV1SpeechToTextPostModelId, g4 as BodySpeechToTextV1SpeechToTextPostNumSpeakers, g5 as BodySpeechToTextV1SpeechToTextPostSeed, g6 as BodySpeechToTextV1SpeechToTextPostTemperature, g7 as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, g8 as BodySpeechToTextV1SpeechToTextPostWebhookId, g9 as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, ga as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, gb as DetectedEntity, gc as DocxExportOptions, gd as DocxExportOptionsFormat, ge as DocxExportOptionsMaxSegmentChars, gf as DocxExportOptionsMaxSegmentDurationS, gg as DocxExportOptionsSegmentOnSilenceLongerThanS, gh as ExportOptions, gi as HtmlExportOptions, gj as HtmlExportOptionsFormat, gk as HtmlExportOptionsMaxSegmentChars, gl as HtmlExportOptionsMaxSegmentDurationS, gm as HtmlExportOptionsSegmentOnSilenceLongerThanS, gn as PdfExportOptions, go as PdfExportOptionsFormat, gp as PdfExportOptionsMaxSegmentChars, gq as PdfExportOptionsMaxSegmentDurationS, gr as PdfExportOptionsSegmentOnSilenceLongerThanS, gs as SegmentedJsonExportOptions, gt as SegmentedJsonExportOptionsFormat, gu as SegmentedJsonExportOptionsMaxSegmentChars, gv as SegmentedJsonExportOptionsMaxSegmentDurationS, gw as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, gx as SpeechToTextCharacterResponseModel, gy as SpeechToTextCharacterResponseModelEnd, gz as SpeechToTextCharacterResponseModelStart, gA as SpeechToTextChunkResponseModelAdditionalFormats, gB as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, gC as SpeechToTextChunkResponseModelChannelIndex, gD as SpeechToTextChunkResponseModelEntities, gE as SpeechToTextChunkResponseModelTranscriptionId, gF as SpeechToTextWordResponseModel, gG as SpeechToTextWordResponseModelCharacters, gH as SpeechToTextWordResponseModelEnd, gI as SpeechToTextWordResponseModelSpeakerId, gJ as SpeechToTextWordResponseModelStart, gK as SpeechToTextWordResponseModelType, gL as SrtExportOptions, gM as SrtExportOptionsFormat, gN as SrtExportOptionsMaxCharactersPerLine, gO as SrtExportOptionsMaxSegmentChars, gP as SrtExportOptionsMaxSegmentDurationS, gQ as SrtExportOptionsSegmentOnSilenceLongerThanS, gR as TxtExportOptions, gS as TxtExportOptionsFormat, gT as TxtExportOptionsMaxCharactersPerLine, gU as TxtExportOptionsMaxSegmentChars, gV as TxtExportOptionsMaxSegmentDurationS, gW as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-BcT1LJSZ.mjs';
2
+ export { g_ as AssemblyAIExtendedData, hp as AssemblyAIUpdateConfiguration, hl as AudioAckEvent, hn as AudioChunk, h4 as BatchOnlyProvider, hw as BeginEvent, hk as ChapterizationEvent, h0 as DeepgramExtendedData, h1 as ElevenLabsExtendedData, hs as ElevenLabsStreamingOptions, hi as EntityEvent, hz as ErrorEvent, g$ as GladiaExtendedData, hm as LifecycleEvent, hc as ListTranscriptsResponse, hq as OpenAIStreamingOptions, h2 as ProviderExtendedDataMap, hd as ProviderRawResponseMap, ht as ProviderStreamingOptions, ho as RawWebSocketMessage, hh as SentimentEvent, h5 as SessionStatus, hr as SonioxStreamingOptions, h6 as Speaker, hf as SpeechEvent, gX as SpeechmaticsOperatingPoint, he as StreamEventType, hA as StreamingEventMessage, hC as StreamingForceEndpoint, hu as StreamingOptionsForProvider, h3 as StreamingProvider, hB as StreamingWord, hj as SummarizationEvent, hy as TerminationEvent, hv as TranscribeStreamParams, hb as TranscriptData, ha as TranscriptMetadata, gZ as TranscriptionLanguage, gY as TranscriptionModel, h9 as TranscriptionStatus, hg as TranslationEvent, hx as TurnEvent, h8 as Utterance, h7 as Word } from './speechToTextChunkResponseModel-BcT1LJSZ.mjs';
3
3
  import { T as TranscriptionProvider, P as ProviderCapabilities, A as AudioInput } from './provider-metadata-MDUUEuqF.mjs';
4
4
  export { k as AllLanguageCodes, o as AllProviders, a as AssemblyAICapabilities, i as AssemblyAILanguageCodes, b as AzureCapabilities, B as BatchOnlyProviderType, q as BatchOnlyProviders, C as CapabilityKeys, f as CapabilityLabels, D as DeepgramCapabilities, j as DeepgramLanguageCodes, E as ElevenLabsCapabilities, G as GladiaCapabilities, h as GladiaLanguageCodes, g as LanguageCode, L as LanguageLabels, O as OpenAICapabilities, d as ProviderCapabilitiesMap, l as ProviderDisplayNames, n as ProviderDocs, m as ProviderWebsites, c as SonioxCapabilities, S as SpeechmaticsCapabilities, e as StreamingProviderType, p as StreamingProviders } from './provider-metadata-MDUUEuqF.mjs';
5
5
  import { AssemblyAIRegionType, DeepgramRegionType, SpeechmaticsRegionType, SonioxModelCode, SonioxRegionType, ElevenLabsModelCode, ElevenLabsRegionType } from './constants.mjs';
6
6
  export { AssemblyAIEncoding, AssemblyAIEncodingType, AssemblyAILanguage, AssemblyAILanguageType, AssemblyAISampleRate, AssemblyAISampleRateType, AssemblyAISpeechModel, AssemblyAISpeechModelType, AssemblyAIStatus, AssemblyAITranscriptionModel, AssemblyAITranscriptionModelType, AzureLocale, AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, AzureLocaleType, AzureLocales, DeepgramCallbackMethod, DeepgramEncoding, DeepgramIntentMode, DeepgramLanguage, DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, DeepgramModelCode as DeepgramModelType, DeepgramRedact, DeepgramRedactType, DeepgramRegion, DeepgramSampleRate, DeepgramStatus, DeepgramTTSContainer, DeepgramTTSEncoding, DeepgramTTSModel, DeepgramTTSSampleRate, DeepgramTopicMode, DeepgramTopicModeType, ElevenLabsLanguageCode, ElevenLabsLanguageCodes, ElevenLabsLanguageLabels, ElevenLabsLanguages, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaRegion, GladiaSampleRate, GladiaStatus, GladiaTranslationLanguage, OpenAILanguageCodes, OpenAIModel, OpenAIRealtimeAudioFormat, OpenAIRealtimeModel, OpenAIRealtimeTranscriptionModel, OpenAIRealtimeTurnDetection, OpenAIResponseFormat, SonioxAsyncModel, SonioxAsyncModelCode, SonioxLanguage, SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, SonioxLanguageType, SonioxLanguages, SonioxModel, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, SonioxRealtimeModelCode, SonioxRegion, SpeechmaticsLanguage, SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsRegion } from './constants.mjs';
7
- export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-t_lVCkE5.mjs';
7
+ export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-DLbrsYTk.mjs';
8
8
  import { z } from 'zod';
9
9
  import { AxiosResponse, AxiosRequestConfig } from 'axios';
10
10
 
@@ -4115,559 +4115,101 @@ declare function createGladiaAdapter(config: ProviderConfig): GladiaAdapter;
4115
4115
  * Generated by orval v7.9.0 🍺
4116
4116
  * Do not edit manually.
4117
4117
  * AssemblyAI API
4118
- * AssemblyAI API
4119
- * OpenAPI spec version: 1.3.4
4120
- */
4121
- type TranscriptParamsAllOf = {
4122
- /** The URL of the audio or video file to transcribe. */
4123
- audio_url: string;
4124
- };
4125
-
4126
- /**
4127
- * Generated by orval v7.9.0 🍺
4128
- * Do not edit manually.
4129
- * AssemblyAI API
4130
- * AssemblyAI API
4131
- * OpenAPI spec version: 1.3.4
4132
- */
4133
-
4134
- /**
4135
- * The parameters for creating a transcript
4136
- */
4137
- type TranscriptParams = TranscriptParamsAllOf & TranscriptOptionalParams;
4138
-
4139
- /**
4140
- * Generated by orval v7.9.0 🍺
4141
- * Do not edit manually.
4142
- * AssemblyAI API
4143
- * AssemblyAI API
4144
- * OpenAPI spec version: 1.3.4
4145
- */
4146
- /**
4147
- * The date and time the transcript was completed
4148
- * @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
4149
- */
4150
- type TranscriptListItemCompleted = string | null;
4151
-
4152
- /**
4153
- * Generated by orval v7.9.0 🍺
4154
- * Do not edit manually.
4155
- * AssemblyAI API
4156
- * AssemblyAI API
4157
- * OpenAPI spec version: 1.3.4
4158
- */
4159
- /**
4160
- * Error message of why the transcript failed
4161
- */
4162
- type TranscriptListItemError = string | null;
4163
-
4164
- /**
4165
- * Generated by orval v7.9.0 🍺
4166
- * Do not edit manually.
4167
- * AssemblyAI API
4168
- * AssemblyAI API
4169
- * OpenAPI spec version: 1.3.4
4170
- */
4171
-
4172
- interface TranscriptListItem {
4173
- /** The unique identifier for the transcript */
4174
- id: string;
4175
- /** The URL to retrieve the transcript */
4176
- resource_url: string;
4177
- /** The status of the transcript */
4178
- status: TranscriptStatus;
4179
- /**
4180
- * The date and time the transcript was created
4181
- * @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
4182
- */
4183
- created: string;
4184
- /**
4185
- * The date and time the transcript was completed
4186
- * @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
4187
- */
4188
- completed: TranscriptListItemCompleted;
4189
- /** The URL to the audio file */
4190
- audio_url: string;
4191
- /** Error message of why the transcript failed */
4192
- error: TranscriptListItemError;
4193
- }
4194
-
4195
- /**
4196
- * Bad request
4197
- */
4198
- type BadRequestResponse = Error;
4199
-
4200
- /**
4201
- * Cannot access uploaded file
4202
- */
4203
- type CannotAccessUploadedFileResponse = Error;
4204
-
4205
- /**
4206
- * Generated by orval v7.9.0 🍺
4207
- * Do not edit manually.
4208
- * AssemblyAI API
4209
- * AssemblyAI API
4210
- * OpenAPI spec version: 1.3.4
4211
- */
4212
- interface CreateRealtimeTemporaryTokenParams {
4213
- /**
4214
- * The amount of time until the token expires in seconds
4215
- * @minimum 60
4216
- */
4217
- expires_in: number;
4218
- }
4219
-
4220
- /**
4221
- * Generated by orval v7.9.0 🍺
4222
- * Do not edit manually.
4223
- * AssemblyAI API
4224
- * AssemblyAI API
4225
- * OpenAPI spec version: 1.3.4
4226
- */
4227
- interface Error$2 {
4228
- /** Error message */
4229
- error: string;
4230
- status?: "error";
4231
- [key: string]: unknown;
4232
- }
4233
-
4234
- /**
4235
- * Generated by orval v7.9.0 🍺
4236
- * Do not edit manually.
4237
- * AssemblyAI API
4238
- * AssemblyAI API
4239
- * OpenAPI spec version: 1.3.4
4240
- */
4241
- /**
4242
- * Gateway timeout
4243
- */
4244
- type GatewayTimeoutResponse = void;
4245
-
4246
- /**
4247
- * Generated by orval v7.9.0 🍺
4248
- * Do not edit manually.
4249
- * AssemblyAI API
4250
- * AssemblyAI API
4251
- * OpenAPI spec version: 1.3.4
4252
- */
4253
- type GetSubtitlesParams = {
4254
- /**
4255
- * The maximum number of characters per caption
4256
- */
4257
- chars_per_caption?: number;
4258
- };
4259
-
4260
- /**
4261
- * An error occurred while processing the request
4262
- */
4263
- type InternalServerErrorResponse = Error;
4264
-
4265
- /**
4266
- * Generated by orval v7.9.0 🍺
4267
- * Do not edit manually.
4268
- * AssemblyAI API
4269
- * AssemblyAI API
4270
- * OpenAPI spec version: 1.3.4
4271
- */
4272
- type LemurBaseParamsContextOneOf = {
4273
- [key: string]: unknown;
4274
- };
4275
-
4276
- /**
4277
- * Generated by orval v7.9.0 🍺
4278
- * Do not edit manually.
4279
- * AssemblyAI API
4280
- * AssemblyAI API
4281
- * OpenAPI spec version: 1.3.4
4282
- */
4283
-
4284
- /**
4285
- * Context to provide the model. This can be a string or a free-form JSON value.
4286
- */
4287
- type LemurBaseParamsContext = string | LemurBaseParamsContextOneOf;
4288
-
4289
- /**
4290
- * Generated by orval v7.9.0 🍺
4291
- * Do not edit manually.
4292
- * AssemblyAI API
4293
- * AssemblyAI API
4294
- * OpenAPI spec version: 1.3.4
4295
- */
4296
- /**
4297
- * The model that is used for the final prompt after compression is performed.
4298
-
4299
- */
4300
- type LemurModel = (typeof LemurModel)[keyof typeof LemurModel];
4301
- declare const LemurModel: {
4302
- readonly "anthropic/claude-3-5-sonnet": "anthropic/claude-3-5-sonnet";
4303
- readonly "anthropic/claude-3-opus": "anthropic/claude-3-opus";
4304
- readonly "anthropic/claude-3-haiku": "anthropic/claude-3-haiku";
4305
- };
4306
-
4307
- /**
4308
- * Generated by orval v7.9.0 🍺
4309
- * Do not edit manually.
4310
- * AssemblyAI API
4311
- * AssemblyAI API
4312
- * OpenAPI spec version: 1.3.4
4313
- */
4314
-
4315
- /**
4316
- * The model that is used for the final prompt after compression is performed.
4317
-
4318
- */
4319
- type LemurBaseParamsFinalModel = LemurModel | string;
4320
-
4321
- /**
4322
- * Generated by orval v7.9.0 🍺
4323
- * Do not edit manually.
4324
- * AssemblyAI API
4325
- * AssemblyAI API
4326
- * OpenAPI spec version: 1.3.4
4327
- */
4328
-
4329
- interface LemurBaseParams {
4330
- /** A list of completed transcripts with text. Up to a maximum of 100 hours of audio.
4331
- Use either transcript_ids or input_text as input into LeMUR.
4332
- */
4333
- transcript_ids?: string[];
4334
- /** Custom formatted transcript data. Maximum size is the context limit of the selected model.
4335
- Use either transcript_ids or input_text as input into LeMUR.
4336
- */
4337
- input_text?: string;
4338
- /** Context to provide the model. This can be a string or a free-form JSON value. */
4339
- context?: LemurBaseParamsContext;
4340
- /** The model that is used for the final prompt after compression is performed.
4341
- */
4342
- final_model: LemurBaseParamsFinalModel;
4343
- /** Max output size in tokens. */
4344
- max_output_size?: number;
4345
- /**
4346
- * The temperature to use for the model.
4347
- Higher values result in answers that are more creative, lower values are more conservative.
4348
- Can be any value between 0.0 and 1.0 inclusive.
4349
-
4350
- * @minimum 0
4351
- * @maximum 1
4352
- */
4353
- temperature?: number;
4354
- }
4355
-
4356
- /**
4357
- * Generated by orval v7.9.0 🍺
4358
- * Do not edit manually.
4359
- * AssemblyAI API
4360
- * AssemblyAI API
4361
- * OpenAPI spec version: 1.3.4
4362
- */
4363
- type LemurActionItemsParamsAllOf = {
4364
- /** How you want the action items to be returned. This can be any text.
4365
- Defaults to "Bullet Points".
4366
- */
4367
- answer_format?: string;
4368
- };
4369
-
4370
- /**
4371
- * Generated by orval v7.9.0 🍺
4372
- * Do not edit manually.
4373
- * AssemblyAI API
4374
- * AssemblyAI API
4375
- * OpenAPI spec version: 1.3.4
4376
- */
4377
-
4378
- type LemurActionItemsParams = LemurBaseParams & LemurActionItemsParamsAllOf;
4379
-
4380
- /**
4381
- * Generated by orval v7.9.0 🍺
4382
- * Do not edit manually.
4383
- * AssemblyAI API
4384
- * AssemblyAI API
4385
- * OpenAPI spec version: 1.3.4
4386
- */
4387
- type LemurStringResponseAllOf = {
4388
- /** The response generated by LeMUR. */
4389
- response: string;
4390
- };
4391
-
4392
- /**
4393
- * Generated by orval v7.9.0 🍺
4394
- * Do not edit manually.
4395
- * AssemblyAI API
4396
- * AssemblyAI API
4397
- * OpenAPI spec version: 1.3.4
4398
- */
4399
- /**
4400
- * The usage numbers for the LeMUR request
4401
- */
4402
- interface LemurUsage {
4403
- /**
4404
- * The number of input tokens used by the model
4405
- * @minimum 0
4406
- */
4407
- input_tokens: number;
4408
- /**
4409
- * The number of output tokens generated by the model
4410
- * @minimum 0
4411
- */
4412
- output_tokens: number;
4413
- }
4414
-
4415
- /**
4416
- * Generated by orval v7.9.0 🍺
4417
- * Do not edit manually.
4418
- * AssemblyAI API
4419
- * AssemblyAI API
4420
- * OpenAPI spec version: 1.3.4
4421
- */
4422
-
4423
- interface LemurBaseResponse {
4424
- /** The ID of the LeMUR request */
4425
- request_id: string;
4426
- /** The usage numbers for the LeMUR request */
4427
- usage: LemurUsage;
4428
- }
4429
-
4430
- /**
4431
- * Generated by orval v7.9.0 🍺
4432
- * Do not edit manually.
4433
- * AssemblyAI API
4434
- * AssemblyAI API
4435
- * OpenAPI spec version: 1.3.4
4436
- */
4437
-
4438
- type LemurStringResponse = LemurStringResponseAllOf & LemurBaseResponse;
4439
-
4440
- /**
4441
- * Generated by orval v7.9.0 🍺
4442
- * Do not edit manually.
4443
- * AssemblyAI API
4444
- * AssemblyAI API
4445
- * OpenAPI spec version: 1.3.4
4446
- */
4447
-
4448
- type LemurActionItemsResponse = LemurStringResponse;
4449
-
4450
- /**
4451
- * Generated by orval v7.9.0 🍺
4452
- * Do not edit manually.
4453
- * AssemblyAI API
4454
- * AssemblyAI API
4455
- * OpenAPI spec version: 1.3.4
4456
- */
4457
- type LemurQuestionContextOneOf = {
4458
- [key: string]: unknown;
4459
- };
4460
-
4461
- /**
4462
- * Generated by orval v7.9.0 🍺
4463
- * Do not edit manually.
4464
- * AssemblyAI API
4465
- * AssemblyAI API
4466
- * OpenAPI spec version: 1.3.4
4467
- */
4468
-
4469
- /**
4470
- * Any context about the transcripts you wish to provide. This can be a string or any object.
4471
- */
4472
- type LemurQuestionContext = string | LemurQuestionContextOneOf;
4473
-
4474
- /**
4475
- * Generated by orval v7.9.0 🍺
4476
- * Do not edit manually.
4477
- * AssemblyAI API
4478
- * AssemblyAI API
4479
- * OpenAPI spec version: 1.3.4
4480
- */
4481
-
4482
- interface LemurQuestion {
4483
- /** The question you wish to ask. For more complex questions use default model. */
4484
- question: string;
4485
- /** Any context about the transcripts you wish to provide. This can be a string or any object. */
4486
- context?: LemurQuestionContext;
4487
- /** How you want the answer to be returned. This can be any text. Can't be used with answer_options. Examples: "short sentence", "bullet points"
4488
- */
4489
- answer_format?: string;
4490
- /** What discrete options to return. Useful for precise responses. Can't be used with answer_format. Example: ["Yes", "No"]
4491
- */
4492
- answer_options?: string[];
4493
- }
4494
-
4495
- /**
4496
- * Generated by orval v7.9.0 🍺
4497
- * Do not edit manually.
4498
- * AssemblyAI API
4499
- * AssemblyAI API
4500
- * OpenAPI spec version: 1.3.4
4501
- */
4502
- /**
4503
- * An answer generated by LeMUR and its question
4504
- */
4505
- interface LemurQuestionAnswer {
4506
- /** The question for LeMUR to answer */
4507
- question: string;
4508
- /** The answer generated by LeMUR */
4509
- answer: string;
4510
- }
4511
-
4512
- /**
4513
- * Generated by orval v7.9.0 🍺
4514
- * Do not edit manually.
4515
- * AssemblyAI API
4516
- * AssemblyAI API
4517
- * OpenAPI spec version: 1.3.4
4518
- */
4519
-
4520
- type LemurQuestionAnswerParamsAllOf = {
4521
- /** A list of questions to ask */
4522
- questions: LemurQuestion[];
4523
- };
4524
-
4525
- /**
4526
- * Generated by orval v7.9.0 🍺
4527
- * Do not edit manually.
4528
- * AssemblyAI API
4529
- * AssemblyAI API
4530
- * OpenAPI spec version: 1.3.4
4531
- */
4532
-
4533
- type LemurQuestionAnswerParams = LemurBaseParams & LemurQuestionAnswerParamsAllOf;
4534
-
4535
- /**
4536
- * Generated by orval v7.9.0 🍺
4537
- * Do not edit manually.
4538
- * AssemblyAI API
4539
- * AssemblyAI API
4540
- * OpenAPI spec version: 1.3.4
4118
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4119
+ * OpenAPI spec version: 1.0.0
4541
4120
  */
4542
-
4543
- type LemurQuestionAnswerResponseAllOf = {
4544
- /** The answers generated by LeMUR and their questions */
4545
- response: LemurQuestionAnswer[];
4546
- };
4547
-
4548
4121
  /**
4549
- * Generated by orval v7.9.0 🍺
4550
- * Do not edit manually.
4551
- * AssemblyAI API
4552
- * AssemblyAI API
4553
- * OpenAPI spec version: 1.3.4
4122
+ * Error message of why the transcript failed
4554
4123
  */
4555
-
4556
- type LemurQuestionAnswerResponse = LemurBaseResponse & LemurQuestionAnswerResponseAllOf;
4124
+ type TranscriptListItemError = string | null;
4557
4125
 
4558
4126
  /**
4559
4127
  * Generated by orval v7.9.0 🍺
4560
4128
  * Do not edit manually.
4561
4129
  * AssemblyAI API
4562
- * AssemblyAI API
4563
- * OpenAPI spec version: 1.3.4
4130
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4131
+ * OpenAPI spec version: 1.0.0
4564
4132
  */
4565
4133
 
4566
- type LemurResponse = LemurStringResponse | LemurQuestionAnswerResponse;
4134
+ interface TranscriptListItem {
4135
+ /** The unique identifier for the transcript */
4136
+ id: string;
4137
+ /** The URL to retrieve the transcript */
4138
+ resource_url: string;
4139
+ /** The status of the transcript */
4140
+ status: TranscriptStatus;
4141
+ /** The date and time the transcript was created */
4142
+ created: string;
4143
+ /** The date and time the transcript was completed */
4144
+ completed?: string;
4145
+ /** The URL to the audio file */
4146
+ audio_url: string;
4147
+ /** Error message of why the transcript failed */
4148
+ error: TranscriptListItemError;
4149
+ }
4567
4150
 
4568
4151
  /**
4569
4152
  * Generated by orval v7.9.0 🍺
4570
4153
  * Do not edit manually.
4571
4154
  * AssemblyAI API
4572
- * AssemblyAI API
4573
- * OpenAPI spec version: 1.3.4
4155
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4156
+ * OpenAPI spec version: 1.0.0
4574
4157
  */
4575
- type LemurSummaryParamsAllOf = {
4576
- /** How you want the summary to be returned. This can be any text. Examples: "TLDR", "bullet points"
4577
- */
4578
- answer_format?: string;
4579
- };
4580
-
4581
4158
  /**
4582
- * Generated by orval v7.9.0 🍺
4583
- * Do not edit manually.
4584
- * AssemblyAI API
4585
- * AssemblyAI API
4586
- * OpenAPI spec version: 1.3.4
4159
+ * Additional error details if available
4587
4160
  */
4588
-
4589
- type LemurSummaryParams = LemurBaseParams & LemurSummaryParamsAllOf;
4161
+ interface ErrorDetails {
4162
+ [key: string]: unknown;
4163
+ }
4590
4164
 
4591
4165
  /**
4592
4166
  * Generated by orval v7.9.0 🍺
4593
4167
  * Do not edit manually.
4594
4168
  * AssemblyAI API
4595
- * AssemblyAI API
4596
- * OpenAPI spec version: 1.3.4
4169
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4170
+ * OpenAPI spec version: 1.0.0
4597
4171
  */
4598
4172
 
4599
- type LemurSummaryResponse = LemurStringResponse;
4173
+ interface Error$2 {
4174
+ /** Error message describing what went wrong */
4175
+ error: string;
4176
+ /** Error code for programmatic handling */
4177
+ code?: string;
4178
+ /** Additional error details if available */
4179
+ details?: ErrorDetails;
4180
+ }
4600
4181
 
4601
4182
  /**
4602
4183
  * Generated by orval v7.9.0 🍺
4603
4184
  * Do not edit manually.
4604
4185
  * AssemblyAI API
4605
- * AssemblyAI API
4606
- * OpenAPI spec version: 1.3.4
4186
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4187
+ * OpenAPI spec version: 1.0.0
4607
4188
  */
4608
- type LemurTaskParamsAllOf = {
4609
- /** Your text to prompt the model to produce a desired output, including any context you want to pass into the model. */
4610
- prompt: string;
4189
+ type GetSubtitles200 = {
4190
+ [key: string]: unknown;
4611
4191
  };
4612
4192
 
4613
4193
  /**
4614
4194
  * Generated by orval v7.9.0 🍺
4615
4195
  * Do not edit manually.
4616
4196
  * AssemblyAI API
4617
- * AssemblyAI API
4618
- * OpenAPI spec version: 1.3.4
4619
- */
4620
-
4621
- type LemurTaskParams = LemurTaskParamsAllOf & LemurBaseParams;
4622
-
4623
- /**
4624
- * Generated by orval v7.9.0 🍺
4625
- * Do not edit manually.
4626
- * AssemblyAI API
4627
- * AssemblyAI API
4628
- * OpenAPI spec version: 1.3.4
4629
- */
4630
-
4631
- type LemurTaskResponse = LemurStringResponse;
4632
-
4633
- /**
4634
- * Generated by orval v7.9.0 🍺
4635
- * Do not edit manually.
4636
- * AssemblyAI API
4637
- * AssemblyAI API
4638
- * OpenAPI spec version: 1.3.4
4197
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4198
+ * OpenAPI spec version: 1.0.0
4639
4199
  */
4640
-
4641
- interface ListTranscriptParams {
4200
+ type GetSubtitlesParams = {
4642
4201
  /**
4643
- * Maximum amount of transcripts to retrieve
4644
- * @minimum 1
4645
- * @maximum 200
4202
+ * The maximum number of characters per caption
4646
4203
  */
4647
- limit?: number;
4648
- /** Filter by transcript status */
4649
- status?: TranscriptStatus;
4650
- /** Only get transcripts created on this date */
4651
- created_on?: string;
4652
- /** Get transcripts that were created before this transcript ID */
4653
- before_id?: string;
4654
- /** Get transcripts that were created after this transcript ID */
4655
- after_id?: string;
4656
- /** Only get throttled transcripts, overrides the status filter */
4657
- throttled_only?: boolean;
4658
- }
4659
-
4660
- /**
4661
- * Not found
4662
- */
4663
- type NotFoundResponse = Error;
4204
+ chars_per_caption?: number;
4205
+ };
4664
4206
 
4665
4207
  /**
4666
4208
  * Generated by orval v7.9.0 🍺
4667
4209
  * Do not edit manually.
4668
4210
  * AssemblyAI API
4669
- * AssemblyAI API
4670
- * OpenAPI spec version: 1.3.4
4211
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4212
+ * OpenAPI spec version: 1.0.0
4671
4213
  */
4672
4214
  /**
4673
4215
  * The URL to the next page of transcripts. The previous URL always points to a page with older transcripts.
@@ -4678,8 +4220,8 @@ type PageDetailsPrevUrl = string | null;
4678
4220
  * Generated by orval v7.9.0 🍺
4679
4221
  * Do not edit manually.
4680
4222
  * AssemblyAI API
4681
- * AssemblyAI API
4682
- * OpenAPI spec version: 1.3.4
4223
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4224
+ * OpenAPI spec version: 1.0.0
4683
4225
  */
4684
4226
  /**
4685
4227
  * The URL to the next page of transcripts. The next URL always points to a page with newer transcripts.
@@ -4690,8 +4232,8 @@ type PageDetailsNextUrl = string | null;
4690
4232
  * Generated by orval v7.9.0 🍺
4691
4233
  * Do not edit manually.
4692
4234
  * AssemblyAI API
4693
- * AssemblyAI API
4694
- * OpenAPI spec version: 1.3.4
4235
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4236
+ * OpenAPI spec version: 1.0.0
4695
4237
  */
4696
4238
 
4697
4239
  /**
@@ -4714,8 +4256,8 @@ interface PageDetails {
4714
4256
  * Generated by orval v7.9.0 🍺
4715
4257
  * Do not edit manually.
4716
4258
  * AssemblyAI API
4717
- * AssemblyAI API
4718
- * OpenAPI spec version: 1.3.4
4259
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4260
+ * OpenAPI spec version: 1.0.0
4719
4261
  */
4720
4262
 
4721
4263
  interface TranscriptParagraph {
@@ -4725,11 +4267,7 @@ interface TranscriptParagraph {
4725
4267
  start: number;
4726
4268
  /** The ending time, in milliseconds, of the paragraph */
4727
4269
  end: number;
4728
- /**
4729
- * The confidence score for the transcript of this paragraph
4730
- * @minimum 0
4731
- * @maximum 1
4732
- */
4270
+ /** The confidence score for the transcript of this paragraph */
4733
4271
  confidence: number;
4734
4272
  /** An array of words in the paragraph */
4735
4273
  words: TranscriptWord[];
@@ -4739,18 +4277,14 @@ interface TranscriptParagraph {
4739
4277
  * Generated by orval v7.9.0 🍺
4740
4278
  * Do not edit manually.
4741
4279
  * AssemblyAI API
4742
- * AssemblyAI API
4743
- * OpenAPI spec version: 1.3.4
4280
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4281
+ * OpenAPI spec version: 1.0.0
4744
4282
  */
4745
4283
 
4746
4284
  interface ParagraphsResponse {
4747
4285
  /** The unique identifier of your transcript */
4748
4286
  id: string;
4749
- /**
4750
- * The confidence score for the transcript
4751
- * @minimum 0
4752
- * @maximum 1
4753
- */
4287
+ /** The confidence score for the transcript */
4754
4288
  confidence: number;
4755
4289
  /** The duration of the audio file in seconds */
4756
4290
  audio_duration: number;
@@ -4762,36 +4296,38 @@ interface ParagraphsResponse {
4762
4296
  * Generated by orval v7.9.0 🍺
4763
4297
  * Do not edit manually.
4764
4298
  * AssemblyAI API
4765
- * AssemblyAI API
4766
- * OpenAPI spec version: 1.3.4
4299
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4300
+ * OpenAPI spec version: 1.0.0
4767
4301
  */
4768
- interface PurgeLemurRequestDataResponse {
4769
- /** The ID of the deletion request of the LeMUR request */
4770
- request_id: string;
4771
- /** The ID of the LeMUR request to purge the data for */
4772
- request_id_to_purge: string;
4773
- /** Whether the request data was deleted */
4774
- deleted: boolean;
4775
- }
4302
+ /**
4303
+ * The status of the redacted audio
4304
+ */
4305
+ type RedactedAudioStatus = (typeof RedactedAudioStatus)[keyof typeof RedactedAudioStatus];
4306
+ declare const RedactedAudioStatus: {
4307
+ readonly redacted_audio_ready: "redacted_audio_ready";
4308
+ };
4776
4309
 
4777
4310
  /**
4778
4311
  * Generated by orval v7.9.0 🍺
4779
4312
  * Do not edit manually.
4780
4313
  * AssemblyAI API
4781
- * AssemblyAI API
4782
- * OpenAPI spec version: 1.3.4
4314
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4315
+ * OpenAPI spec version: 1.0.0
4783
4316
  */
4784
- interface RealtimeTemporaryTokenResponse {
4785
- /** The temporary authentication token for Streaming Speech-to-Text */
4786
- token: string;
4317
+
4318
+ interface RedactedAudioResponse {
4319
+ /** The status of the redacted audio */
4320
+ status: RedactedAudioStatus;
4321
+ /** The URL of the redacted audio file */
4322
+ redacted_audio_url: string;
4787
4323
  }
4788
4324
 
4789
4325
  /**
4790
4326
  * Generated by orval v7.9.0 🍺
4791
4327
  * Do not edit manually.
4792
4328
  * AssemblyAI API
4793
- * AssemblyAI API
4794
- * OpenAPI spec version: 1.3.4
4329
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4330
+ * OpenAPI spec version: 1.0.0
4795
4331
  */
4796
4332
  /**
4797
4333
  * The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
@@ -4802,11 +4338,11 @@ type TranscriptSentenceChannel = string | null;
4802
4338
  * Generated by orval v7.9.0 🍺
4803
4339
  * Do not edit manually.
4804
4340
  * AssemblyAI API
4805
- * AssemblyAI API
4806
- * OpenAPI spec version: 1.3.4
4341
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4342
+ * OpenAPI spec version: 1.0.0
4807
4343
  */
4808
4344
  /**
4809
- * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null
4345
+ * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
4810
4346
  */
4811
4347
  type TranscriptSentenceSpeaker = string | null;
4812
4348
 
@@ -4814,8 +4350,8 @@ type TranscriptSentenceSpeaker = string | null;
4814
4350
  * Generated by orval v7.9.0 🍺
4815
4351
  * Do not edit manually.
4816
4352
  * AssemblyAI API
4817
- * AssemblyAI API
4818
- * OpenAPI spec version: 1.3.4
4353
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4354
+ * OpenAPI spec version: 1.0.0
4819
4355
  */
4820
4356
 
4821
4357
  interface TranscriptSentence {
@@ -4825,17 +4361,13 @@ interface TranscriptSentence {
4825
4361
  start: number;
4826
4362
  /** The ending time, in milliseconds, for the sentence */
4827
4363
  end: number;
4828
- /**
4829
- * The confidence score for the transcript of this sentence
4830
- * @minimum 0
4831
- * @maximum 1
4832
- */
4364
+ /** The confidence score for the transcript of this sentence */
4833
4365
  confidence: number;
4834
4366
  /** An array of words in the sentence */
4835
4367
  words: TranscriptWord[];
4836
4368
  /** The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
4837
4369
  channel?: TranscriptSentenceChannel;
4838
- /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null */
4370
+ /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
4839
4371
  speaker: TranscriptSentenceSpeaker;
4840
4372
  }
4841
4373
 
@@ -4843,18 +4375,14 @@ interface TranscriptSentence {
4843
4375
  * Generated by orval v7.9.0 🍺
4844
4376
  * Do not edit manually.
4845
4377
  * AssemblyAI API
4846
- * AssemblyAI API
4847
- * OpenAPI spec version: 1.3.4
4378
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4379
+ * OpenAPI spec version: 1.0.0
4848
4380
  */
4849
4381
 
4850
4382
  interface SentencesResponse {
4851
4383
  /** The unique identifier for the transcript */
4852
4384
  id: string;
4853
- /**
4854
- * The confidence score for the transcript
4855
- * @minimum 0
4856
- * @maximum 1
4857
- */
4385
+ /** The confidence score for the transcript */
4858
4386
  confidence: number;
4859
4387
  /** The duration of the audio file in seconds */
4860
4388
  audio_duration: number;
@@ -4866,20 +4394,8 @@ interface SentencesResponse {
4866
4394
  * Generated by orval v7.9.0 🍺
4867
4395
  * Do not edit manually.
4868
4396
  * AssemblyAI API
4869
- * AssemblyAI API
4870
- * OpenAPI spec version: 1.3.4
4871
- */
4872
- /**
4873
- * Service unavailable
4874
- */
4875
- type ServiceUnavailableResponse = void;
4876
-
4877
- /**
4878
- * Generated by orval v7.9.0 🍺
4879
- * Do not edit manually.
4880
- * AssemblyAI API
4881
- * AssemblyAI API
4882
- * OpenAPI spec version: 1.3.4
4397
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4398
+ * OpenAPI spec version: 1.0.0
4883
4399
  */
4884
4400
  /**
4885
4401
  * Format of the subtitles
@@ -4890,17 +4406,12 @@ declare const SubtitleFormat: {
4890
4406
  readonly vtt: "vtt";
4891
4407
  };
4892
4408
 
4893
- /**
4894
- * Too many requests
4895
- */
4896
- type TooManyRequestsResponse = Error;
4897
-
4898
4409
  /**
4899
4410
  * Generated by orval v7.9.0 🍺
4900
4411
  * Do not edit manually.
4901
4412
  * AssemblyAI API
4902
- * AssemblyAI API
4903
- * OpenAPI spec version: 1.3.4
4413
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4414
+ * OpenAPI spec version: 1.0.0
4904
4415
  */
4905
4416
 
4906
4417
  /**
@@ -4913,17 +4424,12 @@ interface TranscriptList {
4913
4424
  transcripts: TranscriptListItem[];
4914
4425
  }
4915
4426
 
4916
- /**
4917
- * Unauthorized
4918
- */
4919
- type UnauthorizedResponse = Error;
4920
-
4921
4427
  /**
4922
4428
  * Generated by orval v7.9.0 🍺
4923
4429
  * Do not edit manually.
4924
4430
  * AssemblyAI API
4925
- * AssemblyAI API
4926
- * OpenAPI spec version: 1.3.4
4431
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4432
+ * OpenAPI spec version: 1.0.0
4927
4433
  */
4928
4434
  interface UploadedFile {
4929
4435
  /** A URL that points to your audio file, accessible only by AssemblyAI's servers
@@ -4935,8 +4441,8 @@ interface UploadedFile {
4935
4441
  * Generated by orval v7.9.0 🍺
4936
4442
  * Do not edit manually.
4937
4443
  * AssemblyAI API
4938
- * AssemblyAI API
4939
- * OpenAPI spec version: 1.3.4
4444
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4445
+ * OpenAPI spec version: 1.0.0
4940
4446
  */
4941
4447
  /**
4942
4448
  * An array of timestamps structured as [`start_time`, `end_time`] in milliseconds
@@ -4947,8 +4453,8 @@ type WordSearchTimestamp = number[];
4947
4453
  * Generated by orval v7.9.0 🍺
4948
4454
  * Do not edit manually.
4949
4455
  * AssemblyAI API
4950
- * AssemblyAI API
4951
- * OpenAPI spec version: 1.3.4
4456
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4457
+ * OpenAPI spec version: 1.0.0
4952
4458
  */
4953
4459
 
4954
4460
  interface WordSearchMatch {
@@ -4966,8 +4472,8 @@ interface WordSearchMatch {
4966
4472
  * Generated by orval v7.9.0 🍺
4967
4473
  * Do not edit manually.
4968
4474
  * AssemblyAI API
4969
- * AssemblyAI API
4970
- * OpenAPI spec version: 1.3.4
4475
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4476
+ * OpenAPI spec version: 1.0.0
4971
4477
  */
4972
4478
  type WordSearchParams = {
4973
4479
  /**
@@ -4980,8 +4486,8 @@ type WordSearchParams = {
4980
4486
  * Generated by orval v7.9.0 🍺
4981
4487
  * Do not edit manually.
4982
4488
  * AssemblyAI API
4983
- * AssemblyAI API
4984
- * OpenAPI spec version: 1.3.4
4489
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4490
+ * OpenAPI spec version: 1.0.0
4985
4491
  */
4986
4492
 
4987
4493
  interface WordSearchResponse {
@@ -4997,93 +4503,66 @@ interface WordSearchResponse {
4997
4503
  * Generated by orval v7.9.0 🍺
4998
4504
  * Do not edit manually.
4999
4505
  * AssemblyAI API
5000
- * AssemblyAI API
5001
- * OpenAPI spec version: 1.3.4
4506
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4507
+ * OpenAPI spec version: 1.0.0
5002
4508
  */
5003
4509
 
5004
4510
  declare const index$6_AudioIntelligenceModelStatus: typeof AudioIntelligenceModelStatus;
5005
4511
  declare const index$6_AutoHighlightResult: typeof AutoHighlightResult;
5006
4512
  declare const index$6_AutoHighlightsResult: typeof AutoHighlightsResult;
5007
- type index$6_BadRequestResponse = BadRequestResponse;
5008
- type index$6_CannotAccessUploadedFileResponse = CannotAccessUploadedFileResponse;
5009
4513
  declare const index$6_Chapter: typeof Chapter;
5010
4514
  declare const index$6_ContentSafetyLabel: typeof ContentSafetyLabel;
5011
4515
  declare const index$6_ContentSafetyLabelResult: typeof ContentSafetyLabelResult;
5012
4516
  declare const index$6_ContentSafetyLabelsResult: typeof ContentSafetyLabelsResult;
5013
4517
  declare const index$6_ContentSafetyLabelsResultSeverityScoreSummary: typeof ContentSafetyLabelsResultSeverityScoreSummary;
5014
4518
  declare const index$6_ContentSafetyLabelsResultSummary: typeof ContentSafetyLabelsResultSummary;
5015
- type index$6_CreateRealtimeTemporaryTokenParams = CreateRealtimeTemporaryTokenParams;
4519
+ declare const index$6_CustomFormattingRequestBody: typeof CustomFormattingRequestBody;
4520
+ declare const index$6_CustomFormattingRequestBodyCustomFormatting: typeof CustomFormattingRequestBodyCustomFormatting;
4521
+ declare const index$6_CustomFormattingResponse: typeof CustomFormattingResponse;
4522
+ declare const index$6_CustomFormattingResponseCustomFormatting: typeof CustomFormattingResponseCustomFormatting;
4523
+ declare const index$6_CustomFormattingResponseCustomFormattingMapping: typeof CustomFormattingResponseCustomFormattingMapping;
5016
4524
  declare const index$6_Entity: typeof Entity;
5017
4525
  declare const index$6_EntityType: typeof EntityType;
5018
- type index$6_GatewayTimeoutResponse = GatewayTimeoutResponse;
4526
+ type index$6_ErrorDetails = ErrorDetails;
4527
+ type index$6_GetSubtitles200 = GetSubtitles200;
5019
4528
  type index$6_GetSubtitlesParams = GetSubtitlesParams;
5020
- type index$6_InternalServerErrorResponse = InternalServerErrorResponse;
5021
- type index$6_LemurActionItemsParams = LemurActionItemsParams;
5022
- type index$6_LemurActionItemsParamsAllOf = LemurActionItemsParamsAllOf;
5023
- type index$6_LemurActionItemsResponse = LemurActionItemsResponse;
5024
- type index$6_LemurBaseParams = LemurBaseParams;
5025
- type index$6_LemurBaseParamsContext = LemurBaseParamsContext;
5026
- type index$6_LemurBaseParamsContextOneOf = LemurBaseParamsContextOneOf;
5027
- type index$6_LemurBaseParamsFinalModel = LemurBaseParamsFinalModel;
5028
- type index$6_LemurBaseResponse = LemurBaseResponse;
5029
- declare const index$6_LemurModel: typeof LemurModel;
5030
- type index$6_LemurQuestion = LemurQuestion;
5031
- type index$6_LemurQuestionAnswer = LemurQuestionAnswer;
5032
- type index$6_LemurQuestionAnswerParams = LemurQuestionAnswerParams;
5033
- type index$6_LemurQuestionAnswerParamsAllOf = LemurQuestionAnswerParamsAllOf;
5034
- type index$6_LemurQuestionAnswerResponse = LemurQuestionAnswerResponse;
5035
- type index$6_LemurQuestionAnswerResponseAllOf = LemurQuestionAnswerResponseAllOf;
5036
- type index$6_LemurQuestionContext = LemurQuestionContext;
5037
- type index$6_LemurQuestionContextOneOf = LemurQuestionContextOneOf;
5038
- type index$6_LemurResponse = LemurResponse;
5039
- type index$6_LemurStringResponse = LemurStringResponse;
5040
- type index$6_LemurStringResponseAllOf = LemurStringResponseAllOf;
5041
- type index$6_LemurSummaryParams = LemurSummaryParams;
5042
- type index$6_LemurSummaryParamsAllOf = LemurSummaryParamsAllOf;
5043
- type index$6_LemurSummaryResponse = LemurSummaryResponse;
5044
- type index$6_LemurTaskParams = LemurTaskParams;
5045
- type index$6_LemurTaskParamsAllOf = LemurTaskParamsAllOf;
5046
- type index$6_LemurTaskResponse = LemurTaskResponse;
5047
- type index$6_LemurUsage = LemurUsage;
5048
- type index$6_ListTranscriptParams = ListTranscriptParams;
5049
4529
  declare const index$6_ListTranscriptsParams: typeof ListTranscriptsParams;
5050
- type index$6_NotFoundResponse = NotFoundResponse;
5051
4530
  type index$6_PageDetails = PageDetails;
5052
4531
  type index$6_PageDetailsNextUrl = PageDetailsNextUrl;
5053
4532
  type index$6_PageDetailsPrevUrl = PageDetailsPrevUrl;
5054
4533
  type index$6_ParagraphsResponse = ParagraphsResponse;
5055
4534
  declare const index$6_PiiPolicy: typeof PiiPolicy;
5056
- type index$6_PurgeLemurRequestDataResponse = PurgeLemurRequestDataResponse;
5057
- type index$6_RealtimeTemporaryTokenResponse = RealtimeTemporaryTokenResponse;
5058
4535
  declare const index$6_RedactPiiAudioQuality: typeof RedactPiiAudioQuality;
5059
- declare const index$6_RedactedAudioNotification: typeof RedactedAudioNotification;
5060
- declare const index$6_RedactedAudioResponse: typeof RedactedAudioResponse;
4536
+ type index$6_RedactedAudioResponse = RedactedAudioResponse;
5061
4537
  declare const index$6_RedactedAudioStatus: typeof RedactedAudioStatus;
5062
4538
  type index$6_SentencesResponse = SentencesResponse;
5063
4539
  declare const index$6_Sentiment: typeof Sentiment;
5064
4540
  declare const index$6_SentimentAnalysisResultChannel: typeof SentimentAnalysisResultChannel;
5065
4541
  declare const index$6_SentimentAnalysisResultSpeaker: typeof SentimentAnalysisResultSpeaker;
5066
- type index$6_ServiceUnavailableResponse = ServiceUnavailableResponse;
5067
4542
  declare const index$6_SeverityScoreSummary: typeof SeverityScoreSummary;
4543
+ declare const index$6_SpeakerIdentificationRequestBody: typeof SpeakerIdentificationRequestBody;
4544
+ declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentification: typeof SpeakerIdentificationRequestBodySpeakerIdentification;
4545
+ declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType;
4546
+ declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems;
4547
+ declare const index$6_SpeakerIdentificationResponse: typeof SpeakerIdentificationResponse;
4548
+ declare const index$6_SpeakerIdentificationResponseSpeakerIdentification: typeof SpeakerIdentificationResponseSpeakerIdentification;
4549
+ declare const index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping: typeof SpeakerIdentificationResponseSpeakerIdentificationMapping;
5068
4550
  declare const index$6_SpeechModel: typeof SpeechModel;
5069
4551
  declare const index$6_SubstitutionPolicy: typeof SubstitutionPolicy;
5070
4552
  declare const index$6_SubtitleFormat: typeof SubtitleFormat;
5071
4553
  declare const index$6_SummaryModel: typeof SummaryModel;
5072
4554
  declare const index$6_SummaryType: typeof SummaryType;
5073
4555
  declare const index$6_Timestamp: typeof Timestamp;
5074
- type index$6_TooManyRequestsResponse = TooManyRequestsResponse;
5075
4556
  declare const index$6_TopicDetectionModelResult: typeof TopicDetectionModelResult;
5076
4557
  declare const index$6_TopicDetectionModelResultSummary: typeof TopicDetectionModelResultSummary;
5077
4558
  declare const index$6_TopicDetectionResult: typeof TopicDetectionResult;
5078
- declare const index$6_TopicDetectionResultLabelsItem: typeof TopicDetectionResultLabelsItem;
4559
+ declare const index$6_TopicDetectionResultLabelsItems: typeof TopicDetectionResultLabelsItems;
5079
4560
  declare const index$6_Transcript: typeof Transcript;
5080
4561
  declare const index$6_TranscriptAudioDuration: typeof TranscriptAudioDuration;
5081
4562
  declare const index$6_TranscriptAudioEndAt: typeof TranscriptAudioEndAt;
5082
4563
  declare const index$6_TranscriptAudioStartFrom: typeof TranscriptAudioStartFrom;
5083
4564
  declare const index$6_TranscriptAutoChapters: typeof TranscriptAutoChapters;
5084
4565
  declare const index$6_TranscriptAutoHighlightsResult: typeof TranscriptAutoHighlightsResult;
5085
- declare const index$6_TranscriptBoostParam: typeof TranscriptBoostParam;
5086
- declare const index$6_TranscriptBoostParamProperty: typeof TranscriptBoostParamProperty;
5087
4566
  declare const index$6_TranscriptChapters: typeof TranscriptChapters;
5088
4567
  declare const index$6_TranscriptConfidence: typeof TranscriptConfidence;
5089
4568
  declare const index$6_TranscriptContentSafety: typeof TranscriptContentSafety;
@@ -5092,6 +4571,7 @@ declare const index$6_TranscriptCustomSpelling: typeof TranscriptCustomSpelling;
5092
4571
  declare const index$6_TranscriptCustomSpellingProperty: typeof TranscriptCustomSpellingProperty;
5093
4572
  declare const index$6_TranscriptCustomTopics: typeof TranscriptCustomTopics;
5094
4573
  declare const index$6_TranscriptDisfluencies: typeof TranscriptDisfluencies;
4574
+ declare const index$6_TranscriptDomain: typeof TranscriptDomain;
5095
4575
  declare const index$6_TranscriptEntities: typeof TranscriptEntities;
5096
4576
  declare const index$6_TranscriptEntityDetection: typeof TranscriptEntityDetection;
5097
4577
  declare const index$6_TranscriptFilterProfanity: typeof TranscriptFilterProfanity;
@@ -5099,33 +4579,42 @@ declare const index$6_TranscriptFormatText: typeof TranscriptFormatText;
5099
4579
  declare const index$6_TranscriptIabCategories: typeof TranscriptIabCategories;
5100
4580
  declare const index$6_TranscriptIabCategoriesResult: typeof TranscriptIabCategoriesResult;
5101
4581
  declare const index$6_TranscriptLanguageCode: typeof TranscriptLanguageCode;
5102
- declare const index$6_TranscriptLanguageCodeProperty: typeof TranscriptLanguageCodeProperty;
4582
+ declare const index$6_TranscriptLanguageCodes: typeof TranscriptLanguageCodes;
5103
4583
  declare const index$6_TranscriptLanguageConfidence: typeof TranscriptLanguageConfidence;
5104
4584
  declare const index$6_TranscriptLanguageConfidenceThreshold: typeof TranscriptLanguageConfidenceThreshold;
5105
4585
  declare const index$6_TranscriptLanguageDetection: typeof TranscriptLanguageDetection;
4586
+ declare const index$6_TranscriptLanguageDetectionOptions: typeof TranscriptLanguageDetectionOptions;
5106
4587
  type index$6_TranscriptList = TranscriptList;
5107
4588
  type index$6_TranscriptListItem = TranscriptListItem;
5108
- type index$6_TranscriptListItemCompleted = TranscriptListItemCompleted;
5109
4589
  type index$6_TranscriptListItemError = TranscriptListItemError;
5110
4590
  declare const index$6_TranscriptMultichannel: typeof TranscriptMultichannel;
5111
- declare const index$6_TranscriptOptionalParams: typeof TranscriptOptionalParams;
5112
- declare const index$6_TranscriptOptionalParamsLanguageCode: typeof TranscriptOptionalParamsLanguageCode;
5113
- declare const index$6_TranscriptOptionalParamsLanguageCodeOneOf: typeof TranscriptOptionalParamsLanguageCodeOneOf;
5114
- declare const index$6_TranscriptOptionalParamsRedactPiiSub: typeof TranscriptOptionalParamsRedactPiiSub;
5115
- declare const index$6_TranscriptOptionalParamsSpeakersExpected: typeof TranscriptOptionalParamsSpeakersExpected;
5116
- declare const index$6_TranscriptOptionalParamsSpeechModel: typeof TranscriptOptionalParamsSpeechModel;
5117
- declare const index$6_TranscriptOptionalParamsSpeechThreshold: typeof TranscriptOptionalParamsSpeechThreshold;
5118
- declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderName: typeof TranscriptOptionalParamsWebhookAuthHeaderName;
5119
- declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderValue: typeof TranscriptOptionalParamsWebhookAuthHeaderValue;
4591
+ declare const index$6_TranscriptOptionalParamsLanguageDetectionOptions: typeof TranscriptOptionalParamsLanguageDetectionOptions;
4592
+ declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptions: typeof TranscriptOptionalParamsRedactPiiAudioOptions;
4593
+ declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod;
4594
+ declare const index$6_TranscriptOptionalParamsRemoveAudioTags: typeof TranscriptOptionalParamsRemoveAudioTags;
4595
+ declare const index$6_TranscriptOptionalParamsSpeakerOptions: typeof TranscriptOptionalParamsSpeakerOptions;
4596
+ declare const index$6_TranscriptOptionalParamsSpeechUnderstanding: typeof TranscriptOptionalParamsSpeechUnderstanding;
4597
+ declare const index$6_TranscriptOptionalParamsSpeechUnderstandingRequest: typeof TranscriptOptionalParamsSpeechUnderstandingRequest;
5120
4598
  type index$6_TranscriptParagraph = TranscriptParagraph;
5121
- type index$6_TranscriptParams = TranscriptParams;
5122
- type index$6_TranscriptParamsAllOf = TranscriptParamsAllOf;
4599
+ declare const index$6_TranscriptParams: typeof TranscriptParams;
4600
+ declare const index$6_TranscriptParamsDomain: typeof TranscriptParamsDomain;
4601
+ declare const index$6_TranscriptParamsLanguageCode: typeof TranscriptParamsLanguageCode;
4602
+ declare const index$6_TranscriptParamsLanguageCodes: typeof TranscriptParamsLanguageCodes;
4603
+ declare const index$6_TranscriptParamsRedactPiiSub: typeof TranscriptParamsRedactPiiSub;
4604
+ declare const index$6_TranscriptParamsRemoveAudioTags: typeof TranscriptParamsRemoveAudioTags;
4605
+ declare const index$6_TranscriptParamsSpeakersExpected: typeof TranscriptParamsSpeakersExpected;
4606
+ declare const index$6_TranscriptParamsSpeechModel: typeof TranscriptParamsSpeechModel;
4607
+ declare const index$6_TranscriptParamsSpeechThreshold: typeof TranscriptParamsSpeechThreshold;
4608
+ declare const index$6_TranscriptParamsWebhookAuthHeaderName: typeof TranscriptParamsWebhookAuthHeaderName;
4609
+ declare const index$6_TranscriptParamsWebhookAuthHeaderValue: typeof TranscriptParamsWebhookAuthHeaderValue;
5123
4610
  declare const index$6_TranscriptPunctuate: typeof TranscriptPunctuate;
5124
- declare const index$6_TranscriptReadyNotification: typeof TranscriptReadyNotification;
5125
- declare const index$6_TranscriptReadyStatus: typeof TranscriptReadyStatus;
5126
4611
  declare const index$6_TranscriptRedactPiiAudio: typeof TranscriptRedactPiiAudio;
4612
+ declare const index$6_TranscriptRedactPiiAudioOptions: typeof TranscriptRedactPiiAudioOptions;
4613
+ declare const index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod;
5127
4614
  declare const index$6_TranscriptRedactPiiAudioQuality: typeof TranscriptRedactPiiAudioQuality;
5128
4615
  declare const index$6_TranscriptRedactPiiPolicies: typeof TranscriptRedactPiiPolicies;
4616
+ declare const index$6_TranscriptRemoveAudioTags: typeof TranscriptRemoveAudioTags;
4617
+ declare const index$6_TranscriptRemoveAudioTagsProperty: typeof TranscriptRemoveAudioTagsProperty;
5129
4618
  type index$6_TranscriptSentence = TranscriptSentence;
5130
4619
  type index$6_TranscriptSentenceChannel = TranscriptSentenceChannel;
5131
4620
  type index$6_TranscriptSentenceSpeaker = TranscriptSentenceSpeaker;
@@ -5134,33 +4623,42 @@ declare const index$6_TranscriptSentimentAnalysisResults: typeof TranscriptSenti
5134
4623
  declare const index$6_TranscriptSpeakerLabels: typeof TranscriptSpeakerLabels;
5135
4624
  declare const index$6_TranscriptSpeakersExpected: typeof TranscriptSpeakersExpected;
5136
4625
  declare const index$6_TranscriptSpeechModel: typeof TranscriptSpeechModel;
4626
+ declare const index$6_TranscriptSpeechModels: typeof TranscriptSpeechModels;
5137
4627
  declare const index$6_TranscriptSpeechThreshold: typeof TranscriptSpeechThreshold;
4628
+ declare const index$6_TranscriptSpeechUnderstanding: typeof TranscriptSpeechUnderstanding;
4629
+ declare const index$6_TranscriptSpeechUnderstandingRequest: typeof TranscriptSpeechUnderstandingRequest;
4630
+ declare const index$6_TranscriptSpeechUnderstandingResponse: typeof TranscriptSpeechUnderstandingResponse;
5138
4631
  declare const index$6_TranscriptSpeedBoost: typeof TranscriptSpeedBoost;
5139
4632
  declare const index$6_TranscriptStatus: typeof TranscriptStatus;
5140
4633
  declare const index$6_TranscriptSummary: typeof TranscriptSummary;
5141
4634
  declare const index$6_TranscriptSummaryModel: typeof TranscriptSummaryModel;
5142
4635
  declare const index$6_TranscriptSummaryType: typeof TranscriptSummaryType;
4636
+ declare const index$6_TranscriptTemperature: typeof TranscriptTemperature;
5143
4637
  declare const index$6_TranscriptText: typeof TranscriptText;
5144
4638
  declare const index$6_TranscriptThrottled: typeof TranscriptThrottled;
4639
+ declare const index$6_TranscriptTranslatedTexts: typeof TranscriptTranslatedTexts;
5145
4640
  declare const index$6_TranscriptUtterance: typeof TranscriptUtterance;
5146
4641
  declare const index$6_TranscriptUtteranceChannel: typeof TranscriptUtteranceChannel;
4642
+ declare const index$6_TranscriptUtteranceTranslatedTexts: typeof TranscriptUtteranceTranslatedTexts;
5147
4643
  declare const index$6_TranscriptUtterances: typeof TranscriptUtterances;
5148
4644
  declare const index$6_TranscriptWebhookAuthHeaderName: typeof TranscriptWebhookAuthHeaderName;
5149
- declare const index$6_TranscriptWebhookNotification: typeof TranscriptWebhookNotification;
5150
4645
  declare const index$6_TranscriptWebhookStatusCode: typeof TranscriptWebhookStatusCode;
5151
4646
  declare const index$6_TranscriptWebhookUrl: typeof TranscriptWebhookUrl;
5152
4647
  declare const index$6_TranscriptWord: typeof TranscriptWord;
5153
4648
  declare const index$6_TranscriptWordChannel: typeof TranscriptWordChannel;
5154
4649
  declare const index$6_TranscriptWordSpeaker: typeof TranscriptWordSpeaker;
5155
4650
  declare const index$6_TranscriptWords: typeof TranscriptWords;
5156
- type index$6_UnauthorizedResponse = UnauthorizedResponse;
4651
+ declare const index$6_TranslationRequestBody: typeof TranslationRequestBody;
4652
+ declare const index$6_TranslationRequestBodyTranslation: typeof TranslationRequestBodyTranslation;
4653
+ declare const index$6_TranslationResponse: typeof TranslationResponse;
4654
+ declare const index$6_TranslationResponseTranslation: typeof TranslationResponseTranslation;
5157
4655
  type index$6_UploadedFile = UploadedFile;
5158
4656
  type index$6_WordSearchMatch = WordSearchMatch;
5159
4657
  type index$6_WordSearchParams = WordSearchParams;
5160
4658
  type index$6_WordSearchResponse = WordSearchResponse;
5161
4659
  type index$6_WordSearchTimestamp = WordSearchTimestamp;
5162
4660
  declare namespace index$6 {
5163
- export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, type index$6_BadRequestResponse as BadRequestResponse, type index$6_CannotAccessUploadedFileResponse as CannotAccessUploadedFileResponse, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, type index$6_CreateRealtimeTemporaryTokenParams as CreateRealtimeTemporaryTokenParams, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_GatewayTimeoutResponse as GatewayTimeoutResponse, type index$6_GetSubtitlesParams as GetSubtitlesParams, type index$6_InternalServerErrorResponse as InternalServerErrorResponse, type index$6_LemurActionItemsParams as LemurActionItemsParams, type index$6_LemurActionItemsParamsAllOf as LemurActionItemsParamsAllOf, type index$6_LemurActionItemsResponse as LemurActionItemsResponse, type index$6_LemurBaseParams as LemurBaseParams, type index$6_LemurBaseParamsContext as LemurBaseParamsContext, type index$6_LemurBaseParamsContextOneOf as LemurBaseParamsContextOneOf, type index$6_LemurBaseParamsFinalModel as LemurBaseParamsFinalModel, type index$6_LemurBaseResponse as LemurBaseResponse, index$6_LemurModel as LemurModel, type index$6_LemurQuestion as LemurQuestion, type index$6_LemurQuestionAnswer as LemurQuestionAnswer, type index$6_LemurQuestionAnswerParams as LemurQuestionAnswerParams, type index$6_LemurQuestionAnswerParamsAllOf as LemurQuestionAnswerParamsAllOf, type index$6_LemurQuestionAnswerResponse as LemurQuestionAnswerResponse, type index$6_LemurQuestionAnswerResponseAllOf as LemurQuestionAnswerResponseAllOf, type index$6_LemurQuestionContext as LemurQuestionContext, type index$6_LemurQuestionContextOneOf as LemurQuestionContextOneOf, type index$6_LemurResponse as LemurResponse, type index$6_LemurStringResponse as LemurStringResponse, type index$6_LemurStringResponseAllOf as LemurStringResponseAllOf, type index$6_LemurSummaryParams as LemurSummaryParams, type index$6_LemurSummaryParamsAllOf as LemurSummaryParamsAllOf, type index$6_LemurSummaryResponse as LemurSummaryResponse, type index$6_LemurTaskParams as LemurTaskParams, type index$6_LemurTaskParamsAllOf as LemurTaskParamsAllOf, type index$6_LemurTaskResponse as LemurTaskResponse, type index$6_LemurUsage as LemurUsage, type index$6_ListTranscriptParams as ListTranscriptParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_NotFoundResponse as NotFoundResponse, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, type index$6_PurgeLemurRequestDataResponse as PurgeLemurRequestDataResponse, type index$6_RealtimeTemporaryTokenResponse as RealtimeTemporaryTokenResponse, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, index$6_RedactedAudioNotification as RedactedAudioNotification, index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, type index$6_ServiceUnavailableResponse as ServiceUnavailableResponse, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, type index$6_TooManyRequestsResponse as TooManyRequestsResponse, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItem as TopicDetectionResultLabelsItem, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptBoostParam as TranscriptBoostParam, index$6_TranscriptBoostParamProperty as TranscriptBoostParamProperty, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodeProperty as TranscriptLanguageCodeProperty, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemCompleted as TranscriptListItemCompleted, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParams as TranscriptOptionalParams, index$6_TranscriptOptionalParamsLanguageCode as TranscriptOptionalParamsLanguageCode, index$6_TranscriptOptionalParamsLanguageCodeOneOf as TranscriptOptionalParamsLanguageCodeOneOf, index$6_TranscriptOptionalParamsRedactPiiSub as TranscriptOptionalParamsRedactPiiSub, index$6_TranscriptOptionalParamsSpeakersExpected as TranscriptOptionalParamsSpeakersExpected, index$6_TranscriptOptionalParamsSpeechModel as TranscriptOptionalParamsSpeechModel, index$6_TranscriptOptionalParamsSpeechThreshold as TranscriptOptionalParamsSpeechThreshold, index$6_TranscriptOptionalParamsWebhookAuthHeaderName as TranscriptOptionalParamsWebhookAuthHeaderName, index$6_TranscriptOptionalParamsWebhookAuthHeaderValue as TranscriptOptionalParamsWebhookAuthHeaderValue, type index$6_TranscriptParagraph as TranscriptParagraph, type index$6_TranscriptParams as TranscriptParams, type index$6_TranscriptParamsAllOf as TranscriptParamsAllOf, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptReadyNotification as TranscriptReadyNotification, index$6_TranscriptReadyStatus as TranscriptReadyStatus, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookNotification as TranscriptWebhookNotification, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, type index$6_UnauthorizedResponse as UnauthorizedResponse, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
4661
+ export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, index$6_CustomFormattingRequestBody as CustomFormattingRequestBody, index$6_CustomFormattingRequestBodyCustomFormatting as CustomFormattingRequestBodyCustomFormatting, index$6_CustomFormattingResponse as CustomFormattingResponse, index$6_CustomFormattingResponseCustomFormatting as CustomFormattingResponseCustomFormatting, index$6_CustomFormattingResponseCustomFormattingMapping as CustomFormattingResponseCustomFormattingMapping, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_ErrorDetails as ErrorDetails, type index$6_GetSubtitles200 as GetSubtitles200, type index$6_GetSubtitlesParams as GetSubtitlesParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, type index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeakerIdentificationRequestBody as SpeakerIdentificationRequestBody, index$6_SpeakerIdentificationRequestBodySpeakerIdentification as SpeakerIdentificationRequestBodySpeakerIdentification, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, index$6_SpeakerIdentificationResponse as SpeakerIdentificationResponse, index$6_SpeakerIdentificationResponseSpeakerIdentification as SpeakerIdentificationResponseSpeakerIdentification, index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping as SpeakerIdentificationResponseSpeakerIdentificationMapping, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItems as TopicDetectionResultLabelsItems, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptDomain as TranscriptDomain, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodes as TranscriptLanguageCodes, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, index$6_TranscriptLanguageDetectionOptions as TranscriptLanguageDetectionOptions, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParamsLanguageDetectionOptions as TranscriptOptionalParamsLanguageDetectionOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptions as TranscriptOptionalParamsRedactPiiAudioOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptOptionalParamsRemoveAudioTags as TranscriptOptionalParamsRemoveAudioTags, index$6_TranscriptOptionalParamsSpeakerOptions as TranscriptOptionalParamsSpeakerOptions, index$6_TranscriptOptionalParamsSpeechUnderstanding as TranscriptOptionalParamsSpeechUnderstanding, index$6_TranscriptOptionalParamsSpeechUnderstandingRequest as TranscriptOptionalParamsSpeechUnderstandingRequest, type index$6_TranscriptParagraph as TranscriptParagraph, index$6_TranscriptParams as TranscriptParams, index$6_TranscriptParamsDomain as TranscriptParamsDomain, index$6_TranscriptParamsLanguageCode as TranscriptParamsLanguageCode, index$6_TranscriptParamsLanguageCodes as TranscriptParamsLanguageCodes, index$6_TranscriptParamsRedactPiiSub as TranscriptParamsRedactPiiSub, index$6_TranscriptParamsRemoveAudioTags as TranscriptParamsRemoveAudioTags, index$6_TranscriptParamsSpeakersExpected as TranscriptParamsSpeakersExpected, index$6_TranscriptParamsSpeechModel as TranscriptParamsSpeechModel, index$6_TranscriptParamsSpeechThreshold as TranscriptParamsSpeechThreshold, index$6_TranscriptParamsWebhookAuthHeaderName as TranscriptParamsWebhookAuthHeaderName, index$6_TranscriptParamsWebhookAuthHeaderValue as TranscriptParamsWebhookAuthHeaderValue, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioOptions as TranscriptRedactPiiAudioOptions, index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, index$6_TranscriptRemoveAudioTags as TranscriptRemoveAudioTags, index$6_TranscriptRemoveAudioTagsProperty as TranscriptRemoveAudioTagsProperty, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechModels as TranscriptSpeechModels, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeechUnderstanding as TranscriptSpeechUnderstanding, index$6_TranscriptSpeechUnderstandingRequest as TranscriptSpeechUnderstandingRequest, index$6_TranscriptSpeechUnderstandingResponse as TranscriptSpeechUnderstandingResponse, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptTemperature as TranscriptTemperature, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptTranslatedTexts as TranscriptTranslatedTexts, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtteranceTranslatedTexts as TranscriptUtteranceTranslatedTexts, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, index$6_TranslationRequestBody as TranslationRequestBody, index$6_TranslationRequestBodyTranslation as TranslationRequestBodyTranslation, index$6_TranslationResponse as TranslationResponse, index$6_TranslationResponseTranslation as TranslationResponseTranslation, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
5164
4662
  }
5165
4663
 
5166
4664
  /**
@@ -5177,6 +4675,15 @@ Transcripts are sorted from newest to oldest and can be retrieved for the last 9
5177
4675
 
5178
4676
  If you need to retrieve transcripts from more than 90 days ago please reach out to our Support team at support@assemblyai.com.
5179
4677
 
4678
+ **Pagination**
4679
+
4680
+ This endpoint returns paginated results. The response includes a `page_details` object with the following properties:
4681
+ - `page_details.limit` - Maximum number of transcripts per page.
4682
+ - `page_details.result_count` - Total number of transcripts returned on the current page.
4683
+ - `page_details.current_url` - URL to the current page.
4684
+ - `page_details.prev_url` - URL to the previous page of older transcripts.
4685
+ - `page_details.next_url` - URL to the next page of newer transcripts.
4686
+
5180
4687
  * @summary List transcripts
5181
4688
  */
5182
4689
  declare const listTranscripts: <TData = AxiosResponse<TranscriptList>>(params?: ListTranscriptsParams, options?: AxiosRequestConfig) => Promise<TData>;
@@ -5195,14 +4702,6 @@ Remove the data from the transcript and mark it as deleted.
5195
4702
  * @summary Delete transcript
5196
4703
  */
5197
4704
  declare const deleteTranscript: <TData = AxiosResponse<Transcript>>(transcriptId: string, options?: AxiosRequestConfig) => Promise<TData>;
5198
- /**
5199
- * <Warning>Streaming Speech-to-Text is currently not available on the EU endpoint.</Warning>
5200
- <Note>Any usage associated with a temporary token will be attributed to the API key that generated it.</Note>
5201
- Create a temporary authentication token for Streaming Speech-to-Text
5202
-
5203
- * @summary Create temporary authentication token for Streaming STT
5204
- */
5205
- declare const createTemporaryToken: <TData = AxiosResponse<RealtimeTemporaryTokenResponse>>(createRealtimeTemporaryTokenParams: CreateRealtimeTemporaryTokenParams, options?: AxiosRequestConfig) => Promise<TData>;
5206
4705
 
5207
4706
  /**
5208
4707
  * AssemblyAI transcription provider adapter
@@ -15349,6 +14848,18 @@ type ManageV1FilterVersionParameter = boolean;
15349
14848
  /**
15350
14849
  * ManageV1GroupingParameter type definition
15351
14850
  */
14851
+ /**
14852
+ * ManageV1GroupingParameter type definition
14853
+ */
14854
+ /**
14855
+ * ManageV1GroupingParameter type definition
14856
+ */
14857
+ /**
14858
+ * ManageV1GroupingParameter type definition
14859
+ */
14860
+ /**
14861
+ * ManageV1GroupingParameter type definition
14862
+ */
15352
14863
  type ManageV1GroupingParameter = typeof ManageV1GroupingParameter[keyof typeof ManageV1GroupingParameter];
15353
14864
  declare const ManageV1GroupingParameter: {
15354
14865
  readonly accessor: "accessor";
@@ -15421,6 +14932,18 @@ type ManageV1ModelsListParams = {
15421
14932
  /**
15422
14933
  * ManageV1ProjectKeysStatusParameter type definition
15423
14934
  */
14935
+ /**
14936
+ * ManageV1ProjectKeysStatusParameter type definition
14937
+ */
14938
+ /**
14939
+ * ManageV1ProjectKeysStatusParameter type definition
14940
+ */
14941
+ /**
14942
+ * ManageV1ProjectKeysStatusParameter type definition
14943
+ */
14944
+ /**
14945
+ * ManageV1ProjectKeysStatusParameter type definition
14946
+ */
15424
14947
  type ManageV1ProjectKeysStatusParameter = typeof ManageV1ProjectKeysStatusParameter[keyof typeof ManageV1ProjectKeysStatusParameter];
15425
14948
  declare const ManageV1ProjectKeysStatusParameter: {
15426
14949
  readonly active: "active";
@@ -16306,6 +15829,18 @@ type SelfHostedV1ParamsDistributionScopesParameter = SelfHostedV1ParamsDistribut
16306
15829
  /**
16307
15830
  * SelfHostedV1ParamsDistributionProviderParameter type definition
16308
15831
  */
15832
+ /**
15833
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15834
+ */
15835
+ /**
15836
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15837
+ */
15838
+ /**
15839
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15840
+ */
15841
+ /**
15842
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15843
+ */
16309
15844
  type SelfHostedV1ParamsDistributionProviderParameter = typeof SelfHostedV1ParamsDistributionProviderParameter[keyof typeof SelfHostedV1ParamsDistributionProviderParameter];
16310
15845
  declare const SelfHostedV1ParamsDistributionProviderParameter: {
16311
15846
  readonly quay: "quay";
@@ -19533,8 +19068,8 @@ declare const getJobsResponse: z.ZodObject<{
19533
19068
  }, "strip", z.ZodTypeAny, {
19534
19069
  language: string;
19535
19070
  diarization?: "none" | "speaker" | "channel" | undefined;
19536
- operating_point?: "standard" | "enhanced" | undefined;
19537
19071
  domain?: string | undefined;
19072
+ operating_point?: "standard" | "enhanced" | undefined;
19538
19073
  output_locale?: string | undefined;
19539
19074
  additional_vocab?: {
19540
19075
  content: string;
@@ -19561,8 +19096,8 @@ declare const getJobsResponse: z.ZodObject<{
19561
19096
  }, {
19562
19097
  language: string;
19563
19098
  diarization?: "none" | "speaker" | "channel" | undefined;
19564
- operating_point?: "standard" | "enhanced" | undefined;
19565
19099
  domain?: string | undefined;
19100
+ operating_point?: "standard" | "enhanced" | undefined;
19566
19101
  output_locale?: string | undefined;
19567
19102
  additional_vocab?: {
19568
19103
  content: string;
@@ -19720,8 +19255,8 @@ declare const getJobsResponse: z.ZodObject<{
19720
19255
  transcription_config?: {
19721
19256
  language: string;
19722
19257
  diarization?: "none" | "speaker" | "channel" | undefined;
19723
- operating_point?: "standard" | "enhanced" | undefined;
19724
19258
  domain?: string | undefined;
19259
+ operating_point?: "standard" | "enhanced" | undefined;
19725
19260
  output_locale?: string | undefined;
19726
19261
  additional_vocab?: {
19727
19262
  content: string;
@@ -19801,8 +19336,8 @@ declare const getJobsResponse: z.ZodObject<{
19801
19336
  transcription_config?: {
19802
19337
  language: string;
19803
19338
  diarization?: "none" | "speaker" | "channel" | undefined;
19804
- operating_point?: "standard" | "enhanced" | undefined;
19805
19339
  domain?: string | undefined;
19340
+ operating_point?: "standard" | "enhanced" | undefined;
19806
19341
  output_locale?: string | undefined;
19807
19342
  additional_vocab?: {
19808
19343
  content: string;
@@ -19900,8 +19435,8 @@ declare const getJobsResponse: z.ZodObject<{
19900
19435
  transcription_config?: {
19901
19436
  language: string;
19902
19437
  diarization?: "none" | "speaker" | "channel" | undefined;
19903
- operating_point?: "standard" | "enhanced" | undefined;
19904
19438
  domain?: string | undefined;
19439
+ operating_point?: "standard" | "enhanced" | undefined;
19905
19440
  output_locale?: string | undefined;
19906
19441
  additional_vocab?: {
19907
19442
  content: string;
@@ -19994,8 +19529,8 @@ declare const getJobsResponse: z.ZodObject<{
19994
19529
  transcription_config?: {
19995
19530
  language: string;
19996
19531
  diarization?: "none" | "speaker" | "channel" | undefined;
19997
- operating_point?: "standard" | "enhanced" | undefined;
19998
19532
  domain?: string | undefined;
19533
+ operating_point?: "standard" | "enhanced" | undefined;
19999
19534
  output_locale?: string | undefined;
20000
19535
  additional_vocab?: {
20001
19536
  content: string;
@@ -20090,8 +19625,8 @@ declare const getJobsResponse: z.ZodObject<{
20090
19625
  transcription_config?: {
20091
19626
  language: string;
20092
19627
  diarization?: "none" | "speaker" | "channel" | undefined;
20093
- operating_point?: "standard" | "enhanced" | undefined;
20094
19628
  domain?: string | undefined;
19629
+ operating_point?: "standard" | "enhanced" | undefined;
20095
19630
  output_locale?: string | undefined;
20096
19631
  additional_vocab?: {
20097
19632
  content: string;
@@ -20186,8 +19721,8 @@ declare const getJobsResponse: z.ZodObject<{
20186
19721
  transcription_config?: {
20187
19722
  language: string;
20188
19723
  diarization?: "none" | "speaker" | "channel" | undefined;
20189
- operating_point?: "standard" | "enhanced" | undefined;
20190
19724
  domain?: string | undefined;
19725
+ operating_point?: "standard" | "enhanced" | undefined;
20191
19726
  output_locale?: string | undefined;
20192
19727
  additional_vocab?: {
20193
19728
  content: string;
@@ -20368,8 +19903,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20368
19903
  }, "strip", z.ZodTypeAny, {
20369
19904
  language: string;
20370
19905
  diarization?: "none" | "speaker" | "channel" | undefined;
20371
- operating_point?: "standard" | "enhanced" | undefined;
20372
19906
  domain?: string | undefined;
19907
+ operating_point?: "standard" | "enhanced" | undefined;
20373
19908
  output_locale?: string | undefined;
20374
19909
  additional_vocab?: {
20375
19910
  content: string;
@@ -20396,8 +19931,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20396
19931
  }, {
20397
19932
  language: string;
20398
19933
  diarization?: "none" | "speaker" | "channel" | undefined;
20399
- operating_point?: "standard" | "enhanced" | undefined;
20400
19934
  domain?: string | undefined;
19935
+ operating_point?: "standard" | "enhanced" | undefined;
20401
19936
  output_locale?: string | undefined;
20402
19937
  additional_vocab?: {
20403
19938
  content: string;
@@ -20555,8 +20090,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20555
20090
  transcription_config?: {
20556
20091
  language: string;
20557
20092
  diarization?: "none" | "speaker" | "channel" | undefined;
20558
- operating_point?: "standard" | "enhanced" | undefined;
20559
20093
  domain?: string | undefined;
20094
+ operating_point?: "standard" | "enhanced" | undefined;
20560
20095
  output_locale?: string | undefined;
20561
20096
  additional_vocab?: {
20562
20097
  content: string;
@@ -20636,8 +20171,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20636
20171
  transcription_config?: {
20637
20172
  language: string;
20638
20173
  diarization?: "none" | "speaker" | "channel" | undefined;
20639
- operating_point?: "standard" | "enhanced" | undefined;
20640
20174
  domain?: string | undefined;
20175
+ operating_point?: "standard" | "enhanced" | undefined;
20641
20176
  output_locale?: string | undefined;
20642
20177
  additional_vocab?: {
20643
20178
  content: string;
@@ -20735,8 +20270,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20735
20270
  transcription_config?: {
20736
20271
  language: string;
20737
20272
  diarization?: "none" | "speaker" | "channel" | undefined;
20738
- operating_point?: "standard" | "enhanced" | undefined;
20739
20273
  domain?: string | undefined;
20274
+ operating_point?: "standard" | "enhanced" | undefined;
20740
20275
  output_locale?: string | undefined;
20741
20276
  additional_vocab?: {
20742
20277
  content: string;
@@ -20829,8 +20364,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20829
20364
  transcription_config?: {
20830
20365
  language: string;
20831
20366
  diarization?: "none" | "speaker" | "channel" | undefined;
20832
- operating_point?: "standard" | "enhanced" | undefined;
20833
20367
  domain?: string | undefined;
20368
+ operating_point?: "standard" | "enhanced" | undefined;
20834
20369
  output_locale?: string | undefined;
20835
20370
  additional_vocab?: {
20836
20371
  content: string;
@@ -20925,8 +20460,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20925
20460
  transcription_config?: {
20926
20461
  language: string;
20927
20462
  diarization?: "none" | "speaker" | "channel" | undefined;
20928
- operating_point?: "standard" | "enhanced" | undefined;
20929
20463
  domain?: string | undefined;
20464
+ operating_point?: "standard" | "enhanced" | undefined;
20930
20465
  output_locale?: string | undefined;
20931
20466
  additional_vocab?: {
20932
20467
  content: string;
@@ -21021,8 +20556,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
21021
20556
  transcription_config?: {
21022
20557
  language: string;
21023
20558
  diarization?: "none" | "speaker" | "channel" | undefined;
21024
- operating_point?: "standard" | "enhanced" | undefined;
21025
20559
  domain?: string | undefined;
20560
+ operating_point?: "standard" | "enhanced" | undefined;
21026
20561
  output_locale?: string | undefined;
21027
20562
  additional_vocab?: {
21028
20563
  content: string;
@@ -21210,8 +20745,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21210
20745
  }, "strip", z.ZodTypeAny, {
21211
20746
  language: string;
21212
20747
  diarization?: "none" | "speaker" | "channel" | undefined;
21213
- operating_point?: "standard" | "enhanced" | undefined;
21214
20748
  domain?: string | undefined;
20749
+ operating_point?: "standard" | "enhanced" | undefined;
21215
20750
  output_locale?: string | undefined;
21216
20751
  additional_vocab?: {
21217
20752
  content: string;
@@ -21238,8 +20773,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21238
20773
  }, {
21239
20774
  language: string;
21240
20775
  diarization?: "none" | "speaker" | "channel" | undefined;
21241
- operating_point?: "standard" | "enhanced" | undefined;
21242
20776
  domain?: string | undefined;
20777
+ operating_point?: "standard" | "enhanced" | undefined;
21243
20778
  output_locale?: string | undefined;
21244
20779
  additional_vocab?: {
21245
20780
  content: string;
@@ -21397,8 +20932,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21397
20932
  transcription_config?: {
21398
20933
  language: string;
21399
20934
  diarization?: "none" | "speaker" | "channel" | undefined;
21400
- operating_point?: "standard" | "enhanced" | undefined;
21401
20935
  domain?: string | undefined;
20936
+ operating_point?: "standard" | "enhanced" | undefined;
21402
20937
  output_locale?: string | undefined;
21403
20938
  additional_vocab?: {
21404
20939
  content: string;
@@ -21478,8 +21013,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21478
21013
  transcription_config?: {
21479
21014
  language: string;
21480
21015
  diarization?: "none" | "speaker" | "channel" | undefined;
21481
- operating_point?: "standard" | "enhanced" | undefined;
21482
21016
  domain?: string | undefined;
21017
+ operating_point?: "standard" | "enhanced" | undefined;
21483
21018
  output_locale?: string | undefined;
21484
21019
  additional_vocab?: {
21485
21020
  content: string;
@@ -21577,8 +21112,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21577
21112
  transcription_config?: {
21578
21113
  language: string;
21579
21114
  diarization?: "none" | "speaker" | "channel" | undefined;
21580
- operating_point?: "standard" | "enhanced" | undefined;
21581
21115
  domain?: string | undefined;
21116
+ operating_point?: "standard" | "enhanced" | undefined;
21582
21117
  output_locale?: string | undefined;
21583
21118
  additional_vocab?: {
21584
21119
  content: string;
@@ -21671,8 +21206,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21671
21206
  transcription_config?: {
21672
21207
  language: string;
21673
21208
  diarization?: "none" | "speaker" | "channel" | undefined;
21674
- operating_point?: "standard" | "enhanced" | undefined;
21675
21209
  domain?: string | undefined;
21210
+ operating_point?: "standard" | "enhanced" | undefined;
21676
21211
  output_locale?: string | undefined;
21677
21212
  additional_vocab?: {
21678
21213
  content: string;
@@ -21767,8 +21302,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21767
21302
  transcription_config?: {
21768
21303
  language: string;
21769
21304
  diarization?: "none" | "speaker" | "channel" | undefined;
21770
- operating_point?: "standard" | "enhanced" | undefined;
21771
21305
  domain?: string | undefined;
21306
+ operating_point?: "standard" | "enhanced" | undefined;
21772
21307
  output_locale?: string | undefined;
21773
21308
  additional_vocab?: {
21774
21309
  content: string;
@@ -21863,8 +21398,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21863
21398
  transcription_config?: {
21864
21399
  language: string;
21865
21400
  diarization?: "none" | "speaker" | "channel" | undefined;
21866
- operating_point?: "standard" | "enhanced" | undefined;
21867
21401
  domain?: string | undefined;
21402
+ operating_point?: "standard" | "enhanced" | undefined;
21868
21403
  output_locale?: string | undefined;
21869
21404
  additional_vocab?: {
21870
21405
  content: string;
@@ -22090,8 +21625,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22090
21625
  }, "strip", z.ZodTypeAny, {
22091
21626
  language: string;
22092
21627
  diarization?: "none" | "speaker" | "channel" | undefined;
22093
- operating_point?: "standard" | "enhanced" | undefined;
22094
21628
  domain?: string | undefined;
21629
+ operating_point?: "standard" | "enhanced" | undefined;
22095
21630
  output_locale?: string | undefined;
22096
21631
  additional_vocab?: {
22097
21632
  content: string;
@@ -22118,8 +21653,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22118
21653
  }, {
22119
21654
  language: string;
22120
21655
  diarization?: "none" | "speaker" | "channel" | undefined;
22121
- operating_point?: "standard" | "enhanced" | undefined;
22122
21656
  domain?: string | undefined;
21657
+ operating_point?: "standard" | "enhanced" | undefined;
22123
21658
  output_locale?: string | undefined;
22124
21659
  additional_vocab?: {
22125
21660
  content: string;
@@ -22306,8 +21841,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22306
21841
  transcription_config?: {
22307
21842
  language: string;
22308
21843
  diarization?: "none" | "speaker" | "channel" | undefined;
22309
- operating_point?: "standard" | "enhanced" | undefined;
22310
21844
  domain?: string | undefined;
21845
+ operating_point?: "standard" | "enhanced" | undefined;
22311
21846
  output_locale?: string | undefined;
22312
21847
  additional_vocab?: {
22313
21848
  content: string;
@@ -22387,8 +21922,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22387
21922
  transcription_config?: {
22388
21923
  language: string;
22389
21924
  diarization?: "none" | "speaker" | "channel" | undefined;
22390
- operating_point?: "standard" | "enhanced" | undefined;
22391
21925
  domain?: string | undefined;
21926
+ operating_point?: "standard" | "enhanced" | undefined;
22392
21927
  output_locale?: string | undefined;
22393
21928
  additional_vocab?: {
22394
21929
  content: string;
@@ -23129,8 +22664,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
23129
22664
  transcription_config?: {
23130
22665
  language: string;
23131
22666
  diarization?: "none" | "speaker" | "channel" | undefined;
23132
- operating_point?: "standard" | "enhanced" | undefined;
23133
22667
  domain?: string | undefined;
22668
+ operating_point?: "standard" | "enhanced" | undefined;
23134
22669
  output_locale?: string | undefined;
23135
22670
  additional_vocab?: {
23136
22671
  content: string;
@@ -23353,8 +22888,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
23353
22888
  transcription_config?: {
23354
22889
  language: string;
23355
22890
  diarization?: "none" | "speaker" | "channel" | undefined;
23356
- operating_point?: "standard" | "enhanced" | undefined;
23357
22891
  domain?: string | undefined;
22892
+ operating_point?: "standard" | "enhanced" | undefined;
23358
22893
  output_locale?: string | undefined;
23359
22894
  additional_vocab?: {
23360
22895
  content: string;
@@ -24935,4 +24470,4 @@ declare namespace elevenLabsSpeechToTextAPI_zod {
24935
24470
  export { elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdHeader as deleteTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdParams as deleteTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdResponse as deleteTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdHeader as getTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdParams as getTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdResponse as getTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_speechToTextBody as speechToTextBody, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive as speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsMax as speechToTextBodyAdditionalFormatsMax, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMaxOne as speechToTextBodyDiarizationThresholdMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMinOne as speechToTextBodyDiarizationThresholdMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizeDefault as speechToTextBodyDiarizeDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyFileFormatDefault as speechToTextBodyFileFormatDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyKeytermsDefault as speechToTextBodyKeytermsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNoVerbatimDefault as speechToTextBodyNoVerbatimDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNumSpeakersMaxOne as speechToTextBodyNumSpeakersMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMaxOne as speechToTextBodySeedMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMinOne as speechToTextBodySeedMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTagAudioEventsDefault as speechToTextBodyTagAudioEventsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMaxOne as speechToTextBodyTemperatureMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMinOne as speechToTextBodyTemperatureMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTimestampsGranularityDefault as speechToTextBodyTimestampsGranularityDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyUseMultiChannelDefault as speechToTextBodyUseMultiChannelDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyWebhookDefault as speechToTextBodyWebhookDefault, elevenLabsSpeechToTextAPI_zod_speechToTextHeader as speechToTextHeader, elevenLabsSpeechToTextAPI_zod_speechToTextQueryEnableLoggingDefault as speechToTextQueryEnableLoggingDefault, elevenLabsSpeechToTextAPI_zod_speechToTextQueryParams as speechToTextQueryParams, elevenLabsSpeechToTextAPI_zod_speechToTextResponse as speechToTextResponse };
24936
24471
  }
24937
24472
 
24938
- export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult, TranscriptOptionalParams as AssemblyAIOptions, SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAIStreamingOptions, TopicDetectionModelResult as AssemblyAITopicsResult, index$6 as AssemblyAITypes, type AudioChunkAckMessage, AudioInput, AudioResponseFormat, AzureSTTAdapter, Status as AzureStatus, index$5 as AzureTypes, BaseAdapter, CreateTranscriptionRequest, CreateTranscriptionRequestModel, CreateTranscriptionResponseDiarizedJson, CreateTranscriptionResponseVerboseJson, DeepgramAdapter, type DeepgramConfig, ListenV1ResponseMetadata as DeepgramMetadata, ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramStreamingOptions, index$3 as DeepgramTypes, ElevenLabsAdapter, type ElevenLabsConfig, index as ElevenLabsTypes, elevenLabsSpeechToTextAPI_zod as ElevenLabsZodSchemas, type EndRecordingMessage, type EndSessionMessage, type GetProjectRequestV1Response, GladiaAdapter, AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, AudioToLlmListDTO as GladiaAudioToLlmResult, ChapterizationDTO as GladiaChapters, CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, NamedEntityRecognitionDTO as GladiaEntities, ModerationDTO as GladiaModeration, InitTranscriptionRequest as GladiaOptions, SentimentAnalysisDTO as GladiaSentiment, SpeakerReidentificationDTO as GladiaSpeakerReidentification, GladiaStreamingOptions, StreamingRequest as GladiaStreamingRequest, StructuredDataExtractionDTO as GladiaStructuredData, TranslationDTO as GladiaTranslation, index$7 as GladiaTypes, InitTranscriptionRequest, LanguageIdentificationMode, LanguageIdentificationProperties, type ListProjectRequestsV1Response, type ListTranscriptionResponseItemsItem, ListTranscriptsOptions, ListTranscriptsParams, ListenV1EncodingParameter, ListenV1LanguageParameter, ListenV1MediaTranscribeParams, ListenV1ModelParameter, ListenV1Response, ListenV1ResponseResultsChannelsItemAlternativesItem, ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, ListenV1ResponseResultsUtterancesItem, ListenV1VersionParameter, ManageV1FilterEndpointParameter, ManageV1FilterStatusParameter, ManageV1ProjectsRequestsListParams, type NamedEntityRecognitionMessage, streamingTypes as OpenAIStreamingTypes, index$4 as OpenAITypes, OpenAIWhisperAdapter, CreateTranscriptionRequest as OpenAIWhisperOptions, type PaginatedTranscriptions, type PostChapterizationMessage, type PostFinalTranscriptMessage, type PostSummarizationMessage, type PostTranscriptMessage, PreRecordedResponse, ProfanityFilterMode, type ProjectRequestResponse, ProviderCapabilities, type ProviderConfig, PunctuationMode, type SentimentAnalysisMessage, SonioxAdapter, type SonioxConfig, SonioxModelCode, index$1 as SonioxTypes, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type SpeechEndMessage, type SpeechStartMessage, SpeechmaticsAdapter, type SpeechmaticsConfig, index$2 as SpeechmaticsTypes, speechmaticsASRRESTAPI_zod as SpeechmaticsZodSchemas, type StartRecordingMessage, type StartSessionMessage, type StopRecordingAckMessage, StreamEvent, StreamingCallbacks, StreamingOptions, StreamingRequest, type StreamingResponse, StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedModels, StreamingSupportedSampleRateEnum, StreamingUpdateConfiguration, TranscribeOptions, Transcript, type TranscriptListItem, type TranscriptMessage, TranscriptOptionalParamsSpeechModel, type TranscriptParams, TranscriptStatus, TranscriptUtterance, TranscriptWord, Transcription$1 as Transcription, type TranscriptionAdapter, TranscriptionControllerListV2Params, TranscriptionControllerListV2StatusItem, TranscriptionDTO, TranscriptionLanguageCodeEnum, TranscriptionProperties, TranscriptionProvider, type TranscriptionsListParams, type TranslationMessage, UnifiedTranscriptResponse, UtteranceDTO, VoiceRouter, type VoiceRouterConfig, type WebHook, type WebHookEvents, WordDTO, createAssemblyAIAdapter, createAzureSTTAdapter, createDeepgramAdapter, createElevenLabsAdapter, createGladiaAdapter, createOpenAIWhisperAdapter, createSonioxAdapter, createSpeechmaticsAdapter, createTemporaryToken, createTranscript, createTranscription, createVoiceRouter, deleteTranscript as deleteTranscriptAPI, getTranscript as getTranscriptAPI, listTranscripts as listTranscriptsAPI, preRecordedControllerDeletePreRecordedJobV2, preRecordedControllerGetAudioV2, preRecordedControllerGetPreRecordedJobV2, preRecordedControllerInitPreRecordedJobV2, streamingControllerDeleteStreamingJobV2, streamingControllerGetAudioV2, streamingControllerInitStreamingSessionV2, transcriptionControllerListV2, transcriptionsCreate, transcriptionsDelete, transcriptionsGet, transcriptionsList, transcriptionsListFiles, webHooksCreate, webHooksDelete, webHooksList };
24473
+ export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult, TranscriptParams as AssemblyAIOptions, SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAIStreamingOptions, TopicDetectionModelResult as AssemblyAITopicsResult, index$6 as AssemblyAITypes, type AudioChunkAckMessage, AudioInput, AudioResponseFormat, AzureSTTAdapter, Status as AzureStatus, index$5 as AzureTypes, BaseAdapter, CreateTranscriptionRequest, CreateTranscriptionRequestModel, CreateTranscriptionResponseDiarizedJson, CreateTranscriptionResponseVerboseJson, DeepgramAdapter, type DeepgramConfig, ListenV1ResponseMetadata as DeepgramMetadata, ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramStreamingOptions, index$3 as DeepgramTypes, ElevenLabsAdapter, type ElevenLabsConfig, index as ElevenLabsTypes, elevenLabsSpeechToTextAPI_zod as ElevenLabsZodSchemas, type EndRecordingMessage, type EndSessionMessage, type GetProjectRequestV1Response, GladiaAdapter, AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, AudioToLlmListDTO as GladiaAudioToLlmResult, ChapterizationDTO as GladiaChapters, CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, NamedEntityRecognitionDTO as GladiaEntities, ModerationDTO as GladiaModeration, InitTranscriptionRequest as GladiaOptions, SentimentAnalysisDTO as GladiaSentiment, SpeakerReidentificationDTO as GladiaSpeakerReidentification, GladiaStreamingOptions, StreamingRequest as GladiaStreamingRequest, StructuredDataExtractionDTO as GladiaStructuredData, TranslationDTO as GladiaTranslation, index$7 as GladiaTypes, InitTranscriptionRequest, LanguageIdentificationMode, LanguageIdentificationProperties, type ListProjectRequestsV1Response, type ListTranscriptionResponseItemsItem, ListTranscriptsOptions, ListTranscriptsParams, ListenV1EncodingParameter, ListenV1LanguageParameter, ListenV1MediaTranscribeParams, ListenV1ModelParameter, ListenV1Response, ListenV1ResponseResultsChannelsItemAlternativesItem, ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, ListenV1ResponseResultsUtterancesItem, ListenV1VersionParameter, ManageV1FilterEndpointParameter, ManageV1FilterStatusParameter, ManageV1ProjectsRequestsListParams, type NamedEntityRecognitionMessage, streamingTypes as OpenAIStreamingTypes, index$4 as OpenAITypes, OpenAIWhisperAdapter, CreateTranscriptionRequest as OpenAIWhisperOptions, type PaginatedTranscriptions, type PostChapterizationMessage, type PostFinalTranscriptMessage, type PostSummarizationMessage, type PostTranscriptMessage, PreRecordedResponse, ProfanityFilterMode, type ProjectRequestResponse, ProviderCapabilities, type ProviderConfig, PunctuationMode, type SentimentAnalysisMessage, SonioxAdapter, type SonioxConfig, SonioxModelCode, index$1 as SonioxTypes, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type SpeechEndMessage, SpeechModel, type SpeechStartMessage, SpeechmaticsAdapter, type SpeechmaticsConfig, index$2 as SpeechmaticsTypes, speechmaticsASRRESTAPI_zod as SpeechmaticsZodSchemas, type StartRecordingMessage, type StartSessionMessage, type StopRecordingAckMessage, StreamEvent, StreamingCallbacks, StreamingOptions, StreamingRequest, type StreamingResponse, StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedModels, StreamingSupportedSampleRateEnum, StreamingUpdateConfiguration, TranscribeOptions, Transcript, type TranscriptListItem, type TranscriptMessage, TranscriptParams, TranscriptStatus, TranscriptUtterance, TranscriptWord, Transcription$1 as Transcription, type TranscriptionAdapter, TranscriptionControllerListV2Params, TranscriptionControllerListV2StatusItem, TranscriptionDTO, TranscriptionLanguageCodeEnum, TranscriptionProperties, TranscriptionProvider, type TranscriptionsListParams, type TranslationMessage, UnifiedTranscriptResponse, UtteranceDTO, VoiceRouter, type VoiceRouterConfig, type WebHook, type WebHookEvents, WordDTO, createAssemblyAIAdapter, createAzureSTTAdapter, createDeepgramAdapter, createElevenLabsAdapter, createGladiaAdapter, createOpenAIWhisperAdapter, createSonioxAdapter, createSpeechmaticsAdapter, createTranscript, createTranscription, createVoiceRouter, deleteTranscript as deleteTranscriptAPI, getTranscript as getTranscriptAPI, listTranscripts as listTranscriptsAPI, preRecordedControllerDeletePreRecordedJobV2, preRecordedControllerGetAudioV2, preRecordedControllerGetPreRecordedJobV2, preRecordedControllerInitPreRecordedJobV2, streamingControllerDeleteStreamingJobV2, streamingControllerGetAudioV2, streamingControllerInitStreamingSessionV2, transcriptionControllerListV2, transcriptionsCreate, transcriptionsDelete, transcriptionsGet, transcriptionsList, transcriptionsListFiles, webHooksCreate, webHooksDelete, webHooksList };