voice-router-dev 0.8.3 → 0.8.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.mts CHANGED
@@ -1,10 +1,10 @@
1
- import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptOptionalParams, aQ as TranscriptStatus, aR as TranscriptWord, aS as AudioIntelligenceModelStatus, aT as AutoHighlightResult, aU as AutoHighlightsResult, aV as Chapter, aW as ContentSafetyLabel, aX as ContentSafetyLabelResult, aY as ContentSafetyLabelsResult, aZ as ContentSafetyLabelsResultSeverityScoreSummary, a_ as ContentSafetyLabelsResultSummary, a$ as Entity, b0 as EntityType, b1 as ListTranscriptsParams, b2 as PiiPolicy, b3 as RedactPiiAudioQuality, b4 as RedactedAudioNotification, b5 as RedactedAudioResponse, b6 as RedactedAudioStatus, b7 as Sentiment, b8 as SentimentAnalysisResult$1, b9 as SentimentAnalysisResultChannel, ba as SentimentAnalysisResultSpeaker, bb as SeverityScoreSummary, bc as SpeechModel, bd as SubstitutionPolicy, be as SummaryModel, bf as SummaryType, bg as Timestamp, bh as TopicDetectionModelResult, bi as TopicDetectionModelResultSummary, bj as TopicDetectionResult, bk as TopicDetectionResultLabelsItem, bl as Transcript, bm as TranscriptAudioDuration, bn as TranscriptAudioEndAt, bo as TranscriptAudioStartFrom, bp as TranscriptAutoChapters, bq as TranscriptAutoHighlightsResult, br as TranscriptBoostParam, bs as TranscriptBoostParamProperty, bt as TranscriptChapters, bu as TranscriptConfidence, bv as TranscriptContentSafety, bw as TranscriptContentSafetyLabels, bx as TranscriptCustomSpelling, by as TranscriptCustomSpellingProperty, bz as TranscriptCustomTopics, bA as TranscriptDisfluencies, bB as TranscriptEntities, bC as TranscriptEntityDetection, bD as TranscriptFilterProfanity, bE as TranscriptFormatText, bF as TranscriptIabCategories, bG as TranscriptIabCategoriesResult, bH as TranscriptLanguageCode, bI as TranscriptLanguageCodeProperty, bJ as TranscriptLanguageConfidence, bK as TranscriptLanguageConfidenceThreshold, bL as TranscriptLanguageDetection, bM as TranscriptMultichannel, bN as TranscriptOptionalParamsLanguageCode, bO as TranscriptOptionalParamsLanguageCodeOneOf, bP as TranscriptOptionalParamsRedactPiiSub, bQ as TranscriptOptionalParamsSpeakersExpected, bR as TranscriptOptionalParamsSpeechModel, bS as TranscriptOptionalParamsSpeechThreshold, bT as TranscriptOptionalParamsWebhookAuthHeaderName, bU as TranscriptOptionalParamsWebhookAuthHeaderValue, bV as TranscriptPunctuate, bW as TranscriptReadyNotification, bX as TranscriptReadyStatus, bY as TranscriptRedactPiiAudio, bZ as TranscriptRedactPiiAudioQuality, b_ as TranscriptRedactPiiPolicies, b$ as TranscriptSentimentAnalysis, c0 as TranscriptSentimentAnalysisResults, c1 as TranscriptSpeakerLabels, c2 as TranscriptSpeakersExpected, c3 as TranscriptSpeechModel, c4 as TranscriptSpeechThreshold, c5 as TranscriptSpeedBoost, c6 as TranscriptSummary, c7 as TranscriptSummaryModel, c8 as TranscriptSummaryType, c9 as TranscriptText, ca as TranscriptThrottled, cb as TranscriptUtterance, cc as TranscriptUtteranceChannel, cd as TranscriptUtterances, ce as TranscriptWebhookAuthHeaderName, cf as TranscriptWebhookNotification, cg as TranscriptWebhookStatusCode, ch as TranscriptWebhookUrl, ci as TranscriptWordChannel, cj as TranscriptWordSpeaker, ck as TranscriptWords, cl as StreamingUpdateConfiguration, cm as Transcription$1, cn as EntityError, co as Status, cp as EntityReference, cq as DiarizationProperties, cr as DiarizationSpeakersProperties, cs as LanguageIdentificationMode, ct as LanguageIdentificationProperties, cu as LanguageIdentificationPropertiesSpeechModelMapping, cv as ProfanityFilterMode, cw as PunctuationMode, cx as TranscriptionCustomProperties, cy as TranscriptionLinks, cz as TranscriptionProperties, cA as TranscriptTextUsageTokens, cB as TranscriptionSegment, cC as RealtimeSessionCreateRequestGAModel, cD as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, cE as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, cF as AudioResponseFormat, cG as CreateTranscription200One, cH as CreateTranscriptionRequest, cI as CreateTranscriptionRequestModel, cJ as CreateTranscriptionRequestStream, cK as CreateTranscriptionRequestTimestampGranularitiesItem, cL as CreateTranscriptionResponseDiarizedJson, cM as CreateTranscriptionResponseDiarizedJsonTask, cN as CreateTranscriptionResponseDiarizedJsonUsage, cO as CreateTranscriptionResponseJson, cP as CreateTranscriptionResponseJsonLogprobsItem, cQ as CreateTranscriptionResponseJsonUsage, cR as CreateTranscriptionResponseVerboseJson, cS as TranscriptTextUsageDuration, cT as TranscriptTextUsageDurationType, cU as TranscriptTextUsageTokensInputTokenDetails, cV as TranscriptTextUsageTokensType, cW as TranscriptionChunkingStrategy, cX as TranscriptionChunkingStrategyAnyOf, cY as TranscriptionDiarizedSegment, cZ as TranscriptionDiarizedSegmentType, c_ as TranscriptionInclude, c$ as TranscriptionWord, d0 as VadConfig, d1 as VadConfigType, d2 as ListenV1Response, d3 as ManageV1FilterAccessorParameter, d4 as ManageV1FilterDeploymentParameter, d5 as ManageV1LimitParameter, d6 as ManageV1PageParameter, d7 as ManageV1FilterEndpointParameter, d8 as ManageV1FilterMethodParameter, d9 as SharedTopics, da as SharedIntents, db as SharedSentiments, dc as SharedCallbackParameter, dd as SharedCallbackMethodParameter, de as SharedSentimentParameter, df as SharedSummarizeParameter, dg as SharedTagParameter, dh as SharedTopicsParameter, di as SharedCustomTopicParameter, dj as SharedCustomTopicModeParameter, dk as SharedIntentsParameter, dl as SharedCustomIntentParameter, dm as SharedCustomIntentModeParameter, dn as SharedMipOptOutParameter, dp as ListenV1DetectEntitiesParameter, dq as ListenV1DetectLanguageParameter, dr as ListenV1DiarizeParameter, ds as ListenV1DictationParameter, dt as ListenV1EncodingParameter, du as ListenV1FillerWordsParameter, dv as ListenV1KeytermParameter, dw as ListenV1KeywordsParameter, dx as ListenV1LanguageParameter, dy as ListenV1MeasurementsParameter, dz as ListenV1MediaTranscribeParams, dA as ListenV1ModelParameter, dB as ListenV1MultichannelParameter, dC as ListenV1NumeralsParameter, dD as ListenV1ParagraphsParameter, dE as ListenV1ProfanityFilterParameter, dF as ListenV1PunctuateParameter, dG as ListenV1RedactParameter, dH as ListenV1RedactParameterOneOfItem, dI as ListenV1ReplaceParameter, dJ as ListenV1ResponseMetadata, dK as ListenV1ResponseMetadataIntentsInfo, dL as ListenV1ResponseMetadataModelInfo, dM as ListenV1ResponseMetadataSentimentInfo, dN as ListenV1ResponseMetadataSummaryInfo, dO as ListenV1ResponseMetadataTopicsInfo, dP as ListenV1ResponseResults, dQ as ListenV1ResponseResultsChannels, dR as ListenV1ResponseResultsChannelsItem, dS as ListenV1ResponseResultsChannelsItemAlternativesItem, dT as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, dU as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, dV as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, dW as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, dX as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, dY as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, dZ as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, d_ as ListenV1ResponseResultsChannelsItemSearchItem, d$ as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, e0 as ListenV1ResponseResultsSummary, e1 as ListenV1ResponseResultsUtterances, e2 as ListenV1ResponseResultsUtterancesItem, e3 as ListenV1ResponseResultsUtterancesItemWordsItem, e4 as ListenV1SearchParameter, e5 as ListenV1SmartFormatParameter, e6 as ListenV1UttSplitParameter, e7 as ListenV1UtterancesParameter, e8 as ListenV1VersionParameter, e9 as ManageV1EndDateTimeParameter, ea as ManageV1FilterRequestIdParameter, eb as ManageV1FilterStatusParameter, ec as ManageV1ProjectsRequestsListParams, ed as ManageV1StartDateTimeParameter, ee as SharedExtraParameter, ef as SharedIntentsResults, eg as SharedIntentsResultsIntents, eh as SharedIntentsResultsIntentsSegmentsItem, ei as SharedIntentsResultsIntentsSegmentsItemIntentsItem, ej as SharedSentimentsAverage, ek as SharedSentimentsSegmentsItem, el as SharedTopicsResults, em as SharedTopicsResultsTopics, en as SharedTopicsResultsTopicsSegmentsItem, eo as SharedTopicsResultsTopicsSegmentsItemTopicsItem, ep as JobType, eq as AlignmentConfig, er as TranscriptionConfig, es as TrackingData, et as OutputConfig, eu as OperatingPoint, ev as AudioEventItem, ew as AudioEventSummary, ex as AudioEventSummaryItem, ey as AutoChaptersResult, ez as AutoChaptersResultError, eA as AutoChaptersResultErrorType, eB as Chapter$1, eC as JobInfo, eD as LanguageIdentificationResult, eE as LanguageIdentificationResultAlternative, eF as LanguageIdentificationResultError, eG as LanguageIdentificationResultItem, eH as LanguagePackInfo, eI as LanguagePackInfoWritingDirection, eJ as OutputConfigSrtOverrides, eK as RecognitionAlternative, eL as RecognitionDisplay, eM as RecognitionDisplayDirection, eN as RecognitionMetadata, eO as RecognitionResult, eP as RecognitionResultAttachesTo, eQ as RecognitionResultType, eR as RetrieveTranscriptResponse, eS as RetrieveTranscriptResponseAudioEventSummary, eT as RetrieveTranscriptResponseAudioEventSummaryChannels, eU as RetrieveTranscriptResponseTranslations, eV as SentimentAnalysisError, eW as SentimentAnalysisErrorType, eX as SentimentAnalysisResult$2, eY as SentimentAnalysisResultSentimentAnalysis, eZ as SentimentChannelSummary, e_ as SentimentSegment, e$ as SentimentSpeakerSummary, f0 as SentimentSummary, f1 as SentimentSummaryDetail, f2 as SpokenFormRecognitionResult, f3 as SpokenFormRecognitionResultType, f4 as SummarizationError, f5 as SummarizationErrorType, f6 as SummarizationResult, f7 as TopicDetectionError, f8 as TopicDetectionErrorType, f9 as TopicDetectionResult$1, fa as TopicDetectionSegment, fb as TopicDetectionSegmentTopic, fc as TopicDetectionSummary, fd as TopicDetectionSummaryOverall, fe as TrackingDataDetails, ff as TranscriptionConfigAdditionalVocabItem, fg as TranscriptionConfigDiarization, fh as TranscriptionConfigMaxDelayMode, fi as TranscriptionConfigPunctuationOverrides, fj as TranscriptionConfigSpeakerDiarizationConfig, fk as TranscriptionConfigTranscriptFilteringConfig, fl as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fm as TranslationError, fn as TranslationErrorType, fo as TranslationSentence, fp as WrittenFormRecognitionResult, fq as WrittenFormRecognitionResultType, fr as SpeechToTextChunkResponseModel, fs as AdditionalFormatResponseModel, ft as AdditionalFormats, fu as BodySpeechToTextV1SpeechToTextPost, fv as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fw as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, fx as BodySpeechToTextV1SpeechToTextPostEntityDetection, fy as BodySpeechToTextV1SpeechToTextPostFile, fz as BodySpeechToTextV1SpeechToTextPostFileFormat, fA as BodySpeechToTextV1SpeechToTextPostLanguageCode, fB as BodySpeechToTextV1SpeechToTextPostModelId, fC as BodySpeechToTextV1SpeechToTextPostNumSpeakers, fD as BodySpeechToTextV1SpeechToTextPostSeed, fE as BodySpeechToTextV1SpeechToTextPostTemperature, fF as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, fG as BodySpeechToTextV1SpeechToTextPostWebhookId, fH as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, fI as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, fJ as DetectedEntity, fK as DocxExportOptions, fL as DocxExportOptionsFormat, fM as DocxExportOptionsMaxSegmentChars, fN as DocxExportOptionsMaxSegmentDurationS, fO as DocxExportOptionsSegmentOnSilenceLongerThanS, fP as ExportOptions, fQ as HtmlExportOptions, fR as HtmlExportOptionsFormat, fS as HtmlExportOptionsMaxSegmentChars, fT as HtmlExportOptionsMaxSegmentDurationS, fU as HtmlExportOptionsSegmentOnSilenceLongerThanS, fV as PdfExportOptions, fW as PdfExportOptionsFormat, fX as PdfExportOptionsMaxSegmentChars, fY as PdfExportOptionsMaxSegmentDurationS, fZ as PdfExportOptionsSegmentOnSilenceLongerThanS, f_ as SegmentedJsonExportOptions, f$ as SegmentedJsonExportOptionsFormat, g0 as SegmentedJsonExportOptionsMaxSegmentChars, g1 as SegmentedJsonExportOptionsMaxSegmentDurationS, g2 as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, g3 as SpeechToTextCharacterResponseModel, g4 as SpeechToTextCharacterResponseModelEnd, g5 as SpeechToTextCharacterResponseModelStart, g6 as SpeechToTextChunkResponseModelAdditionalFormats, g7 as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, g8 as SpeechToTextChunkResponseModelChannelIndex, g9 as SpeechToTextChunkResponseModelEntities, ga as SpeechToTextChunkResponseModelTranscriptionId, gb as SpeechToTextWordResponseModel, gc as SpeechToTextWordResponseModelCharacters, gd as SpeechToTextWordResponseModelEnd, ge as SpeechToTextWordResponseModelSpeakerId, gf as SpeechToTextWordResponseModelStart, gg as SpeechToTextWordResponseModelType, gh as SrtExportOptions, gi as SrtExportOptionsFormat, gj as SrtExportOptionsMaxCharactersPerLine, gk as SrtExportOptionsMaxSegmentChars, gl as SrtExportOptionsMaxSegmentDurationS, gm as SrtExportOptionsSegmentOnSilenceLongerThanS, gn as TxtExportOptions, go as TxtExportOptionsFormat, gp as TxtExportOptionsMaxCharactersPerLine, gq as TxtExportOptionsMaxSegmentChars, gr as TxtExportOptionsMaxSegmentDurationS, gs as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-DvIT4xai.mjs';
2
- export { gw as AssemblyAIExtendedData, gX as AssemblyAIUpdateConfiguration, gT as AudioAckEvent, gV as AudioChunk, gC as BatchOnlyProvider, h2 as BeginEvent, gS as ChapterizationEvent, gy as DeepgramExtendedData, gz as ElevenLabsExtendedData, g_ as ElevenLabsStreamingOptions, gQ as EntityEvent, h5 as ErrorEvent, gx as GladiaExtendedData, gU as LifecycleEvent, gK as ListTranscriptsResponse, gY as OpenAIStreamingOptions, gA as ProviderExtendedDataMap, gL as ProviderRawResponseMap, g$ as ProviderStreamingOptions, gW as RawWebSocketMessage, gP as SentimentEvent, gD as SessionStatus, gZ as SonioxStreamingOptions, gE as Speaker, gN as SpeechEvent, gt as SpeechmaticsOperatingPoint, gM as StreamEventType, h6 as StreamingEventMessage, h8 as StreamingForceEndpoint, h0 as StreamingOptionsForProvider, gB as StreamingProvider, h7 as StreamingWord, gR as SummarizationEvent, h4 as TerminationEvent, h1 as TranscribeStreamParams, gJ as TranscriptData, gI as TranscriptMetadata, gv as TranscriptionLanguage, gu as TranscriptionModel, gH as TranscriptionStatus, gO as TranslationEvent, h3 as TurnEvent, gG as Utterance, gF as Word } from './speechToTextChunkResponseModel-DvIT4xai.mjs';
1
+ import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptStatus, aQ as TranscriptWord, aR as AudioIntelligenceModelStatus, aS as AutoHighlightResult, aT as AutoHighlightsResult, aU as Chapter, aV as ContentSafetyLabel, aW as ContentSafetyLabelResult, aX as ContentSafetyLabelsResult, aY as ContentSafetyLabelsResultSeverityScoreSummary, aZ as ContentSafetyLabelsResultSummary, a_ as CustomFormattingRequestBody, a$ as CustomFormattingRequestBodyCustomFormatting, b0 as CustomFormattingResponse, b1 as CustomFormattingResponseCustomFormatting, b2 as CustomFormattingResponseCustomFormattingMapping, b3 as Entity, b4 as EntityType, b5 as ListTranscriptsParams, b6 as PiiPolicy, b7 as RedactPiiAudioQuality, b8 as Sentiment, b9 as SentimentAnalysisResult$1, ba as SentimentAnalysisResultChannel, bb as SentimentAnalysisResultSpeaker, bc as SeverityScoreSummary, bd as SpeakerIdentificationRequestBody, be as SpeakerIdentificationRequestBodySpeakerIdentification, bf as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, bg as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, bh as SpeakerIdentificationResponse, bi as SpeakerIdentificationResponseSpeakerIdentification, bj as SpeakerIdentificationResponseSpeakerIdentificationMapping, bk as SpeechModel, bl as SubstitutionPolicy, bm as SummaryModel, bn as SummaryType, bo as Timestamp, bp as TopicDetectionModelResult, bq as TopicDetectionModelResultSummary, br as TopicDetectionResult, bs as TopicDetectionResultLabelsItems, bt as Transcript, bu as TranscriptAudioDuration, bv as TranscriptAudioEndAt, bw as TranscriptAudioStartFrom, bx as TranscriptAutoChapters, by as TranscriptAutoHighlightsResult, bz as TranscriptChapters, bA as TranscriptConfidence, bB as TranscriptContentSafety, bC as TranscriptContentSafetyLabels, bD as TranscriptCustomSpelling, bE as TranscriptCustomSpellingProperty, bF as TranscriptCustomTopics, bG as TranscriptDisfluencies, bH as TranscriptDomain, bI as TranscriptEntities, bJ as TranscriptEntityDetection, bK as TranscriptFilterProfanity, bL as TranscriptFormatText, bM as TranscriptIabCategories, bN as TranscriptIabCategoriesResult, bO as TranscriptLanguageCode, bP as TranscriptLanguageCodes, bQ as TranscriptLanguageConfidence, bR as TranscriptLanguageConfidenceThreshold, bS as TranscriptLanguageDetection, bT as TranscriptLanguageDetectionOptions, bU as TranscriptMultichannel, bV as TranscriptOptionalParamsLanguageDetectionOptions, bW as TranscriptOptionalParamsRedactPiiAudioOptions, bX as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, bY as TranscriptOptionalParamsRemoveAudioTags, bZ as TranscriptOptionalParamsSpeakerOptions, b_ as TranscriptOptionalParamsSpeechUnderstanding, b$ as TranscriptOptionalParamsSpeechUnderstandingRequest, c0 as TranscriptParams, c1 as TranscriptParamsDomain, c2 as TranscriptParamsLanguageCode, c3 as TranscriptParamsLanguageCodes, c4 as TranscriptParamsRedactPiiSub, c5 as TranscriptParamsRemoveAudioTags, c6 as TranscriptParamsSpeakersExpected, c7 as TranscriptParamsSpeechThreshold, c8 as TranscriptParamsWebhookAuthHeaderName, c9 as TranscriptParamsWebhookAuthHeaderValue, ca as TranscriptPunctuate, cb as TranscriptRedactPiiAudio, cc as TranscriptRedactPiiAudioOptions, cd as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, ce as TranscriptRedactPiiAudioQuality, cf as TranscriptRedactPiiPolicies, cg as TranscriptRemoveAudioTags, ch as TranscriptRemoveAudioTagsProperty, ci as TranscriptSentimentAnalysis, cj as TranscriptSentimentAnalysisResults, ck as TranscriptSpeakerLabels, cl as TranscriptSpeakersExpected, cm as TranscriptSpeechModel, cn as TranscriptSpeechModels, co as TranscriptSpeechThreshold, cp as TranscriptSpeechUnderstanding, cq as TranscriptSpeechUnderstandingRequest, cr as TranscriptSpeechUnderstandingResponse, cs as TranscriptSpeedBoost, ct as TranscriptSummary, cu as TranscriptSummaryModel, cv as TranscriptSummaryType, cw as TranscriptTemperature, cx as TranscriptText, cy as TranscriptThrottled, cz as TranscriptTranslatedTexts, cA as TranscriptUtterance, cB as TranscriptUtteranceChannel, cC as TranscriptUtteranceTranslatedTexts, cD as TranscriptUtterances, cE as TranscriptWebhookAuthHeaderName, cF as TranscriptWebhookStatusCode, cG as TranscriptWebhookUrl, cH as TranscriptWordChannel, cI as TranscriptWordSpeaker, cJ as TranscriptWords, cK as TranslationRequestBody, cL as TranslationRequestBodyTranslation, cM as TranslationResponse, cN as TranslationResponseTranslation, cO as StreamingUpdateConfiguration, cP as Transcription$1, cQ as EntityError, cR as Status, cS as EntityReference, cT as DiarizationProperties, cU as DiarizationSpeakersProperties, cV as LanguageIdentificationMode, cW as LanguageIdentificationProperties, cX as LanguageIdentificationPropertiesSpeechModelMapping, cY as ProfanityFilterMode, cZ as PunctuationMode, c_ as TranscriptionCustomProperties, c$ as TranscriptionLinks, d0 as TranscriptionProperties, d1 as TranscriptTextUsageTokens, d2 as TranscriptionSegment, d3 as RealtimeSessionCreateRequestGAModel, d4 as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, d5 as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, d6 as AudioResponseFormat, d7 as CreateTranscription200One, d8 as CreateTranscriptionRequest, d9 as CreateTranscriptionRequestModel, da as CreateTranscriptionRequestStream, db as CreateTranscriptionRequestTimestampGranularitiesItem, dc as CreateTranscriptionResponseDiarizedJson, dd as CreateTranscriptionResponseDiarizedJsonTask, de as CreateTranscriptionResponseDiarizedJsonUsage, df as CreateTranscriptionResponseJson, dg as CreateTranscriptionResponseJsonLogprobsItem, dh as CreateTranscriptionResponseJsonUsage, di as CreateTranscriptionResponseVerboseJson, dj as TranscriptTextUsageDuration, dk as TranscriptTextUsageDurationType, dl as TranscriptTextUsageTokensInputTokenDetails, dm as TranscriptTextUsageTokensType, dn as TranscriptionChunkingStrategy, dp as TranscriptionChunkingStrategyAnyOf, dq as TranscriptionDiarizedSegment, dr as TranscriptionDiarizedSegmentType, ds as TranscriptionInclude, dt as TranscriptionWord, du as VadConfig, dv as VadConfigType, dw as ListenV1Response, dx as ManageV1FilterAccessorParameter, dy as ManageV1FilterDeploymentParameter, dz as ManageV1LimitParameter, dA as ManageV1PageParameter, dB as ManageV1FilterEndpointParameter, dC as ManageV1FilterMethodParameter, dD as SharedTopics, dE as SharedIntents, dF as SharedSentiments, dG as SharedCallbackParameter, dH as SharedCallbackMethodParameter, dI as SharedSentimentParameter, dJ as SharedSummarizeParameter, dK as SharedTagParameter, dL as SharedTopicsParameter, dM as SharedCustomTopicParameter, dN as SharedCustomTopicModeParameter, dO as SharedIntentsParameter, dP as SharedCustomIntentParameter, dQ as SharedCustomIntentModeParameter, dR as SharedMipOptOutParameter, dS as ListenV1DetectEntitiesParameter, dT as ListenV1DetectLanguageParameter, dU as ListenV1DiarizeParameter, dV as ListenV1DictationParameter, dW as ListenV1EncodingParameter, dX as ListenV1FillerWordsParameter, dY as ListenV1KeytermParameter, dZ as ListenV1KeywordsParameter, d_ as ListenV1LanguageParameter, d$ as ListenV1MeasurementsParameter, e0 as ListenV1MediaTranscribeParams, e1 as ListenV1ModelParameter, e2 as ListenV1MultichannelParameter, e3 as ListenV1NumeralsParameter, e4 as ListenV1ParagraphsParameter, e5 as ListenV1ProfanityFilterParameter, e6 as ListenV1PunctuateParameter, e7 as ListenV1RedactParameter, e8 as ListenV1RedactParameterOneOfItem, e9 as ListenV1ReplaceParameter, ea as ListenV1ResponseMetadata, eb as ListenV1ResponseMetadataIntentsInfo, ec as ListenV1ResponseMetadataModelInfo, ed as ListenV1ResponseMetadataSentimentInfo, ee as ListenV1ResponseMetadataSummaryInfo, ef as ListenV1ResponseMetadataTopicsInfo, eg as ListenV1ResponseResults, eh as ListenV1ResponseResultsChannels, ei as ListenV1ResponseResultsChannelsItem, ej as ListenV1ResponseResultsChannelsItemAlternativesItem, ek as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, el as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, em as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, en as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, eo as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, ep as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, eq as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, er as ListenV1ResponseResultsChannelsItemSearchItem, es as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, et as ListenV1ResponseResultsSummary, eu as ListenV1ResponseResultsUtterances, ev as ListenV1ResponseResultsUtterancesItem, ew as ListenV1ResponseResultsUtterancesItemWordsItem, ex as ListenV1SearchParameter, ey as ListenV1SmartFormatParameter, ez as ListenV1UttSplitParameter, eA as ListenV1UtterancesParameter, eB as ListenV1VersionParameter, eC as ManageV1EndDateTimeParameter, eD as ManageV1FilterRequestIdParameter, eE as ManageV1FilterStatusParameter, eF as ManageV1ProjectsRequestsListParams, eG as ManageV1StartDateTimeParameter, eH as SharedExtraParameter, eI as SharedIntentsResults, eJ as SharedIntentsResultsIntents, eK as SharedIntentsResultsIntentsSegmentsItem, eL as SharedIntentsResultsIntentsSegmentsItemIntentsItem, eM as SharedSentimentsAverage, eN as SharedSentimentsSegmentsItem, eO as SharedTopicsResults, eP as SharedTopicsResultsTopics, eQ as SharedTopicsResultsTopicsSegmentsItem, eR as SharedTopicsResultsTopicsSegmentsItemTopicsItem, eS as JobType, eT as AlignmentConfig, eU as TranscriptionConfig, eV as TrackingData, eW as OutputConfig, eX as OperatingPoint, eY as AudioEventItem, eZ as AudioEventSummary, e_ as AudioEventSummaryItem, e$ as AutoChaptersResult, f0 as AutoChaptersResultError, f1 as AutoChaptersResultErrorType, f2 as Chapter$1, f3 as JobInfo, f4 as LanguageIdentificationResult, f5 as LanguageIdentificationResultAlternative, f6 as LanguageIdentificationResultError, f7 as LanguageIdentificationResultItem, f8 as LanguagePackInfo, f9 as LanguagePackInfoWritingDirection, fa as OutputConfigSrtOverrides, fb as RecognitionAlternative, fc as RecognitionDisplay, fd as RecognitionDisplayDirection, fe as RecognitionMetadata, ff as RecognitionResult, fg as RecognitionResultAttachesTo, fh as RecognitionResultType, fi as RetrieveTranscriptResponse, fj as RetrieveTranscriptResponseAudioEventSummary, fk as RetrieveTranscriptResponseAudioEventSummaryChannels, fl as RetrieveTranscriptResponseTranslations, fm as SentimentAnalysisError, fn as SentimentAnalysisErrorType, fo as SentimentAnalysisResult$2, fp as SentimentAnalysisResultSentimentAnalysis, fq as SentimentChannelSummary, fr as SentimentSegment, fs as SentimentSpeakerSummary, ft as SentimentSummary, fu as SentimentSummaryDetail, fv as SpokenFormRecognitionResult, fw as SpokenFormRecognitionResultType, fx as SummarizationError, fy as SummarizationErrorType, fz as SummarizationResult, fA as TopicDetectionError, fB as TopicDetectionErrorType, fC as TopicDetectionResult$1, fD as TopicDetectionSegment, fE as TopicDetectionSegmentTopic, fF as TopicDetectionSummary, fG as TopicDetectionSummaryOverall, fH as TrackingDataDetails, fI as TranscriptionConfigAdditionalVocabItem, fJ as TranscriptionConfigDiarization, fK as TranscriptionConfigMaxDelayMode, fL as TranscriptionConfigPunctuationOverrides, fM as TranscriptionConfigSpeakerDiarizationConfig, fN as TranscriptionConfigTranscriptFilteringConfig, fO as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fP as TranslationError, fQ as TranslationErrorType, fR as TranslationSentence, fS as WrittenFormRecognitionResult, fT as WrittenFormRecognitionResultType, fU as SpeechToTextChunkResponseModel, fV as AdditionalFormatResponseModel, fW as AdditionalFormats, fX as BodySpeechToTextV1SpeechToTextPost, fY as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fZ as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, f_ as BodySpeechToTextV1SpeechToTextPostEntityDetection, f$ as BodySpeechToTextV1SpeechToTextPostFile, g0 as BodySpeechToTextV1SpeechToTextPostFileFormat, g1 as BodySpeechToTextV1SpeechToTextPostLanguageCode, g2 as BodySpeechToTextV1SpeechToTextPostModelId, g3 as BodySpeechToTextV1SpeechToTextPostNumSpeakers, g4 as BodySpeechToTextV1SpeechToTextPostSeed, g5 as BodySpeechToTextV1SpeechToTextPostTemperature, g6 as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, g7 as BodySpeechToTextV1SpeechToTextPostWebhookId, g8 as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, g9 as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, ga as DetectedEntity, gb as DocxExportOptions, gc as DocxExportOptionsFormat, gd as DocxExportOptionsMaxSegmentChars, ge as DocxExportOptionsMaxSegmentDurationS, gf as DocxExportOptionsSegmentOnSilenceLongerThanS, gg as ExportOptions, gh as HtmlExportOptions, gi as HtmlExportOptionsFormat, gj as HtmlExportOptionsMaxSegmentChars, gk as HtmlExportOptionsMaxSegmentDurationS, gl as HtmlExportOptionsSegmentOnSilenceLongerThanS, gm as PdfExportOptions, gn as PdfExportOptionsFormat, go as PdfExportOptionsMaxSegmentChars, gp as PdfExportOptionsMaxSegmentDurationS, gq as PdfExportOptionsSegmentOnSilenceLongerThanS, gr as SegmentedJsonExportOptions, gs as SegmentedJsonExportOptionsFormat, gt as SegmentedJsonExportOptionsMaxSegmentChars, gu as SegmentedJsonExportOptionsMaxSegmentDurationS, gv as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, gw as SpeechToTextCharacterResponseModel, gx as SpeechToTextCharacterResponseModelEnd, gy as SpeechToTextCharacterResponseModelStart, gz as SpeechToTextChunkResponseModelAdditionalFormats, gA as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, gB as SpeechToTextChunkResponseModelChannelIndex, gC as SpeechToTextChunkResponseModelEntities, gD as SpeechToTextChunkResponseModelTranscriptionId, gE as SpeechToTextWordResponseModel, gF as SpeechToTextWordResponseModelCharacters, gG as SpeechToTextWordResponseModelEnd, gH as SpeechToTextWordResponseModelSpeakerId, gI as SpeechToTextWordResponseModelStart, gJ as SpeechToTextWordResponseModelType, gK as SrtExportOptions, gL as SrtExportOptionsFormat, gM as SrtExportOptionsMaxCharactersPerLine, gN as SrtExportOptionsMaxSegmentChars, gO as SrtExportOptionsMaxSegmentDurationS, gP as SrtExportOptionsSegmentOnSilenceLongerThanS, gQ as TxtExportOptions, gR as TxtExportOptionsFormat, gS as TxtExportOptionsMaxCharactersPerLine, gT as TxtExportOptionsMaxSegmentChars, gU as TxtExportOptionsMaxSegmentDurationS, gV as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-Dj3Weqv1.mjs';
2
+ export { gZ as AssemblyAIExtendedData, ho as AssemblyAIUpdateConfiguration, hk as AudioAckEvent, hm as AudioChunk, h3 as BatchOnlyProvider, hv as BeginEvent, hj as ChapterizationEvent, g$ as DeepgramExtendedData, h0 as ElevenLabsExtendedData, hr as ElevenLabsStreamingOptions, hh as EntityEvent, hy as ErrorEvent, g_ as GladiaExtendedData, hl as LifecycleEvent, hb as ListTranscriptsResponse, hp as OpenAIStreamingOptions, h1 as ProviderExtendedDataMap, hc as ProviderRawResponseMap, hs as ProviderStreamingOptions, hn as RawWebSocketMessage, hg as SentimentEvent, h4 as SessionStatus, hq as SonioxStreamingOptions, h5 as Speaker, he as SpeechEvent, gW as SpeechmaticsOperatingPoint, hd as StreamEventType, hz as StreamingEventMessage, hB as StreamingForceEndpoint, ht as StreamingOptionsForProvider, h2 as StreamingProvider, hA as StreamingWord, hi as SummarizationEvent, hx as TerminationEvent, hu as TranscribeStreamParams, ha as TranscriptData, h9 as TranscriptMetadata, gY as TranscriptionLanguage, gX as TranscriptionModel, h8 as TranscriptionStatus, hf as TranslationEvent, hw as TurnEvent, h7 as Utterance, h6 as Word } from './speechToTextChunkResponseModel-Dj3Weqv1.mjs';
3
3
  import { T as TranscriptionProvider, P as ProviderCapabilities, A as AudioInput } from './provider-metadata-MDUUEuqF.mjs';
4
4
  export { k as AllLanguageCodes, o as AllProviders, a as AssemblyAICapabilities, i as AssemblyAILanguageCodes, b as AzureCapabilities, B as BatchOnlyProviderType, q as BatchOnlyProviders, C as CapabilityKeys, f as CapabilityLabels, D as DeepgramCapabilities, j as DeepgramLanguageCodes, E as ElevenLabsCapabilities, G as GladiaCapabilities, h as GladiaLanguageCodes, g as LanguageCode, L as LanguageLabels, O as OpenAICapabilities, d as ProviderCapabilitiesMap, l as ProviderDisplayNames, n as ProviderDocs, m as ProviderWebsites, c as SonioxCapabilities, S as SpeechmaticsCapabilities, e as StreamingProviderType, p as StreamingProviders } from './provider-metadata-MDUUEuqF.mjs';
5
5
  import { AssemblyAIRegionType, DeepgramRegionType, SpeechmaticsRegionType, SonioxModelCode, SonioxRegionType, ElevenLabsModelCode, ElevenLabsRegionType } from './constants.mjs';
6
6
  export { AssemblyAIEncoding, AssemblyAIEncodingType, AssemblyAILanguage, AssemblyAILanguageType, AssemblyAISampleRate, AssemblyAISampleRateType, AssemblyAISpeechModel, AssemblyAISpeechModelType, AssemblyAIStatus, AssemblyAITranscriptionModel, AssemblyAITranscriptionModelType, AzureLocale, AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, AzureLocaleType, AzureLocales, DeepgramCallbackMethod, DeepgramEncoding, DeepgramIntentMode, DeepgramLanguage, DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, DeepgramModelCode as DeepgramModelType, DeepgramRedact, DeepgramRedactType, DeepgramRegion, DeepgramSampleRate, DeepgramStatus, DeepgramTTSContainer, DeepgramTTSEncoding, DeepgramTTSModel, DeepgramTTSSampleRate, DeepgramTopicMode, DeepgramTopicModeType, ElevenLabsLanguageCode, ElevenLabsLanguageCodes, ElevenLabsLanguageLabels, ElevenLabsLanguages, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaRegion, GladiaSampleRate, GladiaStatus, GladiaTranslationLanguage, OpenAILanguageCodes, OpenAIModel, OpenAIRealtimeAudioFormat, OpenAIRealtimeModel, OpenAIRealtimeTranscriptionModel, OpenAIRealtimeTurnDetection, OpenAIResponseFormat, SonioxAsyncModel, SonioxAsyncModelCode, SonioxLanguage, SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, SonioxLanguageType, SonioxLanguages, SonioxModel, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, SonioxRealtimeModelCode, SonioxRegion, SpeechmaticsLanguage, SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsRegion } from './constants.mjs';
7
- export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-t_lVCkE5.mjs';
7
+ export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-DXAGTUfv.mjs';
8
8
  import { z } from 'zod';
9
9
  import { AxiosResponse, AxiosRequestConfig } from 'axios';
10
10
 
@@ -4115,46 +4115,8 @@ declare function createGladiaAdapter(config: ProviderConfig): GladiaAdapter;
4115
4115
  * Generated by orval v7.9.0 🍺
4116
4116
  * Do not edit manually.
4117
4117
  * AssemblyAI API
4118
- * AssemblyAI API
4119
- * OpenAPI spec version: 1.3.4
4120
- */
4121
- type TranscriptParamsAllOf = {
4122
- /** The URL of the audio or video file to transcribe. */
4123
- audio_url: string;
4124
- };
4125
-
4126
- /**
4127
- * Generated by orval v7.9.0 🍺
4128
- * Do not edit manually.
4129
- * AssemblyAI API
4130
- * AssemblyAI API
4131
- * OpenAPI spec version: 1.3.4
4132
- */
4133
-
4134
- /**
4135
- * The parameters for creating a transcript
4136
- */
4137
- type TranscriptParams = TranscriptParamsAllOf & TranscriptOptionalParams;
4138
-
4139
- /**
4140
- * Generated by orval v7.9.0 🍺
4141
- * Do not edit manually.
4142
- * AssemblyAI API
4143
- * AssemblyAI API
4144
- * OpenAPI spec version: 1.3.4
4145
- */
4146
- /**
4147
- * The date and time the transcript was completed
4148
- * @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
4149
- */
4150
- type TranscriptListItemCompleted = string | null;
4151
-
4152
- /**
4153
- * Generated by orval v7.9.0 🍺
4154
- * Do not edit manually.
4155
- * AssemblyAI API
4156
- * AssemblyAI API
4157
- * OpenAPI spec version: 1.3.4
4118
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4119
+ * OpenAPI spec version: 1.0.0
4158
4120
  */
4159
4121
  /**
4160
4122
  * Error message of why the transcript failed
@@ -4165,509 +4127,89 @@ type TranscriptListItemError = string | null;
4165
4127
  * Generated by orval v7.9.0 🍺
4166
4128
  * Do not edit manually.
4167
4129
  * AssemblyAI API
4168
- * AssemblyAI API
4169
- * OpenAPI spec version: 1.3.4
4170
- */
4171
-
4172
- interface TranscriptListItem {
4173
- /** The unique identifier for the transcript */
4174
- id: string;
4175
- /** The URL to retrieve the transcript */
4176
- resource_url: string;
4177
- /** The status of the transcript */
4178
- status: TranscriptStatus;
4179
- /**
4180
- * The date and time the transcript was created
4181
- * @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
4182
- */
4183
- created: string;
4184
- /**
4185
- * The date and time the transcript was completed
4186
- * @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
4187
- */
4188
- completed: TranscriptListItemCompleted;
4189
- /** The URL to the audio file */
4190
- audio_url: string;
4191
- /** Error message of why the transcript failed */
4192
- error: TranscriptListItemError;
4193
- }
4194
-
4195
- /**
4196
- * Bad request
4197
- */
4198
- type BadRequestResponse = Error;
4199
-
4200
- /**
4201
- * Cannot access uploaded file
4202
- */
4203
- type CannotAccessUploadedFileResponse = Error;
4204
-
4205
- /**
4206
- * Generated by orval v7.9.0 🍺
4207
- * Do not edit manually.
4208
- * AssemblyAI API
4209
- * AssemblyAI API
4210
- * OpenAPI spec version: 1.3.4
4211
- */
4212
- interface CreateRealtimeTemporaryTokenParams {
4213
- /**
4214
- * The amount of time until the token expires in seconds
4215
- * @minimum 60
4216
- */
4217
- expires_in: number;
4218
- }
4219
-
4220
- /**
4221
- * Generated by orval v7.9.0 🍺
4222
- * Do not edit manually.
4223
- * AssemblyAI API
4224
- * AssemblyAI API
4225
- * OpenAPI spec version: 1.3.4
4226
- */
4227
- interface Error$2 {
4228
- /** Error message */
4229
- error: string;
4230
- status?: "error";
4231
- [key: string]: unknown;
4232
- }
4233
-
4234
- /**
4235
- * Generated by orval v7.9.0 🍺
4236
- * Do not edit manually.
4237
- * AssemblyAI API
4238
- * AssemblyAI API
4239
- * OpenAPI spec version: 1.3.4
4240
- */
4241
- /**
4242
- * Gateway timeout
4243
- */
4244
- type GatewayTimeoutResponse = void;
4245
-
4246
- /**
4247
- * Generated by orval v7.9.0 🍺
4248
- * Do not edit manually.
4249
- * AssemblyAI API
4250
- * AssemblyAI API
4251
- * OpenAPI spec version: 1.3.4
4252
- */
4253
- type GetSubtitlesParams = {
4254
- /**
4255
- * The maximum number of characters per caption
4256
- */
4257
- chars_per_caption?: number;
4258
- };
4259
-
4260
- /**
4261
- * An error occurred while processing the request
4262
- */
4263
- type InternalServerErrorResponse = Error;
4264
-
4265
- /**
4266
- * Generated by orval v7.9.0 🍺
4267
- * Do not edit manually.
4268
- * AssemblyAI API
4269
- * AssemblyAI API
4270
- * OpenAPI spec version: 1.3.4
4271
- */
4272
- type LemurBaseParamsContextOneOf = {
4273
- [key: string]: unknown;
4274
- };
4275
-
4276
- /**
4277
- * Generated by orval v7.9.0 🍺
4278
- * Do not edit manually.
4279
- * AssemblyAI API
4280
- * AssemblyAI API
4281
- * OpenAPI spec version: 1.3.4
4282
- */
4283
-
4284
- /**
4285
- * Context to provide the model. This can be a string or a free-form JSON value.
4286
- */
4287
- type LemurBaseParamsContext = string | LemurBaseParamsContextOneOf;
4288
-
4289
- /**
4290
- * Generated by orval v7.9.0 🍺
4291
- * Do not edit manually.
4292
- * AssemblyAI API
4293
- * AssemblyAI API
4294
- * OpenAPI spec version: 1.3.4
4295
- */
4296
- /**
4297
- * The model that is used for the final prompt after compression is performed.
4298
-
4299
- */
4300
- type LemurModel = (typeof LemurModel)[keyof typeof LemurModel];
4301
- declare const LemurModel: {
4302
- readonly "anthropic/claude-3-5-sonnet": "anthropic/claude-3-5-sonnet";
4303
- readonly "anthropic/claude-3-opus": "anthropic/claude-3-opus";
4304
- readonly "anthropic/claude-3-haiku": "anthropic/claude-3-haiku";
4305
- };
4306
-
4307
- /**
4308
- * Generated by orval v7.9.0 🍺
4309
- * Do not edit manually.
4310
- * AssemblyAI API
4311
- * AssemblyAI API
4312
- * OpenAPI spec version: 1.3.4
4313
- */
4314
-
4315
- /**
4316
- * The model that is used for the final prompt after compression is performed.
4317
-
4318
- */
4319
- type LemurBaseParamsFinalModel = LemurModel | string;
4320
-
4321
- /**
4322
- * Generated by orval v7.9.0 🍺
4323
- * Do not edit manually.
4324
- * AssemblyAI API
4325
- * AssemblyAI API
4326
- * OpenAPI spec version: 1.3.4
4327
- */
4328
-
4329
- interface LemurBaseParams {
4330
- /** A list of completed transcripts with text. Up to a maximum of 100 hours of audio.
4331
- Use either transcript_ids or input_text as input into LeMUR.
4332
- */
4333
- transcript_ids?: string[];
4334
- /** Custom formatted transcript data. Maximum size is the context limit of the selected model.
4335
- Use either transcript_ids or input_text as input into LeMUR.
4336
- */
4337
- input_text?: string;
4338
- /** Context to provide the model. This can be a string or a free-form JSON value. */
4339
- context?: LemurBaseParamsContext;
4340
- /** The model that is used for the final prompt after compression is performed.
4341
- */
4342
- final_model: LemurBaseParamsFinalModel;
4343
- /** Max output size in tokens. */
4344
- max_output_size?: number;
4345
- /**
4346
- * The temperature to use for the model.
4347
- Higher values result in answers that are more creative, lower values are more conservative.
4348
- Can be any value between 0.0 and 1.0 inclusive.
4349
-
4350
- * @minimum 0
4351
- * @maximum 1
4352
- */
4353
- temperature?: number;
4354
- }
4355
-
4356
- /**
4357
- * Generated by orval v7.9.0 🍺
4358
- * Do not edit manually.
4359
- * AssemblyAI API
4360
- * AssemblyAI API
4361
- * OpenAPI spec version: 1.3.4
4362
- */
4363
- type LemurActionItemsParamsAllOf = {
4364
- /** How you want the action items to be returned. This can be any text.
4365
- Defaults to "Bullet Points".
4366
- */
4367
- answer_format?: string;
4368
- };
4369
-
4370
- /**
4371
- * Generated by orval v7.9.0 🍺
4372
- * Do not edit manually.
4373
- * AssemblyAI API
4374
- * AssemblyAI API
4375
- * OpenAPI spec version: 1.3.4
4376
- */
4377
-
4378
- type LemurActionItemsParams = LemurBaseParams & LemurActionItemsParamsAllOf;
4379
-
4380
- /**
4381
- * Generated by orval v7.9.0 🍺
4382
- * Do not edit manually.
4383
- * AssemblyAI API
4384
- * AssemblyAI API
4385
- * OpenAPI spec version: 1.3.4
4386
- */
4387
- type LemurStringResponseAllOf = {
4388
- /** The response generated by LeMUR. */
4389
- response: string;
4390
- };
4391
-
4392
- /**
4393
- * Generated by orval v7.9.0 🍺
4394
- * Do not edit manually.
4395
- * AssemblyAI API
4396
- * AssemblyAI API
4397
- * OpenAPI spec version: 1.3.4
4398
- */
4399
- /**
4400
- * The usage numbers for the LeMUR request
4401
- */
4402
- interface LemurUsage {
4403
- /**
4404
- * The number of input tokens used by the model
4405
- * @minimum 0
4406
- */
4407
- input_tokens: number;
4408
- /**
4409
- * The number of output tokens generated by the model
4410
- * @minimum 0
4411
- */
4412
- output_tokens: number;
4413
- }
4414
-
4415
- /**
4416
- * Generated by orval v7.9.0 🍺
4417
- * Do not edit manually.
4418
- * AssemblyAI API
4419
- * AssemblyAI API
4420
- * OpenAPI spec version: 1.3.4
4421
- */
4422
-
4423
- interface LemurBaseResponse {
4424
- /** The ID of the LeMUR request */
4425
- request_id: string;
4426
- /** The usage numbers for the LeMUR request */
4427
- usage: LemurUsage;
4428
- }
4429
-
4430
- /**
4431
- * Generated by orval v7.9.0 🍺
4432
- * Do not edit manually.
4433
- * AssemblyAI API
4434
- * AssemblyAI API
4435
- * OpenAPI spec version: 1.3.4
4436
- */
4437
-
4438
- type LemurStringResponse = LemurStringResponseAllOf & LemurBaseResponse;
4439
-
4440
- /**
4441
- * Generated by orval v7.9.0 🍺
4442
- * Do not edit manually.
4443
- * AssemblyAI API
4444
- * AssemblyAI API
4445
- * OpenAPI spec version: 1.3.4
4446
- */
4447
-
4448
- type LemurActionItemsResponse = LemurStringResponse;
4449
-
4450
- /**
4451
- * Generated by orval v7.9.0 🍺
4452
- * Do not edit manually.
4453
- * AssemblyAI API
4454
- * AssemblyAI API
4455
- * OpenAPI spec version: 1.3.4
4456
- */
4457
- type LemurQuestionContextOneOf = {
4458
- [key: string]: unknown;
4459
- };
4460
-
4461
- /**
4462
- * Generated by orval v7.9.0 🍺
4463
- * Do not edit manually.
4464
- * AssemblyAI API
4465
- * AssemblyAI API
4466
- * OpenAPI spec version: 1.3.4
4467
- */
4468
-
4469
- /**
4470
- * Any context about the transcripts you wish to provide. This can be a string or any object.
4471
- */
4472
- type LemurQuestionContext = string | LemurQuestionContextOneOf;
4473
-
4474
- /**
4475
- * Generated by orval v7.9.0 🍺
4476
- * Do not edit manually.
4477
- * AssemblyAI API
4478
- * AssemblyAI API
4479
- * OpenAPI spec version: 1.3.4
4480
- */
4481
-
4482
- interface LemurQuestion {
4483
- /** The question you wish to ask. For more complex questions use default model. */
4484
- question: string;
4485
- /** Any context about the transcripts you wish to provide. This can be a string or any object. */
4486
- context?: LemurQuestionContext;
4487
- /** How you want the answer to be returned. This can be any text. Can't be used with answer_options. Examples: "short sentence", "bullet points"
4488
- */
4489
- answer_format?: string;
4490
- /** What discrete options to return. Useful for precise responses. Can't be used with answer_format. Example: ["Yes", "No"]
4491
- */
4492
- answer_options?: string[];
4493
- }
4494
-
4495
- /**
4496
- * Generated by orval v7.9.0 🍺
4497
- * Do not edit manually.
4498
- * AssemblyAI API
4499
- * AssemblyAI API
4500
- * OpenAPI spec version: 1.3.4
4501
- */
4502
- /**
4503
- * An answer generated by LeMUR and its question
4504
- */
4505
- interface LemurQuestionAnswer {
4506
- /** The question for LeMUR to answer */
4507
- question: string;
4508
- /** The answer generated by LeMUR */
4509
- answer: string;
4510
- }
4511
-
4512
- /**
4513
- * Generated by orval v7.9.0 🍺
4514
- * Do not edit manually.
4515
- * AssemblyAI API
4516
- * AssemblyAI API
4517
- * OpenAPI spec version: 1.3.4
4518
- */
4519
-
4520
- type LemurQuestionAnswerParamsAllOf = {
4521
- /** A list of questions to ask */
4522
- questions: LemurQuestion[];
4523
- };
4524
-
4525
- /**
4526
- * Generated by orval v7.9.0 🍺
4527
- * Do not edit manually.
4528
- * AssemblyAI API
4529
- * AssemblyAI API
4530
- * OpenAPI spec version: 1.3.4
4531
- */
4532
-
4533
- type LemurQuestionAnswerParams = LemurBaseParams & LemurQuestionAnswerParamsAllOf;
4534
-
4535
- /**
4536
- * Generated by orval v7.9.0 🍺
4537
- * Do not edit manually.
4538
- * AssemblyAI API
4539
- * AssemblyAI API
4540
- * OpenAPI spec version: 1.3.4
4541
- */
4542
-
4543
- type LemurQuestionAnswerResponseAllOf = {
4544
- /** The answers generated by LeMUR and their questions */
4545
- response: LemurQuestionAnswer[];
4546
- };
4547
-
4548
- /**
4549
- * Generated by orval v7.9.0 🍺
4550
- * Do not edit manually.
4551
- * AssemblyAI API
4552
- * AssemblyAI API
4553
- * OpenAPI spec version: 1.3.4
4554
- */
4555
-
4556
- type LemurQuestionAnswerResponse = LemurBaseResponse & LemurQuestionAnswerResponseAllOf;
4557
-
4558
- /**
4559
- * Generated by orval v7.9.0 🍺
4560
- * Do not edit manually.
4561
- * AssemblyAI API
4562
- * AssemblyAI API
4563
- * OpenAPI spec version: 1.3.4
4564
- */
4565
-
4566
- type LemurResponse = LemurStringResponse | LemurQuestionAnswerResponse;
4567
-
4568
- /**
4569
- * Generated by orval v7.9.0 🍺
4570
- * Do not edit manually.
4571
- * AssemblyAI API
4572
- * AssemblyAI API
4573
- * OpenAPI spec version: 1.3.4
4574
- */
4575
- type LemurSummaryParamsAllOf = {
4576
- /** How you want the summary to be returned. This can be any text. Examples: "TLDR", "bullet points"
4577
- */
4578
- answer_format?: string;
4579
- };
4580
-
4581
- /**
4582
- * Generated by orval v7.9.0 🍺
4583
- * Do not edit manually.
4584
- * AssemblyAI API
4585
- * AssemblyAI API
4586
- * OpenAPI spec version: 1.3.4
4587
- */
4588
-
4589
- type LemurSummaryParams = LemurBaseParams & LemurSummaryParamsAllOf;
4590
-
4591
- /**
4592
- * Generated by orval v7.9.0 🍺
4593
- * Do not edit manually.
4594
- * AssemblyAI API
4595
- * AssemblyAI API
4596
- * OpenAPI spec version: 1.3.4
4130
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4131
+ * OpenAPI spec version: 1.0.0
4597
4132
  */
4598
4133
 
4599
- type LemurSummaryResponse = LemurStringResponse;
4134
+ interface TranscriptListItem {
4135
+ /** The unique identifier for the transcript */
4136
+ id: string;
4137
+ /** The URL to retrieve the transcript */
4138
+ resource_url: string;
4139
+ /** The status of the transcript */
4140
+ status: TranscriptStatus;
4141
+ /** The date and time the transcript was created */
4142
+ created: string;
4143
+ /** The date and time the transcript was completed */
4144
+ completed?: string;
4145
+ /** The URL to the audio file */
4146
+ audio_url: string;
4147
+ /** Error message of why the transcript failed */
4148
+ error: TranscriptListItemError;
4149
+ }
4600
4150
 
4601
4151
  /**
4602
4152
  * Generated by orval v7.9.0 🍺
4603
4153
  * Do not edit manually.
4604
4154
  * AssemblyAI API
4605
- * AssemblyAI API
4606
- * OpenAPI spec version: 1.3.4
4155
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4156
+ * OpenAPI spec version: 1.0.0
4607
4157
  */
4608
- type LemurTaskParamsAllOf = {
4609
- /** Your text to prompt the model to produce a desired output, including any context you want to pass into the model. */
4610
- prompt: string;
4611
- };
4158
+ /**
4159
+ * Additional error details if available
4160
+ */
4161
+ interface ErrorDetails {
4162
+ [key: string]: unknown;
4163
+ }
4612
4164
 
4613
4165
  /**
4614
4166
  * Generated by orval v7.9.0 🍺
4615
4167
  * Do not edit manually.
4616
4168
  * AssemblyAI API
4617
- * AssemblyAI API
4618
- * OpenAPI spec version: 1.3.4
4169
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4170
+ * OpenAPI spec version: 1.0.0
4619
4171
  */
4620
4172
 
4621
- type LemurTaskParams = LemurTaskParamsAllOf & LemurBaseParams;
4173
+ interface Error$2 {
4174
+ /** Error message describing what went wrong */
4175
+ error: string;
4176
+ /** Error code for programmatic handling */
4177
+ code?: string;
4178
+ /** Additional error details if available */
4179
+ details?: ErrorDetails;
4180
+ }
4622
4181
 
4623
4182
  /**
4624
4183
  * Generated by orval v7.9.0 🍺
4625
4184
  * Do not edit manually.
4626
4185
  * AssemblyAI API
4627
- * AssemblyAI API
4628
- * OpenAPI spec version: 1.3.4
4186
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4187
+ * OpenAPI spec version: 1.0.0
4629
4188
  */
4630
-
4631
- type LemurTaskResponse = LemurStringResponse;
4189
+ type GetSubtitles200 = {
4190
+ [key: string]: unknown;
4191
+ };
4632
4192
 
4633
4193
  /**
4634
4194
  * Generated by orval v7.9.0 🍺
4635
4195
  * Do not edit manually.
4636
4196
  * AssemblyAI API
4637
- * AssemblyAI API
4638
- * OpenAPI spec version: 1.3.4
4197
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4198
+ * OpenAPI spec version: 1.0.0
4639
4199
  */
4640
-
4641
- interface ListTranscriptParams {
4200
+ type GetSubtitlesParams = {
4642
4201
  /**
4643
- * Maximum amount of transcripts to retrieve
4644
- * @minimum 1
4645
- * @maximum 200
4202
+ * The maximum number of characters per caption
4646
4203
  */
4647
- limit?: number;
4648
- /** Filter by transcript status */
4649
- status?: TranscriptStatus;
4650
- /** Only get transcripts created on this date */
4651
- created_on?: string;
4652
- /** Get transcripts that were created before this transcript ID */
4653
- before_id?: string;
4654
- /** Get transcripts that were created after this transcript ID */
4655
- after_id?: string;
4656
- /** Only get throttled transcripts, overrides the status filter */
4657
- throttled_only?: boolean;
4658
- }
4659
-
4660
- /**
4661
- * Not found
4662
- */
4663
- type NotFoundResponse = Error;
4204
+ chars_per_caption?: number;
4205
+ };
4664
4206
 
4665
4207
  /**
4666
4208
  * Generated by orval v7.9.0 🍺
4667
4209
  * Do not edit manually.
4668
4210
  * AssemblyAI API
4669
- * AssemblyAI API
4670
- * OpenAPI spec version: 1.3.4
4211
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4212
+ * OpenAPI spec version: 1.0.0
4671
4213
  */
4672
4214
  /**
4673
4215
  * The URL to the next page of transcripts. The previous URL always points to a page with older transcripts.
@@ -4678,8 +4220,8 @@ type PageDetailsPrevUrl = string | null;
4678
4220
  * Generated by orval v7.9.0 🍺
4679
4221
  * Do not edit manually.
4680
4222
  * AssemblyAI API
4681
- * AssemblyAI API
4682
- * OpenAPI spec version: 1.3.4
4223
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4224
+ * OpenAPI spec version: 1.0.0
4683
4225
  */
4684
4226
  /**
4685
4227
  * The URL to the next page of transcripts. The next URL always points to a page with newer transcripts.
@@ -4690,8 +4232,8 @@ type PageDetailsNextUrl = string | null;
4690
4232
  * Generated by orval v7.9.0 🍺
4691
4233
  * Do not edit manually.
4692
4234
  * AssemblyAI API
4693
- * AssemblyAI API
4694
- * OpenAPI spec version: 1.3.4
4235
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4236
+ * OpenAPI spec version: 1.0.0
4695
4237
  */
4696
4238
 
4697
4239
  /**
@@ -4714,8 +4256,8 @@ interface PageDetails {
4714
4256
  * Generated by orval v7.9.0 🍺
4715
4257
  * Do not edit manually.
4716
4258
  * AssemblyAI API
4717
- * AssemblyAI API
4718
- * OpenAPI spec version: 1.3.4
4259
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4260
+ * OpenAPI spec version: 1.0.0
4719
4261
  */
4720
4262
 
4721
4263
  interface TranscriptParagraph {
@@ -4725,11 +4267,7 @@ interface TranscriptParagraph {
4725
4267
  start: number;
4726
4268
  /** The ending time, in milliseconds, of the paragraph */
4727
4269
  end: number;
4728
- /**
4729
- * The confidence score for the transcript of this paragraph
4730
- * @minimum 0
4731
- * @maximum 1
4732
- */
4270
+ /** The confidence score for the transcript of this paragraph */
4733
4271
  confidence: number;
4734
4272
  /** An array of words in the paragraph */
4735
4273
  words: TranscriptWord[];
@@ -4739,18 +4277,14 @@ interface TranscriptParagraph {
4739
4277
  * Generated by orval v7.9.0 🍺
4740
4278
  * Do not edit manually.
4741
4279
  * AssemblyAI API
4742
- * AssemblyAI API
4743
- * OpenAPI spec version: 1.3.4
4280
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4281
+ * OpenAPI spec version: 1.0.0
4744
4282
  */
4745
4283
 
4746
4284
  interface ParagraphsResponse {
4747
4285
  /** The unique identifier of your transcript */
4748
4286
  id: string;
4749
- /**
4750
- * The confidence score for the transcript
4751
- * @minimum 0
4752
- * @maximum 1
4753
- */
4287
+ /** The confidence score for the transcript */
4754
4288
  confidence: number;
4755
4289
  /** The duration of the audio file in seconds */
4756
4290
  audio_duration: number;
@@ -4762,36 +4296,38 @@ interface ParagraphsResponse {
4762
4296
  * Generated by orval v7.9.0 🍺
4763
4297
  * Do not edit manually.
4764
4298
  * AssemblyAI API
4765
- * AssemblyAI API
4766
- * OpenAPI spec version: 1.3.4
4299
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4300
+ * OpenAPI spec version: 1.0.0
4767
4301
  */
4768
- interface PurgeLemurRequestDataResponse {
4769
- /** The ID of the deletion request of the LeMUR request */
4770
- request_id: string;
4771
- /** The ID of the LeMUR request to purge the data for */
4772
- request_id_to_purge: string;
4773
- /** Whether the request data was deleted */
4774
- deleted: boolean;
4775
- }
4302
+ /**
4303
+ * The status of the redacted audio
4304
+ */
4305
+ type RedactedAudioStatus = (typeof RedactedAudioStatus)[keyof typeof RedactedAudioStatus];
4306
+ declare const RedactedAudioStatus: {
4307
+ readonly redacted_audio_ready: "redacted_audio_ready";
4308
+ };
4776
4309
 
4777
4310
  /**
4778
4311
  * Generated by orval v7.9.0 🍺
4779
4312
  * Do not edit manually.
4780
4313
  * AssemblyAI API
4781
- * AssemblyAI API
4782
- * OpenAPI spec version: 1.3.4
4314
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4315
+ * OpenAPI spec version: 1.0.0
4783
4316
  */
4784
- interface RealtimeTemporaryTokenResponse {
4785
- /** The temporary authentication token for Streaming Speech-to-Text */
4786
- token: string;
4317
+
4318
+ interface RedactedAudioResponse {
4319
+ /** The status of the redacted audio */
4320
+ status: RedactedAudioStatus;
4321
+ /** The URL of the redacted audio file */
4322
+ redacted_audio_url: string;
4787
4323
  }
4788
4324
 
4789
4325
  /**
4790
4326
  * Generated by orval v7.9.0 🍺
4791
4327
  * Do not edit manually.
4792
4328
  * AssemblyAI API
4793
- * AssemblyAI API
4794
- * OpenAPI spec version: 1.3.4
4329
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4330
+ * OpenAPI spec version: 1.0.0
4795
4331
  */
4796
4332
  /**
4797
4333
  * The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
@@ -4802,11 +4338,11 @@ type TranscriptSentenceChannel = string | null;
4802
4338
  * Generated by orval v7.9.0 🍺
4803
4339
  * Do not edit manually.
4804
4340
  * AssemblyAI API
4805
- * AssemblyAI API
4806
- * OpenAPI spec version: 1.3.4
4341
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4342
+ * OpenAPI spec version: 1.0.0
4807
4343
  */
4808
4344
  /**
4809
- * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null
4345
+ * The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
4810
4346
  */
4811
4347
  type TranscriptSentenceSpeaker = string | null;
4812
4348
 
@@ -4814,8 +4350,8 @@ type TranscriptSentenceSpeaker = string | null;
4814
4350
  * Generated by orval v7.9.0 🍺
4815
4351
  * Do not edit manually.
4816
4352
  * AssemblyAI API
4817
- * AssemblyAI API
4818
- * OpenAPI spec version: 1.3.4
4353
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4354
+ * OpenAPI spec version: 1.0.0
4819
4355
  */
4820
4356
 
4821
4357
  interface TranscriptSentence {
@@ -4825,17 +4361,13 @@ interface TranscriptSentence {
4825
4361
  start: number;
4826
4362
  /** The ending time, in milliseconds, for the sentence */
4827
4363
  end: number;
4828
- /**
4829
- * The confidence score for the transcript of this sentence
4830
- * @minimum 0
4831
- * @maximum 1
4832
- */
4364
+ /** The confidence score for the transcript of this sentence */
4833
4365
  confidence: number;
4834
4366
  /** An array of words in the sentence */
4835
4367
  words: TranscriptWord[];
4836
4368
  /** The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
4837
4369
  channel?: TranscriptSentenceChannel;
4838
- /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/models/speaker-diarization) is enabled, else null */
4370
+ /** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
4839
4371
  speaker: TranscriptSentenceSpeaker;
4840
4372
  }
4841
4373
 
@@ -4843,18 +4375,14 @@ interface TranscriptSentence {
4843
4375
  * Generated by orval v7.9.0 🍺
4844
4376
  * Do not edit manually.
4845
4377
  * AssemblyAI API
4846
- * AssemblyAI API
4847
- * OpenAPI spec version: 1.3.4
4378
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4379
+ * OpenAPI spec version: 1.0.0
4848
4380
  */
4849
4381
 
4850
4382
  interface SentencesResponse {
4851
4383
  /** The unique identifier for the transcript */
4852
4384
  id: string;
4853
- /**
4854
- * The confidence score for the transcript
4855
- * @minimum 0
4856
- * @maximum 1
4857
- */
4385
+ /** The confidence score for the transcript */
4858
4386
  confidence: number;
4859
4387
  /** The duration of the audio file in seconds */
4860
4388
  audio_duration: number;
@@ -4866,20 +4394,8 @@ interface SentencesResponse {
4866
4394
  * Generated by orval v7.9.0 🍺
4867
4395
  * Do not edit manually.
4868
4396
  * AssemblyAI API
4869
- * AssemblyAI API
4870
- * OpenAPI spec version: 1.3.4
4871
- */
4872
- /**
4873
- * Service unavailable
4874
- */
4875
- type ServiceUnavailableResponse = void;
4876
-
4877
- /**
4878
- * Generated by orval v7.9.0 🍺
4879
- * Do not edit manually.
4880
- * AssemblyAI API
4881
- * AssemblyAI API
4882
- * OpenAPI spec version: 1.3.4
4397
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4398
+ * OpenAPI spec version: 1.0.0
4883
4399
  */
4884
4400
  /**
4885
4401
  * Format of the subtitles
@@ -4890,17 +4406,12 @@ declare const SubtitleFormat: {
4890
4406
  readonly vtt: "vtt";
4891
4407
  };
4892
4408
 
4893
- /**
4894
- * Too many requests
4895
- */
4896
- type TooManyRequestsResponse = Error;
4897
-
4898
4409
  /**
4899
4410
  * Generated by orval v7.9.0 🍺
4900
4411
  * Do not edit manually.
4901
4412
  * AssemblyAI API
4902
- * AssemblyAI API
4903
- * OpenAPI spec version: 1.3.4
4413
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4414
+ * OpenAPI spec version: 1.0.0
4904
4415
  */
4905
4416
 
4906
4417
  /**
@@ -4913,17 +4424,12 @@ interface TranscriptList {
4913
4424
  transcripts: TranscriptListItem[];
4914
4425
  }
4915
4426
 
4916
- /**
4917
- * Unauthorized
4918
- */
4919
- type UnauthorizedResponse = Error;
4920
-
4921
4427
  /**
4922
4428
  * Generated by orval v7.9.0 🍺
4923
4429
  * Do not edit manually.
4924
4430
  * AssemblyAI API
4925
- * AssemblyAI API
4926
- * OpenAPI spec version: 1.3.4
4431
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4432
+ * OpenAPI spec version: 1.0.0
4927
4433
  */
4928
4434
  interface UploadedFile {
4929
4435
  /** A URL that points to your audio file, accessible only by AssemblyAI's servers
@@ -4935,8 +4441,8 @@ interface UploadedFile {
4935
4441
  * Generated by orval v7.9.0 🍺
4936
4442
  * Do not edit manually.
4937
4443
  * AssemblyAI API
4938
- * AssemblyAI API
4939
- * OpenAPI spec version: 1.3.4
4444
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4445
+ * OpenAPI spec version: 1.0.0
4940
4446
  */
4941
4447
  /**
4942
4448
  * An array of timestamps structured as [`start_time`, `end_time`] in milliseconds
@@ -4947,8 +4453,8 @@ type WordSearchTimestamp = number[];
4947
4453
  * Generated by orval v7.9.0 🍺
4948
4454
  * Do not edit manually.
4949
4455
  * AssemblyAI API
4950
- * AssemblyAI API
4951
- * OpenAPI spec version: 1.3.4
4456
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4457
+ * OpenAPI spec version: 1.0.0
4952
4458
  */
4953
4459
 
4954
4460
  interface WordSearchMatch {
@@ -4966,8 +4472,8 @@ interface WordSearchMatch {
4966
4472
  * Generated by orval v7.9.0 🍺
4967
4473
  * Do not edit manually.
4968
4474
  * AssemblyAI API
4969
- * AssemblyAI API
4970
- * OpenAPI spec version: 1.3.4
4475
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4476
+ * OpenAPI spec version: 1.0.0
4971
4477
  */
4972
4478
  type WordSearchParams = {
4973
4479
  /**
@@ -4980,8 +4486,8 @@ type WordSearchParams = {
4980
4486
  * Generated by orval v7.9.0 🍺
4981
4487
  * Do not edit manually.
4982
4488
  * AssemblyAI API
4983
- * AssemblyAI API
4984
- * OpenAPI spec version: 1.3.4
4489
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4490
+ * OpenAPI spec version: 1.0.0
4985
4491
  */
4986
4492
 
4987
4493
  interface WordSearchResponse {
@@ -4997,93 +4503,66 @@ interface WordSearchResponse {
4997
4503
  * Generated by orval v7.9.0 🍺
4998
4504
  * Do not edit manually.
4999
4505
  * AssemblyAI API
5000
- * AssemblyAI API
5001
- * OpenAPI spec version: 1.3.4
4506
+ * AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
4507
+ * OpenAPI spec version: 1.0.0
5002
4508
  */
5003
4509
 
5004
4510
  declare const index$6_AudioIntelligenceModelStatus: typeof AudioIntelligenceModelStatus;
5005
4511
  declare const index$6_AutoHighlightResult: typeof AutoHighlightResult;
5006
4512
  declare const index$6_AutoHighlightsResult: typeof AutoHighlightsResult;
5007
- type index$6_BadRequestResponse = BadRequestResponse;
5008
- type index$6_CannotAccessUploadedFileResponse = CannotAccessUploadedFileResponse;
5009
4513
  declare const index$6_Chapter: typeof Chapter;
5010
4514
  declare const index$6_ContentSafetyLabel: typeof ContentSafetyLabel;
5011
4515
  declare const index$6_ContentSafetyLabelResult: typeof ContentSafetyLabelResult;
5012
4516
  declare const index$6_ContentSafetyLabelsResult: typeof ContentSafetyLabelsResult;
5013
4517
  declare const index$6_ContentSafetyLabelsResultSeverityScoreSummary: typeof ContentSafetyLabelsResultSeverityScoreSummary;
5014
4518
  declare const index$6_ContentSafetyLabelsResultSummary: typeof ContentSafetyLabelsResultSummary;
5015
- type index$6_CreateRealtimeTemporaryTokenParams = CreateRealtimeTemporaryTokenParams;
4519
+ declare const index$6_CustomFormattingRequestBody: typeof CustomFormattingRequestBody;
4520
+ declare const index$6_CustomFormattingRequestBodyCustomFormatting: typeof CustomFormattingRequestBodyCustomFormatting;
4521
+ declare const index$6_CustomFormattingResponse: typeof CustomFormattingResponse;
4522
+ declare const index$6_CustomFormattingResponseCustomFormatting: typeof CustomFormattingResponseCustomFormatting;
4523
+ declare const index$6_CustomFormattingResponseCustomFormattingMapping: typeof CustomFormattingResponseCustomFormattingMapping;
5016
4524
  declare const index$6_Entity: typeof Entity;
5017
4525
  declare const index$6_EntityType: typeof EntityType;
5018
- type index$6_GatewayTimeoutResponse = GatewayTimeoutResponse;
4526
+ type index$6_ErrorDetails = ErrorDetails;
4527
+ type index$6_GetSubtitles200 = GetSubtitles200;
5019
4528
  type index$6_GetSubtitlesParams = GetSubtitlesParams;
5020
- type index$6_InternalServerErrorResponse = InternalServerErrorResponse;
5021
- type index$6_LemurActionItemsParams = LemurActionItemsParams;
5022
- type index$6_LemurActionItemsParamsAllOf = LemurActionItemsParamsAllOf;
5023
- type index$6_LemurActionItemsResponse = LemurActionItemsResponse;
5024
- type index$6_LemurBaseParams = LemurBaseParams;
5025
- type index$6_LemurBaseParamsContext = LemurBaseParamsContext;
5026
- type index$6_LemurBaseParamsContextOneOf = LemurBaseParamsContextOneOf;
5027
- type index$6_LemurBaseParamsFinalModel = LemurBaseParamsFinalModel;
5028
- type index$6_LemurBaseResponse = LemurBaseResponse;
5029
- declare const index$6_LemurModel: typeof LemurModel;
5030
- type index$6_LemurQuestion = LemurQuestion;
5031
- type index$6_LemurQuestionAnswer = LemurQuestionAnswer;
5032
- type index$6_LemurQuestionAnswerParams = LemurQuestionAnswerParams;
5033
- type index$6_LemurQuestionAnswerParamsAllOf = LemurQuestionAnswerParamsAllOf;
5034
- type index$6_LemurQuestionAnswerResponse = LemurQuestionAnswerResponse;
5035
- type index$6_LemurQuestionAnswerResponseAllOf = LemurQuestionAnswerResponseAllOf;
5036
- type index$6_LemurQuestionContext = LemurQuestionContext;
5037
- type index$6_LemurQuestionContextOneOf = LemurQuestionContextOneOf;
5038
- type index$6_LemurResponse = LemurResponse;
5039
- type index$6_LemurStringResponse = LemurStringResponse;
5040
- type index$6_LemurStringResponseAllOf = LemurStringResponseAllOf;
5041
- type index$6_LemurSummaryParams = LemurSummaryParams;
5042
- type index$6_LemurSummaryParamsAllOf = LemurSummaryParamsAllOf;
5043
- type index$6_LemurSummaryResponse = LemurSummaryResponse;
5044
- type index$6_LemurTaskParams = LemurTaskParams;
5045
- type index$6_LemurTaskParamsAllOf = LemurTaskParamsAllOf;
5046
- type index$6_LemurTaskResponse = LemurTaskResponse;
5047
- type index$6_LemurUsage = LemurUsage;
5048
- type index$6_ListTranscriptParams = ListTranscriptParams;
5049
4529
  declare const index$6_ListTranscriptsParams: typeof ListTranscriptsParams;
5050
- type index$6_NotFoundResponse = NotFoundResponse;
5051
4530
  type index$6_PageDetails = PageDetails;
5052
4531
  type index$6_PageDetailsNextUrl = PageDetailsNextUrl;
5053
4532
  type index$6_PageDetailsPrevUrl = PageDetailsPrevUrl;
5054
4533
  type index$6_ParagraphsResponse = ParagraphsResponse;
5055
4534
  declare const index$6_PiiPolicy: typeof PiiPolicy;
5056
- type index$6_PurgeLemurRequestDataResponse = PurgeLemurRequestDataResponse;
5057
- type index$6_RealtimeTemporaryTokenResponse = RealtimeTemporaryTokenResponse;
5058
4535
  declare const index$6_RedactPiiAudioQuality: typeof RedactPiiAudioQuality;
5059
- declare const index$6_RedactedAudioNotification: typeof RedactedAudioNotification;
5060
- declare const index$6_RedactedAudioResponse: typeof RedactedAudioResponse;
4536
+ type index$6_RedactedAudioResponse = RedactedAudioResponse;
5061
4537
  declare const index$6_RedactedAudioStatus: typeof RedactedAudioStatus;
5062
4538
  type index$6_SentencesResponse = SentencesResponse;
5063
4539
  declare const index$6_Sentiment: typeof Sentiment;
5064
4540
  declare const index$6_SentimentAnalysisResultChannel: typeof SentimentAnalysisResultChannel;
5065
4541
  declare const index$6_SentimentAnalysisResultSpeaker: typeof SentimentAnalysisResultSpeaker;
5066
- type index$6_ServiceUnavailableResponse = ServiceUnavailableResponse;
5067
4542
  declare const index$6_SeverityScoreSummary: typeof SeverityScoreSummary;
4543
+ declare const index$6_SpeakerIdentificationRequestBody: typeof SpeakerIdentificationRequestBody;
4544
+ declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentification: typeof SpeakerIdentificationRequestBodySpeakerIdentification;
4545
+ declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType;
4546
+ declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems;
4547
+ declare const index$6_SpeakerIdentificationResponse: typeof SpeakerIdentificationResponse;
4548
+ declare const index$6_SpeakerIdentificationResponseSpeakerIdentification: typeof SpeakerIdentificationResponseSpeakerIdentification;
4549
+ declare const index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping: typeof SpeakerIdentificationResponseSpeakerIdentificationMapping;
5068
4550
  declare const index$6_SpeechModel: typeof SpeechModel;
5069
4551
  declare const index$6_SubstitutionPolicy: typeof SubstitutionPolicy;
5070
4552
  declare const index$6_SubtitleFormat: typeof SubtitleFormat;
5071
4553
  declare const index$6_SummaryModel: typeof SummaryModel;
5072
4554
  declare const index$6_SummaryType: typeof SummaryType;
5073
4555
  declare const index$6_Timestamp: typeof Timestamp;
5074
- type index$6_TooManyRequestsResponse = TooManyRequestsResponse;
5075
4556
  declare const index$6_TopicDetectionModelResult: typeof TopicDetectionModelResult;
5076
4557
  declare const index$6_TopicDetectionModelResultSummary: typeof TopicDetectionModelResultSummary;
5077
4558
  declare const index$6_TopicDetectionResult: typeof TopicDetectionResult;
5078
- declare const index$6_TopicDetectionResultLabelsItem: typeof TopicDetectionResultLabelsItem;
4559
+ declare const index$6_TopicDetectionResultLabelsItems: typeof TopicDetectionResultLabelsItems;
5079
4560
  declare const index$6_Transcript: typeof Transcript;
5080
4561
  declare const index$6_TranscriptAudioDuration: typeof TranscriptAudioDuration;
5081
4562
  declare const index$6_TranscriptAudioEndAt: typeof TranscriptAudioEndAt;
5082
4563
  declare const index$6_TranscriptAudioStartFrom: typeof TranscriptAudioStartFrom;
5083
4564
  declare const index$6_TranscriptAutoChapters: typeof TranscriptAutoChapters;
5084
4565
  declare const index$6_TranscriptAutoHighlightsResult: typeof TranscriptAutoHighlightsResult;
5085
- declare const index$6_TranscriptBoostParam: typeof TranscriptBoostParam;
5086
- declare const index$6_TranscriptBoostParamProperty: typeof TranscriptBoostParamProperty;
5087
4566
  declare const index$6_TranscriptChapters: typeof TranscriptChapters;
5088
4567
  declare const index$6_TranscriptConfidence: typeof TranscriptConfidence;
5089
4568
  declare const index$6_TranscriptContentSafety: typeof TranscriptContentSafety;
@@ -5092,6 +4571,7 @@ declare const index$6_TranscriptCustomSpelling: typeof TranscriptCustomSpelling;
5092
4571
  declare const index$6_TranscriptCustomSpellingProperty: typeof TranscriptCustomSpellingProperty;
5093
4572
  declare const index$6_TranscriptCustomTopics: typeof TranscriptCustomTopics;
5094
4573
  declare const index$6_TranscriptDisfluencies: typeof TranscriptDisfluencies;
4574
+ declare const index$6_TranscriptDomain: typeof TranscriptDomain;
5095
4575
  declare const index$6_TranscriptEntities: typeof TranscriptEntities;
5096
4576
  declare const index$6_TranscriptEntityDetection: typeof TranscriptEntityDetection;
5097
4577
  declare const index$6_TranscriptFilterProfanity: typeof TranscriptFilterProfanity;
@@ -5099,33 +4579,41 @@ declare const index$6_TranscriptFormatText: typeof TranscriptFormatText;
5099
4579
  declare const index$6_TranscriptIabCategories: typeof TranscriptIabCategories;
5100
4580
  declare const index$6_TranscriptIabCategoriesResult: typeof TranscriptIabCategoriesResult;
5101
4581
  declare const index$6_TranscriptLanguageCode: typeof TranscriptLanguageCode;
5102
- declare const index$6_TranscriptLanguageCodeProperty: typeof TranscriptLanguageCodeProperty;
4582
+ declare const index$6_TranscriptLanguageCodes: typeof TranscriptLanguageCodes;
5103
4583
  declare const index$6_TranscriptLanguageConfidence: typeof TranscriptLanguageConfidence;
5104
4584
  declare const index$6_TranscriptLanguageConfidenceThreshold: typeof TranscriptLanguageConfidenceThreshold;
5105
4585
  declare const index$6_TranscriptLanguageDetection: typeof TranscriptLanguageDetection;
4586
+ declare const index$6_TranscriptLanguageDetectionOptions: typeof TranscriptLanguageDetectionOptions;
5106
4587
  type index$6_TranscriptList = TranscriptList;
5107
4588
  type index$6_TranscriptListItem = TranscriptListItem;
5108
- type index$6_TranscriptListItemCompleted = TranscriptListItemCompleted;
5109
4589
  type index$6_TranscriptListItemError = TranscriptListItemError;
5110
4590
  declare const index$6_TranscriptMultichannel: typeof TranscriptMultichannel;
5111
- declare const index$6_TranscriptOptionalParams: typeof TranscriptOptionalParams;
5112
- declare const index$6_TranscriptOptionalParamsLanguageCode: typeof TranscriptOptionalParamsLanguageCode;
5113
- declare const index$6_TranscriptOptionalParamsLanguageCodeOneOf: typeof TranscriptOptionalParamsLanguageCodeOneOf;
5114
- declare const index$6_TranscriptOptionalParamsRedactPiiSub: typeof TranscriptOptionalParamsRedactPiiSub;
5115
- declare const index$6_TranscriptOptionalParamsSpeakersExpected: typeof TranscriptOptionalParamsSpeakersExpected;
5116
- declare const index$6_TranscriptOptionalParamsSpeechModel: typeof TranscriptOptionalParamsSpeechModel;
5117
- declare const index$6_TranscriptOptionalParamsSpeechThreshold: typeof TranscriptOptionalParamsSpeechThreshold;
5118
- declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderName: typeof TranscriptOptionalParamsWebhookAuthHeaderName;
5119
- declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderValue: typeof TranscriptOptionalParamsWebhookAuthHeaderValue;
4591
+ declare const index$6_TranscriptOptionalParamsLanguageDetectionOptions: typeof TranscriptOptionalParamsLanguageDetectionOptions;
4592
+ declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptions: typeof TranscriptOptionalParamsRedactPiiAudioOptions;
4593
+ declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod;
4594
+ declare const index$6_TranscriptOptionalParamsRemoveAudioTags: typeof TranscriptOptionalParamsRemoveAudioTags;
4595
+ declare const index$6_TranscriptOptionalParamsSpeakerOptions: typeof TranscriptOptionalParamsSpeakerOptions;
4596
+ declare const index$6_TranscriptOptionalParamsSpeechUnderstanding: typeof TranscriptOptionalParamsSpeechUnderstanding;
4597
+ declare const index$6_TranscriptOptionalParamsSpeechUnderstandingRequest: typeof TranscriptOptionalParamsSpeechUnderstandingRequest;
5120
4598
  type index$6_TranscriptParagraph = TranscriptParagraph;
5121
- type index$6_TranscriptParams = TranscriptParams;
5122
- type index$6_TranscriptParamsAllOf = TranscriptParamsAllOf;
4599
+ declare const index$6_TranscriptParams: typeof TranscriptParams;
4600
+ declare const index$6_TranscriptParamsDomain: typeof TranscriptParamsDomain;
4601
+ declare const index$6_TranscriptParamsLanguageCode: typeof TranscriptParamsLanguageCode;
4602
+ declare const index$6_TranscriptParamsLanguageCodes: typeof TranscriptParamsLanguageCodes;
4603
+ declare const index$6_TranscriptParamsRedactPiiSub: typeof TranscriptParamsRedactPiiSub;
4604
+ declare const index$6_TranscriptParamsRemoveAudioTags: typeof TranscriptParamsRemoveAudioTags;
4605
+ declare const index$6_TranscriptParamsSpeakersExpected: typeof TranscriptParamsSpeakersExpected;
4606
+ declare const index$6_TranscriptParamsSpeechThreshold: typeof TranscriptParamsSpeechThreshold;
4607
+ declare const index$6_TranscriptParamsWebhookAuthHeaderName: typeof TranscriptParamsWebhookAuthHeaderName;
4608
+ declare const index$6_TranscriptParamsWebhookAuthHeaderValue: typeof TranscriptParamsWebhookAuthHeaderValue;
5123
4609
  declare const index$6_TranscriptPunctuate: typeof TranscriptPunctuate;
5124
- declare const index$6_TranscriptReadyNotification: typeof TranscriptReadyNotification;
5125
- declare const index$6_TranscriptReadyStatus: typeof TranscriptReadyStatus;
5126
4610
  declare const index$6_TranscriptRedactPiiAudio: typeof TranscriptRedactPiiAudio;
4611
+ declare const index$6_TranscriptRedactPiiAudioOptions: typeof TranscriptRedactPiiAudioOptions;
4612
+ declare const index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod;
5127
4613
  declare const index$6_TranscriptRedactPiiAudioQuality: typeof TranscriptRedactPiiAudioQuality;
5128
4614
  declare const index$6_TranscriptRedactPiiPolicies: typeof TranscriptRedactPiiPolicies;
4615
+ declare const index$6_TranscriptRemoveAudioTags: typeof TranscriptRemoveAudioTags;
4616
+ declare const index$6_TranscriptRemoveAudioTagsProperty: typeof TranscriptRemoveAudioTagsProperty;
5129
4617
  type index$6_TranscriptSentence = TranscriptSentence;
5130
4618
  type index$6_TranscriptSentenceChannel = TranscriptSentenceChannel;
5131
4619
  type index$6_TranscriptSentenceSpeaker = TranscriptSentenceSpeaker;
@@ -5134,33 +4622,42 @@ declare const index$6_TranscriptSentimentAnalysisResults: typeof TranscriptSenti
5134
4622
  declare const index$6_TranscriptSpeakerLabels: typeof TranscriptSpeakerLabels;
5135
4623
  declare const index$6_TranscriptSpeakersExpected: typeof TranscriptSpeakersExpected;
5136
4624
  declare const index$6_TranscriptSpeechModel: typeof TranscriptSpeechModel;
4625
+ declare const index$6_TranscriptSpeechModels: typeof TranscriptSpeechModels;
5137
4626
  declare const index$6_TranscriptSpeechThreshold: typeof TranscriptSpeechThreshold;
4627
+ declare const index$6_TranscriptSpeechUnderstanding: typeof TranscriptSpeechUnderstanding;
4628
+ declare const index$6_TranscriptSpeechUnderstandingRequest: typeof TranscriptSpeechUnderstandingRequest;
4629
+ declare const index$6_TranscriptSpeechUnderstandingResponse: typeof TranscriptSpeechUnderstandingResponse;
5138
4630
  declare const index$6_TranscriptSpeedBoost: typeof TranscriptSpeedBoost;
5139
4631
  declare const index$6_TranscriptStatus: typeof TranscriptStatus;
5140
4632
  declare const index$6_TranscriptSummary: typeof TranscriptSummary;
5141
4633
  declare const index$6_TranscriptSummaryModel: typeof TranscriptSummaryModel;
5142
4634
  declare const index$6_TranscriptSummaryType: typeof TranscriptSummaryType;
4635
+ declare const index$6_TranscriptTemperature: typeof TranscriptTemperature;
5143
4636
  declare const index$6_TranscriptText: typeof TranscriptText;
5144
4637
  declare const index$6_TranscriptThrottled: typeof TranscriptThrottled;
4638
+ declare const index$6_TranscriptTranslatedTexts: typeof TranscriptTranslatedTexts;
5145
4639
  declare const index$6_TranscriptUtterance: typeof TranscriptUtterance;
5146
4640
  declare const index$6_TranscriptUtteranceChannel: typeof TranscriptUtteranceChannel;
4641
+ declare const index$6_TranscriptUtteranceTranslatedTexts: typeof TranscriptUtteranceTranslatedTexts;
5147
4642
  declare const index$6_TranscriptUtterances: typeof TranscriptUtterances;
5148
4643
  declare const index$6_TranscriptWebhookAuthHeaderName: typeof TranscriptWebhookAuthHeaderName;
5149
- declare const index$6_TranscriptWebhookNotification: typeof TranscriptWebhookNotification;
5150
4644
  declare const index$6_TranscriptWebhookStatusCode: typeof TranscriptWebhookStatusCode;
5151
4645
  declare const index$6_TranscriptWebhookUrl: typeof TranscriptWebhookUrl;
5152
4646
  declare const index$6_TranscriptWord: typeof TranscriptWord;
5153
4647
  declare const index$6_TranscriptWordChannel: typeof TranscriptWordChannel;
5154
4648
  declare const index$6_TranscriptWordSpeaker: typeof TranscriptWordSpeaker;
5155
4649
  declare const index$6_TranscriptWords: typeof TranscriptWords;
5156
- type index$6_UnauthorizedResponse = UnauthorizedResponse;
4650
+ declare const index$6_TranslationRequestBody: typeof TranslationRequestBody;
4651
+ declare const index$6_TranslationRequestBodyTranslation: typeof TranslationRequestBodyTranslation;
4652
+ declare const index$6_TranslationResponse: typeof TranslationResponse;
4653
+ declare const index$6_TranslationResponseTranslation: typeof TranslationResponseTranslation;
5157
4654
  type index$6_UploadedFile = UploadedFile;
5158
4655
  type index$6_WordSearchMatch = WordSearchMatch;
5159
4656
  type index$6_WordSearchParams = WordSearchParams;
5160
4657
  type index$6_WordSearchResponse = WordSearchResponse;
5161
4658
  type index$6_WordSearchTimestamp = WordSearchTimestamp;
5162
4659
  declare namespace index$6 {
5163
- export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, type index$6_BadRequestResponse as BadRequestResponse, type index$6_CannotAccessUploadedFileResponse as CannotAccessUploadedFileResponse, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, type index$6_CreateRealtimeTemporaryTokenParams as CreateRealtimeTemporaryTokenParams, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_GatewayTimeoutResponse as GatewayTimeoutResponse, type index$6_GetSubtitlesParams as GetSubtitlesParams, type index$6_InternalServerErrorResponse as InternalServerErrorResponse, type index$6_LemurActionItemsParams as LemurActionItemsParams, type index$6_LemurActionItemsParamsAllOf as LemurActionItemsParamsAllOf, type index$6_LemurActionItemsResponse as LemurActionItemsResponse, type index$6_LemurBaseParams as LemurBaseParams, type index$6_LemurBaseParamsContext as LemurBaseParamsContext, type index$6_LemurBaseParamsContextOneOf as LemurBaseParamsContextOneOf, type index$6_LemurBaseParamsFinalModel as LemurBaseParamsFinalModel, type index$6_LemurBaseResponse as LemurBaseResponse, index$6_LemurModel as LemurModel, type index$6_LemurQuestion as LemurQuestion, type index$6_LemurQuestionAnswer as LemurQuestionAnswer, type index$6_LemurQuestionAnswerParams as LemurQuestionAnswerParams, type index$6_LemurQuestionAnswerParamsAllOf as LemurQuestionAnswerParamsAllOf, type index$6_LemurQuestionAnswerResponse as LemurQuestionAnswerResponse, type index$6_LemurQuestionAnswerResponseAllOf as LemurQuestionAnswerResponseAllOf, type index$6_LemurQuestionContext as LemurQuestionContext, type index$6_LemurQuestionContextOneOf as LemurQuestionContextOneOf, type index$6_LemurResponse as LemurResponse, type index$6_LemurStringResponse as LemurStringResponse, type index$6_LemurStringResponseAllOf as LemurStringResponseAllOf, type index$6_LemurSummaryParams as LemurSummaryParams, type index$6_LemurSummaryParamsAllOf as LemurSummaryParamsAllOf, type index$6_LemurSummaryResponse as LemurSummaryResponse, type index$6_LemurTaskParams as LemurTaskParams, type index$6_LemurTaskParamsAllOf as LemurTaskParamsAllOf, type index$6_LemurTaskResponse as LemurTaskResponse, type index$6_LemurUsage as LemurUsage, type index$6_ListTranscriptParams as ListTranscriptParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_NotFoundResponse as NotFoundResponse, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, type index$6_PurgeLemurRequestDataResponse as PurgeLemurRequestDataResponse, type index$6_RealtimeTemporaryTokenResponse as RealtimeTemporaryTokenResponse, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, index$6_RedactedAudioNotification as RedactedAudioNotification, index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, type index$6_ServiceUnavailableResponse as ServiceUnavailableResponse, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, type index$6_TooManyRequestsResponse as TooManyRequestsResponse, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItem as TopicDetectionResultLabelsItem, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptBoostParam as TranscriptBoostParam, index$6_TranscriptBoostParamProperty as TranscriptBoostParamProperty, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodeProperty as TranscriptLanguageCodeProperty, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemCompleted as TranscriptListItemCompleted, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParams as TranscriptOptionalParams, index$6_TranscriptOptionalParamsLanguageCode as TranscriptOptionalParamsLanguageCode, index$6_TranscriptOptionalParamsLanguageCodeOneOf as TranscriptOptionalParamsLanguageCodeOneOf, index$6_TranscriptOptionalParamsRedactPiiSub as TranscriptOptionalParamsRedactPiiSub, index$6_TranscriptOptionalParamsSpeakersExpected as TranscriptOptionalParamsSpeakersExpected, index$6_TranscriptOptionalParamsSpeechModel as TranscriptOptionalParamsSpeechModel, index$6_TranscriptOptionalParamsSpeechThreshold as TranscriptOptionalParamsSpeechThreshold, index$6_TranscriptOptionalParamsWebhookAuthHeaderName as TranscriptOptionalParamsWebhookAuthHeaderName, index$6_TranscriptOptionalParamsWebhookAuthHeaderValue as TranscriptOptionalParamsWebhookAuthHeaderValue, type index$6_TranscriptParagraph as TranscriptParagraph, type index$6_TranscriptParams as TranscriptParams, type index$6_TranscriptParamsAllOf as TranscriptParamsAllOf, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptReadyNotification as TranscriptReadyNotification, index$6_TranscriptReadyStatus as TranscriptReadyStatus, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookNotification as TranscriptWebhookNotification, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, type index$6_UnauthorizedResponse as UnauthorizedResponse, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
4660
+ export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, index$6_CustomFormattingRequestBody as CustomFormattingRequestBody, index$6_CustomFormattingRequestBodyCustomFormatting as CustomFormattingRequestBodyCustomFormatting, index$6_CustomFormattingResponse as CustomFormattingResponse, index$6_CustomFormattingResponseCustomFormatting as CustomFormattingResponseCustomFormatting, index$6_CustomFormattingResponseCustomFormattingMapping as CustomFormattingResponseCustomFormattingMapping, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_ErrorDetails as ErrorDetails, type index$6_GetSubtitles200 as GetSubtitles200, type index$6_GetSubtitlesParams as GetSubtitlesParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, type index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeakerIdentificationRequestBody as SpeakerIdentificationRequestBody, index$6_SpeakerIdentificationRequestBodySpeakerIdentification as SpeakerIdentificationRequestBodySpeakerIdentification, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, index$6_SpeakerIdentificationResponse as SpeakerIdentificationResponse, index$6_SpeakerIdentificationResponseSpeakerIdentification as SpeakerIdentificationResponseSpeakerIdentification, index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping as SpeakerIdentificationResponseSpeakerIdentificationMapping, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItems as TopicDetectionResultLabelsItems, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptDomain as TranscriptDomain, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodes as TranscriptLanguageCodes, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, index$6_TranscriptLanguageDetectionOptions as TranscriptLanguageDetectionOptions, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParamsLanguageDetectionOptions as TranscriptOptionalParamsLanguageDetectionOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptions as TranscriptOptionalParamsRedactPiiAudioOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptOptionalParamsRemoveAudioTags as TranscriptOptionalParamsRemoveAudioTags, index$6_TranscriptOptionalParamsSpeakerOptions as TranscriptOptionalParamsSpeakerOptions, index$6_TranscriptOptionalParamsSpeechUnderstanding as TranscriptOptionalParamsSpeechUnderstanding, index$6_TranscriptOptionalParamsSpeechUnderstandingRequest as TranscriptOptionalParamsSpeechUnderstandingRequest, type index$6_TranscriptParagraph as TranscriptParagraph, index$6_TranscriptParams as TranscriptParams, index$6_TranscriptParamsDomain as TranscriptParamsDomain, index$6_TranscriptParamsLanguageCode as TranscriptParamsLanguageCode, index$6_TranscriptParamsLanguageCodes as TranscriptParamsLanguageCodes, index$6_TranscriptParamsRedactPiiSub as TranscriptParamsRedactPiiSub, index$6_TranscriptParamsRemoveAudioTags as TranscriptParamsRemoveAudioTags, index$6_TranscriptParamsSpeakersExpected as TranscriptParamsSpeakersExpected, index$6_TranscriptParamsSpeechThreshold as TranscriptParamsSpeechThreshold, index$6_TranscriptParamsWebhookAuthHeaderName as TranscriptParamsWebhookAuthHeaderName, index$6_TranscriptParamsWebhookAuthHeaderValue as TranscriptParamsWebhookAuthHeaderValue, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioOptions as TranscriptRedactPiiAudioOptions, index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, index$6_TranscriptRemoveAudioTags as TranscriptRemoveAudioTags, index$6_TranscriptRemoveAudioTagsProperty as TranscriptRemoveAudioTagsProperty, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechModels as TranscriptSpeechModels, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeechUnderstanding as TranscriptSpeechUnderstanding, index$6_TranscriptSpeechUnderstandingRequest as TranscriptSpeechUnderstandingRequest, index$6_TranscriptSpeechUnderstandingResponse as TranscriptSpeechUnderstandingResponse, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptTemperature as TranscriptTemperature, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptTranslatedTexts as TranscriptTranslatedTexts, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtteranceTranslatedTexts as TranscriptUtteranceTranslatedTexts, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, index$6_TranslationRequestBody as TranslationRequestBody, index$6_TranslationRequestBodyTranslation as TranslationRequestBodyTranslation, index$6_TranslationResponse as TranslationResponse, index$6_TranslationResponseTranslation as TranslationResponseTranslation, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
5164
4661
  }
5165
4662
 
5166
4663
  /**
@@ -5177,6 +4674,15 @@ Transcripts are sorted from newest to oldest and can be retrieved for the last 9
5177
4674
 
5178
4675
  If you need to retrieve transcripts from more than 90 days ago please reach out to our Support team at support@assemblyai.com.
5179
4676
 
4677
+ **Pagination**
4678
+
4679
+ This endpoint returns paginated results. The response includes a `page_details` object with the following properties:
4680
+ - `page_details.limit` - Maximum number of transcripts per page.
4681
+ - `page_details.result_count` - Total number of transcripts returned on the current page.
4682
+ - `page_details.current_url` - URL to the current page.
4683
+ - `page_details.prev_url` - URL to the previous page of older transcripts.
4684
+ - `page_details.next_url` - URL to the next page of newer transcripts.
4685
+
5180
4686
  * @summary List transcripts
5181
4687
  */
5182
4688
  declare const listTranscripts: <TData = AxiosResponse<TranscriptList>>(params?: ListTranscriptsParams, options?: AxiosRequestConfig) => Promise<TData>;
@@ -5195,14 +4701,6 @@ Remove the data from the transcript and mark it as deleted.
5195
4701
  * @summary Delete transcript
5196
4702
  */
5197
4703
  declare const deleteTranscript: <TData = AxiosResponse<Transcript>>(transcriptId: string, options?: AxiosRequestConfig) => Promise<TData>;
5198
- /**
5199
- * <Warning>Streaming Speech-to-Text is currently not available on the EU endpoint.</Warning>
5200
- <Note>Any usage associated with a temporary token will be attributed to the API key that generated it.</Note>
5201
- Create a temporary authentication token for Streaming Speech-to-Text
5202
-
5203
- * @summary Create temporary authentication token for Streaming STT
5204
- */
5205
- declare const createTemporaryToken: <TData = AxiosResponse<RealtimeTemporaryTokenResponse>>(createRealtimeTemporaryTokenParams: CreateRealtimeTemporaryTokenParams, options?: AxiosRequestConfig) => Promise<TData>;
5206
4704
 
5207
4705
  /**
5208
4706
  * AssemblyAI transcription provider adapter
@@ -15349,6 +14847,24 @@ type ManageV1FilterVersionParameter = boolean;
15349
14847
  /**
15350
14848
  * ManageV1GroupingParameter type definition
15351
14849
  */
14850
+ /**
14851
+ * ManageV1GroupingParameter type definition
14852
+ */
14853
+ /**
14854
+ * ManageV1GroupingParameter type definition
14855
+ */
14856
+ /**
14857
+ * ManageV1GroupingParameter type definition
14858
+ */
14859
+ /**
14860
+ * ManageV1GroupingParameter type definition
14861
+ */
14862
+ /**
14863
+ * ManageV1GroupingParameter type definition
14864
+ */
14865
+ /**
14866
+ * ManageV1GroupingParameter type definition
14867
+ */
15352
14868
  type ManageV1GroupingParameter = typeof ManageV1GroupingParameter[keyof typeof ManageV1GroupingParameter];
15353
14869
  declare const ManageV1GroupingParameter: {
15354
14870
  readonly accessor: "accessor";
@@ -15421,6 +14937,24 @@ type ManageV1ModelsListParams = {
15421
14937
  /**
15422
14938
  * ManageV1ProjectKeysStatusParameter type definition
15423
14939
  */
14940
+ /**
14941
+ * ManageV1ProjectKeysStatusParameter type definition
14942
+ */
14943
+ /**
14944
+ * ManageV1ProjectKeysStatusParameter type definition
14945
+ */
14946
+ /**
14947
+ * ManageV1ProjectKeysStatusParameter type definition
14948
+ */
14949
+ /**
14950
+ * ManageV1ProjectKeysStatusParameter type definition
14951
+ */
14952
+ /**
14953
+ * ManageV1ProjectKeysStatusParameter type definition
14954
+ */
14955
+ /**
14956
+ * ManageV1ProjectKeysStatusParameter type definition
14957
+ */
15424
14958
  type ManageV1ProjectKeysStatusParameter = typeof ManageV1ProjectKeysStatusParameter[keyof typeof ManageV1ProjectKeysStatusParameter];
15425
14959
  declare const ManageV1ProjectKeysStatusParameter: {
15426
14960
  readonly active: "active";
@@ -16306,6 +15840,24 @@ type SelfHostedV1ParamsDistributionScopesParameter = SelfHostedV1ParamsDistribut
16306
15840
  /**
16307
15841
  * SelfHostedV1ParamsDistributionProviderParameter type definition
16308
15842
  */
15843
+ /**
15844
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15845
+ */
15846
+ /**
15847
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15848
+ */
15849
+ /**
15850
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15851
+ */
15852
+ /**
15853
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15854
+ */
15855
+ /**
15856
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15857
+ */
15858
+ /**
15859
+ * SelfHostedV1ParamsDistributionProviderParameter type definition
15860
+ */
16309
15861
  type SelfHostedV1ParamsDistributionProviderParameter = typeof SelfHostedV1ParamsDistributionProviderParameter[keyof typeof SelfHostedV1ParamsDistributionProviderParameter];
16310
15862
  declare const SelfHostedV1ParamsDistributionProviderParameter: {
16311
15863
  readonly quay: "quay";
@@ -19533,8 +19085,8 @@ declare const getJobsResponse: z.ZodObject<{
19533
19085
  }, "strip", z.ZodTypeAny, {
19534
19086
  language: string;
19535
19087
  diarization?: "none" | "speaker" | "channel" | undefined;
19536
- operating_point?: "standard" | "enhanced" | undefined;
19537
19088
  domain?: string | undefined;
19089
+ operating_point?: "standard" | "enhanced" | undefined;
19538
19090
  output_locale?: string | undefined;
19539
19091
  additional_vocab?: {
19540
19092
  content: string;
@@ -19561,8 +19113,8 @@ declare const getJobsResponse: z.ZodObject<{
19561
19113
  }, {
19562
19114
  language: string;
19563
19115
  diarization?: "none" | "speaker" | "channel" | undefined;
19564
- operating_point?: "standard" | "enhanced" | undefined;
19565
19116
  domain?: string | undefined;
19117
+ operating_point?: "standard" | "enhanced" | undefined;
19566
19118
  output_locale?: string | undefined;
19567
19119
  additional_vocab?: {
19568
19120
  content: string;
@@ -19720,8 +19272,8 @@ declare const getJobsResponse: z.ZodObject<{
19720
19272
  transcription_config?: {
19721
19273
  language: string;
19722
19274
  diarization?: "none" | "speaker" | "channel" | undefined;
19723
- operating_point?: "standard" | "enhanced" | undefined;
19724
19275
  domain?: string | undefined;
19276
+ operating_point?: "standard" | "enhanced" | undefined;
19725
19277
  output_locale?: string | undefined;
19726
19278
  additional_vocab?: {
19727
19279
  content: string;
@@ -19801,8 +19353,8 @@ declare const getJobsResponse: z.ZodObject<{
19801
19353
  transcription_config?: {
19802
19354
  language: string;
19803
19355
  diarization?: "none" | "speaker" | "channel" | undefined;
19804
- operating_point?: "standard" | "enhanced" | undefined;
19805
19356
  domain?: string | undefined;
19357
+ operating_point?: "standard" | "enhanced" | undefined;
19806
19358
  output_locale?: string | undefined;
19807
19359
  additional_vocab?: {
19808
19360
  content: string;
@@ -19900,8 +19452,8 @@ declare const getJobsResponse: z.ZodObject<{
19900
19452
  transcription_config?: {
19901
19453
  language: string;
19902
19454
  diarization?: "none" | "speaker" | "channel" | undefined;
19903
- operating_point?: "standard" | "enhanced" | undefined;
19904
19455
  domain?: string | undefined;
19456
+ operating_point?: "standard" | "enhanced" | undefined;
19905
19457
  output_locale?: string | undefined;
19906
19458
  additional_vocab?: {
19907
19459
  content: string;
@@ -19994,8 +19546,8 @@ declare const getJobsResponse: z.ZodObject<{
19994
19546
  transcription_config?: {
19995
19547
  language: string;
19996
19548
  diarization?: "none" | "speaker" | "channel" | undefined;
19997
- operating_point?: "standard" | "enhanced" | undefined;
19998
19549
  domain?: string | undefined;
19550
+ operating_point?: "standard" | "enhanced" | undefined;
19999
19551
  output_locale?: string | undefined;
20000
19552
  additional_vocab?: {
20001
19553
  content: string;
@@ -20090,8 +19642,8 @@ declare const getJobsResponse: z.ZodObject<{
20090
19642
  transcription_config?: {
20091
19643
  language: string;
20092
19644
  diarization?: "none" | "speaker" | "channel" | undefined;
20093
- operating_point?: "standard" | "enhanced" | undefined;
20094
19645
  domain?: string | undefined;
19646
+ operating_point?: "standard" | "enhanced" | undefined;
20095
19647
  output_locale?: string | undefined;
20096
19648
  additional_vocab?: {
20097
19649
  content: string;
@@ -20186,8 +19738,8 @@ declare const getJobsResponse: z.ZodObject<{
20186
19738
  transcription_config?: {
20187
19739
  language: string;
20188
19740
  diarization?: "none" | "speaker" | "channel" | undefined;
20189
- operating_point?: "standard" | "enhanced" | undefined;
20190
19741
  domain?: string | undefined;
19742
+ operating_point?: "standard" | "enhanced" | undefined;
20191
19743
  output_locale?: string | undefined;
20192
19744
  additional_vocab?: {
20193
19745
  content: string;
@@ -20368,8 +19920,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20368
19920
  }, "strip", z.ZodTypeAny, {
20369
19921
  language: string;
20370
19922
  diarization?: "none" | "speaker" | "channel" | undefined;
20371
- operating_point?: "standard" | "enhanced" | undefined;
20372
19923
  domain?: string | undefined;
19924
+ operating_point?: "standard" | "enhanced" | undefined;
20373
19925
  output_locale?: string | undefined;
20374
19926
  additional_vocab?: {
20375
19927
  content: string;
@@ -20396,8 +19948,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20396
19948
  }, {
20397
19949
  language: string;
20398
19950
  diarization?: "none" | "speaker" | "channel" | undefined;
20399
- operating_point?: "standard" | "enhanced" | undefined;
20400
19951
  domain?: string | undefined;
19952
+ operating_point?: "standard" | "enhanced" | undefined;
20401
19953
  output_locale?: string | undefined;
20402
19954
  additional_vocab?: {
20403
19955
  content: string;
@@ -20555,8 +20107,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20555
20107
  transcription_config?: {
20556
20108
  language: string;
20557
20109
  diarization?: "none" | "speaker" | "channel" | undefined;
20558
- operating_point?: "standard" | "enhanced" | undefined;
20559
20110
  domain?: string | undefined;
20111
+ operating_point?: "standard" | "enhanced" | undefined;
20560
20112
  output_locale?: string | undefined;
20561
20113
  additional_vocab?: {
20562
20114
  content: string;
@@ -20636,8 +20188,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20636
20188
  transcription_config?: {
20637
20189
  language: string;
20638
20190
  diarization?: "none" | "speaker" | "channel" | undefined;
20639
- operating_point?: "standard" | "enhanced" | undefined;
20640
20191
  domain?: string | undefined;
20192
+ operating_point?: "standard" | "enhanced" | undefined;
20641
20193
  output_locale?: string | undefined;
20642
20194
  additional_vocab?: {
20643
20195
  content: string;
@@ -20735,8 +20287,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20735
20287
  transcription_config?: {
20736
20288
  language: string;
20737
20289
  diarization?: "none" | "speaker" | "channel" | undefined;
20738
- operating_point?: "standard" | "enhanced" | undefined;
20739
20290
  domain?: string | undefined;
20291
+ operating_point?: "standard" | "enhanced" | undefined;
20740
20292
  output_locale?: string | undefined;
20741
20293
  additional_vocab?: {
20742
20294
  content: string;
@@ -20829,8 +20381,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20829
20381
  transcription_config?: {
20830
20382
  language: string;
20831
20383
  diarization?: "none" | "speaker" | "channel" | undefined;
20832
- operating_point?: "standard" | "enhanced" | undefined;
20833
20384
  domain?: string | undefined;
20385
+ operating_point?: "standard" | "enhanced" | undefined;
20834
20386
  output_locale?: string | undefined;
20835
20387
  additional_vocab?: {
20836
20388
  content: string;
@@ -20925,8 +20477,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
20925
20477
  transcription_config?: {
20926
20478
  language: string;
20927
20479
  diarization?: "none" | "speaker" | "channel" | undefined;
20928
- operating_point?: "standard" | "enhanced" | undefined;
20929
20480
  domain?: string | undefined;
20481
+ operating_point?: "standard" | "enhanced" | undefined;
20930
20482
  output_locale?: string | undefined;
20931
20483
  additional_vocab?: {
20932
20484
  content: string;
@@ -21021,8 +20573,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
21021
20573
  transcription_config?: {
21022
20574
  language: string;
21023
20575
  diarization?: "none" | "speaker" | "channel" | undefined;
21024
- operating_point?: "standard" | "enhanced" | undefined;
21025
20576
  domain?: string | undefined;
20577
+ operating_point?: "standard" | "enhanced" | undefined;
21026
20578
  output_locale?: string | undefined;
21027
20579
  additional_vocab?: {
21028
20580
  content: string;
@@ -21210,8 +20762,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21210
20762
  }, "strip", z.ZodTypeAny, {
21211
20763
  language: string;
21212
20764
  diarization?: "none" | "speaker" | "channel" | undefined;
21213
- operating_point?: "standard" | "enhanced" | undefined;
21214
20765
  domain?: string | undefined;
20766
+ operating_point?: "standard" | "enhanced" | undefined;
21215
20767
  output_locale?: string | undefined;
21216
20768
  additional_vocab?: {
21217
20769
  content: string;
@@ -21238,8 +20790,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21238
20790
  }, {
21239
20791
  language: string;
21240
20792
  diarization?: "none" | "speaker" | "channel" | undefined;
21241
- operating_point?: "standard" | "enhanced" | undefined;
21242
20793
  domain?: string | undefined;
20794
+ operating_point?: "standard" | "enhanced" | undefined;
21243
20795
  output_locale?: string | undefined;
21244
20796
  additional_vocab?: {
21245
20797
  content: string;
@@ -21397,8 +20949,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21397
20949
  transcription_config?: {
21398
20950
  language: string;
21399
20951
  diarization?: "none" | "speaker" | "channel" | undefined;
21400
- operating_point?: "standard" | "enhanced" | undefined;
21401
20952
  domain?: string | undefined;
20953
+ operating_point?: "standard" | "enhanced" | undefined;
21402
20954
  output_locale?: string | undefined;
21403
20955
  additional_vocab?: {
21404
20956
  content: string;
@@ -21478,8 +21030,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21478
21030
  transcription_config?: {
21479
21031
  language: string;
21480
21032
  diarization?: "none" | "speaker" | "channel" | undefined;
21481
- operating_point?: "standard" | "enhanced" | undefined;
21482
21033
  domain?: string | undefined;
21034
+ operating_point?: "standard" | "enhanced" | undefined;
21483
21035
  output_locale?: string | undefined;
21484
21036
  additional_vocab?: {
21485
21037
  content: string;
@@ -21577,8 +21129,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21577
21129
  transcription_config?: {
21578
21130
  language: string;
21579
21131
  diarization?: "none" | "speaker" | "channel" | undefined;
21580
- operating_point?: "standard" | "enhanced" | undefined;
21581
21132
  domain?: string | undefined;
21133
+ operating_point?: "standard" | "enhanced" | undefined;
21582
21134
  output_locale?: string | undefined;
21583
21135
  additional_vocab?: {
21584
21136
  content: string;
@@ -21671,8 +21223,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21671
21223
  transcription_config?: {
21672
21224
  language: string;
21673
21225
  diarization?: "none" | "speaker" | "channel" | undefined;
21674
- operating_point?: "standard" | "enhanced" | undefined;
21675
21226
  domain?: string | undefined;
21227
+ operating_point?: "standard" | "enhanced" | undefined;
21676
21228
  output_locale?: string | undefined;
21677
21229
  additional_vocab?: {
21678
21230
  content: string;
@@ -21767,8 +21319,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21767
21319
  transcription_config?: {
21768
21320
  language: string;
21769
21321
  diarization?: "none" | "speaker" | "channel" | undefined;
21770
- operating_point?: "standard" | "enhanced" | undefined;
21771
21322
  domain?: string | undefined;
21323
+ operating_point?: "standard" | "enhanced" | undefined;
21772
21324
  output_locale?: string | undefined;
21773
21325
  additional_vocab?: {
21774
21326
  content: string;
@@ -21863,8 +21415,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
21863
21415
  transcription_config?: {
21864
21416
  language: string;
21865
21417
  diarization?: "none" | "speaker" | "channel" | undefined;
21866
- operating_point?: "standard" | "enhanced" | undefined;
21867
21418
  domain?: string | undefined;
21419
+ operating_point?: "standard" | "enhanced" | undefined;
21868
21420
  output_locale?: string | undefined;
21869
21421
  additional_vocab?: {
21870
21422
  content: string;
@@ -22090,8 +21642,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22090
21642
  }, "strip", z.ZodTypeAny, {
22091
21643
  language: string;
22092
21644
  diarization?: "none" | "speaker" | "channel" | undefined;
22093
- operating_point?: "standard" | "enhanced" | undefined;
22094
21645
  domain?: string | undefined;
21646
+ operating_point?: "standard" | "enhanced" | undefined;
22095
21647
  output_locale?: string | undefined;
22096
21648
  additional_vocab?: {
22097
21649
  content: string;
@@ -22118,8 +21670,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22118
21670
  }, {
22119
21671
  language: string;
22120
21672
  diarization?: "none" | "speaker" | "channel" | undefined;
22121
- operating_point?: "standard" | "enhanced" | undefined;
22122
21673
  domain?: string | undefined;
21674
+ operating_point?: "standard" | "enhanced" | undefined;
22123
21675
  output_locale?: string | undefined;
22124
21676
  additional_vocab?: {
22125
21677
  content: string;
@@ -22306,8 +21858,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22306
21858
  transcription_config?: {
22307
21859
  language: string;
22308
21860
  diarization?: "none" | "speaker" | "channel" | undefined;
22309
- operating_point?: "standard" | "enhanced" | undefined;
22310
21861
  domain?: string | undefined;
21862
+ operating_point?: "standard" | "enhanced" | undefined;
22311
21863
  output_locale?: string | undefined;
22312
21864
  additional_vocab?: {
22313
21865
  content: string;
@@ -22387,8 +21939,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
22387
21939
  transcription_config?: {
22388
21940
  language: string;
22389
21941
  diarization?: "none" | "speaker" | "channel" | undefined;
22390
- operating_point?: "standard" | "enhanced" | undefined;
22391
21942
  domain?: string | undefined;
21943
+ operating_point?: "standard" | "enhanced" | undefined;
22392
21944
  output_locale?: string | undefined;
22393
21945
  additional_vocab?: {
22394
21946
  content: string;
@@ -23129,8 +22681,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
23129
22681
  transcription_config?: {
23130
22682
  language: string;
23131
22683
  diarization?: "none" | "speaker" | "channel" | undefined;
23132
- operating_point?: "standard" | "enhanced" | undefined;
23133
22684
  domain?: string | undefined;
22685
+ operating_point?: "standard" | "enhanced" | undefined;
23134
22686
  output_locale?: string | undefined;
23135
22687
  additional_vocab?: {
23136
22688
  content: string;
@@ -23353,8 +22905,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
23353
22905
  transcription_config?: {
23354
22906
  language: string;
23355
22907
  diarization?: "none" | "speaker" | "channel" | undefined;
23356
- operating_point?: "standard" | "enhanced" | undefined;
23357
22908
  domain?: string | undefined;
22909
+ operating_point?: "standard" | "enhanced" | undefined;
23358
22910
  output_locale?: string | undefined;
23359
22911
  additional_vocab?: {
23360
22912
  content: string;
@@ -23622,15 +23174,15 @@ declare const getUsageResponse: z.ZodObject<{
23622
23174
  duration_hrs: z.ZodNumber;
23623
23175
  }, "strip", z.ZodTypeAny, {
23624
23176
  type: "transcription" | "alignment";
23625
- count: number;
23626
23177
  mode: "batch";
23178
+ count: number;
23627
23179
  duration_hrs: number;
23628
23180
  language?: string | undefined;
23629
23181
  operating_point?: "standard" | "enhanced" | undefined;
23630
23182
  }, {
23631
23183
  type: "transcription" | "alignment";
23632
- count: number;
23633
23184
  mode: "batch";
23185
+ count: number;
23634
23186
  duration_hrs: number;
23635
23187
  language?: string | undefined;
23636
23188
  operating_point?: "standard" | "enhanced" | undefined;
@@ -23644,15 +23196,15 @@ declare const getUsageResponse: z.ZodObject<{
23644
23196
  duration_hrs: z.ZodNumber;
23645
23197
  }, "strip", z.ZodTypeAny, {
23646
23198
  type: "transcription" | "alignment";
23647
- count: number;
23648
23199
  mode: "batch";
23200
+ count: number;
23649
23201
  duration_hrs: number;
23650
23202
  language?: string | undefined;
23651
23203
  operating_point?: "standard" | "enhanced" | undefined;
23652
23204
  }, {
23653
23205
  type: "transcription" | "alignment";
23654
- count: number;
23655
23206
  mode: "batch";
23207
+ count: number;
23656
23208
  duration_hrs: number;
23657
23209
  language?: string | undefined;
23658
23210
  operating_point?: "standard" | "enhanced" | undefined;
@@ -23660,16 +23212,16 @@ declare const getUsageResponse: z.ZodObject<{
23660
23212
  }, "strip", z.ZodTypeAny, {
23661
23213
  summary: {
23662
23214
  type: "transcription" | "alignment";
23663
- count: number;
23664
23215
  mode: "batch";
23216
+ count: number;
23665
23217
  duration_hrs: number;
23666
23218
  language?: string | undefined;
23667
23219
  operating_point?: "standard" | "enhanced" | undefined;
23668
23220
  }[];
23669
23221
  details: {
23670
23222
  type: "transcription" | "alignment";
23671
- count: number;
23672
23223
  mode: "batch";
23224
+ count: number;
23673
23225
  duration_hrs: number;
23674
23226
  language?: string | undefined;
23675
23227
  operating_point?: "standard" | "enhanced" | undefined;
@@ -23679,16 +23231,16 @@ declare const getUsageResponse: z.ZodObject<{
23679
23231
  }, {
23680
23232
  summary: {
23681
23233
  type: "transcription" | "alignment";
23682
- count: number;
23683
23234
  mode: "batch";
23235
+ count: number;
23684
23236
  duration_hrs: number;
23685
23237
  language?: string | undefined;
23686
23238
  operating_point?: "standard" | "enhanced" | undefined;
23687
23239
  }[];
23688
23240
  details: {
23689
23241
  type: "transcription" | "alignment";
23690
- count: number;
23691
23242
  mode: "batch";
23243
+ count: number;
23692
23244
  duration_hrs: number;
23693
23245
  language?: string | undefined;
23694
23246
  operating_point?: "standard" | "enhanced" | undefined;
@@ -24935,4 +24487,4 @@ declare namespace elevenLabsSpeechToTextAPI_zod {
24935
24487
  export { elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdHeader as deleteTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdParams as deleteTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdResponse as deleteTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdHeader as getTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdParams as getTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdResponse as getTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_speechToTextBody as speechToTextBody, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive as speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsMax as speechToTextBodyAdditionalFormatsMax, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMaxOne as speechToTextBodyDiarizationThresholdMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMinOne as speechToTextBodyDiarizationThresholdMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizeDefault as speechToTextBodyDiarizeDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyFileFormatDefault as speechToTextBodyFileFormatDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyKeytermsDefault as speechToTextBodyKeytermsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNoVerbatimDefault as speechToTextBodyNoVerbatimDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNumSpeakersMaxOne as speechToTextBodyNumSpeakersMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMaxOne as speechToTextBodySeedMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMinOne as speechToTextBodySeedMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTagAudioEventsDefault as speechToTextBodyTagAudioEventsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMaxOne as speechToTextBodyTemperatureMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMinOne as speechToTextBodyTemperatureMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTimestampsGranularityDefault as speechToTextBodyTimestampsGranularityDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyUseMultiChannelDefault as speechToTextBodyUseMultiChannelDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyWebhookDefault as speechToTextBodyWebhookDefault, elevenLabsSpeechToTextAPI_zod_speechToTextHeader as speechToTextHeader, elevenLabsSpeechToTextAPI_zod_speechToTextQueryEnableLoggingDefault as speechToTextQueryEnableLoggingDefault, elevenLabsSpeechToTextAPI_zod_speechToTextQueryParams as speechToTextQueryParams, elevenLabsSpeechToTextAPI_zod_speechToTextResponse as speechToTextResponse };
24936
24488
  }
24937
24489
 
24938
- export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult, TranscriptOptionalParams as AssemblyAIOptions, SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAIStreamingOptions, TopicDetectionModelResult as AssemblyAITopicsResult, index$6 as AssemblyAITypes, type AudioChunkAckMessage, AudioInput, AudioResponseFormat, AzureSTTAdapter, Status as AzureStatus, index$5 as AzureTypes, BaseAdapter, CreateTranscriptionRequest, CreateTranscriptionRequestModel, CreateTranscriptionResponseDiarizedJson, CreateTranscriptionResponseVerboseJson, DeepgramAdapter, type DeepgramConfig, ListenV1ResponseMetadata as DeepgramMetadata, ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramStreamingOptions, index$3 as DeepgramTypes, ElevenLabsAdapter, type ElevenLabsConfig, index as ElevenLabsTypes, elevenLabsSpeechToTextAPI_zod as ElevenLabsZodSchemas, type EndRecordingMessage, type EndSessionMessage, type GetProjectRequestV1Response, GladiaAdapter, AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, AudioToLlmListDTO as GladiaAudioToLlmResult, ChapterizationDTO as GladiaChapters, CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, NamedEntityRecognitionDTO as GladiaEntities, ModerationDTO as GladiaModeration, InitTranscriptionRequest as GladiaOptions, SentimentAnalysisDTO as GladiaSentiment, SpeakerReidentificationDTO as GladiaSpeakerReidentification, GladiaStreamingOptions, StreamingRequest as GladiaStreamingRequest, StructuredDataExtractionDTO as GladiaStructuredData, TranslationDTO as GladiaTranslation, index$7 as GladiaTypes, InitTranscriptionRequest, LanguageIdentificationMode, LanguageIdentificationProperties, type ListProjectRequestsV1Response, type ListTranscriptionResponseItemsItem, ListTranscriptsOptions, ListTranscriptsParams, ListenV1EncodingParameter, ListenV1LanguageParameter, ListenV1MediaTranscribeParams, ListenV1ModelParameter, ListenV1Response, ListenV1ResponseResultsChannelsItemAlternativesItem, ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, ListenV1ResponseResultsUtterancesItem, ListenV1VersionParameter, ManageV1FilterEndpointParameter, ManageV1FilterStatusParameter, ManageV1ProjectsRequestsListParams, type NamedEntityRecognitionMessage, streamingTypes as OpenAIStreamingTypes, index$4 as OpenAITypes, OpenAIWhisperAdapter, CreateTranscriptionRequest as OpenAIWhisperOptions, type PaginatedTranscriptions, type PostChapterizationMessage, type PostFinalTranscriptMessage, type PostSummarizationMessage, type PostTranscriptMessage, PreRecordedResponse, ProfanityFilterMode, type ProjectRequestResponse, ProviderCapabilities, type ProviderConfig, PunctuationMode, type SentimentAnalysisMessage, SonioxAdapter, type SonioxConfig, SonioxModelCode, index$1 as SonioxTypes, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type SpeechEndMessage, type SpeechStartMessage, SpeechmaticsAdapter, type SpeechmaticsConfig, index$2 as SpeechmaticsTypes, speechmaticsASRRESTAPI_zod as SpeechmaticsZodSchemas, type StartRecordingMessage, type StartSessionMessage, type StopRecordingAckMessage, StreamEvent, StreamingCallbacks, StreamingOptions, StreamingRequest, type StreamingResponse, StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedModels, StreamingSupportedSampleRateEnum, StreamingUpdateConfiguration, TranscribeOptions, Transcript, type TranscriptListItem, type TranscriptMessage, TranscriptOptionalParamsSpeechModel, type TranscriptParams, TranscriptStatus, TranscriptUtterance, TranscriptWord, Transcription$1 as Transcription, type TranscriptionAdapter, TranscriptionControllerListV2Params, TranscriptionControllerListV2StatusItem, TranscriptionDTO, TranscriptionLanguageCodeEnum, TranscriptionProperties, TranscriptionProvider, type TranscriptionsListParams, type TranslationMessage, UnifiedTranscriptResponse, UtteranceDTO, VoiceRouter, type VoiceRouterConfig, type WebHook, type WebHookEvents, WordDTO, createAssemblyAIAdapter, createAzureSTTAdapter, createDeepgramAdapter, createElevenLabsAdapter, createGladiaAdapter, createOpenAIWhisperAdapter, createSonioxAdapter, createSpeechmaticsAdapter, createTemporaryToken, createTranscript, createTranscription, createVoiceRouter, deleteTranscript as deleteTranscriptAPI, getTranscript as getTranscriptAPI, listTranscripts as listTranscriptsAPI, preRecordedControllerDeletePreRecordedJobV2, preRecordedControllerGetAudioV2, preRecordedControllerGetPreRecordedJobV2, preRecordedControllerInitPreRecordedJobV2, streamingControllerDeleteStreamingJobV2, streamingControllerGetAudioV2, streamingControllerInitStreamingSessionV2, transcriptionControllerListV2, transcriptionsCreate, transcriptionsDelete, transcriptionsGet, transcriptionsList, transcriptionsListFiles, webHooksCreate, webHooksDelete, webHooksList };
24490
+ export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult, TranscriptParams as AssemblyAIOptions, SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAIStreamingOptions, TopicDetectionModelResult as AssemblyAITopicsResult, index$6 as AssemblyAITypes, type AudioChunkAckMessage, AudioInput, AudioResponseFormat, AzureSTTAdapter, Status as AzureStatus, index$5 as AzureTypes, BaseAdapter, CreateTranscriptionRequest, CreateTranscriptionRequestModel, CreateTranscriptionResponseDiarizedJson, CreateTranscriptionResponseVerboseJson, DeepgramAdapter, type DeepgramConfig, ListenV1ResponseMetadata as DeepgramMetadata, ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramStreamingOptions, index$3 as DeepgramTypes, ElevenLabsAdapter, type ElevenLabsConfig, index as ElevenLabsTypes, elevenLabsSpeechToTextAPI_zod as ElevenLabsZodSchemas, type EndRecordingMessage, type EndSessionMessage, type GetProjectRequestV1Response, GladiaAdapter, AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, AudioToLlmListDTO as GladiaAudioToLlmResult, ChapterizationDTO as GladiaChapters, CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, NamedEntityRecognitionDTO as GladiaEntities, ModerationDTO as GladiaModeration, InitTranscriptionRequest as GladiaOptions, SentimentAnalysisDTO as GladiaSentiment, SpeakerReidentificationDTO as GladiaSpeakerReidentification, GladiaStreamingOptions, StreamingRequest as GladiaStreamingRequest, StructuredDataExtractionDTO as GladiaStructuredData, TranslationDTO as GladiaTranslation, index$7 as GladiaTypes, InitTranscriptionRequest, LanguageIdentificationMode, LanguageIdentificationProperties, type ListProjectRequestsV1Response, type ListTranscriptionResponseItemsItem, ListTranscriptsOptions, ListTranscriptsParams, ListenV1EncodingParameter, ListenV1LanguageParameter, ListenV1MediaTranscribeParams, ListenV1ModelParameter, ListenV1Response, ListenV1ResponseResultsChannelsItemAlternativesItem, ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, ListenV1ResponseResultsUtterancesItem, ListenV1VersionParameter, ManageV1FilterEndpointParameter, ManageV1FilterStatusParameter, ManageV1ProjectsRequestsListParams, type NamedEntityRecognitionMessage, streamingTypes as OpenAIStreamingTypes, index$4 as OpenAITypes, OpenAIWhisperAdapter, CreateTranscriptionRequest as OpenAIWhisperOptions, type PaginatedTranscriptions, type PostChapterizationMessage, type PostFinalTranscriptMessage, type PostSummarizationMessage, type PostTranscriptMessage, PreRecordedResponse, ProfanityFilterMode, type ProjectRequestResponse, ProviderCapabilities, type ProviderConfig, PunctuationMode, type SentimentAnalysisMessage, SonioxAdapter, type SonioxConfig, SonioxModelCode, index$1 as SonioxTypes, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type SpeechEndMessage, SpeechModel, type SpeechStartMessage, SpeechmaticsAdapter, type SpeechmaticsConfig, index$2 as SpeechmaticsTypes, speechmaticsASRRESTAPI_zod as SpeechmaticsZodSchemas, type StartRecordingMessage, type StartSessionMessage, type StopRecordingAckMessage, StreamEvent, StreamingCallbacks, StreamingOptions, StreamingRequest, type StreamingResponse, StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedModels, StreamingSupportedSampleRateEnum, StreamingUpdateConfiguration, TranscribeOptions, Transcript, type TranscriptListItem, type TranscriptMessage, TranscriptParams, TranscriptStatus, TranscriptUtterance, TranscriptWord, Transcription$1 as Transcription, type TranscriptionAdapter, TranscriptionControllerListV2Params, TranscriptionControllerListV2StatusItem, TranscriptionDTO, TranscriptionLanguageCodeEnum, TranscriptionProperties, TranscriptionProvider, type TranscriptionsListParams, type TranslationMessage, UnifiedTranscriptResponse, UtteranceDTO, VoiceRouter, type VoiceRouterConfig, type WebHook, type WebHookEvents, WordDTO, createAssemblyAIAdapter, createAzureSTTAdapter, createDeepgramAdapter, createElevenLabsAdapter, createGladiaAdapter, createOpenAIWhisperAdapter, createSonioxAdapter, createSpeechmaticsAdapter, createTranscript, createTranscription, createVoiceRouter, deleteTranscript as deleteTranscriptAPI, getTranscript as getTranscriptAPI, listTranscripts as listTranscriptsAPI, preRecordedControllerDeletePreRecordedJobV2, preRecordedControllerGetAudioV2, preRecordedControllerGetPreRecordedJobV2, preRecordedControllerInitPreRecordedJobV2, streamingControllerDeleteStreamingJobV2, streamingControllerGetAudioV2, streamingControllerInitStreamingSessionV2, transcriptionControllerListV2, transcriptionsCreate, transcriptionsDelete, transcriptionsGet, transcriptionsList, transcriptionsListFiles, webHooksCreate, webHooksDelete, webHooksList };