voice-router-dev 0.8.6 → 0.8.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -1,10 +1,10 @@
1
- import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, t as NamedEntityRecognitionResult, W as WordDTO, u as StreamingSupportedRegions, v as AddonErrorDTO, w as AudioToLlmDTO, x as AudioToLlmDTOError, y as AudioToLlmDTOResults, z as AudioToLlmListConfigDTO, B as AudioToLlmListDTO, E as AudioToLlmListDTOError, H as AudioToLlmResultDTO, I as CallbackConfigDto, J as CallbackMethodEnum, K as CallbackTranscriptionErrorPayload, O as CallbackTranscriptionErrorPayloadCustomMetadata, Q as CallbackTranscriptionErrorPayloadEvent, V as CallbackTranscriptionSuccessPayload, X as CallbackTranscriptionSuccessPayloadCustomMetadata, Y as CallbackTranscriptionSuccessPayloadEvent, Z as ChapterizationDTOError, _ as ChapterizationDTOResults, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PiiRedactionConfigDTO, ai as PiiRedactionConfigDTOProcessedTextType, aj as PiiRedactionEntityTypeEnum, ak as PreRecordedRequestParamsResponse, al as PreRecordedResponseCustomMetadata, am as PreRecordedResponseFile, an as PreRecordedResponseKind, ao as PreRecordedResponsePostSessionMetadata, ap as PreRecordedResponseRequestParams, aq as PreRecordedResponseResult, ar as PreRecordedResponseStatus, as as SentencesDTO, at as SentencesDTOError, au as SentimentAnalysisDTOError, av as SpeakerReidentificationDTO, aw as SpeakerReidentificationDTOError, ax as StreamingRequest, ay as StreamingRequestCustomMetadata, az as StructuredDataExtractionDTO, aA as StructuredDataExtractionDTOError, aB as SubtitleDTO, aC as SubtitlesConfigDTO, aD as SubtitlesFormatEnum, aE as SubtitlesStyleEnum, aF as SummarizationConfigDTO, aG as SummarizationDTOError, aH as SummaryTypesEnum, aI as TranscriptionControllerListV2KindItem, aJ as TranscriptionControllerListV2Params, aK as TranscriptionControllerListV2StatusItem, aL as TranscriptionResultDTO, aM as TranslationConfigDTO, aN as TranslationDTOError, aO as TranslationModelEnum, aP as TranslationResultDTO, aQ as TranslationResultDTOError, aR as TranscriptStatus, aS as TranscriptWord, aT as AudioIntelligenceModelStatus, aU as AutoHighlightResult, aV as AutoHighlightsResult, aW as Chapter, aX as ContentSafetyLabel, aY as ContentSafetyLabelResult, aZ as ContentSafetyLabelsResult, a_ as ContentSafetyLabelsResultSeverityScoreSummary, a$ as ContentSafetyLabelsResultSummary, b0 as CustomFormattingRequestBody, b1 as CustomFormattingRequestBodyCustomFormatting, b2 as CustomFormattingResponse, b3 as CustomFormattingResponseCustomFormatting, b4 as CustomFormattingResponseCustomFormattingMapping, b5 as Entity, b6 as EntityType, b7 as ListTranscriptsParams, b8 as PiiPolicy, b9 as RedactPiiAudioQuality, ba as Sentiment, bb as SentimentAnalysisResult$1, bc as SentimentAnalysisResultChannel, bd as SentimentAnalysisResultSpeaker, be as SeverityScoreSummary, bf as SpeakerIdentificationRequestBody, bg as SpeakerIdentificationRequestBodySpeakerIdentification, bh as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, bi as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, bj as SpeakerIdentificationResponse, bk as SpeakerIdentificationResponseSpeakerIdentification, bl as SpeakerIdentificationResponseSpeakerIdentificationMapping, bm as SpeechModel, bn as SubstitutionPolicy, bo as SummaryModel, bp as SummaryType, bq as Timestamp, br as TopicDetectionModelResult, bs as TopicDetectionModelResultSummary, bt as TopicDetectionResult, bu as TopicDetectionResultLabelsItems, bv as Transcript, bw as TranscriptAudioDuration, bx as TranscriptAudioEndAt, by as TranscriptAudioStartFrom, bz as TranscriptAutoChapters, bA as TranscriptAutoHighlightsResult, bB as TranscriptChapters, bC as TranscriptConfidence, bD as TranscriptContentSafety, bE as TranscriptContentSafetyLabels, bF as TranscriptCustomSpelling, bG as TranscriptCustomSpellingProperty, bH as TranscriptCustomTopics, bI as TranscriptDisfluencies, bJ as TranscriptDomain, bK as TranscriptEntities, bL as TranscriptEntityDetection, bM as TranscriptFilterProfanity, bN as TranscriptFormatText, bO as TranscriptIabCategories, bP as TranscriptIabCategoriesResult, bQ as TranscriptLanguageCode, bR as TranscriptLanguageCodes, bS as TranscriptLanguageConfidence, bT as TranscriptLanguageConfidenceThreshold, bU as TranscriptLanguageDetection, bV as TranscriptLanguageDetectionOptions, bW as TranscriptMultichannel, bX as TranscriptOptionalParamsLanguageDetectionOptions, bY as TranscriptOptionalParamsRedactPiiAudioOptions, bZ as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, b_ as TranscriptOptionalParamsRemoveAudioTags, b$ as TranscriptOptionalParamsSpeakerOptions, c0 as TranscriptOptionalParamsSpeechUnderstanding, c1 as TranscriptOptionalParamsSpeechUnderstandingRequest, c2 as TranscriptParams, c3 as TranscriptParamsDomain, c4 as TranscriptParamsLanguageCode, c5 as TranscriptParamsLanguageCodes, c6 as TranscriptParamsRedactPiiSub, c7 as TranscriptParamsRemoveAudioTags, c8 as TranscriptParamsSpeakersExpected, c9 as TranscriptParamsSpeechThreshold, ca as TranscriptParamsWebhookAuthHeaderName, cb as TranscriptParamsWebhookAuthHeaderValue, cc as TranscriptPunctuate, cd as TranscriptRedactPiiAudio, ce as TranscriptRedactPiiAudioOptions, cf as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, cg as TranscriptRedactPiiAudioQuality, ch as TranscriptRedactPiiPolicies, ci as TranscriptRemoveAudioTags, cj as TranscriptRemoveAudioTagsProperty, ck as TranscriptSentimentAnalysis, cl as TranscriptSentimentAnalysisResults, cm as TranscriptSpeakerLabels, cn as TranscriptSpeakersExpected, co as TranscriptSpeechModel, cp as TranscriptSpeechModels, cq as TranscriptSpeechThreshold, cr as TranscriptSpeechUnderstanding, cs as TranscriptSpeechUnderstandingRequest, ct as TranscriptSpeechUnderstandingResponse, cu as TranscriptSpeedBoost, cv as TranscriptSummary, cw as TranscriptSummaryModel, cx as TranscriptSummaryType, cy as TranscriptText, cz as TranscriptThrottled, cA as TranscriptTranslatedTexts, cB as TranscriptUtterance, cC as TranscriptUtteranceChannel, cD as TranscriptUtteranceTranslatedTexts, cE as TranscriptUtterances, cF as TranscriptWebhookAuthHeaderName, cG as TranscriptWebhookStatusCode, cH as TranscriptWebhookUrl, cI as TranscriptWordChannel, cJ as TranscriptWordSpeaker, cK as TranscriptWords, cL as TranslationRequestBody, cM as TranslationRequestBodyTranslation, cN as TranslationResponse, cO as TranslationResponseTranslation, cP as StreamingUpdateConfiguration, cQ as Transcription$1, cR as EntityError, cS as Status, cT as EntityReference, cU as DiarizationProperties, cV as DiarizationSpeakersProperties, cW as LanguageIdentificationMode, cX as LanguageIdentificationProperties, cY as LanguageIdentificationPropertiesSpeechModelMapping, cZ as ProfanityFilterMode, c_ as PunctuationMode, c$ as TranscriptionCustomProperties, d0 as TranscriptionLinks, d1 as TranscriptionProperties, d2 as TranscriptTextUsageTokens, d3 as TranscriptionSegment, d4 as RealtimeSessionCreateRequestGAModel, d5 as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, d6 as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, d7 as AudioResponseFormat, d8 as CreateTranscription200One, d9 as CreateTranscriptionRequest, da as CreateTranscriptionRequestChunkingStrategy, db as CreateTranscriptionRequestChunkingStrategyAnyOf, dc as CreateTranscriptionRequestModel, dd as CreateTranscriptionRequestStream, de as CreateTranscriptionRequestTimestampGranularitiesItem, df as CreateTranscriptionResponseDiarizedJson, dg as CreateTranscriptionResponseDiarizedJsonTask, dh as CreateTranscriptionResponseDiarizedJsonUsage, di as CreateTranscriptionResponseJson, dj as CreateTranscriptionResponseJsonLogprobsItem, dk as CreateTranscriptionResponseJsonUsage, dl as CreateTranscriptionResponseVerboseJson, dm as TranscriptTextUsageDuration, dn as TranscriptTextUsageDurationType, dp as TranscriptTextUsageTokensInputTokenDetails, dq as TranscriptTextUsageTokensType, dr as TranscriptionDiarizedSegment, ds as TranscriptionDiarizedSegmentType, dt as TranscriptionInclude, du as TranscriptionWord, dv as VadConfig, dw as VadConfigType, dx as ListenV1Response, dy as SharedTopics, dz as SharedIntents, dA as SharedSentiments, dB as ListProjectRequestsParams, dC as ListenTranscribeParams, dD as ListenV1ResponseMetadata, dE as ListenV1ResponseMetadataIntentsInfo, dF as ListenV1ResponseMetadataModelInfo, dG as ListenV1ResponseMetadataSentimentInfo, dH as ListenV1ResponseMetadataSummaryInfo, dI as ListenV1ResponseMetadataTopicsInfo, dJ as ListenV1ResponseResults, dK as ListenV1ResponseResultsChannels, dL as ListenV1ResponseResultsChannelsItems, dM as ListenV1ResponseResultsChannelsItemsAlternativesItems, dN as ListenV1ResponseResultsChannelsItemsAlternativesItemsEntitiesItems, dO as ListenV1ResponseResultsChannelsItemsAlternativesItemsParagraphs, dP as ListenV1ResponseResultsChannelsItemsAlternativesItemsParagraphsParagraphsItems, dQ as ListenV1ResponseResultsChannelsItemsAlternativesItemsParagraphsParagraphsItemsSentencesItems, dR as ListenV1ResponseResultsChannelsItemsAlternativesItemsSummariesItems, dS as ListenV1ResponseResultsChannelsItemsAlternativesItemsTopicsItems, dT as ListenV1ResponseResultsChannelsItemsAlternativesItemsWordsItems, dU as ListenV1ResponseResultsChannelsItemsSearchItems, dV as ListenV1ResponseResultsChannelsItemsSearchItemsHitsItems, dW as ListenV1ResponseResultsSummary, dX as ListenV1ResponseResultsUtterances, dY as ListenV1ResponseResultsUtterancesItems, dZ as ListenV1ResponseResultsUtterancesItemsWordsItems, d_ as SharedIntentsResults, d$ as SharedIntentsResultsIntents, e0 as SharedIntentsResultsIntentsSegmentsItems, e1 as SharedIntentsResultsIntentsSegmentsItemsIntentsItems, e2 as SharedSentimentsAverage, e3 as SharedSentimentsSegmentsItems, e4 as SharedTopicsResults, e5 as SharedTopicsResultsTopics, e6 as SharedTopicsResultsTopicsSegmentsItems, e7 as SharedTopicsResultsTopicsSegmentsItemsTopicsItems, e8 as V1ListenPostParametersCallbackMethod, e9 as V1ListenPostParametersCustomIntent, ea as V1ListenPostParametersCustomIntentMode, eb as V1ListenPostParametersCustomTopic, ec as V1ListenPostParametersCustomTopicMode, ed as V1ListenPostParametersDetectLanguage, ee as V1ListenPostParametersEncoding, ef as V1ListenPostParametersExtra, eg as V1ListenPostParametersKeywords, eh as V1ListenPostParametersModel, ei as V1ListenPostParametersModel0, ej as V1ListenPostParametersRedact, ek as V1ListenPostParametersRedact1, el as V1ListenPostParametersRedactSchemaOneOf1Items, em as V1ListenPostParametersReplace, en as V1ListenPostParametersSearch, eo as V1ListenPostParametersSummarize, ep as V1ListenPostParametersSummarize0, eq as V1ListenPostParametersTag, er as V1ListenPostParametersVersion, es as V1ListenPostParametersVersion0, et as V1ProjectsProjectIdRequestsGetParametersDeployment, eu as V1ProjectsProjectIdRequestsGetParametersEndpoint, ev as V1ProjectsProjectIdRequestsGetParametersMethod, ew as V1ProjectsProjectIdRequestsGetParametersStatus, ex as JobType, ey as AlignmentConfig, ez as TranscriptionConfig, eA as TrackingData, eB as OutputConfig, eC as OperatingPoint, eD as AudioEventItem, eE as AudioEventSummary, eF as AudioEventSummaryItem, eG as AutoChaptersResult, eH as AutoChaptersResultError, eI as AutoChaptersResultErrorType, eJ as Chapter$1, eK as JobInfo, eL as LanguageIdentificationResult, eM as LanguageIdentificationResultAlternative, eN as LanguageIdentificationResultError, eO as LanguageIdentificationResultItem, eP as LanguagePackInfo, eQ as LanguagePackInfoWritingDirection, eR as OutputConfigSrtOverrides, eS as RecognitionAlternative, eT as RecognitionDisplay, eU as RecognitionDisplayDirection, eV as RecognitionMetadata, eW as RecognitionResult, eX as RecognitionResultAttachesTo, eY as RecognitionResultType, eZ as RetrieveTranscriptResponse, e_ as RetrieveTranscriptResponseAudioEventSummary, e$ as RetrieveTranscriptResponseAudioEventSummaryChannels, f0 as RetrieveTranscriptResponseTranslations, f1 as SentimentAnalysisError, f2 as SentimentAnalysisErrorType, f3 as SentimentAnalysisResult$2, f4 as SentimentAnalysisResultSentimentAnalysis, f5 as SentimentChannelSummary, f6 as SentimentSegment, f7 as SentimentSpeakerSummary, f8 as SentimentSummary, f9 as SentimentSummaryDetail, fa as SpokenFormRecognitionResult, fb as SpokenFormRecognitionResultType, fc as SummarizationError, fd as SummarizationErrorType, fe as SummarizationResult, ff as TopicDetectionError, fg as TopicDetectionErrorType, fh as TopicDetectionResult$1, fi as TopicDetectionSegment, fj as TopicDetectionSegmentTopic, fk as TopicDetectionSummary, fl as TopicDetectionSummaryOverall, fm as TrackingDataDetails, fn as TranscriptionConfigAdditionalVocabItem, fo as TranscriptionConfigDiarization, fp as TranscriptionConfigMaxDelayMode, fq as TranscriptionConfigPunctuationOverrides, fr as TranscriptionConfigSpeakerDiarizationConfig, fs as TranscriptionConfigTranscriptFilteringConfig, ft as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fu as TranslationError, fv as TranslationErrorType, fw as TranslationSentence, fx as WrittenFormRecognitionResult, fy as WrittenFormRecognitionResultType, fz as SpeechToTextChunkResponseModel, fA as AdditionalFormatResponseModel, fB as AdditionalFormats, fC as BodySpeechToTextV1SpeechToTextPost, fD as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fE as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, fF as BodySpeechToTextV1SpeechToTextPostEntityDetection, fG as BodySpeechToTextV1SpeechToTextPostEntityRedaction, fH as BodySpeechToTextV1SpeechToTextPostFile, fI as BodySpeechToTextV1SpeechToTextPostFileFormat, fJ as BodySpeechToTextV1SpeechToTextPostLanguageCode, fK as BodySpeechToTextV1SpeechToTextPostModelId, fL as BodySpeechToTextV1SpeechToTextPostNumSpeakers, fM as BodySpeechToTextV1SpeechToTextPostSeed, fN as BodySpeechToTextV1SpeechToTextPostSourceUrl, fO as BodySpeechToTextV1SpeechToTextPostTemperature, fP as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, fQ as BodySpeechToTextV1SpeechToTextPostWebhookId, fR as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, fS as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, fT as DetectedEntity, fU as DocxExportOptions, fV as DocxExportOptionsFormat, fW as DocxExportOptionsMaxSegmentChars, fX as DocxExportOptionsMaxSegmentDurationS, fY as DocxExportOptionsSegmentOnSilenceLongerThanS, fZ as ExportOptions, f_ as HtmlExportOptions, f$ as HtmlExportOptionsFormat, g0 as HtmlExportOptionsMaxSegmentChars, g1 as HtmlExportOptionsMaxSegmentDurationS, g2 as HtmlExportOptionsSegmentOnSilenceLongerThanS, g3 as PdfExportOptions, g4 as PdfExportOptionsFormat, g5 as PdfExportOptionsMaxSegmentChars, g6 as PdfExportOptionsMaxSegmentDurationS, g7 as PdfExportOptionsSegmentOnSilenceLongerThanS, g8 as SegmentedJsonExportOptions, g9 as SegmentedJsonExportOptionsFormat, ga as SegmentedJsonExportOptionsMaxSegmentChars, gb as SegmentedJsonExportOptionsMaxSegmentDurationS, gc as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, gd as SpeechToTextCharacterResponseModel, ge as SpeechToTextCharacterResponseModelEnd, gf as SpeechToTextCharacterResponseModelStart, gg as SpeechToTextChunkResponseModelAdditionalFormats, gh as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, gi as SpeechToTextChunkResponseModelAudioDurationSecs, gj as SpeechToTextChunkResponseModelChannelIndex, gk as SpeechToTextChunkResponseModelEntities, gl as SpeechToTextChunkResponseModelTranscriptionId, gm as SpeechToTextWordResponseModel, gn as SpeechToTextWordResponseModelCharacters, go as SpeechToTextWordResponseModelEnd, gp as SpeechToTextWordResponseModelSpeakerId, gq as SpeechToTextWordResponseModelStart, gr as SpeechToTextWordResponseModelType, gs as SrtExportOptions, gt as SrtExportOptionsFormat, gu as SrtExportOptionsMaxCharactersPerLine, gv as SrtExportOptionsMaxSegmentChars, gw as SrtExportOptionsMaxSegmentDurationS, gx as SrtExportOptionsSegmentOnSilenceLongerThanS, gy as TxtExportOptions, gz as TxtExportOptionsFormat, gA as TxtExportOptionsMaxCharactersPerLine, gB as TxtExportOptionsMaxSegmentChars, gC as TxtExportOptionsMaxSegmentDurationS, gD as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-BZSxrijj.js';
2
- export { gH as AssemblyAIExtendedData, h6 as AssemblyAIUpdateConfiguration, h2 as AudioAckEvent, h4 as AudioChunk, gN as BatchOnlyProvider, hd as BeginEvent, h1 as ChapterizationEvent, gJ as DeepgramExtendedData, gK as ElevenLabsExtendedData, h9 as ElevenLabsStreamingOptions, g$ as EntityEvent, hg as ErrorEvent, gI as GladiaExtendedData, h3 as LifecycleEvent, gV as ListTranscriptsResponse, h7 as OpenAIStreamingOptions, gL as ProviderExtendedDataMap, gW as ProviderRawResponseMap, ha as ProviderStreamingOptions, h5 as RawWebSocketMessage, g_ as SentimentEvent, gO as SessionStatus, h8 as SonioxStreamingOptions, gP as Speaker, gY as SpeechEvent, gE as SpeechmaticsOperatingPoint, gX as StreamEventType, hh as StreamingEventMessage, hj as StreamingForceEndpoint, hb as StreamingOptionsForProvider, gM as StreamingProvider, hi as StreamingWord, h0 as SummarizationEvent, hf as TerminationEvent, hc as TranscribeStreamParams, gU as TranscriptData, gT as TranscriptMetadata, gG as TranscriptionLanguage, gF as TranscriptionModel, gS as TranscriptionStatus, gZ as TranslationEvent, he as TurnEvent, gR as Utterance, gQ as Word } from './speechToTextChunkResponseModel-BZSxrijj.js';
3
- import { T as TranscriptionProvider, P as ProviderCapabilities, A as AudioInput } from './provider-metadata-DbsSGAO7.js';
4
- export { k as AllLanguageCodes, o as AllProviders, a as AssemblyAICapabilities, i as AssemblyAILanguageCodes, b as AzureCapabilities, B as BatchOnlyProviderType, q as BatchOnlyProviders, C as CapabilityKeys, f as CapabilityLabels, D as DeepgramCapabilities, j as DeepgramLanguageCodes, E as ElevenLabsCapabilities, G as GladiaCapabilities, h as GladiaLanguageCodes, g as LanguageCode, L as LanguageLabels, O as OpenAICapabilities, d as ProviderCapabilitiesMap, l as ProviderDisplayNames, n as ProviderDocs, m as ProviderWebsites, c as SonioxCapabilities, S as SpeechmaticsCapabilities, e as StreamingProviderType, p as StreamingProviders } from './provider-metadata-DbsSGAO7.js';
1
+ import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, t as NamedEntityRecognitionResult, W as WordDTO, u as StreamingSupportedRegions, v as AddonErrorDTO, w as AudioToLlmDTO, x as AudioToLlmDTOError, y as AudioToLlmDTOResults, z as AudioToLlmListConfigDTO, B as AudioToLlmListDTO, E as AudioToLlmListDTOError, H as AudioToLlmResultDTO, I as CallbackConfigDto, J as CallbackMethodEnum, K as CallbackTranscriptionErrorPayload, O as CallbackTranscriptionErrorPayloadCustomMetadata, Q as CallbackTranscriptionErrorPayloadEvent, V as CallbackTranscriptionSuccessPayload, X as CallbackTranscriptionSuccessPayloadCustomMetadata, Y as CallbackTranscriptionSuccessPayloadEvent, Z as ChapterizationDTOError, _ as ChapterizationDTOResults, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PiiRedactionConfigDTO, ai as PiiRedactionConfigDTOProcessedTextType, aj as PiiRedactionEntityTypeEnum, ak as PreRecordedRequestParamsResponse, al as PreRecordedResponseCustomMetadata, am as PreRecordedResponseFile, an as PreRecordedResponseKind, ao as PreRecordedResponsePostSessionMetadata, ap as PreRecordedResponseRequestParams, aq as PreRecordedResponseResult, ar as PreRecordedResponseStatus, as as SentencesDTO, at as SentencesDTOError, au as SentimentAnalysisDTOError, av as SpeakerReidentificationDTO, aw as SpeakerReidentificationDTOError, ax as StreamingRequest, ay as StreamingRequestCustomMetadata, az as StructuredDataExtractionDTO, aA as StructuredDataExtractionDTOError, aB as SubtitleDTO, aC as SubtitlesConfigDTO, aD as SubtitlesFormatEnum, aE as SubtitlesStyleEnum, aF as SummarizationConfigDTO, aG as SummarizationDTOError, aH as SummaryTypesEnum, aI as TranscriptionControllerListV2KindItem, aJ as TranscriptionControllerListV2Params, aK as TranscriptionControllerListV2StatusItem, aL as TranscriptionResultDTO, aM as TranslationConfigDTO, aN as TranslationDTOError, aO as TranslationModelEnum, aP as TranslationResultDTO, aQ as TranslationResultDTOError, aR as TranscriptStatus, aS as TranscriptWord, aT as AudioIntelligenceModelStatus, aU as AutoHighlightResult, aV as AutoHighlightsResult, aW as Chapter, aX as ContentSafetyLabel, aY as ContentSafetyLabelResult, aZ as ContentSafetyLabelsResult, a_ as ContentSafetyLabelsResultSeverityScoreSummary, a$ as ContentSafetyLabelsResultSummary, b0 as CustomFormattingRequestBody, b1 as CustomFormattingRequestBodyCustomFormatting, b2 as CustomFormattingResponse, b3 as CustomFormattingResponseCustomFormatting, b4 as CustomFormattingResponseCustomFormattingMapping, b5 as Entity, b6 as EntityType, b7 as ListTranscriptsParams, b8 as PiiPolicy, b9 as RedactPiiAudioQuality, ba as Sentiment, bb as SentimentAnalysisResult$1, bc as SentimentAnalysisResultChannel, bd as SentimentAnalysisResultSpeaker, be as SeverityScoreSummary, bf as SpeakerIdentificationRequestBody, bg as SpeakerIdentificationRequestBodySpeakerIdentification, bh as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, bi as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, bj as SpeakerIdentificationResponse, bk as SpeakerIdentificationResponseSpeakerIdentification, bl as SpeakerIdentificationResponseSpeakerIdentificationMapping, bm as SpeechModel, bn as SubstitutionPolicy, bo as SummaryModel, bp as SummaryType, bq as Timestamp, br as TopicDetectionModelResult, bs as TopicDetectionModelResultSummary, bt as TopicDetectionResult, bu as TopicDetectionResultLabelsItems, bv as Transcript, bw as TranscriptAudioDuration, bx as TranscriptAudioEndAt, by as TranscriptAudioStartFrom, bz as TranscriptAutoChapters, bA as TranscriptAutoHighlightsResult, bB as TranscriptChapters, bC as TranscriptConfidence, bD as TranscriptContentSafety, bE as TranscriptContentSafetyLabels, bF as TranscriptCustomSpelling, bG as TranscriptCustomSpellingProperty, bH as TranscriptCustomTopics, bI as TranscriptDisfluencies, bJ as TranscriptDomain, bK as TranscriptEntities, bL as TranscriptEntityDetection, bM as TranscriptFilterProfanity, bN as TranscriptFormatText, bO as TranscriptIabCategories, bP as TranscriptIabCategoriesResult, bQ as TranscriptLanguageCode, bR as TranscriptLanguageCodes, bS as TranscriptLanguageConfidence, bT as TranscriptLanguageConfidenceThreshold, bU as TranscriptLanguageDetection, bV as TranscriptLanguageDetectionOptions, bW as TranscriptMultichannel, bX as TranscriptOptionalParamsLanguageDetectionOptions, bY as TranscriptOptionalParamsRedactPiiAudioOptions, bZ as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, b_ as TranscriptOptionalParamsRemoveAudioTags, b$ as TranscriptOptionalParamsSpeakerOptions, c0 as TranscriptOptionalParamsSpeechUnderstanding, c1 as TranscriptOptionalParamsSpeechUnderstandingRequest, c2 as TranscriptParams, c3 as TranscriptParamsDomain, c4 as TranscriptParamsLanguageCode, c5 as TranscriptParamsLanguageCodes, c6 as TranscriptParamsRedactPiiSub, c7 as TranscriptParamsRemoveAudioTags, c8 as TranscriptParamsSpeakersExpected, c9 as TranscriptParamsSpeechThreshold, ca as TranscriptParamsWebhookAuthHeaderName, cb as TranscriptParamsWebhookAuthHeaderValue, cc as TranscriptPunctuate, cd as TranscriptRedactPiiAudio, ce as TranscriptRedactPiiAudioOptions, cf as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, cg as TranscriptRedactPiiAudioQuality, ch as TranscriptRedactPiiPolicies, ci as TranscriptRemoveAudioTags, cj as TranscriptRemoveAudioTagsProperty, ck as TranscriptSentimentAnalysis, cl as TranscriptSentimentAnalysisResults, cm as TranscriptSpeakerLabels, cn as TranscriptSpeakersExpected, co as TranscriptSpeechModel, cp as TranscriptSpeechModels, cq as TranscriptSpeechThreshold, cr as TranscriptSpeechUnderstanding, cs as TranscriptSpeechUnderstandingRequest, ct as TranscriptSpeechUnderstandingResponse, cu as TranscriptSpeedBoost, cv as TranscriptSummary, cw as TranscriptSummaryModel, cx as TranscriptSummaryType, cy as TranscriptText, cz as TranscriptThrottled, cA as TranscriptTranslatedTexts, cB as TranscriptUtterance, cC as TranscriptUtteranceChannel, cD as TranscriptUtteranceTranslatedTexts, cE as TranscriptUtterances, cF as TranscriptWebhookAuthHeaderName, cG as TranscriptWebhookStatusCode, cH as TranscriptWebhookUrl, cI as TranscriptWordChannel, cJ as TranscriptWordSpeaker, cK as TranscriptWords, cL as TranslationRequestBody, cM as TranslationRequestBodyTranslation, cN as TranslationResponse, cO as TranslationResponseTranslation, cP as StreamingUpdateConfiguration, cQ as Transcription$1, cR as EntityError, cS as Status, cT as EntityReference, cU as DiarizationProperties, cV as DiarizationSpeakersProperties, cW as LanguageIdentificationMode, cX as LanguageIdentificationProperties, cY as LanguageIdentificationPropertiesSpeechModelMapping, cZ as ProfanityFilterMode, c_ as PunctuationMode, c$ as TranscriptionCustomProperties, d0 as TranscriptionLinks, d1 as TranscriptionProperties, d2 as TranscriptTextUsageTokens, d3 as TranscriptionSegment, d4 as RealtimeSessionCreateRequestGAModel, d5 as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, d6 as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, d7 as AudioResponseFormat, d8 as CreateTranscription200One, d9 as CreateTranscriptionRequest, da as CreateTranscriptionRequestChunkingStrategy, db as CreateTranscriptionRequestChunkingStrategyAnyOf, dc as CreateTranscriptionRequestModel, dd as CreateTranscriptionRequestStream, de as CreateTranscriptionRequestTimestampGranularitiesItem, df as CreateTranscriptionResponseDiarizedJson, dg as CreateTranscriptionResponseDiarizedJsonTask, dh as CreateTranscriptionResponseDiarizedJsonUsage, di as CreateTranscriptionResponseJson, dj as CreateTranscriptionResponseJsonLogprobsItem, dk as CreateTranscriptionResponseJsonUsage, dl as CreateTranscriptionResponseVerboseJson, dm as TranscriptTextUsageDuration, dn as TranscriptTextUsageDurationType, dp as TranscriptTextUsageTokensInputTokenDetails, dq as TranscriptTextUsageTokensType, dr as TranscriptionDiarizedSegment, ds as TranscriptionDiarizedSegmentType, dt as TranscriptionInclude, du as TranscriptionWord, dv as VadConfig, dw as VadConfigType, dx as ListenV1Response, dy as SharedTopics, dz as SharedIntents, dA as SharedSentiments, dB as ListProjectRequestsParams, dC as ListenTranscribeParams, dD as ListenV1ResponseMetadata, dE as ListenV1ResponseMetadataIntentsInfo, dF as ListenV1ResponseMetadataModelInfo, dG as ListenV1ResponseMetadataSentimentInfo, dH as ListenV1ResponseMetadataSummaryInfo, dI as ListenV1ResponseMetadataTopicsInfo, dJ as ListenV1ResponseResults, dK as ListenV1ResponseResultsChannels, dL as ListenV1ResponseResultsChannelsItems, dM as ListenV1ResponseResultsChannelsItemsAlternativesItems, dN as ListenV1ResponseResultsChannelsItemsAlternativesItemsEntitiesItems, dO as ListenV1ResponseResultsChannelsItemsAlternativesItemsParagraphs, dP as ListenV1ResponseResultsChannelsItemsAlternativesItemsParagraphsParagraphsItems, dQ as ListenV1ResponseResultsChannelsItemsAlternativesItemsParagraphsParagraphsItemsSentencesItems, dR as ListenV1ResponseResultsChannelsItemsAlternativesItemsSummariesItems, dS as ListenV1ResponseResultsChannelsItemsAlternativesItemsTopicsItems, dT as ListenV1ResponseResultsChannelsItemsAlternativesItemsWordsItems, dU as ListenV1ResponseResultsChannelsItemsSearchItems, dV as ListenV1ResponseResultsChannelsItemsSearchItemsHitsItems, dW as ListenV1ResponseResultsSummary, dX as ListenV1ResponseResultsUtterances, dY as ListenV1ResponseResultsUtterancesItems, dZ as ListenV1ResponseResultsUtterancesItemsWordsItems, d_ as SharedIntentsResults, d$ as SharedIntentsResultsIntents, e0 as SharedIntentsResultsIntentsSegmentsItems, e1 as SharedIntentsResultsIntentsSegmentsItemsIntentsItems, e2 as SharedSentimentsAverage, e3 as SharedSentimentsSegmentsItems, e4 as SharedTopicsResults, e5 as SharedTopicsResultsTopics, e6 as SharedTopicsResultsTopicsSegmentsItems, e7 as SharedTopicsResultsTopicsSegmentsItemsTopicsItems, e8 as V1ListenPostParametersCallbackMethod, e9 as V1ListenPostParametersCustomIntent, ea as V1ListenPostParametersCustomIntentMode, eb as V1ListenPostParametersCustomTopic, ec as V1ListenPostParametersCustomTopicMode, ed as V1ListenPostParametersDetectLanguage, ee as V1ListenPostParametersEncoding, ef as V1ListenPostParametersExtra, eg as V1ListenPostParametersKeywords, eh as V1ListenPostParametersModel, ei as V1ListenPostParametersModel0, ej as V1ListenPostParametersRedact, ek as V1ListenPostParametersRedact1, el as V1ListenPostParametersRedactSchemaOneOf1Items, em as V1ListenPostParametersReplace, en as V1ListenPostParametersSearch, eo as V1ListenPostParametersSummarize, ep as V1ListenPostParametersSummarize0, eq as V1ListenPostParametersTag, er as V1ListenPostParametersVersion, es as V1ListenPostParametersVersion0, et as V1ProjectsProjectIdRequestsGetParametersDeployment, eu as V1ProjectsProjectIdRequestsGetParametersEndpoint, ev as V1ProjectsProjectIdRequestsGetParametersMethod, ew as V1ProjectsProjectIdRequestsGetParametersStatus, ex as JobType, ey as AlignmentConfig, ez as TranscriptionConfig, eA as TrackingData, eB as OutputConfig, eC as OperatingPoint, eD as AudioEventItem, eE as AudioEventSummary, eF as AudioEventSummaryItem, eG as AutoChaptersResult, eH as AutoChaptersResultError, eI as AutoChaptersResultErrorType, eJ as Chapter$1, eK as JobInfo, eL as LanguageIdentificationResult, eM as LanguageIdentificationResultAlternative, eN as LanguageIdentificationResultError, eO as LanguageIdentificationResultItem, eP as LanguagePackInfo, eQ as LanguagePackInfoWritingDirection, eR as OutputConfigSrtOverrides, eS as RecognitionAlternative, eT as RecognitionDisplay, eU as RecognitionDisplayDirection, eV as RecognitionMetadata, eW as RecognitionResult, eX as RecognitionResultAttachesTo, eY as RecognitionResultType, eZ as RetrieveTranscriptResponse, e_ as RetrieveTranscriptResponseAudioEventSummary, e$ as RetrieveTranscriptResponseAudioEventSummaryChannels, f0 as RetrieveTranscriptResponseTranslations, f1 as SentimentAnalysisError, f2 as SentimentAnalysisErrorType, f3 as SentimentAnalysisResult$2, f4 as SentimentAnalysisResultSentimentAnalysis, f5 as SentimentChannelSummary, f6 as SentimentSegment, f7 as SentimentSpeakerSummary, f8 as SentimentSummary, f9 as SentimentSummaryDetail, fa as SpokenFormRecognitionResult, fb as SpokenFormRecognitionResultType, fc as SummarizationError, fd as SummarizationErrorType, fe as SummarizationResult, ff as TopicDetectionError, fg as TopicDetectionErrorType, fh as TopicDetectionResult$1, fi as TopicDetectionSegment, fj as TopicDetectionSegmentTopic, fk as TopicDetectionSummary, fl as TopicDetectionSummaryOverall, fm as TrackingDataDetails, fn as TranscriptionConfigAdditionalVocabItem, fo as TranscriptionConfigDiarization, fp as TranscriptionConfigMaxDelayMode, fq as TranscriptionConfigPunctuationOverrides, fr as TranscriptionConfigSpeakerDiarizationConfig, fs as TranscriptionConfigTranscriptFilteringConfig, ft as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fu as TranslationError, fv as TranslationErrorType, fw as TranslationSentence, fx as WrittenFormRecognitionResult, fy as WrittenFormRecognitionResultType, fz as SpeechToTextChunkResponseModel, fA as AdditionalFormatResponseModel, fB as AdditionalFormats, fC as BodySpeechToTextV1SpeechToTextPost, fD as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fE as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, fF as BodySpeechToTextV1SpeechToTextPostEntityDetection, fG as BodySpeechToTextV1SpeechToTextPostEntityRedaction, fH as BodySpeechToTextV1SpeechToTextPostFile, fI as BodySpeechToTextV1SpeechToTextPostFileFormat, fJ as BodySpeechToTextV1SpeechToTextPostLanguageCode, fK as BodySpeechToTextV1SpeechToTextPostModelId, fL as BodySpeechToTextV1SpeechToTextPostNumSpeakers, fM as BodySpeechToTextV1SpeechToTextPostSeed, fN as BodySpeechToTextV1SpeechToTextPostSourceUrl, fO as BodySpeechToTextV1SpeechToTextPostTemperature, fP as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, fQ as BodySpeechToTextV1SpeechToTextPostWebhookId, fR as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, fS as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, fT as DetectedEntity, fU as DocxExportOptions, fV as DocxExportOptionsFormat, fW as DocxExportOptionsMaxSegmentChars, fX as DocxExportOptionsMaxSegmentDurationS, fY as DocxExportOptionsSegmentOnSilenceLongerThanS, fZ as ExportOptions, f_ as HtmlExportOptions, f$ as HtmlExportOptionsFormat, g0 as HtmlExportOptionsMaxSegmentChars, g1 as HtmlExportOptionsMaxSegmentDurationS, g2 as HtmlExportOptionsSegmentOnSilenceLongerThanS, g3 as PdfExportOptions, g4 as PdfExportOptionsFormat, g5 as PdfExportOptionsMaxSegmentChars, g6 as PdfExportOptionsMaxSegmentDurationS, g7 as PdfExportOptionsSegmentOnSilenceLongerThanS, g8 as SegmentedJsonExportOptions, g9 as SegmentedJsonExportOptionsFormat, ga as SegmentedJsonExportOptionsMaxSegmentChars, gb as SegmentedJsonExportOptionsMaxSegmentDurationS, gc as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, gd as SpeechToTextCharacterResponseModel, ge as SpeechToTextCharacterResponseModelEnd, gf as SpeechToTextCharacterResponseModelStart, gg as SpeechToTextChunkResponseModelAdditionalFormats, gh as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, gi as SpeechToTextChunkResponseModelAudioDurationSecs, gj as SpeechToTextChunkResponseModelChannelIndex, gk as SpeechToTextChunkResponseModelEntities, gl as SpeechToTextChunkResponseModelTranscriptionId, gm as SpeechToTextWordResponseModel, gn as SpeechToTextWordResponseModelCharacters, go as SpeechToTextWordResponseModelEnd, gp as SpeechToTextWordResponseModelSpeakerId, gq as SpeechToTextWordResponseModelStart, gr as SpeechToTextWordResponseModelType, gs as SrtExportOptions, gt as SrtExportOptionsFormat, gu as SrtExportOptionsMaxCharactersPerLine, gv as SrtExportOptionsMaxSegmentChars, gw as SrtExportOptionsMaxSegmentDurationS, gx as SrtExportOptionsSegmentOnSilenceLongerThanS, gy as TxtExportOptions, gz as TxtExportOptionsFormat, gA as TxtExportOptionsMaxCharactersPerLine, gB as TxtExportOptionsMaxSegmentChars, gC as TxtExportOptionsMaxSegmentDurationS, gD as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-o8_dfC4c.js';
2
+ export { gH as AssemblyAIExtendedData, h6 as AssemblyAIUpdateConfiguration, h2 as AudioAckEvent, h4 as AudioChunk, gN as BatchOnlyProvider, hd as BeginEvent, h1 as ChapterizationEvent, gJ as DeepgramExtendedData, gK as ElevenLabsExtendedData, h9 as ElevenLabsStreamingOptions, g$ as EntityEvent, hg as ErrorEvent, gI as GladiaExtendedData, h3 as LifecycleEvent, gV as ListTranscriptsResponse, h7 as OpenAIStreamingOptions, gL as ProviderExtendedDataMap, gW as ProviderRawResponseMap, ha as ProviderStreamingOptions, h5 as RawWebSocketMessage, g_ as SentimentEvent, gO as SessionStatus, h8 as SonioxStreamingOptions, gP as Speaker, gY as SpeechEvent, gE as SpeechmaticsOperatingPoint, gX as StreamEventType, hh as StreamingEventMessage, hj as StreamingForceEndpoint, hb as StreamingOptionsForProvider, gM as StreamingProvider, hi as StreamingWord, h0 as SummarizationEvent, hf as TerminationEvent, hc as TranscribeStreamParams, gU as TranscriptData, gT as TranscriptMetadata, gG as TranscriptionLanguage, gF as TranscriptionModel, gS as TranscriptionStatus, gZ as TranslationEvent, he as TurnEvent, gR as Utterance, gQ as Word } from './speechToTextChunkResponseModel-o8_dfC4c.js';
3
+ import { T as TranscriptionProvider, P as ProviderCapabilities, A as AudioInput } from './provider-metadata-_gUWlRXS.js';
4
+ export { k as AllLanguageCodes, o as AllProviders, a as AssemblyAICapabilities, i as AssemblyAILanguageCodes, b as AzureCapabilities, B as BatchOnlyProviderType, q as BatchOnlyProviders, C as CapabilityKeys, f as CapabilityLabels, D as DeepgramCapabilities, j as DeepgramLanguageCodes, E as ElevenLabsCapabilities, G as GladiaCapabilities, h as GladiaLanguageCodes, g as LanguageCode, L as LanguageLabels, O as OpenAICapabilities, d as ProviderCapabilitiesMap, l as ProviderDisplayNames, n as ProviderDocs, m as ProviderWebsites, c as SonioxCapabilities, S as SpeechmaticsCapabilities, e as StreamingProviderType, p as StreamingProviders } from './provider-metadata-_gUWlRXS.js';
5
5
  import { AssemblyAIRegionType, DeepgramRegionType, SpeechmaticsRegionType, SonioxModelCode, SonioxRegionType, ElevenLabsModelCode, ElevenLabsRegionType } from './constants.js';
6
6
  export { AssemblyAIEncoding, AssemblyAIEncodingType, AssemblyAILanguage, AssemblyAILanguageType, AssemblyAISampleRate, AssemblyAISampleRateType, AssemblyAISpeechModel, AssemblyAISpeechModelType, AssemblyAIStatus, AssemblyAITranscriptionModel, AssemblyAITranscriptionModelType, AzureLocale, AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, AzureLocaleType, AzureLocales, DeepgramCallbackMethod, DeepgramEncoding, DeepgramIntentMode, DeepgramLanguage, DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, DeepgramModelCode as DeepgramModelType, DeepgramRedact, DeepgramRedactType, DeepgramRegion, DeepgramSampleRate, DeepgramStatus, DeepgramTTSContainer, DeepgramTTSEncoding, DeepgramTTSModel, DeepgramTTSSampleRate, DeepgramTopicMode, DeepgramTopicModeType, ElevenLabsLanguageCode, ElevenLabsLanguageCodes, ElevenLabsLanguageLabels, ElevenLabsLanguages, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaRegion, GladiaSampleRate, GladiaStatus, GladiaTranslationLanguage, OpenAILanguageCodes, OpenAIModel, OpenAIRealtimeAudioFormat, OpenAIRealtimeModel, OpenAIRealtimeTranscriptionModel, OpenAIRealtimeTurnDetection, OpenAIResponseFormat, SonioxAsyncModel, SonioxAsyncModelCode, SonioxLanguage, SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, SonioxLanguageType, SonioxLanguages, SonioxModel, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, SonioxRealtimeModelCode, SonioxRegion, SpeechmaticsLanguage, SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsRegion } from './constants.js';
7
- export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-D1RCJSmr.js';
7
+ export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-2c1-pid1.js';
8
8
  import { z } from 'zod';
9
9
  import { AxiosResponse, AxiosRequestConfig } from 'axios';
10
10
 
@@ -12856,8 +12856,9 @@ declare class SonioxAdapter extends BaseAdapter {
12856
12856
  /**
12857
12857
  * Get transcription result by ID
12858
12858
  *
12859
- * Soniox batch transcription is synchronous (returns immediately),
12860
- * but this method can be used for consistency with other providers.
12859
+ * Checks job status via GET /v1/transcriptions/{id}, then fetches
12860
+ * the full transcript via GET /v1/transcriptions/{id}/transcript
12861
+ * when completed.
12861
12862
  *
12862
12863
  * @param transcriptId - Transcript ID
12863
12864
  * @returns Transcription response
@@ -16797,7 +16798,7 @@ declare const getJobsResponse: z.ZodObject<{
16797
16798
  language: string;
16798
16799
  diarization?: "none" | "speaker" | "channel" | undefined;
16799
16800
  domain?: string | undefined;
16800
- operating_point?: "enhanced" | "standard" | undefined;
16801
+ operating_point?: "standard" | "enhanced" | undefined;
16801
16802
  output_locale?: string | undefined;
16802
16803
  additional_vocab?: {
16803
16804
  content: string;
@@ -16825,7 +16826,7 @@ declare const getJobsResponse: z.ZodObject<{
16825
16826
  language: string;
16826
16827
  diarization?: "none" | "speaker" | "channel" | undefined;
16827
16828
  domain?: string | undefined;
16828
- operating_point?: "enhanced" | "standard" | undefined;
16829
+ operating_point?: "standard" | "enhanced" | undefined;
16829
16830
  output_locale?: string | undefined;
16830
16831
  additional_vocab?: {
16831
16832
  content: string;
@@ -16984,7 +16985,7 @@ declare const getJobsResponse: z.ZodObject<{
16984
16985
  language: string;
16985
16986
  diarization?: "none" | "speaker" | "channel" | undefined;
16986
16987
  domain?: string | undefined;
16987
- operating_point?: "enhanced" | "standard" | undefined;
16988
+ operating_point?: "standard" | "enhanced" | undefined;
16988
16989
  output_locale?: string | undefined;
16989
16990
  additional_vocab?: {
16990
16991
  content: string;
@@ -17065,7 +17066,7 @@ declare const getJobsResponse: z.ZodObject<{
17065
17066
  language: string;
17066
17067
  diarization?: "none" | "speaker" | "channel" | undefined;
17067
17068
  domain?: string | undefined;
17068
- operating_point?: "enhanced" | "standard" | undefined;
17069
+ operating_point?: "standard" | "enhanced" | undefined;
17069
17070
  output_locale?: string | undefined;
17070
17071
  additional_vocab?: {
17071
17072
  content: string;
@@ -17164,7 +17165,7 @@ declare const getJobsResponse: z.ZodObject<{
17164
17165
  language: string;
17165
17166
  diarization?: "none" | "speaker" | "channel" | undefined;
17166
17167
  domain?: string | undefined;
17167
- operating_point?: "enhanced" | "standard" | undefined;
17168
+ operating_point?: "standard" | "enhanced" | undefined;
17168
17169
  output_locale?: string | undefined;
17169
17170
  additional_vocab?: {
17170
17171
  content: string;
@@ -17258,7 +17259,7 @@ declare const getJobsResponse: z.ZodObject<{
17258
17259
  language: string;
17259
17260
  diarization?: "none" | "speaker" | "channel" | undefined;
17260
17261
  domain?: string | undefined;
17261
- operating_point?: "enhanced" | "standard" | undefined;
17262
+ operating_point?: "standard" | "enhanced" | undefined;
17262
17263
  output_locale?: string | undefined;
17263
17264
  additional_vocab?: {
17264
17265
  content: string;
@@ -17354,7 +17355,7 @@ declare const getJobsResponse: z.ZodObject<{
17354
17355
  language: string;
17355
17356
  diarization?: "none" | "speaker" | "channel" | undefined;
17356
17357
  domain?: string | undefined;
17357
- operating_point?: "enhanced" | "standard" | undefined;
17358
+ operating_point?: "standard" | "enhanced" | undefined;
17358
17359
  output_locale?: string | undefined;
17359
17360
  additional_vocab?: {
17360
17361
  content: string;
@@ -17450,7 +17451,7 @@ declare const getJobsResponse: z.ZodObject<{
17450
17451
  language: string;
17451
17452
  diarization?: "none" | "speaker" | "channel" | undefined;
17452
17453
  domain?: string | undefined;
17453
- operating_point?: "enhanced" | "standard" | undefined;
17454
+ operating_point?: "standard" | "enhanced" | undefined;
17454
17455
  output_locale?: string | undefined;
17455
17456
  additional_vocab?: {
17456
17457
  content: string;
@@ -17632,7 +17633,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
17632
17633
  language: string;
17633
17634
  diarization?: "none" | "speaker" | "channel" | undefined;
17634
17635
  domain?: string | undefined;
17635
- operating_point?: "enhanced" | "standard" | undefined;
17636
+ operating_point?: "standard" | "enhanced" | undefined;
17636
17637
  output_locale?: string | undefined;
17637
17638
  additional_vocab?: {
17638
17639
  content: string;
@@ -17660,7 +17661,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
17660
17661
  language: string;
17661
17662
  diarization?: "none" | "speaker" | "channel" | undefined;
17662
17663
  domain?: string | undefined;
17663
- operating_point?: "enhanced" | "standard" | undefined;
17664
+ operating_point?: "standard" | "enhanced" | undefined;
17664
17665
  output_locale?: string | undefined;
17665
17666
  additional_vocab?: {
17666
17667
  content: string;
@@ -17819,7 +17820,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
17819
17820
  language: string;
17820
17821
  diarization?: "none" | "speaker" | "channel" | undefined;
17821
17822
  domain?: string | undefined;
17822
- operating_point?: "enhanced" | "standard" | undefined;
17823
+ operating_point?: "standard" | "enhanced" | undefined;
17823
17824
  output_locale?: string | undefined;
17824
17825
  additional_vocab?: {
17825
17826
  content: string;
@@ -17900,7 +17901,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
17900
17901
  language: string;
17901
17902
  diarization?: "none" | "speaker" | "channel" | undefined;
17902
17903
  domain?: string | undefined;
17903
- operating_point?: "enhanced" | "standard" | undefined;
17904
+ operating_point?: "standard" | "enhanced" | undefined;
17904
17905
  output_locale?: string | undefined;
17905
17906
  additional_vocab?: {
17906
17907
  content: string;
@@ -17999,7 +18000,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
17999
18000
  language: string;
18000
18001
  diarization?: "none" | "speaker" | "channel" | undefined;
18001
18002
  domain?: string | undefined;
18002
- operating_point?: "enhanced" | "standard" | undefined;
18003
+ operating_point?: "standard" | "enhanced" | undefined;
18003
18004
  output_locale?: string | undefined;
18004
18005
  additional_vocab?: {
18005
18006
  content: string;
@@ -18093,7 +18094,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
18093
18094
  language: string;
18094
18095
  diarization?: "none" | "speaker" | "channel" | undefined;
18095
18096
  domain?: string | undefined;
18096
- operating_point?: "enhanced" | "standard" | undefined;
18097
+ operating_point?: "standard" | "enhanced" | undefined;
18097
18098
  output_locale?: string | undefined;
18098
18099
  additional_vocab?: {
18099
18100
  content: string;
@@ -18189,7 +18190,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
18189
18190
  language: string;
18190
18191
  diarization?: "none" | "speaker" | "channel" | undefined;
18191
18192
  domain?: string | undefined;
18192
- operating_point?: "enhanced" | "standard" | undefined;
18193
+ operating_point?: "standard" | "enhanced" | undefined;
18193
18194
  output_locale?: string | undefined;
18194
18195
  additional_vocab?: {
18195
18196
  content: string;
@@ -18285,7 +18286,7 @@ declare const getJobsJobidResponse: z.ZodObject<{
18285
18286
  language: string;
18286
18287
  diarization?: "none" | "speaker" | "channel" | undefined;
18287
18288
  domain?: string | undefined;
18288
- operating_point?: "enhanced" | "standard" | undefined;
18289
+ operating_point?: "standard" | "enhanced" | undefined;
18289
18290
  output_locale?: string | undefined;
18290
18291
  additional_vocab?: {
18291
18292
  content: string;
@@ -18474,7 +18475,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
18474
18475
  language: string;
18475
18476
  diarization?: "none" | "speaker" | "channel" | undefined;
18476
18477
  domain?: string | undefined;
18477
- operating_point?: "enhanced" | "standard" | undefined;
18478
+ operating_point?: "standard" | "enhanced" | undefined;
18478
18479
  output_locale?: string | undefined;
18479
18480
  additional_vocab?: {
18480
18481
  content: string;
@@ -18502,7 +18503,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
18502
18503
  language: string;
18503
18504
  diarization?: "none" | "speaker" | "channel" | undefined;
18504
18505
  domain?: string | undefined;
18505
- operating_point?: "enhanced" | "standard" | undefined;
18506
+ operating_point?: "standard" | "enhanced" | undefined;
18506
18507
  output_locale?: string | undefined;
18507
18508
  additional_vocab?: {
18508
18509
  content: string;
@@ -18661,7 +18662,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
18661
18662
  language: string;
18662
18663
  diarization?: "none" | "speaker" | "channel" | undefined;
18663
18664
  domain?: string | undefined;
18664
- operating_point?: "enhanced" | "standard" | undefined;
18665
+ operating_point?: "standard" | "enhanced" | undefined;
18665
18666
  output_locale?: string | undefined;
18666
18667
  additional_vocab?: {
18667
18668
  content: string;
@@ -18742,7 +18743,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
18742
18743
  language: string;
18743
18744
  diarization?: "none" | "speaker" | "channel" | undefined;
18744
18745
  domain?: string | undefined;
18745
- operating_point?: "enhanced" | "standard" | undefined;
18746
+ operating_point?: "standard" | "enhanced" | undefined;
18746
18747
  output_locale?: string | undefined;
18747
18748
  additional_vocab?: {
18748
18749
  content: string;
@@ -18841,7 +18842,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
18841
18842
  language: string;
18842
18843
  diarization?: "none" | "speaker" | "channel" | undefined;
18843
18844
  domain?: string | undefined;
18844
- operating_point?: "enhanced" | "standard" | undefined;
18845
+ operating_point?: "standard" | "enhanced" | undefined;
18845
18846
  output_locale?: string | undefined;
18846
18847
  additional_vocab?: {
18847
18848
  content: string;
@@ -18935,7 +18936,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
18935
18936
  language: string;
18936
18937
  diarization?: "none" | "speaker" | "channel" | undefined;
18937
18938
  domain?: string | undefined;
18938
- operating_point?: "enhanced" | "standard" | undefined;
18939
+ operating_point?: "standard" | "enhanced" | undefined;
18939
18940
  output_locale?: string | undefined;
18940
18941
  additional_vocab?: {
18941
18942
  content: string;
@@ -19031,7 +19032,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
19031
19032
  language: string;
19032
19033
  diarization?: "none" | "speaker" | "channel" | undefined;
19033
19034
  domain?: string | undefined;
19034
- operating_point?: "enhanced" | "standard" | undefined;
19035
+ operating_point?: "standard" | "enhanced" | undefined;
19035
19036
  output_locale?: string | undefined;
19036
19037
  additional_vocab?: {
19037
19038
  content: string;
@@ -19127,7 +19128,7 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
19127
19128
  language: string;
19128
19129
  diarization?: "none" | "speaker" | "channel" | undefined;
19129
19130
  domain?: string | undefined;
19130
- operating_point?: "enhanced" | "standard" | undefined;
19131
+ operating_point?: "standard" | "enhanced" | undefined;
19131
19132
  output_locale?: string | undefined;
19132
19133
  additional_vocab?: {
19133
19134
  content: string;
@@ -19354,7 +19355,7 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
19354
19355
  language: string;
19355
19356
  diarization?: "none" | "speaker" | "channel" | undefined;
19356
19357
  domain?: string | undefined;
19357
- operating_point?: "enhanced" | "standard" | undefined;
19358
+ operating_point?: "standard" | "enhanced" | undefined;
19358
19359
  output_locale?: string | undefined;
19359
19360
  additional_vocab?: {
19360
19361
  content: string;
@@ -19382,7 +19383,7 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
19382
19383
  language: string;
19383
19384
  diarization?: "none" | "speaker" | "channel" | undefined;
19384
19385
  domain?: string | undefined;
19385
- operating_point?: "enhanced" | "standard" | undefined;
19386
+ operating_point?: "standard" | "enhanced" | undefined;
19386
19387
  output_locale?: string | undefined;
19387
19388
  additional_vocab?: {
19388
19389
  content: string;
@@ -19570,7 +19571,7 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
19570
19571
  language: string;
19571
19572
  diarization?: "none" | "speaker" | "channel" | undefined;
19572
19573
  domain?: string | undefined;
19573
- operating_point?: "enhanced" | "standard" | undefined;
19574
+ operating_point?: "standard" | "enhanced" | undefined;
19574
19575
  output_locale?: string | undefined;
19575
19576
  additional_vocab?: {
19576
19577
  content: string;
@@ -19651,7 +19652,7 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
19651
19652
  language: string;
19652
19653
  diarization?: "none" | "speaker" | "channel" | undefined;
19653
19654
  domain?: string | undefined;
19654
- operating_point?: "enhanced" | "standard" | undefined;
19655
+ operating_point?: "standard" | "enhanced" | undefined;
19655
19656
  output_locale?: string | undefined;
19656
19657
  additional_vocab?: {
19657
19658
  content: string;
@@ -20393,7 +20394,7 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
20393
20394
  language: string;
20394
20395
  diarization?: "none" | "speaker" | "channel" | undefined;
20395
20396
  domain?: string | undefined;
20396
- operating_point?: "enhanced" | "standard" | undefined;
20397
+ operating_point?: "standard" | "enhanced" | undefined;
20397
20398
  output_locale?: string | undefined;
20398
20399
  additional_vocab?: {
20399
20400
  content: string;
@@ -20617,7 +20618,7 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
20617
20618
  language: string;
20618
20619
  diarization?: "none" | "speaker" | "channel" | undefined;
20619
20620
  domain?: string | undefined;
20620
- operating_point?: "enhanced" | "standard" | undefined;
20621
+ operating_point?: "standard" | "enhanced" | undefined;
20621
20622
  output_locale?: string | undefined;
20622
20623
  additional_vocab?: {
20623
20624
  content: string;
@@ -20889,14 +20890,14 @@ declare const getUsageResponse: z.ZodObject<{
20889
20890
  count: number;
20890
20891
  duration_hrs: number;
20891
20892
  language?: string | undefined;
20892
- operating_point?: "enhanced" | "standard" | undefined;
20893
+ operating_point?: "standard" | "enhanced" | undefined;
20893
20894
  }, {
20894
20895
  type: "transcription" | "alignment";
20895
20896
  mode: "batch";
20896
20897
  count: number;
20897
20898
  duration_hrs: number;
20898
20899
  language?: string | undefined;
20899
- operating_point?: "enhanced" | "standard" | undefined;
20900
+ operating_point?: "standard" | "enhanced" | undefined;
20900
20901
  }>, "many">;
20901
20902
  details: z.ZodArray<z.ZodObject<{
20902
20903
  mode: z.ZodEnum<["batch"]>;
@@ -20911,14 +20912,14 @@ declare const getUsageResponse: z.ZodObject<{
20911
20912
  count: number;
20912
20913
  duration_hrs: number;
20913
20914
  language?: string | undefined;
20914
- operating_point?: "enhanced" | "standard" | undefined;
20915
+ operating_point?: "standard" | "enhanced" | undefined;
20915
20916
  }, {
20916
20917
  type: "transcription" | "alignment";
20917
20918
  mode: "batch";
20918
20919
  count: number;
20919
20920
  duration_hrs: number;
20920
20921
  language?: string | undefined;
20921
- operating_point?: "enhanced" | "standard" | undefined;
20922
+ operating_point?: "standard" | "enhanced" | undefined;
20922
20923
  }>, "many">;
20923
20924
  }, "strip", z.ZodTypeAny, {
20924
20925
  summary: {
@@ -20927,7 +20928,7 @@ declare const getUsageResponse: z.ZodObject<{
20927
20928
  count: number;
20928
20929
  duration_hrs: number;
20929
20930
  language?: string | undefined;
20930
- operating_point?: "enhanced" | "standard" | undefined;
20931
+ operating_point?: "standard" | "enhanced" | undefined;
20931
20932
  }[];
20932
20933
  details: {
20933
20934
  type: "transcription" | "alignment";
@@ -20935,7 +20936,7 @@ declare const getUsageResponse: z.ZodObject<{
20935
20936
  count: number;
20936
20937
  duration_hrs: number;
20937
20938
  language?: string | undefined;
20938
- operating_point?: "enhanced" | "standard" | undefined;
20939
+ operating_point?: "standard" | "enhanced" | undefined;
20939
20940
  }[];
20940
20941
  since: string;
20941
20942
  until: string;
@@ -20946,7 +20947,7 @@ declare const getUsageResponse: z.ZodObject<{
20946
20947
  count: number;
20947
20948
  duration_hrs: number;
20948
20949
  language?: string | undefined;
20949
- operating_point?: "enhanced" | "standard" | undefined;
20950
+ operating_point?: "standard" | "enhanced" | undefined;
20950
20951
  }[];
20951
20952
  details: {
20952
20953
  type: "transcription" | "alignment";
@@ -20954,7 +20955,7 @@ declare const getUsageResponse: z.ZodObject<{
20954
20955
  count: number;
20955
20956
  duration_hrs: number;
20956
20957
  language?: string | undefined;
20957
- operating_point?: "enhanced" | "standard" | undefined;
20958
+ operating_point?: "standard" | "enhanced" | undefined;
20958
20959
  }[];
20959
20960
  since: string;
20960
20961
  until: string;
package/dist/index.js CHANGED
@@ -8270,6 +8270,7 @@ var AzureSTTAdapter = class extends BaseAdapter {
8270
8270
  id: String(speakerId),
8271
8271
  label: `Speaker ${speakerId}`
8272
8272
  })) : void 0;
8273
+ const utterances = words.length > 0 ? buildUtterancesFromWords(words) : void 0;
8273
8274
  const transcriptionId = transcription.self?.split("/").pop() || "";
8274
8275
  return {
8275
8276
  success: true,
@@ -8283,6 +8284,7 @@ var AzureSTTAdapter = class extends BaseAdapter {
8283
8284
  duration: transcriptionData.duration ? transcriptionData.duration / 1e7 : void 0,
8284
8285
  speakers,
8285
8286
  words: words.length > 0 ? words : void 0,
8287
+ utterances: utterances && utterances.length > 0 ? utterances : void 0,
8286
8288
  createdAt: transcription.createdDateTime,
8287
8289
  completedAt: transcription.lastActionDateTime
8288
8290
  },
@@ -9103,13 +9105,16 @@ var SpeechmaticsAdapter = class extends BaseAdapter {
9103
9105
  jobConfig.fetch_data = {
9104
9106
  url: audio.url
9105
9107
  };
9106
- requestBody = { config: JSON.stringify(jobConfig) };
9107
- headers = { "Content-Type": "application/json" };
9108
+ const formData = new FormData();
9109
+ formData.append("config", JSON.stringify(jobConfig));
9110
+ requestBody = formData;
9111
+ headers = { "Content-Type": "multipart/form-data" };
9108
9112
  } else if (audio.type === "file") {
9109
- requestBody = {
9110
- config: JSON.stringify(jobConfig),
9111
- data_file: audio.file
9112
- };
9113
+ const formData = new FormData();
9114
+ formData.append("config", JSON.stringify(jobConfig));
9115
+ const audioBlob = audio.file instanceof Blob ? audio.file : new Blob([audio.file], { type: audio.mimeType || "audio/wav" });
9116
+ formData.append("data_file", audioBlob, audio.filename || "audio.wav");
9117
+ requestBody = formData;
9113
9118
  headers = { "Content-Type": "multipart/form-data" };
9114
9119
  } else {
9115
9120
  return {
@@ -9432,26 +9437,13 @@ var SonioxAdapter = class extends BaseAdapter {
9432
9437
  } else if (audio.type === "file") {
9433
9438
  const formData = new FormData();
9434
9439
  const audioBlob = audio.file instanceof Blob ? audio.file : new Blob([audio.file], { type: audio.mimeType || "audio/wav" });
9435
- formData.append("audio", audioBlob, audio.filename || "audio.wav");
9436
- formData.append("model", requestBody.model);
9437
- if (options?.language) {
9438
- formData.append("language_hints", JSON.stringify([options.language]));
9439
- }
9440
- if (options?.diarization) {
9441
- formData.append("enable_speaker_diarization", "true");
9442
- }
9443
- if (options?.languageDetection) {
9444
- formData.append("enable_language_identification", "true");
9445
- }
9446
- if (options?.customVocabulary) {
9447
- formData.append("context", JSON.stringify({ terms: options.customVocabulary }));
9448
- }
9449
- const response2 = await this.client.post("/speech/transcribe", formData, {
9440
+ formData.append("file", audioBlob, audio.filename || "audio.wav");
9441
+ const uploadResponse = await this.client.post("/files", formData, {
9450
9442
  headers: {
9451
9443
  "Content-Type": "multipart/form-data"
9452
9444
  }
9453
9445
  });
9454
- return this.normalizeResponse(response2.data);
9446
+ requestBody.file_id = uploadResponse.data.id;
9455
9447
  } else {
9456
9448
  return {
9457
9449
  success: false,
@@ -9476,8 +9468,9 @@ var SonioxAdapter = class extends BaseAdapter {
9476
9468
  terms: options.customVocabulary
9477
9469
  };
9478
9470
  }
9479
- const response = await this.client.post("/speech/transcribe", requestBody);
9480
- return this.normalizeResponse(response.data);
9471
+ const response = await this.client.post("/transcriptions", requestBody);
9472
+ const transcriptionId = response.data.id;
9473
+ return await this.pollForCompletion(transcriptionId);
9481
9474
  } catch (error) {
9482
9475
  return this.createErrorResponse(error);
9483
9476
  }
@@ -9485,8 +9478,9 @@ var SonioxAdapter = class extends BaseAdapter {
9485
9478
  /**
9486
9479
  * Get transcription result by ID
9487
9480
  *
9488
- * Soniox batch transcription is synchronous (returns immediately),
9489
- * but this method can be used for consistency with other providers.
9481
+ * Checks job status via GET /v1/transcriptions/{id}, then fetches
9482
+ * the full transcript via GET /v1/transcriptions/{id}/transcript
9483
+ * when completed.
9490
9484
  *
9491
9485
  * @param transcriptId - Transcript ID
9492
9486
  * @returns Transcription response
@@ -9494,8 +9488,39 @@ var SonioxAdapter = class extends BaseAdapter {
9494
9488
  async getTranscript(transcriptId) {
9495
9489
  this.validateConfig();
9496
9490
  try {
9497
- const response = await this.client.get(`/speech/transcripts/${transcriptId}`);
9498
- return this.normalizeResponse(response.data);
9491
+ const statusResponse = await this.client.get(`/transcriptions/${transcriptId}`);
9492
+ const job = statusResponse.data;
9493
+ if (job.status === "error") {
9494
+ return {
9495
+ success: false,
9496
+ provider: this.name,
9497
+ error: {
9498
+ code: "TRANSCRIPTION_ERROR",
9499
+ message: job.error_message || "Transcription failed"
9500
+ }
9501
+ };
9502
+ }
9503
+ if (job.status !== "completed") {
9504
+ return {
9505
+ success: true,
9506
+ provider: this.name,
9507
+ data: {
9508
+ id: job.id,
9509
+ text: "",
9510
+ status: job.status
9511
+ },
9512
+ raw: job
9513
+ };
9514
+ }
9515
+ const transcriptResponse = await this.client.get(
9516
+ `/transcriptions/${transcriptId}/transcript`
9517
+ );
9518
+ return this.normalizeResponse({
9519
+ ...transcriptResponse.data,
9520
+ // Carry over job metadata
9521
+ id: job.id,
9522
+ audio_duration_ms: job.audio_duration_ms
9523
+ });
9499
9524
  } catch (error) {
9500
9525
  return this.createErrorResponse(error);
9501
9526
  }
@@ -9787,8 +9812,10 @@ var SonioxAdapter = class extends BaseAdapter {
9787
9812
  * Normalize Soniox response to unified format
9788
9813
  */
9789
9814
  normalizeResponse(response) {
9790
- const text = response.text || (response.tokens ? response.tokens.filter((t) => t.is_final).map((t) => t.text).join("") : "");
9791
- const words = response.tokens ? response.tokens.filter((t) => t.is_final && t.start_ms !== void 0 && t.end_ms !== void 0).map((token) => ({
9815
+ const text = response.text || (response.tokens ? response.tokens.filter((t) => t.is_final !== false).map((t) => t.text).join("") : "");
9816
+ const words = response.tokens ? response.tokens.filter(
9817
+ (t) => t.is_final !== false && t.start_ms !== void 0 && t.end_ms !== void 0
9818
+ ).map((token) => ({
9792
9819
  word: token.text,
9793
9820
  start: token.start_ms / 1e3,
9794
9821
  end: token.end_ms / 1e3,
@@ -9805,7 +9832,8 @@ var SonioxAdapter = class extends BaseAdapter {
9805
9832
  id,
9806
9833
  label: `Speaker ${id}`
9807
9834
  })) : void 0;
9808
- const utterances = response.tokens ? this.buildUtterancesFromTokens(response.tokens.filter((t) => t.is_final)) : [];
9835
+ const tokens = response.tokens ? response.tokens.filter((t) => t.is_final !== false) : [];
9836
+ const utterances = tokens.length > 0 ? this.buildUtterancesFromTokens(tokens) : [];
9809
9837
  const language = response.tokens?.find((t) => t.language)?.language;
9810
9838
  return {
9811
9839
  success: true,
@@ -9815,7 +9843,7 @@ var SonioxAdapter = class extends BaseAdapter {
9815
9843
  text,
9816
9844
  status: TranscriptionStatus.completed,
9817
9845
  language,
9818
- duration: response.total_audio_proc_ms ? response.total_audio_proc_ms / 1e3 : void 0,
9846
+ duration: response.audio_duration_ms ? response.audio_duration_ms / 1e3 : response.total_audio_proc_ms ? response.total_audio_proc_ms / 1e3 : void 0,
9819
9847
  speakers,
9820
9848
  words: words.length > 0 ? words : void 0,
9821
9849
  utterances: utterances.length > 0 ? utterances : void 0