voice-router-dev 0.8.3 → 0.8.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +107 -1
- package/dist/constants.d.mts +7 -8
- package/dist/constants.d.ts +7 -8
- package/dist/constants.js +4 -8
- package/dist/constants.mjs +4 -8
- package/dist/{field-configs-t_lVCkE5.d.mts → field-configs-DXAGTUfv.d.mts} +3135 -1956
- package/dist/{field-configs-t_lVCkE5.d.ts → field-configs-DXAGTUfv.d.ts} +3135 -1956
- package/dist/field-configs.d.mts +1 -1
- package/dist/field-configs.d.ts +1 -1
- package/dist/field-configs.js +2320 -1634
- package/dist/field-configs.mjs +2320 -1634
- package/dist/index.d.mts +263 -711
- package/dist/index.d.ts +263 -711
- package/dist/index.js +2080 -1532
- package/dist/index.mjs +2080 -1531
- package/dist/{speechToTextChunkResponseModel-DjL2ncnf.d.ts → speechToTextChunkResponseModel-CmxWFIjf.d.ts} +1667 -1172
- package/dist/{speechToTextChunkResponseModel-DvIT4xai.d.mts → speechToTextChunkResponseModel-Dj3Weqv1.d.mts} +1667 -1172
- package/dist/webhooks.d.mts +9 -3
- package/dist/webhooks.d.ts +9 -3
- package/package.json +6 -6
package/dist/index.d.mts
CHANGED
|
@@ -1,10 +1,10 @@
|
|
|
1
|
-
import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptOptionalParams, aQ as TranscriptStatus, aR as TranscriptWord, aS as AudioIntelligenceModelStatus, aT as AutoHighlightResult, aU as AutoHighlightsResult, aV as Chapter, aW as ContentSafetyLabel, aX as ContentSafetyLabelResult, aY as ContentSafetyLabelsResult, aZ as ContentSafetyLabelsResultSeverityScoreSummary, a_ as ContentSafetyLabelsResultSummary, a$ as Entity, b0 as EntityType, b1 as ListTranscriptsParams, b2 as PiiPolicy, b3 as RedactPiiAudioQuality, b4 as RedactedAudioNotification, b5 as RedactedAudioResponse, b6 as RedactedAudioStatus, b7 as Sentiment, b8 as SentimentAnalysisResult$1, b9 as SentimentAnalysisResultChannel, ba as SentimentAnalysisResultSpeaker, bb as SeverityScoreSummary, bc as SpeechModel, bd as SubstitutionPolicy, be as SummaryModel, bf as SummaryType, bg as Timestamp, bh as TopicDetectionModelResult, bi as TopicDetectionModelResultSummary, bj as TopicDetectionResult, bk as TopicDetectionResultLabelsItem, bl as Transcript, bm as TranscriptAudioDuration, bn as TranscriptAudioEndAt, bo as TranscriptAudioStartFrom, bp as TranscriptAutoChapters, bq as TranscriptAutoHighlightsResult, br as TranscriptBoostParam, bs as TranscriptBoostParamProperty, bt as TranscriptChapters, bu as TranscriptConfidence, bv as TranscriptContentSafety, bw as TranscriptContentSafetyLabels, bx as TranscriptCustomSpelling, by as TranscriptCustomSpellingProperty, bz as TranscriptCustomTopics, bA as TranscriptDisfluencies, bB as TranscriptEntities, bC as TranscriptEntityDetection, bD as TranscriptFilterProfanity, bE as TranscriptFormatText, bF as TranscriptIabCategories, bG as TranscriptIabCategoriesResult, bH as TranscriptLanguageCode, bI as TranscriptLanguageCodeProperty, bJ as TranscriptLanguageConfidence, bK as TranscriptLanguageConfidenceThreshold, bL as TranscriptLanguageDetection, bM as TranscriptMultichannel, bN as TranscriptOptionalParamsLanguageCode, bO as TranscriptOptionalParamsLanguageCodeOneOf, bP as TranscriptOptionalParamsRedactPiiSub, bQ as TranscriptOptionalParamsSpeakersExpected, bR as TranscriptOptionalParamsSpeechModel, bS as TranscriptOptionalParamsSpeechThreshold, bT as TranscriptOptionalParamsWebhookAuthHeaderName, bU as TranscriptOptionalParamsWebhookAuthHeaderValue, bV as TranscriptPunctuate, bW as TranscriptReadyNotification, bX as TranscriptReadyStatus, bY as TranscriptRedactPiiAudio, bZ as TranscriptRedactPiiAudioQuality, b_ as TranscriptRedactPiiPolicies, b$ as TranscriptSentimentAnalysis, c0 as TranscriptSentimentAnalysisResults, c1 as TranscriptSpeakerLabels, c2 as TranscriptSpeakersExpected, c3 as TranscriptSpeechModel, c4 as TranscriptSpeechThreshold, c5 as TranscriptSpeedBoost, c6 as TranscriptSummary, c7 as TranscriptSummaryModel, c8 as TranscriptSummaryType, c9 as TranscriptText, ca as TranscriptThrottled, cb as TranscriptUtterance, cc as TranscriptUtteranceChannel, cd as TranscriptUtterances, ce as TranscriptWebhookAuthHeaderName, cf as TranscriptWebhookNotification, cg as TranscriptWebhookStatusCode, ch as TranscriptWebhookUrl, ci as TranscriptWordChannel, cj as TranscriptWordSpeaker, ck as TranscriptWords, cl as StreamingUpdateConfiguration, cm as Transcription$1, cn as EntityError, co as Status, cp as EntityReference, cq as DiarizationProperties, cr as DiarizationSpeakersProperties, cs as LanguageIdentificationMode, ct as LanguageIdentificationProperties, cu as LanguageIdentificationPropertiesSpeechModelMapping, cv as ProfanityFilterMode, cw as PunctuationMode, cx as TranscriptionCustomProperties, cy as TranscriptionLinks, cz as TranscriptionProperties, cA as TranscriptTextUsageTokens, cB as TranscriptionSegment, cC as RealtimeSessionCreateRequestGAModel, cD as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, cE as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, cF as AudioResponseFormat, cG as CreateTranscription200One, cH as CreateTranscriptionRequest, cI as CreateTranscriptionRequestModel, cJ as CreateTranscriptionRequestStream, cK as CreateTranscriptionRequestTimestampGranularitiesItem, cL as CreateTranscriptionResponseDiarizedJson, cM as CreateTranscriptionResponseDiarizedJsonTask, cN as CreateTranscriptionResponseDiarizedJsonUsage, cO as CreateTranscriptionResponseJson, cP as CreateTranscriptionResponseJsonLogprobsItem, cQ as CreateTranscriptionResponseJsonUsage, cR as CreateTranscriptionResponseVerboseJson, cS as TranscriptTextUsageDuration, cT as TranscriptTextUsageDurationType, cU as TranscriptTextUsageTokensInputTokenDetails, cV as TranscriptTextUsageTokensType, cW as TranscriptionChunkingStrategy, cX as TranscriptionChunkingStrategyAnyOf, cY as TranscriptionDiarizedSegment, cZ as TranscriptionDiarizedSegmentType, c_ as TranscriptionInclude, c$ as TranscriptionWord, d0 as VadConfig, d1 as VadConfigType, d2 as ListenV1Response, d3 as ManageV1FilterAccessorParameter, d4 as ManageV1FilterDeploymentParameter, d5 as ManageV1LimitParameter, d6 as ManageV1PageParameter, d7 as ManageV1FilterEndpointParameter, d8 as ManageV1FilterMethodParameter, d9 as SharedTopics, da as SharedIntents, db as SharedSentiments, dc as SharedCallbackParameter, dd as SharedCallbackMethodParameter, de as SharedSentimentParameter, df as SharedSummarizeParameter, dg as SharedTagParameter, dh as SharedTopicsParameter, di as SharedCustomTopicParameter, dj as SharedCustomTopicModeParameter, dk as SharedIntentsParameter, dl as SharedCustomIntentParameter, dm as SharedCustomIntentModeParameter, dn as SharedMipOptOutParameter, dp as ListenV1DetectEntitiesParameter, dq as ListenV1DetectLanguageParameter, dr as ListenV1DiarizeParameter, ds as ListenV1DictationParameter, dt as ListenV1EncodingParameter, du as ListenV1FillerWordsParameter, dv as ListenV1KeytermParameter, dw as ListenV1KeywordsParameter, dx as ListenV1LanguageParameter, dy as ListenV1MeasurementsParameter, dz as ListenV1MediaTranscribeParams, dA as ListenV1ModelParameter, dB as ListenV1MultichannelParameter, dC as ListenV1NumeralsParameter, dD as ListenV1ParagraphsParameter, dE as ListenV1ProfanityFilterParameter, dF as ListenV1PunctuateParameter, dG as ListenV1RedactParameter, dH as ListenV1RedactParameterOneOfItem, dI as ListenV1ReplaceParameter, dJ as ListenV1ResponseMetadata, dK as ListenV1ResponseMetadataIntentsInfo, dL as ListenV1ResponseMetadataModelInfo, dM as ListenV1ResponseMetadataSentimentInfo, dN as ListenV1ResponseMetadataSummaryInfo, dO as ListenV1ResponseMetadataTopicsInfo, dP as ListenV1ResponseResults, dQ as ListenV1ResponseResultsChannels, dR as ListenV1ResponseResultsChannelsItem, dS as ListenV1ResponseResultsChannelsItemAlternativesItem, dT as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, dU as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, dV as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, dW as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, dX as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, dY as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, dZ as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, d_ as ListenV1ResponseResultsChannelsItemSearchItem, d$ as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, e0 as ListenV1ResponseResultsSummary, e1 as ListenV1ResponseResultsUtterances, e2 as ListenV1ResponseResultsUtterancesItem, e3 as ListenV1ResponseResultsUtterancesItemWordsItem, e4 as ListenV1SearchParameter, e5 as ListenV1SmartFormatParameter, e6 as ListenV1UttSplitParameter, e7 as ListenV1UtterancesParameter, e8 as ListenV1VersionParameter, e9 as ManageV1EndDateTimeParameter, ea as ManageV1FilterRequestIdParameter, eb as ManageV1FilterStatusParameter, ec as ManageV1ProjectsRequestsListParams, ed as ManageV1StartDateTimeParameter, ee as SharedExtraParameter, ef as SharedIntentsResults, eg as SharedIntentsResultsIntents, eh as SharedIntentsResultsIntentsSegmentsItem, ei as SharedIntentsResultsIntentsSegmentsItemIntentsItem, ej as SharedSentimentsAverage, ek as SharedSentimentsSegmentsItem, el as SharedTopicsResults, em as SharedTopicsResultsTopics, en as SharedTopicsResultsTopicsSegmentsItem, eo as SharedTopicsResultsTopicsSegmentsItemTopicsItem, ep as JobType, eq as AlignmentConfig, er as TranscriptionConfig, es as TrackingData, et as OutputConfig, eu as OperatingPoint, ev as AudioEventItem, ew as AudioEventSummary, ex as AudioEventSummaryItem, ey as AutoChaptersResult, ez as AutoChaptersResultError, eA as AutoChaptersResultErrorType, eB as Chapter$1, eC as JobInfo, eD as LanguageIdentificationResult, eE as LanguageIdentificationResultAlternative, eF as LanguageIdentificationResultError, eG as LanguageIdentificationResultItem, eH as LanguagePackInfo, eI as LanguagePackInfoWritingDirection, eJ as OutputConfigSrtOverrides, eK as RecognitionAlternative, eL as RecognitionDisplay, eM as RecognitionDisplayDirection, eN as RecognitionMetadata, eO as RecognitionResult, eP as RecognitionResultAttachesTo, eQ as RecognitionResultType, eR as RetrieveTranscriptResponse, eS as RetrieveTranscriptResponseAudioEventSummary, eT as RetrieveTranscriptResponseAudioEventSummaryChannels, eU as RetrieveTranscriptResponseTranslations, eV as SentimentAnalysisError, eW as SentimentAnalysisErrorType, eX as SentimentAnalysisResult$2, eY as SentimentAnalysisResultSentimentAnalysis, eZ as SentimentChannelSummary, e_ as SentimentSegment, e$ as SentimentSpeakerSummary, f0 as SentimentSummary, f1 as SentimentSummaryDetail, f2 as SpokenFormRecognitionResult, f3 as SpokenFormRecognitionResultType, f4 as SummarizationError, f5 as SummarizationErrorType, f6 as SummarizationResult, f7 as TopicDetectionError, f8 as TopicDetectionErrorType, f9 as TopicDetectionResult$1, fa as TopicDetectionSegment, fb as TopicDetectionSegmentTopic, fc as TopicDetectionSummary, fd as TopicDetectionSummaryOverall, fe as TrackingDataDetails, ff as TranscriptionConfigAdditionalVocabItem, fg as TranscriptionConfigDiarization, fh as TranscriptionConfigMaxDelayMode, fi as TranscriptionConfigPunctuationOverrides, fj as TranscriptionConfigSpeakerDiarizationConfig, fk as TranscriptionConfigTranscriptFilteringConfig, fl as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fm as TranslationError, fn as TranslationErrorType, fo as TranslationSentence, fp as WrittenFormRecognitionResult, fq as WrittenFormRecognitionResultType, fr as SpeechToTextChunkResponseModel, fs as AdditionalFormatResponseModel, ft as AdditionalFormats, fu as BodySpeechToTextV1SpeechToTextPost, fv as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fw as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, fx as BodySpeechToTextV1SpeechToTextPostEntityDetection, fy as BodySpeechToTextV1SpeechToTextPostFile, fz as BodySpeechToTextV1SpeechToTextPostFileFormat, fA as BodySpeechToTextV1SpeechToTextPostLanguageCode, fB as BodySpeechToTextV1SpeechToTextPostModelId, fC as BodySpeechToTextV1SpeechToTextPostNumSpeakers, fD as BodySpeechToTextV1SpeechToTextPostSeed, fE as BodySpeechToTextV1SpeechToTextPostTemperature, fF as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, fG as BodySpeechToTextV1SpeechToTextPostWebhookId, fH as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, fI as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, fJ as DetectedEntity, fK as DocxExportOptions, fL as DocxExportOptionsFormat, fM as DocxExportOptionsMaxSegmentChars, fN as DocxExportOptionsMaxSegmentDurationS, fO as DocxExportOptionsSegmentOnSilenceLongerThanS, fP as ExportOptions, fQ as HtmlExportOptions, fR as HtmlExportOptionsFormat, fS as HtmlExportOptionsMaxSegmentChars, fT as HtmlExportOptionsMaxSegmentDurationS, fU as HtmlExportOptionsSegmentOnSilenceLongerThanS, fV as PdfExportOptions, fW as PdfExportOptionsFormat, fX as PdfExportOptionsMaxSegmentChars, fY as PdfExportOptionsMaxSegmentDurationS, fZ as PdfExportOptionsSegmentOnSilenceLongerThanS, f_ as SegmentedJsonExportOptions, f$ as SegmentedJsonExportOptionsFormat, g0 as SegmentedJsonExportOptionsMaxSegmentChars, g1 as SegmentedJsonExportOptionsMaxSegmentDurationS, g2 as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, g3 as SpeechToTextCharacterResponseModel, g4 as SpeechToTextCharacterResponseModelEnd, g5 as SpeechToTextCharacterResponseModelStart, g6 as SpeechToTextChunkResponseModelAdditionalFormats, g7 as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, g8 as SpeechToTextChunkResponseModelChannelIndex, g9 as SpeechToTextChunkResponseModelEntities, ga as SpeechToTextChunkResponseModelTranscriptionId, gb as SpeechToTextWordResponseModel, gc as SpeechToTextWordResponseModelCharacters, gd as SpeechToTextWordResponseModelEnd, ge as SpeechToTextWordResponseModelSpeakerId, gf as SpeechToTextWordResponseModelStart, gg as SpeechToTextWordResponseModelType, gh as SrtExportOptions, gi as SrtExportOptionsFormat, gj as SrtExportOptionsMaxCharactersPerLine, gk as SrtExportOptionsMaxSegmentChars, gl as SrtExportOptionsMaxSegmentDurationS, gm as SrtExportOptionsSegmentOnSilenceLongerThanS, gn as TxtExportOptions, go as TxtExportOptionsFormat, gp as TxtExportOptionsMaxCharactersPerLine, gq as TxtExportOptionsMaxSegmentChars, gr as TxtExportOptionsMaxSegmentDurationS, gs as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-DvIT4xai.mjs';
|
|
2
|
-
export {
|
|
1
|
+
import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptStatus, aQ as TranscriptWord, aR as AudioIntelligenceModelStatus, aS as AutoHighlightResult, aT as AutoHighlightsResult, aU as Chapter, aV as ContentSafetyLabel, aW as ContentSafetyLabelResult, aX as ContentSafetyLabelsResult, aY as ContentSafetyLabelsResultSeverityScoreSummary, aZ as ContentSafetyLabelsResultSummary, a_ as CustomFormattingRequestBody, a$ as CustomFormattingRequestBodyCustomFormatting, b0 as CustomFormattingResponse, b1 as CustomFormattingResponseCustomFormatting, b2 as CustomFormattingResponseCustomFormattingMapping, b3 as Entity, b4 as EntityType, b5 as ListTranscriptsParams, b6 as PiiPolicy, b7 as RedactPiiAudioQuality, b8 as Sentiment, b9 as SentimentAnalysisResult$1, ba as SentimentAnalysisResultChannel, bb as SentimentAnalysisResultSpeaker, bc as SeverityScoreSummary, bd as SpeakerIdentificationRequestBody, be as SpeakerIdentificationRequestBodySpeakerIdentification, bf as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, bg as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, bh as SpeakerIdentificationResponse, bi as SpeakerIdentificationResponseSpeakerIdentification, bj as SpeakerIdentificationResponseSpeakerIdentificationMapping, bk as SpeechModel, bl as SubstitutionPolicy, bm as SummaryModel, bn as SummaryType, bo as Timestamp, bp as TopicDetectionModelResult, bq as TopicDetectionModelResultSummary, br as TopicDetectionResult, bs as TopicDetectionResultLabelsItems, bt as Transcript, bu as TranscriptAudioDuration, bv as TranscriptAudioEndAt, bw as TranscriptAudioStartFrom, bx as TranscriptAutoChapters, by as TranscriptAutoHighlightsResult, bz as TranscriptChapters, bA as TranscriptConfidence, bB as TranscriptContentSafety, bC as TranscriptContentSafetyLabels, bD as TranscriptCustomSpelling, bE as TranscriptCustomSpellingProperty, bF as TranscriptCustomTopics, bG as TranscriptDisfluencies, bH as TranscriptDomain, bI as TranscriptEntities, bJ as TranscriptEntityDetection, bK as TranscriptFilterProfanity, bL as TranscriptFormatText, bM as TranscriptIabCategories, bN as TranscriptIabCategoriesResult, bO as TranscriptLanguageCode, bP as TranscriptLanguageCodes, bQ as TranscriptLanguageConfidence, bR as TranscriptLanguageConfidenceThreshold, bS as TranscriptLanguageDetection, bT as TranscriptLanguageDetectionOptions, bU as TranscriptMultichannel, bV as TranscriptOptionalParamsLanguageDetectionOptions, bW as TranscriptOptionalParamsRedactPiiAudioOptions, bX as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, bY as TranscriptOptionalParamsRemoveAudioTags, bZ as TranscriptOptionalParamsSpeakerOptions, b_ as TranscriptOptionalParamsSpeechUnderstanding, b$ as TranscriptOptionalParamsSpeechUnderstandingRequest, c0 as TranscriptParams, c1 as TranscriptParamsDomain, c2 as TranscriptParamsLanguageCode, c3 as TranscriptParamsLanguageCodes, c4 as TranscriptParamsRedactPiiSub, c5 as TranscriptParamsRemoveAudioTags, c6 as TranscriptParamsSpeakersExpected, c7 as TranscriptParamsSpeechThreshold, c8 as TranscriptParamsWebhookAuthHeaderName, c9 as TranscriptParamsWebhookAuthHeaderValue, ca as TranscriptPunctuate, cb as TranscriptRedactPiiAudio, cc as TranscriptRedactPiiAudioOptions, cd as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, ce as TranscriptRedactPiiAudioQuality, cf as TranscriptRedactPiiPolicies, cg as TranscriptRemoveAudioTags, ch as TranscriptRemoveAudioTagsProperty, ci as TranscriptSentimentAnalysis, cj as TranscriptSentimentAnalysisResults, ck as TranscriptSpeakerLabels, cl as TranscriptSpeakersExpected, cm as TranscriptSpeechModel, cn as TranscriptSpeechModels, co as TranscriptSpeechThreshold, cp as TranscriptSpeechUnderstanding, cq as TranscriptSpeechUnderstandingRequest, cr as TranscriptSpeechUnderstandingResponse, cs as TranscriptSpeedBoost, ct as TranscriptSummary, cu as TranscriptSummaryModel, cv as TranscriptSummaryType, cw as TranscriptTemperature, cx as TranscriptText, cy as TranscriptThrottled, cz as TranscriptTranslatedTexts, cA as TranscriptUtterance, cB as TranscriptUtteranceChannel, cC as TranscriptUtteranceTranslatedTexts, cD as TranscriptUtterances, cE as TranscriptWebhookAuthHeaderName, cF as TranscriptWebhookStatusCode, cG as TranscriptWebhookUrl, cH as TranscriptWordChannel, cI as TranscriptWordSpeaker, cJ as TranscriptWords, cK as TranslationRequestBody, cL as TranslationRequestBodyTranslation, cM as TranslationResponse, cN as TranslationResponseTranslation, cO as StreamingUpdateConfiguration, cP as Transcription$1, cQ as EntityError, cR as Status, cS as EntityReference, cT as DiarizationProperties, cU as DiarizationSpeakersProperties, cV as LanguageIdentificationMode, cW as LanguageIdentificationProperties, cX as LanguageIdentificationPropertiesSpeechModelMapping, cY as ProfanityFilterMode, cZ as PunctuationMode, c_ as TranscriptionCustomProperties, c$ as TranscriptionLinks, d0 as TranscriptionProperties, d1 as TranscriptTextUsageTokens, d2 as TranscriptionSegment, d3 as RealtimeSessionCreateRequestGAModel, d4 as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, d5 as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, d6 as AudioResponseFormat, d7 as CreateTranscription200One, d8 as CreateTranscriptionRequest, d9 as CreateTranscriptionRequestModel, da as CreateTranscriptionRequestStream, db as CreateTranscriptionRequestTimestampGranularitiesItem, dc as CreateTranscriptionResponseDiarizedJson, dd as CreateTranscriptionResponseDiarizedJsonTask, de as CreateTranscriptionResponseDiarizedJsonUsage, df as CreateTranscriptionResponseJson, dg as CreateTranscriptionResponseJsonLogprobsItem, dh as CreateTranscriptionResponseJsonUsage, di as CreateTranscriptionResponseVerboseJson, dj as TranscriptTextUsageDuration, dk as TranscriptTextUsageDurationType, dl as TranscriptTextUsageTokensInputTokenDetails, dm as TranscriptTextUsageTokensType, dn as TranscriptionChunkingStrategy, dp as TranscriptionChunkingStrategyAnyOf, dq as TranscriptionDiarizedSegment, dr as TranscriptionDiarizedSegmentType, ds as TranscriptionInclude, dt as TranscriptionWord, du as VadConfig, dv as VadConfigType, dw as ListenV1Response, dx as ManageV1FilterAccessorParameter, dy as ManageV1FilterDeploymentParameter, dz as ManageV1LimitParameter, dA as ManageV1PageParameter, dB as ManageV1FilterEndpointParameter, dC as ManageV1FilterMethodParameter, dD as SharedTopics, dE as SharedIntents, dF as SharedSentiments, dG as SharedCallbackParameter, dH as SharedCallbackMethodParameter, dI as SharedSentimentParameter, dJ as SharedSummarizeParameter, dK as SharedTagParameter, dL as SharedTopicsParameter, dM as SharedCustomTopicParameter, dN as SharedCustomTopicModeParameter, dO as SharedIntentsParameter, dP as SharedCustomIntentParameter, dQ as SharedCustomIntentModeParameter, dR as SharedMipOptOutParameter, dS as ListenV1DetectEntitiesParameter, dT as ListenV1DetectLanguageParameter, dU as ListenV1DiarizeParameter, dV as ListenV1DictationParameter, dW as ListenV1EncodingParameter, dX as ListenV1FillerWordsParameter, dY as ListenV1KeytermParameter, dZ as ListenV1KeywordsParameter, d_ as ListenV1LanguageParameter, d$ as ListenV1MeasurementsParameter, e0 as ListenV1MediaTranscribeParams, e1 as ListenV1ModelParameter, e2 as ListenV1MultichannelParameter, e3 as ListenV1NumeralsParameter, e4 as ListenV1ParagraphsParameter, e5 as ListenV1ProfanityFilterParameter, e6 as ListenV1PunctuateParameter, e7 as ListenV1RedactParameter, e8 as ListenV1RedactParameterOneOfItem, e9 as ListenV1ReplaceParameter, ea as ListenV1ResponseMetadata, eb as ListenV1ResponseMetadataIntentsInfo, ec as ListenV1ResponseMetadataModelInfo, ed as ListenV1ResponseMetadataSentimentInfo, ee as ListenV1ResponseMetadataSummaryInfo, ef as ListenV1ResponseMetadataTopicsInfo, eg as ListenV1ResponseResults, eh as ListenV1ResponseResultsChannels, ei as ListenV1ResponseResultsChannelsItem, ej as ListenV1ResponseResultsChannelsItemAlternativesItem, ek as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, el as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, em as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, en as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, eo as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, ep as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, eq as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, er as ListenV1ResponseResultsChannelsItemSearchItem, es as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, et as ListenV1ResponseResultsSummary, eu as ListenV1ResponseResultsUtterances, ev as ListenV1ResponseResultsUtterancesItem, ew as ListenV1ResponseResultsUtterancesItemWordsItem, ex as ListenV1SearchParameter, ey as ListenV1SmartFormatParameter, ez as ListenV1UttSplitParameter, eA as ListenV1UtterancesParameter, eB as ListenV1VersionParameter, eC as ManageV1EndDateTimeParameter, eD as ManageV1FilterRequestIdParameter, eE as ManageV1FilterStatusParameter, eF as ManageV1ProjectsRequestsListParams, eG as ManageV1StartDateTimeParameter, eH as SharedExtraParameter, eI as SharedIntentsResults, eJ as SharedIntentsResultsIntents, eK as SharedIntentsResultsIntentsSegmentsItem, eL as SharedIntentsResultsIntentsSegmentsItemIntentsItem, eM as SharedSentimentsAverage, eN as SharedSentimentsSegmentsItem, eO as SharedTopicsResults, eP as SharedTopicsResultsTopics, eQ as SharedTopicsResultsTopicsSegmentsItem, eR as SharedTopicsResultsTopicsSegmentsItemTopicsItem, eS as JobType, eT as AlignmentConfig, eU as TranscriptionConfig, eV as TrackingData, eW as OutputConfig, eX as OperatingPoint, eY as AudioEventItem, eZ as AudioEventSummary, e_ as AudioEventSummaryItem, e$ as AutoChaptersResult, f0 as AutoChaptersResultError, f1 as AutoChaptersResultErrorType, f2 as Chapter$1, f3 as JobInfo, f4 as LanguageIdentificationResult, f5 as LanguageIdentificationResultAlternative, f6 as LanguageIdentificationResultError, f7 as LanguageIdentificationResultItem, f8 as LanguagePackInfo, f9 as LanguagePackInfoWritingDirection, fa as OutputConfigSrtOverrides, fb as RecognitionAlternative, fc as RecognitionDisplay, fd as RecognitionDisplayDirection, fe as RecognitionMetadata, ff as RecognitionResult, fg as RecognitionResultAttachesTo, fh as RecognitionResultType, fi as RetrieveTranscriptResponse, fj as RetrieveTranscriptResponseAudioEventSummary, fk as RetrieveTranscriptResponseAudioEventSummaryChannels, fl as RetrieveTranscriptResponseTranslations, fm as SentimentAnalysisError, fn as SentimentAnalysisErrorType, fo as SentimentAnalysisResult$2, fp as SentimentAnalysisResultSentimentAnalysis, fq as SentimentChannelSummary, fr as SentimentSegment, fs as SentimentSpeakerSummary, ft as SentimentSummary, fu as SentimentSummaryDetail, fv as SpokenFormRecognitionResult, fw as SpokenFormRecognitionResultType, fx as SummarizationError, fy as SummarizationErrorType, fz as SummarizationResult, fA as TopicDetectionError, fB as TopicDetectionErrorType, fC as TopicDetectionResult$1, fD as TopicDetectionSegment, fE as TopicDetectionSegmentTopic, fF as TopicDetectionSummary, fG as TopicDetectionSummaryOverall, fH as TrackingDataDetails, fI as TranscriptionConfigAdditionalVocabItem, fJ as TranscriptionConfigDiarization, fK as TranscriptionConfigMaxDelayMode, fL as TranscriptionConfigPunctuationOverrides, fM as TranscriptionConfigSpeakerDiarizationConfig, fN as TranscriptionConfigTranscriptFilteringConfig, fO as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fP as TranslationError, fQ as TranslationErrorType, fR as TranslationSentence, fS as WrittenFormRecognitionResult, fT as WrittenFormRecognitionResultType, fU as SpeechToTextChunkResponseModel, fV as AdditionalFormatResponseModel, fW as AdditionalFormats, fX as BodySpeechToTextV1SpeechToTextPost, fY as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fZ as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, f_ as BodySpeechToTextV1SpeechToTextPostEntityDetection, f$ as BodySpeechToTextV1SpeechToTextPostFile, g0 as BodySpeechToTextV1SpeechToTextPostFileFormat, g1 as BodySpeechToTextV1SpeechToTextPostLanguageCode, g2 as BodySpeechToTextV1SpeechToTextPostModelId, g3 as BodySpeechToTextV1SpeechToTextPostNumSpeakers, g4 as BodySpeechToTextV1SpeechToTextPostSeed, g5 as BodySpeechToTextV1SpeechToTextPostTemperature, g6 as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, g7 as BodySpeechToTextV1SpeechToTextPostWebhookId, g8 as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, g9 as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, ga as DetectedEntity, gb as DocxExportOptions, gc as DocxExportOptionsFormat, gd as DocxExportOptionsMaxSegmentChars, ge as DocxExportOptionsMaxSegmentDurationS, gf as DocxExportOptionsSegmentOnSilenceLongerThanS, gg as ExportOptions, gh as HtmlExportOptions, gi as HtmlExportOptionsFormat, gj as HtmlExportOptionsMaxSegmentChars, gk as HtmlExportOptionsMaxSegmentDurationS, gl as HtmlExportOptionsSegmentOnSilenceLongerThanS, gm as PdfExportOptions, gn as PdfExportOptionsFormat, go as PdfExportOptionsMaxSegmentChars, gp as PdfExportOptionsMaxSegmentDurationS, gq as PdfExportOptionsSegmentOnSilenceLongerThanS, gr as SegmentedJsonExportOptions, gs as SegmentedJsonExportOptionsFormat, gt as SegmentedJsonExportOptionsMaxSegmentChars, gu as SegmentedJsonExportOptionsMaxSegmentDurationS, gv as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, gw as SpeechToTextCharacterResponseModel, gx as SpeechToTextCharacterResponseModelEnd, gy as SpeechToTextCharacterResponseModelStart, gz as SpeechToTextChunkResponseModelAdditionalFormats, gA as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, gB as SpeechToTextChunkResponseModelChannelIndex, gC as SpeechToTextChunkResponseModelEntities, gD as SpeechToTextChunkResponseModelTranscriptionId, gE as SpeechToTextWordResponseModel, gF as SpeechToTextWordResponseModelCharacters, gG as SpeechToTextWordResponseModelEnd, gH as SpeechToTextWordResponseModelSpeakerId, gI as SpeechToTextWordResponseModelStart, gJ as SpeechToTextWordResponseModelType, gK as SrtExportOptions, gL as SrtExportOptionsFormat, gM as SrtExportOptionsMaxCharactersPerLine, gN as SrtExportOptionsMaxSegmentChars, gO as SrtExportOptionsMaxSegmentDurationS, gP as SrtExportOptionsSegmentOnSilenceLongerThanS, gQ as TxtExportOptions, gR as TxtExportOptionsFormat, gS as TxtExportOptionsMaxCharactersPerLine, gT as TxtExportOptionsMaxSegmentChars, gU as TxtExportOptionsMaxSegmentDurationS, gV as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-Dj3Weqv1.mjs';
|
|
2
|
+
export { gZ as AssemblyAIExtendedData, ho as AssemblyAIUpdateConfiguration, hk as AudioAckEvent, hm as AudioChunk, h3 as BatchOnlyProvider, hv as BeginEvent, hj as ChapterizationEvent, g$ as DeepgramExtendedData, h0 as ElevenLabsExtendedData, hr as ElevenLabsStreamingOptions, hh as EntityEvent, hy as ErrorEvent, g_ as GladiaExtendedData, hl as LifecycleEvent, hb as ListTranscriptsResponse, hp as OpenAIStreamingOptions, h1 as ProviderExtendedDataMap, hc as ProviderRawResponseMap, hs as ProviderStreamingOptions, hn as RawWebSocketMessage, hg as SentimentEvent, h4 as SessionStatus, hq as SonioxStreamingOptions, h5 as Speaker, he as SpeechEvent, gW as SpeechmaticsOperatingPoint, hd as StreamEventType, hz as StreamingEventMessage, hB as StreamingForceEndpoint, ht as StreamingOptionsForProvider, h2 as StreamingProvider, hA as StreamingWord, hi as SummarizationEvent, hx as TerminationEvent, hu as TranscribeStreamParams, ha as TranscriptData, h9 as TranscriptMetadata, gY as TranscriptionLanguage, gX as TranscriptionModel, h8 as TranscriptionStatus, hf as TranslationEvent, hw as TurnEvent, h7 as Utterance, h6 as Word } from './speechToTextChunkResponseModel-Dj3Weqv1.mjs';
|
|
3
3
|
import { T as TranscriptionProvider, P as ProviderCapabilities, A as AudioInput } from './provider-metadata-MDUUEuqF.mjs';
|
|
4
4
|
export { k as AllLanguageCodes, o as AllProviders, a as AssemblyAICapabilities, i as AssemblyAILanguageCodes, b as AzureCapabilities, B as BatchOnlyProviderType, q as BatchOnlyProviders, C as CapabilityKeys, f as CapabilityLabels, D as DeepgramCapabilities, j as DeepgramLanguageCodes, E as ElevenLabsCapabilities, G as GladiaCapabilities, h as GladiaLanguageCodes, g as LanguageCode, L as LanguageLabels, O as OpenAICapabilities, d as ProviderCapabilitiesMap, l as ProviderDisplayNames, n as ProviderDocs, m as ProviderWebsites, c as SonioxCapabilities, S as SpeechmaticsCapabilities, e as StreamingProviderType, p as StreamingProviders } from './provider-metadata-MDUUEuqF.mjs';
|
|
5
5
|
import { AssemblyAIRegionType, DeepgramRegionType, SpeechmaticsRegionType, SonioxModelCode, SonioxRegionType, ElevenLabsModelCode, ElevenLabsRegionType } from './constants.mjs';
|
|
6
6
|
export { AssemblyAIEncoding, AssemblyAIEncodingType, AssemblyAILanguage, AssemblyAILanguageType, AssemblyAISampleRate, AssemblyAISampleRateType, AssemblyAISpeechModel, AssemblyAISpeechModelType, AssemblyAIStatus, AssemblyAITranscriptionModel, AssemblyAITranscriptionModelType, AzureLocale, AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, AzureLocaleType, AzureLocales, DeepgramCallbackMethod, DeepgramEncoding, DeepgramIntentMode, DeepgramLanguage, DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, DeepgramModelCode as DeepgramModelType, DeepgramRedact, DeepgramRedactType, DeepgramRegion, DeepgramSampleRate, DeepgramStatus, DeepgramTTSContainer, DeepgramTTSEncoding, DeepgramTTSModel, DeepgramTTSSampleRate, DeepgramTopicMode, DeepgramTopicModeType, ElevenLabsLanguageCode, ElevenLabsLanguageCodes, ElevenLabsLanguageLabels, ElevenLabsLanguages, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaRegion, GladiaSampleRate, GladiaStatus, GladiaTranslationLanguage, OpenAILanguageCodes, OpenAIModel, OpenAIRealtimeAudioFormat, OpenAIRealtimeModel, OpenAIRealtimeTranscriptionModel, OpenAIRealtimeTurnDetection, OpenAIResponseFormat, SonioxAsyncModel, SonioxAsyncModelCode, SonioxLanguage, SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, SonioxLanguageType, SonioxLanguages, SonioxModel, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, SonioxRealtimeModelCode, SonioxRegion, SpeechmaticsLanguage, SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsRegion } from './constants.mjs';
|
|
7
|
-
export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-
|
|
7
|
+
export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-DXAGTUfv.mjs';
|
|
8
8
|
import { z } from 'zod';
|
|
9
9
|
import { AxiosResponse, AxiosRequestConfig } from 'axios';
|
|
10
10
|
|
|
@@ -4115,46 +4115,8 @@ declare function createGladiaAdapter(config: ProviderConfig): GladiaAdapter;
|
|
|
4115
4115
|
* Generated by orval v7.9.0 🍺
|
|
4116
4116
|
* Do not edit manually.
|
|
4117
4117
|
* AssemblyAI API
|
|
4118
|
-
* AssemblyAI API
|
|
4119
|
-
* OpenAPI spec version: 1.
|
|
4120
|
-
*/
|
|
4121
|
-
type TranscriptParamsAllOf = {
|
|
4122
|
-
/** The URL of the audio or video file to transcribe. */
|
|
4123
|
-
audio_url: string;
|
|
4124
|
-
};
|
|
4125
|
-
|
|
4126
|
-
/**
|
|
4127
|
-
* Generated by orval v7.9.0 🍺
|
|
4128
|
-
* Do not edit manually.
|
|
4129
|
-
* AssemblyAI API
|
|
4130
|
-
* AssemblyAI API
|
|
4131
|
-
* OpenAPI spec version: 1.3.4
|
|
4132
|
-
*/
|
|
4133
|
-
|
|
4134
|
-
/**
|
|
4135
|
-
* The parameters for creating a transcript
|
|
4136
|
-
*/
|
|
4137
|
-
type TranscriptParams = TranscriptParamsAllOf & TranscriptOptionalParams;
|
|
4138
|
-
|
|
4139
|
-
/**
|
|
4140
|
-
* Generated by orval v7.9.0 🍺
|
|
4141
|
-
* Do not edit manually.
|
|
4142
|
-
* AssemblyAI API
|
|
4143
|
-
* AssemblyAI API
|
|
4144
|
-
* OpenAPI spec version: 1.3.4
|
|
4145
|
-
*/
|
|
4146
|
-
/**
|
|
4147
|
-
* The date and time the transcript was completed
|
|
4148
|
-
* @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
|
|
4149
|
-
*/
|
|
4150
|
-
type TranscriptListItemCompleted = string | null;
|
|
4151
|
-
|
|
4152
|
-
/**
|
|
4153
|
-
* Generated by orval v7.9.0 🍺
|
|
4154
|
-
* Do not edit manually.
|
|
4155
|
-
* AssemblyAI API
|
|
4156
|
-
* AssemblyAI API
|
|
4157
|
-
* OpenAPI spec version: 1.3.4
|
|
4118
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4119
|
+
* OpenAPI spec version: 1.0.0
|
|
4158
4120
|
*/
|
|
4159
4121
|
/**
|
|
4160
4122
|
* Error message of why the transcript failed
|
|
@@ -4165,509 +4127,89 @@ type TranscriptListItemError = string | null;
|
|
|
4165
4127
|
* Generated by orval v7.9.0 🍺
|
|
4166
4128
|
* Do not edit manually.
|
|
4167
4129
|
* AssemblyAI API
|
|
4168
|
-
* AssemblyAI API
|
|
4169
|
-
* OpenAPI spec version: 1.
|
|
4170
|
-
*/
|
|
4171
|
-
|
|
4172
|
-
interface TranscriptListItem {
|
|
4173
|
-
/** The unique identifier for the transcript */
|
|
4174
|
-
id: string;
|
|
4175
|
-
/** The URL to retrieve the transcript */
|
|
4176
|
-
resource_url: string;
|
|
4177
|
-
/** The status of the transcript */
|
|
4178
|
-
status: TranscriptStatus;
|
|
4179
|
-
/**
|
|
4180
|
-
* The date and time the transcript was created
|
|
4181
|
-
* @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
|
|
4182
|
-
*/
|
|
4183
|
-
created: string;
|
|
4184
|
-
/**
|
|
4185
|
-
* The date and time the transcript was completed
|
|
4186
|
-
* @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
|
|
4187
|
-
*/
|
|
4188
|
-
completed: TranscriptListItemCompleted;
|
|
4189
|
-
/** The URL to the audio file */
|
|
4190
|
-
audio_url: string;
|
|
4191
|
-
/** Error message of why the transcript failed */
|
|
4192
|
-
error: TranscriptListItemError;
|
|
4193
|
-
}
|
|
4194
|
-
|
|
4195
|
-
/**
|
|
4196
|
-
* Bad request
|
|
4197
|
-
*/
|
|
4198
|
-
type BadRequestResponse = Error;
|
|
4199
|
-
|
|
4200
|
-
/**
|
|
4201
|
-
* Cannot access uploaded file
|
|
4202
|
-
*/
|
|
4203
|
-
type CannotAccessUploadedFileResponse = Error;
|
|
4204
|
-
|
|
4205
|
-
/**
|
|
4206
|
-
* Generated by orval v7.9.0 🍺
|
|
4207
|
-
* Do not edit manually.
|
|
4208
|
-
* AssemblyAI API
|
|
4209
|
-
* AssemblyAI API
|
|
4210
|
-
* OpenAPI spec version: 1.3.4
|
|
4211
|
-
*/
|
|
4212
|
-
interface CreateRealtimeTemporaryTokenParams {
|
|
4213
|
-
/**
|
|
4214
|
-
* The amount of time until the token expires in seconds
|
|
4215
|
-
* @minimum 60
|
|
4216
|
-
*/
|
|
4217
|
-
expires_in: number;
|
|
4218
|
-
}
|
|
4219
|
-
|
|
4220
|
-
/**
|
|
4221
|
-
* Generated by orval v7.9.0 🍺
|
|
4222
|
-
* Do not edit manually.
|
|
4223
|
-
* AssemblyAI API
|
|
4224
|
-
* AssemblyAI API
|
|
4225
|
-
* OpenAPI spec version: 1.3.4
|
|
4226
|
-
*/
|
|
4227
|
-
interface Error$2 {
|
|
4228
|
-
/** Error message */
|
|
4229
|
-
error: string;
|
|
4230
|
-
status?: "error";
|
|
4231
|
-
[key: string]: unknown;
|
|
4232
|
-
}
|
|
4233
|
-
|
|
4234
|
-
/**
|
|
4235
|
-
* Generated by orval v7.9.0 🍺
|
|
4236
|
-
* Do not edit manually.
|
|
4237
|
-
* AssemblyAI API
|
|
4238
|
-
* AssemblyAI API
|
|
4239
|
-
* OpenAPI spec version: 1.3.4
|
|
4240
|
-
*/
|
|
4241
|
-
/**
|
|
4242
|
-
* Gateway timeout
|
|
4243
|
-
*/
|
|
4244
|
-
type GatewayTimeoutResponse = void;
|
|
4245
|
-
|
|
4246
|
-
/**
|
|
4247
|
-
* Generated by orval v7.9.0 🍺
|
|
4248
|
-
* Do not edit manually.
|
|
4249
|
-
* AssemblyAI API
|
|
4250
|
-
* AssemblyAI API
|
|
4251
|
-
* OpenAPI spec version: 1.3.4
|
|
4252
|
-
*/
|
|
4253
|
-
type GetSubtitlesParams = {
|
|
4254
|
-
/**
|
|
4255
|
-
* The maximum number of characters per caption
|
|
4256
|
-
*/
|
|
4257
|
-
chars_per_caption?: number;
|
|
4258
|
-
};
|
|
4259
|
-
|
|
4260
|
-
/**
|
|
4261
|
-
* An error occurred while processing the request
|
|
4262
|
-
*/
|
|
4263
|
-
type InternalServerErrorResponse = Error;
|
|
4264
|
-
|
|
4265
|
-
/**
|
|
4266
|
-
* Generated by orval v7.9.0 🍺
|
|
4267
|
-
* Do not edit manually.
|
|
4268
|
-
* AssemblyAI API
|
|
4269
|
-
* AssemblyAI API
|
|
4270
|
-
* OpenAPI spec version: 1.3.4
|
|
4271
|
-
*/
|
|
4272
|
-
type LemurBaseParamsContextOneOf = {
|
|
4273
|
-
[key: string]: unknown;
|
|
4274
|
-
};
|
|
4275
|
-
|
|
4276
|
-
/**
|
|
4277
|
-
* Generated by orval v7.9.0 🍺
|
|
4278
|
-
* Do not edit manually.
|
|
4279
|
-
* AssemblyAI API
|
|
4280
|
-
* AssemblyAI API
|
|
4281
|
-
* OpenAPI spec version: 1.3.4
|
|
4282
|
-
*/
|
|
4283
|
-
|
|
4284
|
-
/**
|
|
4285
|
-
* Context to provide the model. This can be a string or a free-form JSON value.
|
|
4286
|
-
*/
|
|
4287
|
-
type LemurBaseParamsContext = string | LemurBaseParamsContextOneOf;
|
|
4288
|
-
|
|
4289
|
-
/**
|
|
4290
|
-
* Generated by orval v7.9.0 🍺
|
|
4291
|
-
* Do not edit manually.
|
|
4292
|
-
* AssemblyAI API
|
|
4293
|
-
* AssemblyAI API
|
|
4294
|
-
* OpenAPI spec version: 1.3.4
|
|
4295
|
-
*/
|
|
4296
|
-
/**
|
|
4297
|
-
* The model that is used for the final prompt after compression is performed.
|
|
4298
|
-
|
|
4299
|
-
*/
|
|
4300
|
-
type LemurModel = (typeof LemurModel)[keyof typeof LemurModel];
|
|
4301
|
-
declare const LemurModel: {
|
|
4302
|
-
readonly "anthropic/claude-3-5-sonnet": "anthropic/claude-3-5-sonnet";
|
|
4303
|
-
readonly "anthropic/claude-3-opus": "anthropic/claude-3-opus";
|
|
4304
|
-
readonly "anthropic/claude-3-haiku": "anthropic/claude-3-haiku";
|
|
4305
|
-
};
|
|
4306
|
-
|
|
4307
|
-
/**
|
|
4308
|
-
* Generated by orval v7.9.0 🍺
|
|
4309
|
-
* Do not edit manually.
|
|
4310
|
-
* AssemblyAI API
|
|
4311
|
-
* AssemblyAI API
|
|
4312
|
-
* OpenAPI spec version: 1.3.4
|
|
4313
|
-
*/
|
|
4314
|
-
|
|
4315
|
-
/**
|
|
4316
|
-
* The model that is used for the final prompt after compression is performed.
|
|
4317
|
-
|
|
4318
|
-
*/
|
|
4319
|
-
type LemurBaseParamsFinalModel = LemurModel | string;
|
|
4320
|
-
|
|
4321
|
-
/**
|
|
4322
|
-
* Generated by orval v7.9.0 🍺
|
|
4323
|
-
* Do not edit manually.
|
|
4324
|
-
* AssemblyAI API
|
|
4325
|
-
* AssemblyAI API
|
|
4326
|
-
* OpenAPI spec version: 1.3.4
|
|
4327
|
-
*/
|
|
4328
|
-
|
|
4329
|
-
interface LemurBaseParams {
|
|
4330
|
-
/** A list of completed transcripts with text. Up to a maximum of 100 hours of audio.
|
|
4331
|
-
Use either transcript_ids or input_text as input into LeMUR.
|
|
4332
|
-
*/
|
|
4333
|
-
transcript_ids?: string[];
|
|
4334
|
-
/** Custom formatted transcript data. Maximum size is the context limit of the selected model.
|
|
4335
|
-
Use either transcript_ids or input_text as input into LeMUR.
|
|
4336
|
-
*/
|
|
4337
|
-
input_text?: string;
|
|
4338
|
-
/** Context to provide the model. This can be a string or a free-form JSON value. */
|
|
4339
|
-
context?: LemurBaseParamsContext;
|
|
4340
|
-
/** The model that is used for the final prompt after compression is performed.
|
|
4341
|
-
*/
|
|
4342
|
-
final_model: LemurBaseParamsFinalModel;
|
|
4343
|
-
/** Max output size in tokens. */
|
|
4344
|
-
max_output_size?: number;
|
|
4345
|
-
/**
|
|
4346
|
-
* The temperature to use for the model.
|
|
4347
|
-
Higher values result in answers that are more creative, lower values are more conservative.
|
|
4348
|
-
Can be any value between 0.0 and 1.0 inclusive.
|
|
4349
|
-
|
|
4350
|
-
* @minimum 0
|
|
4351
|
-
* @maximum 1
|
|
4352
|
-
*/
|
|
4353
|
-
temperature?: number;
|
|
4354
|
-
}
|
|
4355
|
-
|
|
4356
|
-
/**
|
|
4357
|
-
* Generated by orval v7.9.0 🍺
|
|
4358
|
-
* Do not edit manually.
|
|
4359
|
-
* AssemblyAI API
|
|
4360
|
-
* AssemblyAI API
|
|
4361
|
-
* OpenAPI spec version: 1.3.4
|
|
4362
|
-
*/
|
|
4363
|
-
type LemurActionItemsParamsAllOf = {
|
|
4364
|
-
/** How you want the action items to be returned. This can be any text.
|
|
4365
|
-
Defaults to "Bullet Points".
|
|
4366
|
-
*/
|
|
4367
|
-
answer_format?: string;
|
|
4368
|
-
};
|
|
4369
|
-
|
|
4370
|
-
/**
|
|
4371
|
-
* Generated by orval v7.9.0 🍺
|
|
4372
|
-
* Do not edit manually.
|
|
4373
|
-
* AssemblyAI API
|
|
4374
|
-
* AssemblyAI API
|
|
4375
|
-
* OpenAPI spec version: 1.3.4
|
|
4376
|
-
*/
|
|
4377
|
-
|
|
4378
|
-
type LemurActionItemsParams = LemurBaseParams & LemurActionItemsParamsAllOf;
|
|
4379
|
-
|
|
4380
|
-
/**
|
|
4381
|
-
* Generated by orval v7.9.0 🍺
|
|
4382
|
-
* Do not edit manually.
|
|
4383
|
-
* AssemblyAI API
|
|
4384
|
-
* AssemblyAI API
|
|
4385
|
-
* OpenAPI spec version: 1.3.4
|
|
4386
|
-
*/
|
|
4387
|
-
type LemurStringResponseAllOf = {
|
|
4388
|
-
/** The response generated by LeMUR. */
|
|
4389
|
-
response: string;
|
|
4390
|
-
};
|
|
4391
|
-
|
|
4392
|
-
/**
|
|
4393
|
-
* Generated by orval v7.9.0 🍺
|
|
4394
|
-
* Do not edit manually.
|
|
4395
|
-
* AssemblyAI API
|
|
4396
|
-
* AssemblyAI API
|
|
4397
|
-
* OpenAPI spec version: 1.3.4
|
|
4398
|
-
*/
|
|
4399
|
-
/**
|
|
4400
|
-
* The usage numbers for the LeMUR request
|
|
4401
|
-
*/
|
|
4402
|
-
interface LemurUsage {
|
|
4403
|
-
/**
|
|
4404
|
-
* The number of input tokens used by the model
|
|
4405
|
-
* @minimum 0
|
|
4406
|
-
*/
|
|
4407
|
-
input_tokens: number;
|
|
4408
|
-
/**
|
|
4409
|
-
* The number of output tokens generated by the model
|
|
4410
|
-
* @minimum 0
|
|
4411
|
-
*/
|
|
4412
|
-
output_tokens: number;
|
|
4413
|
-
}
|
|
4414
|
-
|
|
4415
|
-
/**
|
|
4416
|
-
* Generated by orval v7.9.0 🍺
|
|
4417
|
-
* Do not edit manually.
|
|
4418
|
-
* AssemblyAI API
|
|
4419
|
-
* AssemblyAI API
|
|
4420
|
-
* OpenAPI spec version: 1.3.4
|
|
4421
|
-
*/
|
|
4422
|
-
|
|
4423
|
-
interface LemurBaseResponse {
|
|
4424
|
-
/** The ID of the LeMUR request */
|
|
4425
|
-
request_id: string;
|
|
4426
|
-
/** The usage numbers for the LeMUR request */
|
|
4427
|
-
usage: LemurUsage;
|
|
4428
|
-
}
|
|
4429
|
-
|
|
4430
|
-
/**
|
|
4431
|
-
* Generated by orval v7.9.0 🍺
|
|
4432
|
-
* Do not edit manually.
|
|
4433
|
-
* AssemblyAI API
|
|
4434
|
-
* AssemblyAI API
|
|
4435
|
-
* OpenAPI spec version: 1.3.4
|
|
4436
|
-
*/
|
|
4437
|
-
|
|
4438
|
-
type LemurStringResponse = LemurStringResponseAllOf & LemurBaseResponse;
|
|
4439
|
-
|
|
4440
|
-
/**
|
|
4441
|
-
* Generated by orval v7.9.0 🍺
|
|
4442
|
-
* Do not edit manually.
|
|
4443
|
-
* AssemblyAI API
|
|
4444
|
-
* AssemblyAI API
|
|
4445
|
-
* OpenAPI spec version: 1.3.4
|
|
4446
|
-
*/
|
|
4447
|
-
|
|
4448
|
-
type LemurActionItemsResponse = LemurStringResponse;
|
|
4449
|
-
|
|
4450
|
-
/**
|
|
4451
|
-
* Generated by orval v7.9.0 🍺
|
|
4452
|
-
* Do not edit manually.
|
|
4453
|
-
* AssemblyAI API
|
|
4454
|
-
* AssemblyAI API
|
|
4455
|
-
* OpenAPI spec version: 1.3.4
|
|
4456
|
-
*/
|
|
4457
|
-
type LemurQuestionContextOneOf = {
|
|
4458
|
-
[key: string]: unknown;
|
|
4459
|
-
};
|
|
4460
|
-
|
|
4461
|
-
/**
|
|
4462
|
-
* Generated by orval v7.9.0 🍺
|
|
4463
|
-
* Do not edit manually.
|
|
4464
|
-
* AssemblyAI API
|
|
4465
|
-
* AssemblyAI API
|
|
4466
|
-
* OpenAPI spec version: 1.3.4
|
|
4467
|
-
*/
|
|
4468
|
-
|
|
4469
|
-
/**
|
|
4470
|
-
* Any context about the transcripts you wish to provide. This can be a string or any object.
|
|
4471
|
-
*/
|
|
4472
|
-
type LemurQuestionContext = string | LemurQuestionContextOneOf;
|
|
4473
|
-
|
|
4474
|
-
/**
|
|
4475
|
-
* Generated by orval v7.9.0 🍺
|
|
4476
|
-
* Do not edit manually.
|
|
4477
|
-
* AssemblyAI API
|
|
4478
|
-
* AssemblyAI API
|
|
4479
|
-
* OpenAPI spec version: 1.3.4
|
|
4480
|
-
*/
|
|
4481
|
-
|
|
4482
|
-
interface LemurQuestion {
|
|
4483
|
-
/** The question you wish to ask. For more complex questions use default model. */
|
|
4484
|
-
question: string;
|
|
4485
|
-
/** Any context about the transcripts you wish to provide. This can be a string or any object. */
|
|
4486
|
-
context?: LemurQuestionContext;
|
|
4487
|
-
/** How you want the answer to be returned. This can be any text. Can't be used with answer_options. Examples: "short sentence", "bullet points"
|
|
4488
|
-
*/
|
|
4489
|
-
answer_format?: string;
|
|
4490
|
-
/** What discrete options to return. Useful for precise responses. Can't be used with answer_format. Example: ["Yes", "No"]
|
|
4491
|
-
*/
|
|
4492
|
-
answer_options?: string[];
|
|
4493
|
-
}
|
|
4494
|
-
|
|
4495
|
-
/**
|
|
4496
|
-
* Generated by orval v7.9.0 🍺
|
|
4497
|
-
* Do not edit manually.
|
|
4498
|
-
* AssemblyAI API
|
|
4499
|
-
* AssemblyAI API
|
|
4500
|
-
* OpenAPI spec version: 1.3.4
|
|
4501
|
-
*/
|
|
4502
|
-
/**
|
|
4503
|
-
* An answer generated by LeMUR and its question
|
|
4504
|
-
*/
|
|
4505
|
-
interface LemurQuestionAnswer {
|
|
4506
|
-
/** The question for LeMUR to answer */
|
|
4507
|
-
question: string;
|
|
4508
|
-
/** The answer generated by LeMUR */
|
|
4509
|
-
answer: string;
|
|
4510
|
-
}
|
|
4511
|
-
|
|
4512
|
-
/**
|
|
4513
|
-
* Generated by orval v7.9.0 🍺
|
|
4514
|
-
* Do not edit manually.
|
|
4515
|
-
* AssemblyAI API
|
|
4516
|
-
* AssemblyAI API
|
|
4517
|
-
* OpenAPI spec version: 1.3.4
|
|
4518
|
-
*/
|
|
4519
|
-
|
|
4520
|
-
type LemurQuestionAnswerParamsAllOf = {
|
|
4521
|
-
/** A list of questions to ask */
|
|
4522
|
-
questions: LemurQuestion[];
|
|
4523
|
-
};
|
|
4524
|
-
|
|
4525
|
-
/**
|
|
4526
|
-
* Generated by orval v7.9.0 🍺
|
|
4527
|
-
* Do not edit manually.
|
|
4528
|
-
* AssemblyAI API
|
|
4529
|
-
* AssemblyAI API
|
|
4530
|
-
* OpenAPI spec version: 1.3.4
|
|
4531
|
-
*/
|
|
4532
|
-
|
|
4533
|
-
type LemurQuestionAnswerParams = LemurBaseParams & LemurQuestionAnswerParamsAllOf;
|
|
4534
|
-
|
|
4535
|
-
/**
|
|
4536
|
-
* Generated by orval v7.9.0 🍺
|
|
4537
|
-
* Do not edit manually.
|
|
4538
|
-
* AssemblyAI API
|
|
4539
|
-
* AssemblyAI API
|
|
4540
|
-
* OpenAPI spec version: 1.3.4
|
|
4541
|
-
*/
|
|
4542
|
-
|
|
4543
|
-
type LemurQuestionAnswerResponseAllOf = {
|
|
4544
|
-
/** The answers generated by LeMUR and their questions */
|
|
4545
|
-
response: LemurQuestionAnswer[];
|
|
4546
|
-
};
|
|
4547
|
-
|
|
4548
|
-
/**
|
|
4549
|
-
* Generated by orval v7.9.0 🍺
|
|
4550
|
-
* Do not edit manually.
|
|
4551
|
-
* AssemblyAI API
|
|
4552
|
-
* AssemblyAI API
|
|
4553
|
-
* OpenAPI spec version: 1.3.4
|
|
4554
|
-
*/
|
|
4555
|
-
|
|
4556
|
-
type LemurQuestionAnswerResponse = LemurBaseResponse & LemurQuestionAnswerResponseAllOf;
|
|
4557
|
-
|
|
4558
|
-
/**
|
|
4559
|
-
* Generated by orval v7.9.0 🍺
|
|
4560
|
-
* Do not edit manually.
|
|
4561
|
-
* AssemblyAI API
|
|
4562
|
-
* AssemblyAI API
|
|
4563
|
-
* OpenAPI spec version: 1.3.4
|
|
4564
|
-
*/
|
|
4565
|
-
|
|
4566
|
-
type LemurResponse = LemurStringResponse | LemurQuestionAnswerResponse;
|
|
4567
|
-
|
|
4568
|
-
/**
|
|
4569
|
-
* Generated by orval v7.9.0 🍺
|
|
4570
|
-
* Do not edit manually.
|
|
4571
|
-
* AssemblyAI API
|
|
4572
|
-
* AssemblyAI API
|
|
4573
|
-
* OpenAPI spec version: 1.3.4
|
|
4574
|
-
*/
|
|
4575
|
-
type LemurSummaryParamsAllOf = {
|
|
4576
|
-
/** How you want the summary to be returned. This can be any text. Examples: "TLDR", "bullet points"
|
|
4577
|
-
*/
|
|
4578
|
-
answer_format?: string;
|
|
4579
|
-
};
|
|
4580
|
-
|
|
4581
|
-
/**
|
|
4582
|
-
* Generated by orval v7.9.0 🍺
|
|
4583
|
-
* Do not edit manually.
|
|
4584
|
-
* AssemblyAI API
|
|
4585
|
-
* AssemblyAI API
|
|
4586
|
-
* OpenAPI spec version: 1.3.4
|
|
4587
|
-
*/
|
|
4588
|
-
|
|
4589
|
-
type LemurSummaryParams = LemurBaseParams & LemurSummaryParamsAllOf;
|
|
4590
|
-
|
|
4591
|
-
/**
|
|
4592
|
-
* Generated by orval v7.9.0 🍺
|
|
4593
|
-
* Do not edit manually.
|
|
4594
|
-
* AssemblyAI API
|
|
4595
|
-
* AssemblyAI API
|
|
4596
|
-
* OpenAPI spec version: 1.3.4
|
|
4130
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4131
|
+
* OpenAPI spec version: 1.0.0
|
|
4597
4132
|
*/
|
|
4598
4133
|
|
|
4599
|
-
|
|
4134
|
+
interface TranscriptListItem {
|
|
4135
|
+
/** The unique identifier for the transcript */
|
|
4136
|
+
id: string;
|
|
4137
|
+
/** The URL to retrieve the transcript */
|
|
4138
|
+
resource_url: string;
|
|
4139
|
+
/** The status of the transcript */
|
|
4140
|
+
status: TranscriptStatus;
|
|
4141
|
+
/** The date and time the transcript was created */
|
|
4142
|
+
created: string;
|
|
4143
|
+
/** The date and time the transcript was completed */
|
|
4144
|
+
completed?: string;
|
|
4145
|
+
/** The URL to the audio file */
|
|
4146
|
+
audio_url: string;
|
|
4147
|
+
/** Error message of why the transcript failed */
|
|
4148
|
+
error: TranscriptListItemError;
|
|
4149
|
+
}
|
|
4600
4150
|
|
|
4601
4151
|
/**
|
|
4602
4152
|
* Generated by orval v7.9.0 🍺
|
|
4603
4153
|
* Do not edit manually.
|
|
4604
4154
|
* AssemblyAI API
|
|
4605
|
-
* AssemblyAI API
|
|
4606
|
-
* OpenAPI spec version: 1.
|
|
4155
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4156
|
+
* OpenAPI spec version: 1.0.0
|
|
4607
4157
|
*/
|
|
4608
|
-
|
|
4609
|
-
|
|
4610
|
-
|
|
4611
|
-
|
|
4158
|
+
/**
|
|
4159
|
+
* Additional error details if available
|
|
4160
|
+
*/
|
|
4161
|
+
interface ErrorDetails {
|
|
4162
|
+
[key: string]: unknown;
|
|
4163
|
+
}
|
|
4612
4164
|
|
|
4613
4165
|
/**
|
|
4614
4166
|
* Generated by orval v7.9.0 🍺
|
|
4615
4167
|
* Do not edit manually.
|
|
4616
4168
|
* AssemblyAI API
|
|
4617
|
-
* AssemblyAI API
|
|
4618
|
-
* OpenAPI spec version: 1.
|
|
4169
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4170
|
+
* OpenAPI spec version: 1.0.0
|
|
4619
4171
|
*/
|
|
4620
4172
|
|
|
4621
|
-
|
|
4173
|
+
interface Error$2 {
|
|
4174
|
+
/** Error message describing what went wrong */
|
|
4175
|
+
error: string;
|
|
4176
|
+
/** Error code for programmatic handling */
|
|
4177
|
+
code?: string;
|
|
4178
|
+
/** Additional error details if available */
|
|
4179
|
+
details?: ErrorDetails;
|
|
4180
|
+
}
|
|
4622
4181
|
|
|
4623
4182
|
/**
|
|
4624
4183
|
* Generated by orval v7.9.0 🍺
|
|
4625
4184
|
* Do not edit manually.
|
|
4626
4185
|
* AssemblyAI API
|
|
4627
|
-
* AssemblyAI API
|
|
4628
|
-
* OpenAPI spec version: 1.
|
|
4186
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4187
|
+
* OpenAPI spec version: 1.0.0
|
|
4629
4188
|
*/
|
|
4630
|
-
|
|
4631
|
-
|
|
4189
|
+
type GetSubtitles200 = {
|
|
4190
|
+
[key: string]: unknown;
|
|
4191
|
+
};
|
|
4632
4192
|
|
|
4633
4193
|
/**
|
|
4634
4194
|
* Generated by orval v7.9.0 🍺
|
|
4635
4195
|
* Do not edit manually.
|
|
4636
4196
|
* AssemblyAI API
|
|
4637
|
-
* AssemblyAI API
|
|
4638
|
-
* OpenAPI spec version: 1.
|
|
4197
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4198
|
+
* OpenAPI spec version: 1.0.0
|
|
4639
4199
|
*/
|
|
4640
|
-
|
|
4641
|
-
interface ListTranscriptParams {
|
|
4200
|
+
type GetSubtitlesParams = {
|
|
4642
4201
|
/**
|
|
4643
|
-
*
|
|
4644
|
-
* @minimum 1
|
|
4645
|
-
* @maximum 200
|
|
4202
|
+
* The maximum number of characters per caption
|
|
4646
4203
|
*/
|
|
4647
|
-
|
|
4648
|
-
|
|
4649
|
-
status?: TranscriptStatus;
|
|
4650
|
-
/** Only get transcripts created on this date */
|
|
4651
|
-
created_on?: string;
|
|
4652
|
-
/** Get transcripts that were created before this transcript ID */
|
|
4653
|
-
before_id?: string;
|
|
4654
|
-
/** Get transcripts that were created after this transcript ID */
|
|
4655
|
-
after_id?: string;
|
|
4656
|
-
/** Only get throttled transcripts, overrides the status filter */
|
|
4657
|
-
throttled_only?: boolean;
|
|
4658
|
-
}
|
|
4659
|
-
|
|
4660
|
-
/**
|
|
4661
|
-
* Not found
|
|
4662
|
-
*/
|
|
4663
|
-
type NotFoundResponse = Error;
|
|
4204
|
+
chars_per_caption?: number;
|
|
4205
|
+
};
|
|
4664
4206
|
|
|
4665
4207
|
/**
|
|
4666
4208
|
* Generated by orval v7.9.0 🍺
|
|
4667
4209
|
* Do not edit manually.
|
|
4668
4210
|
* AssemblyAI API
|
|
4669
|
-
* AssemblyAI API
|
|
4670
|
-
* OpenAPI spec version: 1.
|
|
4211
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4212
|
+
* OpenAPI spec version: 1.0.0
|
|
4671
4213
|
*/
|
|
4672
4214
|
/**
|
|
4673
4215
|
* The URL to the next page of transcripts. The previous URL always points to a page with older transcripts.
|
|
@@ -4678,8 +4220,8 @@ type PageDetailsPrevUrl = string | null;
|
|
|
4678
4220
|
* Generated by orval v7.9.0 🍺
|
|
4679
4221
|
* Do not edit manually.
|
|
4680
4222
|
* AssemblyAI API
|
|
4681
|
-
* AssemblyAI API
|
|
4682
|
-
* OpenAPI spec version: 1.
|
|
4223
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4224
|
+
* OpenAPI spec version: 1.0.0
|
|
4683
4225
|
*/
|
|
4684
4226
|
/**
|
|
4685
4227
|
* The URL to the next page of transcripts. The next URL always points to a page with newer transcripts.
|
|
@@ -4690,8 +4232,8 @@ type PageDetailsNextUrl = string | null;
|
|
|
4690
4232
|
* Generated by orval v7.9.0 🍺
|
|
4691
4233
|
* Do not edit manually.
|
|
4692
4234
|
* AssemblyAI API
|
|
4693
|
-
* AssemblyAI API
|
|
4694
|
-
* OpenAPI spec version: 1.
|
|
4235
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4236
|
+
* OpenAPI spec version: 1.0.0
|
|
4695
4237
|
*/
|
|
4696
4238
|
|
|
4697
4239
|
/**
|
|
@@ -4714,8 +4256,8 @@ interface PageDetails {
|
|
|
4714
4256
|
* Generated by orval v7.9.0 🍺
|
|
4715
4257
|
* Do not edit manually.
|
|
4716
4258
|
* AssemblyAI API
|
|
4717
|
-
* AssemblyAI API
|
|
4718
|
-
* OpenAPI spec version: 1.
|
|
4259
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4260
|
+
* OpenAPI spec version: 1.0.0
|
|
4719
4261
|
*/
|
|
4720
4262
|
|
|
4721
4263
|
interface TranscriptParagraph {
|
|
@@ -4725,11 +4267,7 @@ interface TranscriptParagraph {
|
|
|
4725
4267
|
start: number;
|
|
4726
4268
|
/** The ending time, in milliseconds, of the paragraph */
|
|
4727
4269
|
end: number;
|
|
4728
|
-
/**
|
|
4729
|
-
* The confidence score for the transcript of this paragraph
|
|
4730
|
-
* @minimum 0
|
|
4731
|
-
* @maximum 1
|
|
4732
|
-
*/
|
|
4270
|
+
/** The confidence score for the transcript of this paragraph */
|
|
4733
4271
|
confidence: number;
|
|
4734
4272
|
/** An array of words in the paragraph */
|
|
4735
4273
|
words: TranscriptWord[];
|
|
@@ -4739,18 +4277,14 @@ interface TranscriptParagraph {
|
|
|
4739
4277
|
* Generated by orval v7.9.0 🍺
|
|
4740
4278
|
* Do not edit manually.
|
|
4741
4279
|
* AssemblyAI API
|
|
4742
|
-
* AssemblyAI API
|
|
4743
|
-
* OpenAPI spec version: 1.
|
|
4280
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4281
|
+
* OpenAPI spec version: 1.0.0
|
|
4744
4282
|
*/
|
|
4745
4283
|
|
|
4746
4284
|
interface ParagraphsResponse {
|
|
4747
4285
|
/** The unique identifier of your transcript */
|
|
4748
4286
|
id: string;
|
|
4749
|
-
/**
|
|
4750
|
-
* The confidence score for the transcript
|
|
4751
|
-
* @minimum 0
|
|
4752
|
-
* @maximum 1
|
|
4753
|
-
*/
|
|
4287
|
+
/** The confidence score for the transcript */
|
|
4754
4288
|
confidence: number;
|
|
4755
4289
|
/** The duration of the audio file in seconds */
|
|
4756
4290
|
audio_duration: number;
|
|
@@ -4762,36 +4296,38 @@ interface ParagraphsResponse {
|
|
|
4762
4296
|
* Generated by orval v7.9.0 🍺
|
|
4763
4297
|
* Do not edit manually.
|
|
4764
4298
|
* AssemblyAI API
|
|
4765
|
-
* AssemblyAI API
|
|
4766
|
-
* OpenAPI spec version: 1.
|
|
4299
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4300
|
+
* OpenAPI spec version: 1.0.0
|
|
4767
4301
|
*/
|
|
4768
|
-
|
|
4769
|
-
|
|
4770
|
-
|
|
4771
|
-
|
|
4772
|
-
|
|
4773
|
-
|
|
4774
|
-
|
|
4775
|
-
}
|
|
4302
|
+
/**
|
|
4303
|
+
* The status of the redacted audio
|
|
4304
|
+
*/
|
|
4305
|
+
type RedactedAudioStatus = (typeof RedactedAudioStatus)[keyof typeof RedactedAudioStatus];
|
|
4306
|
+
declare const RedactedAudioStatus: {
|
|
4307
|
+
readonly redacted_audio_ready: "redacted_audio_ready";
|
|
4308
|
+
};
|
|
4776
4309
|
|
|
4777
4310
|
/**
|
|
4778
4311
|
* Generated by orval v7.9.0 🍺
|
|
4779
4312
|
* Do not edit manually.
|
|
4780
4313
|
* AssemblyAI API
|
|
4781
|
-
* AssemblyAI API
|
|
4782
|
-
* OpenAPI spec version: 1.
|
|
4314
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4315
|
+
* OpenAPI spec version: 1.0.0
|
|
4783
4316
|
*/
|
|
4784
|
-
|
|
4785
|
-
|
|
4786
|
-
|
|
4317
|
+
|
|
4318
|
+
interface RedactedAudioResponse {
|
|
4319
|
+
/** The status of the redacted audio */
|
|
4320
|
+
status: RedactedAudioStatus;
|
|
4321
|
+
/** The URL of the redacted audio file */
|
|
4322
|
+
redacted_audio_url: string;
|
|
4787
4323
|
}
|
|
4788
4324
|
|
|
4789
4325
|
/**
|
|
4790
4326
|
* Generated by orval v7.9.0 🍺
|
|
4791
4327
|
* Do not edit manually.
|
|
4792
4328
|
* AssemblyAI API
|
|
4793
|
-
* AssemblyAI API
|
|
4794
|
-
* OpenAPI spec version: 1.
|
|
4329
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4330
|
+
* OpenAPI spec version: 1.0.0
|
|
4795
4331
|
*/
|
|
4796
4332
|
/**
|
|
4797
4333
|
* The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
|
|
@@ -4802,11 +4338,11 @@ type TranscriptSentenceChannel = string | null;
|
|
|
4802
4338
|
* Generated by orval v7.9.0 🍺
|
|
4803
4339
|
* Do not edit manually.
|
|
4804
4340
|
* AssemblyAI API
|
|
4805
|
-
* AssemblyAI API
|
|
4806
|
-
* OpenAPI spec version: 1.
|
|
4341
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4342
|
+
* OpenAPI spec version: 1.0.0
|
|
4807
4343
|
*/
|
|
4808
4344
|
/**
|
|
4809
|
-
* The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/
|
|
4345
|
+
* The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
|
|
4810
4346
|
*/
|
|
4811
4347
|
type TranscriptSentenceSpeaker = string | null;
|
|
4812
4348
|
|
|
@@ -4814,8 +4350,8 @@ type TranscriptSentenceSpeaker = string | null;
|
|
|
4814
4350
|
* Generated by orval v7.9.0 🍺
|
|
4815
4351
|
* Do not edit manually.
|
|
4816
4352
|
* AssemblyAI API
|
|
4817
|
-
* AssemblyAI API
|
|
4818
|
-
* OpenAPI spec version: 1.
|
|
4353
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4354
|
+
* OpenAPI spec version: 1.0.0
|
|
4819
4355
|
*/
|
|
4820
4356
|
|
|
4821
4357
|
interface TranscriptSentence {
|
|
@@ -4825,17 +4361,13 @@ interface TranscriptSentence {
|
|
|
4825
4361
|
start: number;
|
|
4826
4362
|
/** The ending time, in milliseconds, for the sentence */
|
|
4827
4363
|
end: number;
|
|
4828
|
-
/**
|
|
4829
|
-
* The confidence score for the transcript of this sentence
|
|
4830
|
-
* @minimum 0
|
|
4831
|
-
* @maximum 1
|
|
4832
|
-
*/
|
|
4364
|
+
/** The confidence score for the transcript of this sentence */
|
|
4833
4365
|
confidence: number;
|
|
4834
4366
|
/** An array of words in the sentence */
|
|
4835
4367
|
words: TranscriptWord[];
|
|
4836
4368
|
/** The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
|
|
4837
4369
|
channel?: TranscriptSentenceChannel;
|
|
4838
|
-
/** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/
|
|
4370
|
+
/** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
|
|
4839
4371
|
speaker: TranscriptSentenceSpeaker;
|
|
4840
4372
|
}
|
|
4841
4373
|
|
|
@@ -4843,18 +4375,14 @@ interface TranscriptSentence {
|
|
|
4843
4375
|
* Generated by orval v7.9.0 🍺
|
|
4844
4376
|
* Do not edit manually.
|
|
4845
4377
|
* AssemblyAI API
|
|
4846
|
-
* AssemblyAI API
|
|
4847
|
-
* OpenAPI spec version: 1.
|
|
4378
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4379
|
+
* OpenAPI spec version: 1.0.0
|
|
4848
4380
|
*/
|
|
4849
4381
|
|
|
4850
4382
|
interface SentencesResponse {
|
|
4851
4383
|
/** The unique identifier for the transcript */
|
|
4852
4384
|
id: string;
|
|
4853
|
-
/**
|
|
4854
|
-
* The confidence score for the transcript
|
|
4855
|
-
* @minimum 0
|
|
4856
|
-
* @maximum 1
|
|
4857
|
-
*/
|
|
4385
|
+
/** The confidence score for the transcript */
|
|
4858
4386
|
confidence: number;
|
|
4859
4387
|
/** The duration of the audio file in seconds */
|
|
4860
4388
|
audio_duration: number;
|
|
@@ -4866,20 +4394,8 @@ interface SentencesResponse {
|
|
|
4866
4394
|
* Generated by orval v7.9.0 🍺
|
|
4867
4395
|
* Do not edit manually.
|
|
4868
4396
|
* AssemblyAI API
|
|
4869
|
-
* AssemblyAI API
|
|
4870
|
-
* OpenAPI spec version: 1.
|
|
4871
|
-
*/
|
|
4872
|
-
/**
|
|
4873
|
-
* Service unavailable
|
|
4874
|
-
*/
|
|
4875
|
-
type ServiceUnavailableResponse = void;
|
|
4876
|
-
|
|
4877
|
-
/**
|
|
4878
|
-
* Generated by orval v7.9.0 🍺
|
|
4879
|
-
* Do not edit manually.
|
|
4880
|
-
* AssemblyAI API
|
|
4881
|
-
* AssemblyAI API
|
|
4882
|
-
* OpenAPI spec version: 1.3.4
|
|
4397
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4398
|
+
* OpenAPI spec version: 1.0.0
|
|
4883
4399
|
*/
|
|
4884
4400
|
/**
|
|
4885
4401
|
* Format of the subtitles
|
|
@@ -4890,17 +4406,12 @@ declare const SubtitleFormat: {
|
|
|
4890
4406
|
readonly vtt: "vtt";
|
|
4891
4407
|
};
|
|
4892
4408
|
|
|
4893
|
-
/**
|
|
4894
|
-
* Too many requests
|
|
4895
|
-
*/
|
|
4896
|
-
type TooManyRequestsResponse = Error;
|
|
4897
|
-
|
|
4898
4409
|
/**
|
|
4899
4410
|
* Generated by orval v7.9.0 🍺
|
|
4900
4411
|
* Do not edit manually.
|
|
4901
4412
|
* AssemblyAI API
|
|
4902
|
-
* AssemblyAI API
|
|
4903
|
-
* OpenAPI spec version: 1.
|
|
4413
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4414
|
+
* OpenAPI spec version: 1.0.0
|
|
4904
4415
|
*/
|
|
4905
4416
|
|
|
4906
4417
|
/**
|
|
@@ -4913,17 +4424,12 @@ interface TranscriptList {
|
|
|
4913
4424
|
transcripts: TranscriptListItem[];
|
|
4914
4425
|
}
|
|
4915
4426
|
|
|
4916
|
-
/**
|
|
4917
|
-
* Unauthorized
|
|
4918
|
-
*/
|
|
4919
|
-
type UnauthorizedResponse = Error;
|
|
4920
|
-
|
|
4921
4427
|
/**
|
|
4922
4428
|
* Generated by orval v7.9.0 🍺
|
|
4923
4429
|
* Do not edit manually.
|
|
4924
4430
|
* AssemblyAI API
|
|
4925
|
-
* AssemblyAI API
|
|
4926
|
-
* OpenAPI spec version: 1.
|
|
4431
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4432
|
+
* OpenAPI spec version: 1.0.0
|
|
4927
4433
|
*/
|
|
4928
4434
|
interface UploadedFile {
|
|
4929
4435
|
/** A URL that points to your audio file, accessible only by AssemblyAI's servers
|
|
@@ -4935,8 +4441,8 @@ interface UploadedFile {
|
|
|
4935
4441
|
* Generated by orval v7.9.0 🍺
|
|
4936
4442
|
* Do not edit manually.
|
|
4937
4443
|
* AssemblyAI API
|
|
4938
|
-
* AssemblyAI API
|
|
4939
|
-
* OpenAPI spec version: 1.
|
|
4444
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4445
|
+
* OpenAPI spec version: 1.0.0
|
|
4940
4446
|
*/
|
|
4941
4447
|
/**
|
|
4942
4448
|
* An array of timestamps structured as [`start_time`, `end_time`] in milliseconds
|
|
@@ -4947,8 +4453,8 @@ type WordSearchTimestamp = number[];
|
|
|
4947
4453
|
* Generated by orval v7.9.0 🍺
|
|
4948
4454
|
* Do not edit manually.
|
|
4949
4455
|
* AssemblyAI API
|
|
4950
|
-
* AssemblyAI API
|
|
4951
|
-
* OpenAPI spec version: 1.
|
|
4456
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4457
|
+
* OpenAPI spec version: 1.0.0
|
|
4952
4458
|
*/
|
|
4953
4459
|
|
|
4954
4460
|
interface WordSearchMatch {
|
|
@@ -4966,8 +4472,8 @@ interface WordSearchMatch {
|
|
|
4966
4472
|
* Generated by orval v7.9.0 🍺
|
|
4967
4473
|
* Do not edit manually.
|
|
4968
4474
|
* AssemblyAI API
|
|
4969
|
-
* AssemblyAI API
|
|
4970
|
-
* OpenAPI spec version: 1.
|
|
4475
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4476
|
+
* OpenAPI spec version: 1.0.0
|
|
4971
4477
|
*/
|
|
4972
4478
|
type WordSearchParams = {
|
|
4973
4479
|
/**
|
|
@@ -4980,8 +4486,8 @@ type WordSearchParams = {
|
|
|
4980
4486
|
* Generated by orval v7.9.0 🍺
|
|
4981
4487
|
* Do not edit manually.
|
|
4982
4488
|
* AssemblyAI API
|
|
4983
|
-
* AssemblyAI API
|
|
4984
|
-
* OpenAPI spec version: 1.
|
|
4489
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4490
|
+
* OpenAPI spec version: 1.0.0
|
|
4985
4491
|
*/
|
|
4986
4492
|
|
|
4987
4493
|
interface WordSearchResponse {
|
|
@@ -4997,93 +4503,66 @@ interface WordSearchResponse {
|
|
|
4997
4503
|
* Generated by orval v7.9.0 🍺
|
|
4998
4504
|
* Do not edit manually.
|
|
4999
4505
|
* AssemblyAI API
|
|
5000
|
-
* AssemblyAI API
|
|
5001
|
-
* OpenAPI spec version: 1.
|
|
4506
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4507
|
+
* OpenAPI spec version: 1.0.0
|
|
5002
4508
|
*/
|
|
5003
4509
|
|
|
5004
4510
|
declare const index$6_AudioIntelligenceModelStatus: typeof AudioIntelligenceModelStatus;
|
|
5005
4511
|
declare const index$6_AutoHighlightResult: typeof AutoHighlightResult;
|
|
5006
4512
|
declare const index$6_AutoHighlightsResult: typeof AutoHighlightsResult;
|
|
5007
|
-
type index$6_BadRequestResponse = BadRequestResponse;
|
|
5008
|
-
type index$6_CannotAccessUploadedFileResponse = CannotAccessUploadedFileResponse;
|
|
5009
4513
|
declare const index$6_Chapter: typeof Chapter;
|
|
5010
4514
|
declare const index$6_ContentSafetyLabel: typeof ContentSafetyLabel;
|
|
5011
4515
|
declare const index$6_ContentSafetyLabelResult: typeof ContentSafetyLabelResult;
|
|
5012
4516
|
declare const index$6_ContentSafetyLabelsResult: typeof ContentSafetyLabelsResult;
|
|
5013
4517
|
declare const index$6_ContentSafetyLabelsResultSeverityScoreSummary: typeof ContentSafetyLabelsResultSeverityScoreSummary;
|
|
5014
4518
|
declare const index$6_ContentSafetyLabelsResultSummary: typeof ContentSafetyLabelsResultSummary;
|
|
5015
|
-
|
|
4519
|
+
declare const index$6_CustomFormattingRequestBody: typeof CustomFormattingRequestBody;
|
|
4520
|
+
declare const index$6_CustomFormattingRequestBodyCustomFormatting: typeof CustomFormattingRequestBodyCustomFormatting;
|
|
4521
|
+
declare const index$6_CustomFormattingResponse: typeof CustomFormattingResponse;
|
|
4522
|
+
declare const index$6_CustomFormattingResponseCustomFormatting: typeof CustomFormattingResponseCustomFormatting;
|
|
4523
|
+
declare const index$6_CustomFormattingResponseCustomFormattingMapping: typeof CustomFormattingResponseCustomFormattingMapping;
|
|
5016
4524
|
declare const index$6_Entity: typeof Entity;
|
|
5017
4525
|
declare const index$6_EntityType: typeof EntityType;
|
|
5018
|
-
type index$
|
|
4526
|
+
type index$6_ErrorDetails = ErrorDetails;
|
|
4527
|
+
type index$6_GetSubtitles200 = GetSubtitles200;
|
|
5019
4528
|
type index$6_GetSubtitlesParams = GetSubtitlesParams;
|
|
5020
|
-
type index$6_InternalServerErrorResponse = InternalServerErrorResponse;
|
|
5021
|
-
type index$6_LemurActionItemsParams = LemurActionItemsParams;
|
|
5022
|
-
type index$6_LemurActionItemsParamsAllOf = LemurActionItemsParamsAllOf;
|
|
5023
|
-
type index$6_LemurActionItemsResponse = LemurActionItemsResponse;
|
|
5024
|
-
type index$6_LemurBaseParams = LemurBaseParams;
|
|
5025
|
-
type index$6_LemurBaseParamsContext = LemurBaseParamsContext;
|
|
5026
|
-
type index$6_LemurBaseParamsContextOneOf = LemurBaseParamsContextOneOf;
|
|
5027
|
-
type index$6_LemurBaseParamsFinalModel = LemurBaseParamsFinalModel;
|
|
5028
|
-
type index$6_LemurBaseResponse = LemurBaseResponse;
|
|
5029
|
-
declare const index$6_LemurModel: typeof LemurModel;
|
|
5030
|
-
type index$6_LemurQuestion = LemurQuestion;
|
|
5031
|
-
type index$6_LemurQuestionAnswer = LemurQuestionAnswer;
|
|
5032
|
-
type index$6_LemurQuestionAnswerParams = LemurQuestionAnswerParams;
|
|
5033
|
-
type index$6_LemurQuestionAnswerParamsAllOf = LemurQuestionAnswerParamsAllOf;
|
|
5034
|
-
type index$6_LemurQuestionAnswerResponse = LemurQuestionAnswerResponse;
|
|
5035
|
-
type index$6_LemurQuestionAnswerResponseAllOf = LemurQuestionAnswerResponseAllOf;
|
|
5036
|
-
type index$6_LemurQuestionContext = LemurQuestionContext;
|
|
5037
|
-
type index$6_LemurQuestionContextOneOf = LemurQuestionContextOneOf;
|
|
5038
|
-
type index$6_LemurResponse = LemurResponse;
|
|
5039
|
-
type index$6_LemurStringResponse = LemurStringResponse;
|
|
5040
|
-
type index$6_LemurStringResponseAllOf = LemurStringResponseAllOf;
|
|
5041
|
-
type index$6_LemurSummaryParams = LemurSummaryParams;
|
|
5042
|
-
type index$6_LemurSummaryParamsAllOf = LemurSummaryParamsAllOf;
|
|
5043
|
-
type index$6_LemurSummaryResponse = LemurSummaryResponse;
|
|
5044
|
-
type index$6_LemurTaskParams = LemurTaskParams;
|
|
5045
|
-
type index$6_LemurTaskParamsAllOf = LemurTaskParamsAllOf;
|
|
5046
|
-
type index$6_LemurTaskResponse = LemurTaskResponse;
|
|
5047
|
-
type index$6_LemurUsage = LemurUsage;
|
|
5048
|
-
type index$6_ListTranscriptParams = ListTranscriptParams;
|
|
5049
4529
|
declare const index$6_ListTranscriptsParams: typeof ListTranscriptsParams;
|
|
5050
|
-
type index$6_NotFoundResponse = NotFoundResponse;
|
|
5051
4530
|
type index$6_PageDetails = PageDetails;
|
|
5052
4531
|
type index$6_PageDetailsNextUrl = PageDetailsNextUrl;
|
|
5053
4532
|
type index$6_PageDetailsPrevUrl = PageDetailsPrevUrl;
|
|
5054
4533
|
type index$6_ParagraphsResponse = ParagraphsResponse;
|
|
5055
4534
|
declare const index$6_PiiPolicy: typeof PiiPolicy;
|
|
5056
|
-
type index$6_PurgeLemurRequestDataResponse = PurgeLemurRequestDataResponse;
|
|
5057
|
-
type index$6_RealtimeTemporaryTokenResponse = RealtimeTemporaryTokenResponse;
|
|
5058
4535
|
declare const index$6_RedactPiiAudioQuality: typeof RedactPiiAudioQuality;
|
|
5059
|
-
|
|
5060
|
-
declare const index$6_RedactedAudioResponse: typeof RedactedAudioResponse;
|
|
4536
|
+
type index$6_RedactedAudioResponse = RedactedAudioResponse;
|
|
5061
4537
|
declare const index$6_RedactedAudioStatus: typeof RedactedAudioStatus;
|
|
5062
4538
|
type index$6_SentencesResponse = SentencesResponse;
|
|
5063
4539
|
declare const index$6_Sentiment: typeof Sentiment;
|
|
5064
4540
|
declare const index$6_SentimentAnalysisResultChannel: typeof SentimentAnalysisResultChannel;
|
|
5065
4541
|
declare const index$6_SentimentAnalysisResultSpeaker: typeof SentimentAnalysisResultSpeaker;
|
|
5066
|
-
type index$6_ServiceUnavailableResponse = ServiceUnavailableResponse;
|
|
5067
4542
|
declare const index$6_SeverityScoreSummary: typeof SeverityScoreSummary;
|
|
4543
|
+
declare const index$6_SpeakerIdentificationRequestBody: typeof SpeakerIdentificationRequestBody;
|
|
4544
|
+
declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentification: typeof SpeakerIdentificationRequestBodySpeakerIdentification;
|
|
4545
|
+
declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType;
|
|
4546
|
+
declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems;
|
|
4547
|
+
declare const index$6_SpeakerIdentificationResponse: typeof SpeakerIdentificationResponse;
|
|
4548
|
+
declare const index$6_SpeakerIdentificationResponseSpeakerIdentification: typeof SpeakerIdentificationResponseSpeakerIdentification;
|
|
4549
|
+
declare const index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping: typeof SpeakerIdentificationResponseSpeakerIdentificationMapping;
|
|
5068
4550
|
declare const index$6_SpeechModel: typeof SpeechModel;
|
|
5069
4551
|
declare const index$6_SubstitutionPolicy: typeof SubstitutionPolicy;
|
|
5070
4552
|
declare const index$6_SubtitleFormat: typeof SubtitleFormat;
|
|
5071
4553
|
declare const index$6_SummaryModel: typeof SummaryModel;
|
|
5072
4554
|
declare const index$6_SummaryType: typeof SummaryType;
|
|
5073
4555
|
declare const index$6_Timestamp: typeof Timestamp;
|
|
5074
|
-
type index$6_TooManyRequestsResponse = TooManyRequestsResponse;
|
|
5075
4556
|
declare const index$6_TopicDetectionModelResult: typeof TopicDetectionModelResult;
|
|
5076
4557
|
declare const index$6_TopicDetectionModelResultSummary: typeof TopicDetectionModelResultSummary;
|
|
5077
4558
|
declare const index$6_TopicDetectionResult: typeof TopicDetectionResult;
|
|
5078
|
-
declare const index$
|
|
4559
|
+
declare const index$6_TopicDetectionResultLabelsItems: typeof TopicDetectionResultLabelsItems;
|
|
5079
4560
|
declare const index$6_Transcript: typeof Transcript;
|
|
5080
4561
|
declare const index$6_TranscriptAudioDuration: typeof TranscriptAudioDuration;
|
|
5081
4562
|
declare const index$6_TranscriptAudioEndAt: typeof TranscriptAudioEndAt;
|
|
5082
4563
|
declare const index$6_TranscriptAudioStartFrom: typeof TranscriptAudioStartFrom;
|
|
5083
4564
|
declare const index$6_TranscriptAutoChapters: typeof TranscriptAutoChapters;
|
|
5084
4565
|
declare const index$6_TranscriptAutoHighlightsResult: typeof TranscriptAutoHighlightsResult;
|
|
5085
|
-
declare const index$6_TranscriptBoostParam: typeof TranscriptBoostParam;
|
|
5086
|
-
declare const index$6_TranscriptBoostParamProperty: typeof TranscriptBoostParamProperty;
|
|
5087
4566
|
declare const index$6_TranscriptChapters: typeof TranscriptChapters;
|
|
5088
4567
|
declare const index$6_TranscriptConfidence: typeof TranscriptConfidence;
|
|
5089
4568
|
declare const index$6_TranscriptContentSafety: typeof TranscriptContentSafety;
|
|
@@ -5092,6 +4571,7 @@ declare const index$6_TranscriptCustomSpelling: typeof TranscriptCustomSpelling;
|
|
|
5092
4571
|
declare const index$6_TranscriptCustomSpellingProperty: typeof TranscriptCustomSpellingProperty;
|
|
5093
4572
|
declare const index$6_TranscriptCustomTopics: typeof TranscriptCustomTopics;
|
|
5094
4573
|
declare const index$6_TranscriptDisfluencies: typeof TranscriptDisfluencies;
|
|
4574
|
+
declare const index$6_TranscriptDomain: typeof TranscriptDomain;
|
|
5095
4575
|
declare const index$6_TranscriptEntities: typeof TranscriptEntities;
|
|
5096
4576
|
declare const index$6_TranscriptEntityDetection: typeof TranscriptEntityDetection;
|
|
5097
4577
|
declare const index$6_TranscriptFilterProfanity: typeof TranscriptFilterProfanity;
|
|
@@ -5099,33 +4579,41 @@ declare const index$6_TranscriptFormatText: typeof TranscriptFormatText;
|
|
|
5099
4579
|
declare const index$6_TranscriptIabCategories: typeof TranscriptIabCategories;
|
|
5100
4580
|
declare const index$6_TranscriptIabCategoriesResult: typeof TranscriptIabCategoriesResult;
|
|
5101
4581
|
declare const index$6_TranscriptLanguageCode: typeof TranscriptLanguageCode;
|
|
5102
|
-
declare const index$
|
|
4582
|
+
declare const index$6_TranscriptLanguageCodes: typeof TranscriptLanguageCodes;
|
|
5103
4583
|
declare const index$6_TranscriptLanguageConfidence: typeof TranscriptLanguageConfidence;
|
|
5104
4584
|
declare const index$6_TranscriptLanguageConfidenceThreshold: typeof TranscriptLanguageConfidenceThreshold;
|
|
5105
4585
|
declare const index$6_TranscriptLanguageDetection: typeof TranscriptLanguageDetection;
|
|
4586
|
+
declare const index$6_TranscriptLanguageDetectionOptions: typeof TranscriptLanguageDetectionOptions;
|
|
5106
4587
|
type index$6_TranscriptList = TranscriptList;
|
|
5107
4588
|
type index$6_TranscriptListItem = TranscriptListItem;
|
|
5108
|
-
type index$6_TranscriptListItemCompleted = TranscriptListItemCompleted;
|
|
5109
4589
|
type index$6_TranscriptListItemError = TranscriptListItemError;
|
|
5110
4590
|
declare const index$6_TranscriptMultichannel: typeof TranscriptMultichannel;
|
|
5111
|
-
declare const index$
|
|
5112
|
-
declare const index$
|
|
5113
|
-
declare const index$
|
|
5114
|
-
declare const index$
|
|
5115
|
-
declare const index$
|
|
5116
|
-
declare const index$
|
|
5117
|
-
declare const index$
|
|
5118
|
-
declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderName: typeof TranscriptOptionalParamsWebhookAuthHeaderName;
|
|
5119
|
-
declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderValue: typeof TranscriptOptionalParamsWebhookAuthHeaderValue;
|
|
4591
|
+
declare const index$6_TranscriptOptionalParamsLanguageDetectionOptions: typeof TranscriptOptionalParamsLanguageDetectionOptions;
|
|
4592
|
+
declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptions: typeof TranscriptOptionalParamsRedactPiiAudioOptions;
|
|
4593
|
+
declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod;
|
|
4594
|
+
declare const index$6_TranscriptOptionalParamsRemoveAudioTags: typeof TranscriptOptionalParamsRemoveAudioTags;
|
|
4595
|
+
declare const index$6_TranscriptOptionalParamsSpeakerOptions: typeof TranscriptOptionalParamsSpeakerOptions;
|
|
4596
|
+
declare const index$6_TranscriptOptionalParamsSpeechUnderstanding: typeof TranscriptOptionalParamsSpeechUnderstanding;
|
|
4597
|
+
declare const index$6_TranscriptOptionalParamsSpeechUnderstandingRequest: typeof TranscriptOptionalParamsSpeechUnderstandingRequest;
|
|
5120
4598
|
type index$6_TranscriptParagraph = TranscriptParagraph;
|
|
5121
|
-
|
|
5122
|
-
|
|
4599
|
+
declare const index$6_TranscriptParams: typeof TranscriptParams;
|
|
4600
|
+
declare const index$6_TranscriptParamsDomain: typeof TranscriptParamsDomain;
|
|
4601
|
+
declare const index$6_TranscriptParamsLanguageCode: typeof TranscriptParamsLanguageCode;
|
|
4602
|
+
declare const index$6_TranscriptParamsLanguageCodes: typeof TranscriptParamsLanguageCodes;
|
|
4603
|
+
declare const index$6_TranscriptParamsRedactPiiSub: typeof TranscriptParamsRedactPiiSub;
|
|
4604
|
+
declare const index$6_TranscriptParamsRemoveAudioTags: typeof TranscriptParamsRemoveAudioTags;
|
|
4605
|
+
declare const index$6_TranscriptParamsSpeakersExpected: typeof TranscriptParamsSpeakersExpected;
|
|
4606
|
+
declare const index$6_TranscriptParamsSpeechThreshold: typeof TranscriptParamsSpeechThreshold;
|
|
4607
|
+
declare const index$6_TranscriptParamsWebhookAuthHeaderName: typeof TranscriptParamsWebhookAuthHeaderName;
|
|
4608
|
+
declare const index$6_TranscriptParamsWebhookAuthHeaderValue: typeof TranscriptParamsWebhookAuthHeaderValue;
|
|
5123
4609
|
declare const index$6_TranscriptPunctuate: typeof TranscriptPunctuate;
|
|
5124
|
-
declare const index$6_TranscriptReadyNotification: typeof TranscriptReadyNotification;
|
|
5125
|
-
declare const index$6_TranscriptReadyStatus: typeof TranscriptReadyStatus;
|
|
5126
4610
|
declare const index$6_TranscriptRedactPiiAudio: typeof TranscriptRedactPiiAudio;
|
|
4611
|
+
declare const index$6_TranscriptRedactPiiAudioOptions: typeof TranscriptRedactPiiAudioOptions;
|
|
4612
|
+
declare const index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod;
|
|
5127
4613
|
declare const index$6_TranscriptRedactPiiAudioQuality: typeof TranscriptRedactPiiAudioQuality;
|
|
5128
4614
|
declare const index$6_TranscriptRedactPiiPolicies: typeof TranscriptRedactPiiPolicies;
|
|
4615
|
+
declare const index$6_TranscriptRemoveAudioTags: typeof TranscriptRemoveAudioTags;
|
|
4616
|
+
declare const index$6_TranscriptRemoveAudioTagsProperty: typeof TranscriptRemoveAudioTagsProperty;
|
|
5129
4617
|
type index$6_TranscriptSentence = TranscriptSentence;
|
|
5130
4618
|
type index$6_TranscriptSentenceChannel = TranscriptSentenceChannel;
|
|
5131
4619
|
type index$6_TranscriptSentenceSpeaker = TranscriptSentenceSpeaker;
|
|
@@ -5134,33 +4622,42 @@ declare const index$6_TranscriptSentimentAnalysisResults: typeof TranscriptSenti
|
|
|
5134
4622
|
declare const index$6_TranscriptSpeakerLabels: typeof TranscriptSpeakerLabels;
|
|
5135
4623
|
declare const index$6_TranscriptSpeakersExpected: typeof TranscriptSpeakersExpected;
|
|
5136
4624
|
declare const index$6_TranscriptSpeechModel: typeof TranscriptSpeechModel;
|
|
4625
|
+
declare const index$6_TranscriptSpeechModels: typeof TranscriptSpeechModels;
|
|
5137
4626
|
declare const index$6_TranscriptSpeechThreshold: typeof TranscriptSpeechThreshold;
|
|
4627
|
+
declare const index$6_TranscriptSpeechUnderstanding: typeof TranscriptSpeechUnderstanding;
|
|
4628
|
+
declare const index$6_TranscriptSpeechUnderstandingRequest: typeof TranscriptSpeechUnderstandingRequest;
|
|
4629
|
+
declare const index$6_TranscriptSpeechUnderstandingResponse: typeof TranscriptSpeechUnderstandingResponse;
|
|
5138
4630
|
declare const index$6_TranscriptSpeedBoost: typeof TranscriptSpeedBoost;
|
|
5139
4631
|
declare const index$6_TranscriptStatus: typeof TranscriptStatus;
|
|
5140
4632
|
declare const index$6_TranscriptSummary: typeof TranscriptSummary;
|
|
5141
4633
|
declare const index$6_TranscriptSummaryModel: typeof TranscriptSummaryModel;
|
|
5142
4634
|
declare const index$6_TranscriptSummaryType: typeof TranscriptSummaryType;
|
|
4635
|
+
declare const index$6_TranscriptTemperature: typeof TranscriptTemperature;
|
|
5143
4636
|
declare const index$6_TranscriptText: typeof TranscriptText;
|
|
5144
4637
|
declare const index$6_TranscriptThrottled: typeof TranscriptThrottled;
|
|
4638
|
+
declare const index$6_TranscriptTranslatedTexts: typeof TranscriptTranslatedTexts;
|
|
5145
4639
|
declare const index$6_TranscriptUtterance: typeof TranscriptUtterance;
|
|
5146
4640
|
declare const index$6_TranscriptUtteranceChannel: typeof TranscriptUtteranceChannel;
|
|
4641
|
+
declare const index$6_TranscriptUtteranceTranslatedTexts: typeof TranscriptUtteranceTranslatedTexts;
|
|
5147
4642
|
declare const index$6_TranscriptUtterances: typeof TranscriptUtterances;
|
|
5148
4643
|
declare const index$6_TranscriptWebhookAuthHeaderName: typeof TranscriptWebhookAuthHeaderName;
|
|
5149
|
-
declare const index$6_TranscriptWebhookNotification: typeof TranscriptWebhookNotification;
|
|
5150
4644
|
declare const index$6_TranscriptWebhookStatusCode: typeof TranscriptWebhookStatusCode;
|
|
5151
4645
|
declare const index$6_TranscriptWebhookUrl: typeof TranscriptWebhookUrl;
|
|
5152
4646
|
declare const index$6_TranscriptWord: typeof TranscriptWord;
|
|
5153
4647
|
declare const index$6_TranscriptWordChannel: typeof TranscriptWordChannel;
|
|
5154
4648
|
declare const index$6_TranscriptWordSpeaker: typeof TranscriptWordSpeaker;
|
|
5155
4649
|
declare const index$6_TranscriptWords: typeof TranscriptWords;
|
|
5156
|
-
|
|
4650
|
+
declare const index$6_TranslationRequestBody: typeof TranslationRequestBody;
|
|
4651
|
+
declare const index$6_TranslationRequestBodyTranslation: typeof TranslationRequestBodyTranslation;
|
|
4652
|
+
declare const index$6_TranslationResponse: typeof TranslationResponse;
|
|
4653
|
+
declare const index$6_TranslationResponseTranslation: typeof TranslationResponseTranslation;
|
|
5157
4654
|
type index$6_UploadedFile = UploadedFile;
|
|
5158
4655
|
type index$6_WordSearchMatch = WordSearchMatch;
|
|
5159
4656
|
type index$6_WordSearchParams = WordSearchParams;
|
|
5160
4657
|
type index$6_WordSearchResponse = WordSearchResponse;
|
|
5161
4658
|
type index$6_WordSearchTimestamp = WordSearchTimestamp;
|
|
5162
4659
|
declare namespace index$6 {
|
|
5163
|
-
export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult,
|
|
4660
|
+
export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, index$6_CustomFormattingRequestBody as CustomFormattingRequestBody, index$6_CustomFormattingRequestBodyCustomFormatting as CustomFormattingRequestBodyCustomFormatting, index$6_CustomFormattingResponse as CustomFormattingResponse, index$6_CustomFormattingResponseCustomFormatting as CustomFormattingResponseCustomFormatting, index$6_CustomFormattingResponseCustomFormattingMapping as CustomFormattingResponseCustomFormattingMapping, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_ErrorDetails as ErrorDetails, type index$6_GetSubtitles200 as GetSubtitles200, type index$6_GetSubtitlesParams as GetSubtitlesParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, type index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeakerIdentificationRequestBody as SpeakerIdentificationRequestBody, index$6_SpeakerIdentificationRequestBodySpeakerIdentification as SpeakerIdentificationRequestBodySpeakerIdentification, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, index$6_SpeakerIdentificationResponse as SpeakerIdentificationResponse, index$6_SpeakerIdentificationResponseSpeakerIdentification as SpeakerIdentificationResponseSpeakerIdentification, index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping as SpeakerIdentificationResponseSpeakerIdentificationMapping, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItems as TopicDetectionResultLabelsItems, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptDomain as TranscriptDomain, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodes as TranscriptLanguageCodes, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, index$6_TranscriptLanguageDetectionOptions as TranscriptLanguageDetectionOptions, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParamsLanguageDetectionOptions as TranscriptOptionalParamsLanguageDetectionOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptions as TranscriptOptionalParamsRedactPiiAudioOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptOptionalParamsRemoveAudioTags as TranscriptOptionalParamsRemoveAudioTags, index$6_TranscriptOptionalParamsSpeakerOptions as TranscriptOptionalParamsSpeakerOptions, index$6_TranscriptOptionalParamsSpeechUnderstanding as TranscriptOptionalParamsSpeechUnderstanding, index$6_TranscriptOptionalParamsSpeechUnderstandingRequest as TranscriptOptionalParamsSpeechUnderstandingRequest, type index$6_TranscriptParagraph as TranscriptParagraph, index$6_TranscriptParams as TranscriptParams, index$6_TranscriptParamsDomain as TranscriptParamsDomain, index$6_TranscriptParamsLanguageCode as TranscriptParamsLanguageCode, index$6_TranscriptParamsLanguageCodes as TranscriptParamsLanguageCodes, index$6_TranscriptParamsRedactPiiSub as TranscriptParamsRedactPiiSub, index$6_TranscriptParamsRemoveAudioTags as TranscriptParamsRemoveAudioTags, index$6_TranscriptParamsSpeakersExpected as TranscriptParamsSpeakersExpected, index$6_TranscriptParamsSpeechThreshold as TranscriptParamsSpeechThreshold, index$6_TranscriptParamsWebhookAuthHeaderName as TranscriptParamsWebhookAuthHeaderName, index$6_TranscriptParamsWebhookAuthHeaderValue as TranscriptParamsWebhookAuthHeaderValue, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioOptions as TranscriptRedactPiiAudioOptions, index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, index$6_TranscriptRemoveAudioTags as TranscriptRemoveAudioTags, index$6_TranscriptRemoveAudioTagsProperty as TranscriptRemoveAudioTagsProperty, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechModels as TranscriptSpeechModels, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeechUnderstanding as TranscriptSpeechUnderstanding, index$6_TranscriptSpeechUnderstandingRequest as TranscriptSpeechUnderstandingRequest, index$6_TranscriptSpeechUnderstandingResponse as TranscriptSpeechUnderstandingResponse, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptTemperature as TranscriptTemperature, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptTranslatedTexts as TranscriptTranslatedTexts, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtteranceTranslatedTexts as TranscriptUtteranceTranslatedTexts, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, index$6_TranslationRequestBody as TranslationRequestBody, index$6_TranslationRequestBodyTranslation as TranslationRequestBodyTranslation, index$6_TranslationResponse as TranslationResponse, index$6_TranslationResponseTranslation as TranslationResponseTranslation, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
|
|
5164
4661
|
}
|
|
5165
4662
|
|
|
5166
4663
|
/**
|
|
@@ -5177,6 +4674,15 @@ Transcripts are sorted from newest to oldest and can be retrieved for the last 9
|
|
|
5177
4674
|
|
|
5178
4675
|
If you need to retrieve transcripts from more than 90 days ago please reach out to our Support team at support@assemblyai.com.
|
|
5179
4676
|
|
|
4677
|
+
**Pagination**
|
|
4678
|
+
|
|
4679
|
+
This endpoint returns paginated results. The response includes a `page_details` object with the following properties:
|
|
4680
|
+
- `page_details.limit` - Maximum number of transcripts per page.
|
|
4681
|
+
- `page_details.result_count` - Total number of transcripts returned on the current page.
|
|
4682
|
+
- `page_details.current_url` - URL to the current page.
|
|
4683
|
+
- `page_details.prev_url` - URL to the previous page of older transcripts.
|
|
4684
|
+
- `page_details.next_url` - URL to the next page of newer transcripts.
|
|
4685
|
+
|
|
5180
4686
|
* @summary List transcripts
|
|
5181
4687
|
*/
|
|
5182
4688
|
declare const listTranscripts: <TData = AxiosResponse<TranscriptList>>(params?: ListTranscriptsParams, options?: AxiosRequestConfig) => Promise<TData>;
|
|
@@ -5195,14 +4701,6 @@ Remove the data from the transcript and mark it as deleted.
|
|
|
5195
4701
|
* @summary Delete transcript
|
|
5196
4702
|
*/
|
|
5197
4703
|
declare const deleteTranscript: <TData = AxiosResponse<Transcript>>(transcriptId: string, options?: AxiosRequestConfig) => Promise<TData>;
|
|
5198
|
-
/**
|
|
5199
|
-
* <Warning>Streaming Speech-to-Text is currently not available on the EU endpoint.</Warning>
|
|
5200
|
-
<Note>Any usage associated with a temporary token will be attributed to the API key that generated it.</Note>
|
|
5201
|
-
Create a temporary authentication token for Streaming Speech-to-Text
|
|
5202
|
-
|
|
5203
|
-
* @summary Create temporary authentication token for Streaming STT
|
|
5204
|
-
*/
|
|
5205
|
-
declare const createTemporaryToken: <TData = AxiosResponse<RealtimeTemporaryTokenResponse>>(createRealtimeTemporaryTokenParams: CreateRealtimeTemporaryTokenParams, options?: AxiosRequestConfig) => Promise<TData>;
|
|
5206
4704
|
|
|
5207
4705
|
/**
|
|
5208
4706
|
* AssemblyAI transcription provider adapter
|
|
@@ -15349,6 +14847,24 @@ type ManageV1FilterVersionParameter = boolean;
|
|
|
15349
14847
|
/**
|
|
15350
14848
|
* ManageV1GroupingParameter type definition
|
|
15351
14849
|
*/
|
|
14850
|
+
/**
|
|
14851
|
+
* ManageV1GroupingParameter type definition
|
|
14852
|
+
*/
|
|
14853
|
+
/**
|
|
14854
|
+
* ManageV1GroupingParameter type definition
|
|
14855
|
+
*/
|
|
14856
|
+
/**
|
|
14857
|
+
* ManageV1GroupingParameter type definition
|
|
14858
|
+
*/
|
|
14859
|
+
/**
|
|
14860
|
+
* ManageV1GroupingParameter type definition
|
|
14861
|
+
*/
|
|
14862
|
+
/**
|
|
14863
|
+
* ManageV1GroupingParameter type definition
|
|
14864
|
+
*/
|
|
14865
|
+
/**
|
|
14866
|
+
* ManageV1GroupingParameter type definition
|
|
14867
|
+
*/
|
|
15352
14868
|
type ManageV1GroupingParameter = typeof ManageV1GroupingParameter[keyof typeof ManageV1GroupingParameter];
|
|
15353
14869
|
declare const ManageV1GroupingParameter: {
|
|
15354
14870
|
readonly accessor: "accessor";
|
|
@@ -15421,6 +14937,24 @@ type ManageV1ModelsListParams = {
|
|
|
15421
14937
|
/**
|
|
15422
14938
|
* ManageV1ProjectKeysStatusParameter type definition
|
|
15423
14939
|
*/
|
|
14940
|
+
/**
|
|
14941
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14942
|
+
*/
|
|
14943
|
+
/**
|
|
14944
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14945
|
+
*/
|
|
14946
|
+
/**
|
|
14947
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14948
|
+
*/
|
|
14949
|
+
/**
|
|
14950
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14951
|
+
*/
|
|
14952
|
+
/**
|
|
14953
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14954
|
+
*/
|
|
14955
|
+
/**
|
|
14956
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14957
|
+
*/
|
|
15424
14958
|
type ManageV1ProjectKeysStatusParameter = typeof ManageV1ProjectKeysStatusParameter[keyof typeof ManageV1ProjectKeysStatusParameter];
|
|
15425
14959
|
declare const ManageV1ProjectKeysStatusParameter: {
|
|
15426
14960
|
readonly active: "active";
|
|
@@ -16306,6 +15840,24 @@ type SelfHostedV1ParamsDistributionScopesParameter = SelfHostedV1ParamsDistribut
|
|
|
16306
15840
|
/**
|
|
16307
15841
|
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
16308
15842
|
*/
|
|
15843
|
+
/**
|
|
15844
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15845
|
+
*/
|
|
15846
|
+
/**
|
|
15847
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15848
|
+
*/
|
|
15849
|
+
/**
|
|
15850
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15851
|
+
*/
|
|
15852
|
+
/**
|
|
15853
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15854
|
+
*/
|
|
15855
|
+
/**
|
|
15856
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15857
|
+
*/
|
|
15858
|
+
/**
|
|
15859
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15860
|
+
*/
|
|
16309
15861
|
type SelfHostedV1ParamsDistributionProviderParameter = typeof SelfHostedV1ParamsDistributionProviderParameter[keyof typeof SelfHostedV1ParamsDistributionProviderParameter];
|
|
16310
15862
|
declare const SelfHostedV1ParamsDistributionProviderParameter: {
|
|
16311
15863
|
readonly quay: "quay";
|
|
@@ -19533,8 +19085,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19533
19085
|
}, "strip", z.ZodTypeAny, {
|
|
19534
19086
|
language: string;
|
|
19535
19087
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19536
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19537
19088
|
domain?: string | undefined;
|
|
19089
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19538
19090
|
output_locale?: string | undefined;
|
|
19539
19091
|
additional_vocab?: {
|
|
19540
19092
|
content: string;
|
|
@@ -19561,8 +19113,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19561
19113
|
}, {
|
|
19562
19114
|
language: string;
|
|
19563
19115
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19564
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19565
19116
|
domain?: string | undefined;
|
|
19117
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19566
19118
|
output_locale?: string | undefined;
|
|
19567
19119
|
additional_vocab?: {
|
|
19568
19120
|
content: string;
|
|
@@ -19720,8 +19272,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19720
19272
|
transcription_config?: {
|
|
19721
19273
|
language: string;
|
|
19722
19274
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19723
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19724
19275
|
domain?: string | undefined;
|
|
19276
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19725
19277
|
output_locale?: string | undefined;
|
|
19726
19278
|
additional_vocab?: {
|
|
19727
19279
|
content: string;
|
|
@@ -19801,8 +19353,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19801
19353
|
transcription_config?: {
|
|
19802
19354
|
language: string;
|
|
19803
19355
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19804
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19805
19356
|
domain?: string | undefined;
|
|
19357
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19806
19358
|
output_locale?: string | undefined;
|
|
19807
19359
|
additional_vocab?: {
|
|
19808
19360
|
content: string;
|
|
@@ -19900,8 +19452,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19900
19452
|
transcription_config?: {
|
|
19901
19453
|
language: string;
|
|
19902
19454
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19903
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19904
19455
|
domain?: string | undefined;
|
|
19456
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19905
19457
|
output_locale?: string | undefined;
|
|
19906
19458
|
additional_vocab?: {
|
|
19907
19459
|
content: string;
|
|
@@ -19994,8 +19546,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19994
19546
|
transcription_config?: {
|
|
19995
19547
|
language: string;
|
|
19996
19548
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19997
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19998
19549
|
domain?: string | undefined;
|
|
19550
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19999
19551
|
output_locale?: string | undefined;
|
|
20000
19552
|
additional_vocab?: {
|
|
20001
19553
|
content: string;
|
|
@@ -20090,8 +19642,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
20090
19642
|
transcription_config?: {
|
|
20091
19643
|
language: string;
|
|
20092
19644
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20093
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20094
19645
|
domain?: string | undefined;
|
|
19646
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20095
19647
|
output_locale?: string | undefined;
|
|
20096
19648
|
additional_vocab?: {
|
|
20097
19649
|
content: string;
|
|
@@ -20186,8 +19738,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
20186
19738
|
transcription_config?: {
|
|
20187
19739
|
language: string;
|
|
20188
19740
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20189
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20190
19741
|
domain?: string | undefined;
|
|
19742
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20191
19743
|
output_locale?: string | undefined;
|
|
20192
19744
|
additional_vocab?: {
|
|
20193
19745
|
content: string;
|
|
@@ -20368,8 +19920,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20368
19920
|
}, "strip", z.ZodTypeAny, {
|
|
20369
19921
|
language: string;
|
|
20370
19922
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20371
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20372
19923
|
domain?: string | undefined;
|
|
19924
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20373
19925
|
output_locale?: string | undefined;
|
|
20374
19926
|
additional_vocab?: {
|
|
20375
19927
|
content: string;
|
|
@@ -20396,8 +19948,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20396
19948
|
}, {
|
|
20397
19949
|
language: string;
|
|
20398
19950
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20399
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20400
19951
|
domain?: string | undefined;
|
|
19952
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20401
19953
|
output_locale?: string | undefined;
|
|
20402
19954
|
additional_vocab?: {
|
|
20403
19955
|
content: string;
|
|
@@ -20555,8 +20107,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20555
20107
|
transcription_config?: {
|
|
20556
20108
|
language: string;
|
|
20557
20109
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20558
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20559
20110
|
domain?: string | undefined;
|
|
20111
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20560
20112
|
output_locale?: string | undefined;
|
|
20561
20113
|
additional_vocab?: {
|
|
20562
20114
|
content: string;
|
|
@@ -20636,8 +20188,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20636
20188
|
transcription_config?: {
|
|
20637
20189
|
language: string;
|
|
20638
20190
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20639
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20640
20191
|
domain?: string | undefined;
|
|
20192
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20641
20193
|
output_locale?: string | undefined;
|
|
20642
20194
|
additional_vocab?: {
|
|
20643
20195
|
content: string;
|
|
@@ -20735,8 +20287,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20735
20287
|
transcription_config?: {
|
|
20736
20288
|
language: string;
|
|
20737
20289
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20738
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20739
20290
|
domain?: string | undefined;
|
|
20291
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20740
20292
|
output_locale?: string | undefined;
|
|
20741
20293
|
additional_vocab?: {
|
|
20742
20294
|
content: string;
|
|
@@ -20829,8 +20381,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20829
20381
|
transcription_config?: {
|
|
20830
20382
|
language: string;
|
|
20831
20383
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20832
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20833
20384
|
domain?: string | undefined;
|
|
20385
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20834
20386
|
output_locale?: string | undefined;
|
|
20835
20387
|
additional_vocab?: {
|
|
20836
20388
|
content: string;
|
|
@@ -20925,8 +20477,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20925
20477
|
transcription_config?: {
|
|
20926
20478
|
language: string;
|
|
20927
20479
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20928
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20929
20480
|
domain?: string | undefined;
|
|
20481
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20930
20482
|
output_locale?: string | undefined;
|
|
20931
20483
|
additional_vocab?: {
|
|
20932
20484
|
content: string;
|
|
@@ -21021,8 +20573,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
21021
20573
|
transcription_config?: {
|
|
21022
20574
|
language: string;
|
|
21023
20575
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21024
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21025
20576
|
domain?: string | undefined;
|
|
20577
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21026
20578
|
output_locale?: string | undefined;
|
|
21027
20579
|
additional_vocab?: {
|
|
21028
20580
|
content: string;
|
|
@@ -21210,8 +20762,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21210
20762
|
}, "strip", z.ZodTypeAny, {
|
|
21211
20763
|
language: string;
|
|
21212
20764
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21213
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21214
20765
|
domain?: string | undefined;
|
|
20766
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21215
20767
|
output_locale?: string | undefined;
|
|
21216
20768
|
additional_vocab?: {
|
|
21217
20769
|
content: string;
|
|
@@ -21238,8 +20790,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21238
20790
|
}, {
|
|
21239
20791
|
language: string;
|
|
21240
20792
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21241
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21242
20793
|
domain?: string | undefined;
|
|
20794
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21243
20795
|
output_locale?: string | undefined;
|
|
21244
20796
|
additional_vocab?: {
|
|
21245
20797
|
content: string;
|
|
@@ -21397,8 +20949,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21397
20949
|
transcription_config?: {
|
|
21398
20950
|
language: string;
|
|
21399
20951
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21400
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21401
20952
|
domain?: string | undefined;
|
|
20953
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21402
20954
|
output_locale?: string | undefined;
|
|
21403
20955
|
additional_vocab?: {
|
|
21404
20956
|
content: string;
|
|
@@ -21478,8 +21030,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21478
21030
|
transcription_config?: {
|
|
21479
21031
|
language: string;
|
|
21480
21032
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21481
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21482
21033
|
domain?: string | undefined;
|
|
21034
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21483
21035
|
output_locale?: string | undefined;
|
|
21484
21036
|
additional_vocab?: {
|
|
21485
21037
|
content: string;
|
|
@@ -21577,8 +21129,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21577
21129
|
transcription_config?: {
|
|
21578
21130
|
language: string;
|
|
21579
21131
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21580
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21581
21132
|
domain?: string | undefined;
|
|
21133
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21582
21134
|
output_locale?: string | undefined;
|
|
21583
21135
|
additional_vocab?: {
|
|
21584
21136
|
content: string;
|
|
@@ -21671,8 +21223,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21671
21223
|
transcription_config?: {
|
|
21672
21224
|
language: string;
|
|
21673
21225
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21674
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21675
21226
|
domain?: string | undefined;
|
|
21227
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21676
21228
|
output_locale?: string | undefined;
|
|
21677
21229
|
additional_vocab?: {
|
|
21678
21230
|
content: string;
|
|
@@ -21767,8 +21319,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21767
21319
|
transcription_config?: {
|
|
21768
21320
|
language: string;
|
|
21769
21321
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21770
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21771
21322
|
domain?: string | undefined;
|
|
21323
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21772
21324
|
output_locale?: string | undefined;
|
|
21773
21325
|
additional_vocab?: {
|
|
21774
21326
|
content: string;
|
|
@@ -21863,8 +21415,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21863
21415
|
transcription_config?: {
|
|
21864
21416
|
language: string;
|
|
21865
21417
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21866
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21867
21418
|
domain?: string | undefined;
|
|
21419
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21868
21420
|
output_locale?: string | undefined;
|
|
21869
21421
|
additional_vocab?: {
|
|
21870
21422
|
content: string;
|
|
@@ -22090,8 +21642,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22090
21642
|
}, "strip", z.ZodTypeAny, {
|
|
22091
21643
|
language: string;
|
|
22092
21644
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22093
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22094
21645
|
domain?: string | undefined;
|
|
21646
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22095
21647
|
output_locale?: string | undefined;
|
|
22096
21648
|
additional_vocab?: {
|
|
22097
21649
|
content: string;
|
|
@@ -22118,8 +21670,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22118
21670
|
}, {
|
|
22119
21671
|
language: string;
|
|
22120
21672
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22121
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22122
21673
|
domain?: string | undefined;
|
|
21674
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22123
21675
|
output_locale?: string | undefined;
|
|
22124
21676
|
additional_vocab?: {
|
|
22125
21677
|
content: string;
|
|
@@ -22306,8 +21858,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22306
21858
|
transcription_config?: {
|
|
22307
21859
|
language: string;
|
|
22308
21860
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22309
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22310
21861
|
domain?: string | undefined;
|
|
21862
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22311
21863
|
output_locale?: string | undefined;
|
|
22312
21864
|
additional_vocab?: {
|
|
22313
21865
|
content: string;
|
|
@@ -22387,8 +21939,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22387
21939
|
transcription_config?: {
|
|
22388
21940
|
language: string;
|
|
22389
21941
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22390
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22391
21942
|
domain?: string | undefined;
|
|
21943
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22392
21944
|
output_locale?: string | undefined;
|
|
22393
21945
|
additional_vocab?: {
|
|
22394
21946
|
content: string;
|
|
@@ -23129,8 +22681,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
23129
22681
|
transcription_config?: {
|
|
23130
22682
|
language: string;
|
|
23131
22683
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
23132
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
23133
22684
|
domain?: string | undefined;
|
|
22685
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
23134
22686
|
output_locale?: string | undefined;
|
|
23135
22687
|
additional_vocab?: {
|
|
23136
22688
|
content: string;
|
|
@@ -23353,8 +22905,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
23353
22905
|
transcription_config?: {
|
|
23354
22906
|
language: string;
|
|
23355
22907
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
23356
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
23357
22908
|
domain?: string | undefined;
|
|
22909
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
23358
22910
|
output_locale?: string | undefined;
|
|
23359
22911
|
additional_vocab?: {
|
|
23360
22912
|
content: string;
|
|
@@ -23622,15 +23174,15 @@ declare const getUsageResponse: z.ZodObject<{
|
|
|
23622
23174
|
duration_hrs: z.ZodNumber;
|
|
23623
23175
|
}, "strip", z.ZodTypeAny, {
|
|
23624
23176
|
type: "transcription" | "alignment";
|
|
23625
|
-
count: number;
|
|
23626
23177
|
mode: "batch";
|
|
23178
|
+
count: number;
|
|
23627
23179
|
duration_hrs: number;
|
|
23628
23180
|
language?: string | undefined;
|
|
23629
23181
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
23630
23182
|
}, {
|
|
23631
23183
|
type: "transcription" | "alignment";
|
|
23632
|
-
count: number;
|
|
23633
23184
|
mode: "batch";
|
|
23185
|
+
count: number;
|
|
23634
23186
|
duration_hrs: number;
|
|
23635
23187
|
language?: string | undefined;
|
|
23636
23188
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
@@ -23644,15 +23196,15 @@ declare const getUsageResponse: z.ZodObject<{
|
|
|
23644
23196
|
duration_hrs: z.ZodNumber;
|
|
23645
23197
|
}, "strip", z.ZodTypeAny, {
|
|
23646
23198
|
type: "transcription" | "alignment";
|
|
23647
|
-
count: number;
|
|
23648
23199
|
mode: "batch";
|
|
23200
|
+
count: number;
|
|
23649
23201
|
duration_hrs: number;
|
|
23650
23202
|
language?: string | undefined;
|
|
23651
23203
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
23652
23204
|
}, {
|
|
23653
23205
|
type: "transcription" | "alignment";
|
|
23654
|
-
count: number;
|
|
23655
23206
|
mode: "batch";
|
|
23207
|
+
count: number;
|
|
23656
23208
|
duration_hrs: number;
|
|
23657
23209
|
language?: string | undefined;
|
|
23658
23210
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
@@ -23660,16 +23212,16 @@ declare const getUsageResponse: z.ZodObject<{
|
|
|
23660
23212
|
}, "strip", z.ZodTypeAny, {
|
|
23661
23213
|
summary: {
|
|
23662
23214
|
type: "transcription" | "alignment";
|
|
23663
|
-
count: number;
|
|
23664
23215
|
mode: "batch";
|
|
23216
|
+
count: number;
|
|
23665
23217
|
duration_hrs: number;
|
|
23666
23218
|
language?: string | undefined;
|
|
23667
23219
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
23668
23220
|
}[];
|
|
23669
23221
|
details: {
|
|
23670
23222
|
type: "transcription" | "alignment";
|
|
23671
|
-
count: number;
|
|
23672
23223
|
mode: "batch";
|
|
23224
|
+
count: number;
|
|
23673
23225
|
duration_hrs: number;
|
|
23674
23226
|
language?: string | undefined;
|
|
23675
23227
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
@@ -23679,16 +23231,16 @@ declare const getUsageResponse: z.ZodObject<{
|
|
|
23679
23231
|
}, {
|
|
23680
23232
|
summary: {
|
|
23681
23233
|
type: "transcription" | "alignment";
|
|
23682
|
-
count: number;
|
|
23683
23234
|
mode: "batch";
|
|
23235
|
+
count: number;
|
|
23684
23236
|
duration_hrs: number;
|
|
23685
23237
|
language?: string | undefined;
|
|
23686
23238
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
23687
23239
|
}[];
|
|
23688
23240
|
details: {
|
|
23689
23241
|
type: "transcription" | "alignment";
|
|
23690
|
-
count: number;
|
|
23691
23242
|
mode: "batch";
|
|
23243
|
+
count: number;
|
|
23692
23244
|
duration_hrs: number;
|
|
23693
23245
|
language?: string | undefined;
|
|
23694
23246
|
operating_point?: "standard" | "enhanced" | undefined;
|
|
@@ -24935,4 +24487,4 @@ declare namespace elevenLabsSpeechToTextAPI_zod {
|
|
|
24935
24487
|
export { elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdHeader as deleteTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdParams as deleteTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdResponse as deleteTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdHeader as getTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdParams as getTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdResponse as getTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_speechToTextBody as speechToTextBody, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive as speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsMax as speechToTextBodyAdditionalFormatsMax, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMaxOne as speechToTextBodyDiarizationThresholdMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMinOne as speechToTextBodyDiarizationThresholdMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizeDefault as speechToTextBodyDiarizeDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyFileFormatDefault as speechToTextBodyFileFormatDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyKeytermsDefault as speechToTextBodyKeytermsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNoVerbatimDefault as speechToTextBodyNoVerbatimDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNumSpeakersMaxOne as speechToTextBodyNumSpeakersMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMaxOne as speechToTextBodySeedMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMinOne as speechToTextBodySeedMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTagAudioEventsDefault as speechToTextBodyTagAudioEventsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMaxOne as speechToTextBodyTemperatureMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMinOne as speechToTextBodyTemperatureMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTimestampsGranularityDefault as speechToTextBodyTimestampsGranularityDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyUseMultiChannelDefault as speechToTextBodyUseMultiChannelDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyWebhookDefault as speechToTextBodyWebhookDefault, elevenLabsSpeechToTextAPI_zod_speechToTextHeader as speechToTextHeader, elevenLabsSpeechToTextAPI_zod_speechToTextQueryEnableLoggingDefault as speechToTextQueryEnableLoggingDefault, elevenLabsSpeechToTextAPI_zod_speechToTextQueryParams as speechToTextQueryParams, elevenLabsSpeechToTextAPI_zod_speechToTextResponse as speechToTextResponse };
|
|
24936
24488
|
}
|
|
24937
24489
|
|
|
24938
|
-
export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult,
|
|
24490
|
+
export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult, TranscriptParams as AssemblyAIOptions, SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAIStreamingOptions, TopicDetectionModelResult as AssemblyAITopicsResult, index$6 as AssemblyAITypes, type AudioChunkAckMessage, AudioInput, AudioResponseFormat, AzureSTTAdapter, Status as AzureStatus, index$5 as AzureTypes, BaseAdapter, CreateTranscriptionRequest, CreateTranscriptionRequestModel, CreateTranscriptionResponseDiarizedJson, CreateTranscriptionResponseVerboseJson, DeepgramAdapter, type DeepgramConfig, ListenV1ResponseMetadata as DeepgramMetadata, ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramStreamingOptions, index$3 as DeepgramTypes, ElevenLabsAdapter, type ElevenLabsConfig, index as ElevenLabsTypes, elevenLabsSpeechToTextAPI_zod as ElevenLabsZodSchemas, type EndRecordingMessage, type EndSessionMessage, type GetProjectRequestV1Response, GladiaAdapter, AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, AudioToLlmListDTO as GladiaAudioToLlmResult, ChapterizationDTO as GladiaChapters, CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, NamedEntityRecognitionDTO as GladiaEntities, ModerationDTO as GladiaModeration, InitTranscriptionRequest as GladiaOptions, SentimentAnalysisDTO as GladiaSentiment, SpeakerReidentificationDTO as GladiaSpeakerReidentification, GladiaStreamingOptions, StreamingRequest as GladiaStreamingRequest, StructuredDataExtractionDTO as GladiaStructuredData, TranslationDTO as GladiaTranslation, index$7 as GladiaTypes, InitTranscriptionRequest, LanguageIdentificationMode, LanguageIdentificationProperties, type ListProjectRequestsV1Response, type ListTranscriptionResponseItemsItem, ListTranscriptsOptions, ListTranscriptsParams, ListenV1EncodingParameter, ListenV1LanguageParameter, ListenV1MediaTranscribeParams, ListenV1ModelParameter, ListenV1Response, ListenV1ResponseResultsChannelsItemAlternativesItem, ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, ListenV1ResponseResultsUtterancesItem, ListenV1VersionParameter, ManageV1FilterEndpointParameter, ManageV1FilterStatusParameter, ManageV1ProjectsRequestsListParams, type NamedEntityRecognitionMessage, streamingTypes as OpenAIStreamingTypes, index$4 as OpenAITypes, OpenAIWhisperAdapter, CreateTranscriptionRequest as OpenAIWhisperOptions, type PaginatedTranscriptions, type PostChapterizationMessage, type PostFinalTranscriptMessage, type PostSummarizationMessage, type PostTranscriptMessage, PreRecordedResponse, ProfanityFilterMode, type ProjectRequestResponse, ProviderCapabilities, type ProviderConfig, PunctuationMode, type SentimentAnalysisMessage, SonioxAdapter, type SonioxConfig, SonioxModelCode, index$1 as SonioxTypes, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type SpeechEndMessage, SpeechModel, type SpeechStartMessage, SpeechmaticsAdapter, type SpeechmaticsConfig, index$2 as SpeechmaticsTypes, speechmaticsASRRESTAPI_zod as SpeechmaticsZodSchemas, type StartRecordingMessage, type StartSessionMessage, type StopRecordingAckMessage, StreamEvent, StreamingCallbacks, StreamingOptions, StreamingRequest, type StreamingResponse, StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedModels, StreamingSupportedSampleRateEnum, StreamingUpdateConfiguration, TranscribeOptions, Transcript, type TranscriptListItem, type TranscriptMessage, TranscriptParams, TranscriptStatus, TranscriptUtterance, TranscriptWord, Transcription$1 as Transcription, type TranscriptionAdapter, TranscriptionControllerListV2Params, TranscriptionControllerListV2StatusItem, TranscriptionDTO, TranscriptionLanguageCodeEnum, TranscriptionProperties, TranscriptionProvider, type TranscriptionsListParams, type TranslationMessage, UnifiedTranscriptResponse, UtteranceDTO, VoiceRouter, type VoiceRouterConfig, type WebHook, type WebHookEvents, WordDTO, createAssemblyAIAdapter, createAzureSTTAdapter, createDeepgramAdapter, createElevenLabsAdapter, createGladiaAdapter, createOpenAIWhisperAdapter, createSonioxAdapter, createSpeechmaticsAdapter, createTranscript, createTranscription, createVoiceRouter, deleteTranscript as deleteTranscriptAPI, getTranscript as getTranscriptAPI, listTranscripts as listTranscriptsAPI, preRecordedControllerDeletePreRecordedJobV2, preRecordedControllerGetAudioV2, preRecordedControllerGetPreRecordedJobV2, preRecordedControllerInitPreRecordedJobV2, streamingControllerDeleteStreamingJobV2, streamingControllerGetAudioV2, streamingControllerInitStreamingSessionV2, transcriptionControllerListV2, transcriptionsCreate, transcriptionsDelete, transcriptionsGet, transcriptionsList, transcriptionsListFiles, webHooksCreate, webHooksDelete, webHooksList };
|