voice-router-dev 0.8.3 → 0.8.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +42 -1
- package/dist/constants.d.mts +7 -8
- package/dist/constants.d.ts +7 -8
- package/dist/constants.js +4 -8
- package/dist/constants.mjs +4 -8
- package/dist/{field-configs-t_lVCkE5.d.mts → field-configs-DLbrsYTk.d.mts} +3299 -2114
- package/dist/{field-configs-t_lVCkE5.d.ts → field-configs-DLbrsYTk.d.ts} +3299 -2114
- package/dist/field-configs.d.mts +1 -1
- package/dist/field-configs.d.ts +1 -1
- package/dist/field-configs.js +2325 -1634
- package/dist/field-configs.mjs +2325 -1634
- package/dist/index.d.mts +237 -702
- package/dist/index.d.ts +237 -702
- package/dist/index.js +2073 -1529
- package/dist/index.mjs +2073 -1528
- package/dist/{speechToTextChunkResponseModel-DvIT4xai.d.mts → speechToTextChunkResponseModel-BcT1LJSZ.d.mts} +1630 -1166
- package/dist/{speechToTextChunkResponseModel-DjL2ncnf.d.ts → speechToTextChunkResponseModel-eq8eLKEA.d.ts} +1630 -1166
- package/dist/webhooks.d.mts +9 -3
- package/dist/webhooks.d.ts +9 -3
- package/package.json +5 -5
package/dist/index.d.ts
CHANGED
|
@@ -1,10 +1,10 @@
|
|
|
1
|
-
import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptOptionalParams, aQ as TranscriptStatus, aR as TranscriptWord, aS as AudioIntelligenceModelStatus, aT as AutoHighlightResult, aU as AutoHighlightsResult, aV as Chapter, aW as ContentSafetyLabel, aX as ContentSafetyLabelResult, aY as ContentSafetyLabelsResult, aZ as ContentSafetyLabelsResultSeverityScoreSummary, a_ as ContentSafetyLabelsResultSummary, a$ as Entity, b0 as EntityType, b1 as ListTranscriptsParams, b2 as PiiPolicy, b3 as RedactPiiAudioQuality, b4 as RedactedAudioNotification, b5 as RedactedAudioResponse, b6 as RedactedAudioStatus, b7 as Sentiment, b8 as SentimentAnalysisResult$1, b9 as SentimentAnalysisResultChannel, ba as SentimentAnalysisResultSpeaker, bb as SeverityScoreSummary, bc as SpeechModel, bd as SubstitutionPolicy, be as SummaryModel, bf as SummaryType, bg as Timestamp, bh as TopicDetectionModelResult, bi as TopicDetectionModelResultSummary, bj as TopicDetectionResult, bk as TopicDetectionResultLabelsItem, bl as Transcript, bm as TranscriptAudioDuration, bn as TranscriptAudioEndAt, bo as TranscriptAudioStartFrom, bp as TranscriptAutoChapters, bq as TranscriptAutoHighlightsResult, br as TranscriptBoostParam, bs as TranscriptBoostParamProperty, bt as TranscriptChapters, bu as TranscriptConfidence, bv as TranscriptContentSafety, bw as TranscriptContentSafetyLabels, bx as TranscriptCustomSpelling, by as TranscriptCustomSpellingProperty, bz as TranscriptCustomTopics, bA as TranscriptDisfluencies, bB as TranscriptEntities, bC as TranscriptEntityDetection, bD as TranscriptFilterProfanity, bE as TranscriptFormatText, bF as TranscriptIabCategories, bG as TranscriptIabCategoriesResult, bH as TranscriptLanguageCode, bI as TranscriptLanguageCodeProperty, bJ as TranscriptLanguageConfidence, bK as TranscriptLanguageConfidenceThreshold, bL as TranscriptLanguageDetection, bM as TranscriptMultichannel, bN as TranscriptOptionalParamsLanguageCode, bO as TranscriptOptionalParamsLanguageCodeOneOf, bP as TranscriptOptionalParamsRedactPiiSub, bQ as TranscriptOptionalParamsSpeakersExpected, bR as TranscriptOptionalParamsSpeechModel, bS as TranscriptOptionalParamsSpeechThreshold, bT as TranscriptOptionalParamsWebhookAuthHeaderName, bU as TranscriptOptionalParamsWebhookAuthHeaderValue, bV as TranscriptPunctuate, bW as TranscriptReadyNotification, bX as TranscriptReadyStatus, bY as TranscriptRedactPiiAudio, bZ as TranscriptRedactPiiAudioQuality, b_ as TranscriptRedactPiiPolicies, b$ as TranscriptSentimentAnalysis, c0 as TranscriptSentimentAnalysisResults, c1 as TranscriptSpeakerLabels, c2 as TranscriptSpeakersExpected, c3 as TranscriptSpeechModel, c4 as TranscriptSpeechThreshold, c5 as TranscriptSpeedBoost, c6 as TranscriptSummary, c7 as TranscriptSummaryModel, c8 as TranscriptSummaryType, c9 as TranscriptText, ca as TranscriptThrottled, cb as TranscriptUtterance, cc as TranscriptUtteranceChannel, cd as TranscriptUtterances, ce as TranscriptWebhookAuthHeaderName, cf as TranscriptWebhookNotification, cg as TranscriptWebhookStatusCode, ch as TranscriptWebhookUrl, ci as TranscriptWordChannel, cj as TranscriptWordSpeaker, ck as TranscriptWords, cl as StreamingUpdateConfiguration, cm as Transcription$1, cn as EntityError, co as Status, cp as EntityReference, cq as DiarizationProperties, cr as DiarizationSpeakersProperties, cs as LanguageIdentificationMode, ct as LanguageIdentificationProperties, cu as LanguageIdentificationPropertiesSpeechModelMapping, cv as ProfanityFilterMode, cw as PunctuationMode, cx as TranscriptionCustomProperties, cy as TranscriptionLinks, cz as TranscriptionProperties, cA as TranscriptTextUsageTokens, cB as TranscriptionSegment, cC as RealtimeSessionCreateRequestGAModel, cD as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, cE as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, cF as AudioResponseFormat, cG as CreateTranscription200One, cH as CreateTranscriptionRequest, cI as CreateTranscriptionRequestModel, cJ as CreateTranscriptionRequestStream, cK as CreateTranscriptionRequestTimestampGranularitiesItem, cL as CreateTranscriptionResponseDiarizedJson, cM as CreateTranscriptionResponseDiarizedJsonTask, cN as CreateTranscriptionResponseDiarizedJsonUsage, cO as CreateTranscriptionResponseJson, cP as CreateTranscriptionResponseJsonLogprobsItem, cQ as CreateTranscriptionResponseJsonUsage, cR as CreateTranscriptionResponseVerboseJson, cS as TranscriptTextUsageDuration, cT as TranscriptTextUsageDurationType, cU as TranscriptTextUsageTokensInputTokenDetails, cV as TranscriptTextUsageTokensType, cW as TranscriptionChunkingStrategy, cX as TranscriptionChunkingStrategyAnyOf, cY as TranscriptionDiarizedSegment, cZ as TranscriptionDiarizedSegmentType, c_ as TranscriptionInclude, c$ as TranscriptionWord, d0 as VadConfig, d1 as VadConfigType, d2 as ListenV1Response, d3 as ManageV1FilterAccessorParameter, d4 as ManageV1FilterDeploymentParameter, d5 as ManageV1LimitParameter, d6 as ManageV1PageParameter, d7 as ManageV1FilterEndpointParameter, d8 as ManageV1FilterMethodParameter, d9 as SharedTopics, da as SharedIntents, db as SharedSentiments, dc as SharedCallbackParameter, dd as SharedCallbackMethodParameter, de as SharedSentimentParameter, df as SharedSummarizeParameter, dg as SharedTagParameter, dh as SharedTopicsParameter, di as SharedCustomTopicParameter, dj as SharedCustomTopicModeParameter, dk as SharedIntentsParameter, dl as SharedCustomIntentParameter, dm as SharedCustomIntentModeParameter, dn as SharedMipOptOutParameter, dp as ListenV1DetectEntitiesParameter, dq as ListenV1DetectLanguageParameter, dr as ListenV1DiarizeParameter, ds as ListenV1DictationParameter, dt as ListenV1EncodingParameter, du as ListenV1FillerWordsParameter, dv as ListenV1KeytermParameter, dw as ListenV1KeywordsParameter, dx as ListenV1LanguageParameter, dy as ListenV1MeasurementsParameter, dz as ListenV1MediaTranscribeParams, dA as ListenV1ModelParameter, dB as ListenV1MultichannelParameter, dC as ListenV1NumeralsParameter, dD as ListenV1ParagraphsParameter, dE as ListenV1ProfanityFilterParameter, dF as ListenV1PunctuateParameter, dG as ListenV1RedactParameter, dH as ListenV1RedactParameterOneOfItem, dI as ListenV1ReplaceParameter, dJ as ListenV1ResponseMetadata, dK as ListenV1ResponseMetadataIntentsInfo, dL as ListenV1ResponseMetadataModelInfo, dM as ListenV1ResponseMetadataSentimentInfo, dN as ListenV1ResponseMetadataSummaryInfo, dO as ListenV1ResponseMetadataTopicsInfo, dP as ListenV1ResponseResults, dQ as ListenV1ResponseResultsChannels, dR as ListenV1ResponseResultsChannelsItem, dS as ListenV1ResponseResultsChannelsItemAlternativesItem, dT as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, dU as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, dV as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, dW as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, dX as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, dY as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, dZ as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, d_ as ListenV1ResponseResultsChannelsItemSearchItem, d$ as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, e0 as ListenV1ResponseResultsSummary, e1 as ListenV1ResponseResultsUtterances, e2 as ListenV1ResponseResultsUtterancesItem, e3 as ListenV1ResponseResultsUtterancesItemWordsItem, e4 as ListenV1SearchParameter, e5 as ListenV1SmartFormatParameter, e6 as ListenV1UttSplitParameter, e7 as ListenV1UtterancesParameter, e8 as ListenV1VersionParameter, e9 as ManageV1EndDateTimeParameter, ea as ManageV1FilterRequestIdParameter, eb as ManageV1FilterStatusParameter, ec as ManageV1ProjectsRequestsListParams, ed as ManageV1StartDateTimeParameter, ee as SharedExtraParameter, ef as SharedIntentsResults, eg as SharedIntentsResultsIntents, eh as SharedIntentsResultsIntentsSegmentsItem, ei as SharedIntentsResultsIntentsSegmentsItemIntentsItem, ej as SharedSentimentsAverage, ek as SharedSentimentsSegmentsItem, el as SharedTopicsResults, em as SharedTopicsResultsTopics, en as SharedTopicsResultsTopicsSegmentsItem, eo as SharedTopicsResultsTopicsSegmentsItemTopicsItem, ep as JobType, eq as AlignmentConfig, er as TranscriptionConfig, es as TrackingData, et as OutputConfig, eu as OperatingPoint, ev as AudioEventItem, ew as AudioEventSummary, ex as AudioEventSummaryItem, ey as AutoChaptersResult, ez as AutoChaptersResultError, eA as AutoChaptersResultErrorType, eB as Chapter$1, eC as JobInfo, eD as LanguageIdentificationResult, eE as LanguageIdentificationResultAlternative, eF as LanguageIdentificationResultError, eG as LanguageIdentificationResultItem, eH as LanguagePackInfo, eI as LanguagePackInfoWritingDirection, eJ as OutputConfigSrtOverrides, eK as RecognitionAlternative, eL as RecognitionDisplay, eM as RecognitionDisplayDirection, eN as RecognitionMetadata, eO as RecognitionResult, eP as RecognitionResultAttachesTo, eQ as RecognitionResultType, eR as RetrieveTranscriptResponse, eS as RetrieveTranscriptResponseAudioEventSummary, eT as RetrieveTranscriptResponseAudioEventSummaryChannels, eU as RetrieveTranscriptResponseTranslations, eV as SentimentAnalysisError, eW as SentimentAnalysisErrorType, eX as SentimentAnalysisResult$2, eY as SentimentAnalysisResultSentimentAnalysis, eZ as SentimentChannelSummary, e_ as SentimentSegment, e$ as SentimentSpeakerSummary, f0 as SentimentSummary, f1 as SentimentSummaryDetail, f2 as SpokenFormRecognitionResult, f3 as SpokenFormRecognitionResultType, f4 as SummarizationError, f5 as SummarizationErrorType, f6 as SummarizationResult, f7 as TopicDetectionError, f8 as TopicDetectionErrorType, f9 as TopicDetectionResult$1, fa as TopicDetectionSegment, fb as TopicDetectionSegmentTopic, fc as TopicDetectionSummary, fd as TopicDetectionSummaryOverall, fe as TrackingDataDetails, ff as TranscriptionConfigAdditionalVocabItem, fg as TranscriptionConfigDiarization, fh as TranscriptionConfigMaxDelayMode, fi as TranscriptionConfigPunctuationOverrides, fj as TranscriptionConfigSpeakerDiarizationConfig, fk as TranscriptionConfigTranscriptFilteringConfig, fl as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fm as TranslationError, fn as TranslationErrorType, fo as TranslationSentence, fp as WrittenFormRecognitionResult, fq as WrittenFormRecognitionResultType, fr as SpeechToTextChunkResponseModel, fs as AdditionalFormatResponseModel, ft as AdditionalFormats, fu as BodySpeechToTextV1SpeechToTextPost, fv as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, fw as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, fx as BodySpeechToTextV1SpeechToTextPostEntityDetection, fy as BodySpeechToTextV1SpeechToTextPostFile, fz as BodySpeechToTextV1SpeechToTextPostFileFormat, fA as BodySpeechToTextV1SpeechToTextPostLanguageCode, fB as BodySpeechToTextV1SpeechToTextPostModelId, fC as BodySpeechToTextV1SpeechToTextPostNumSpeakers, fD as BodySpeechToTextV1SpeechToTextPostSeed, fE as BodySpeechToTextV1SpeechToTextPostTemperature, fF as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, fG as BodySpeechToTextV1SpeechToTextPostWebhookId, fH as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, fI as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, fJ as DetectedEntity, fK as DocxExportOptions, fL as DocxExportOptionsFormat, fM as DocxExportOptionsMaxSegmentChars, fN as DocxExportOptionsMaxSegmentDurationS, fO as DocxExportOptionsSegmentOnSilenceLongerThanS, fP as ExportOptions, fQ as HtmlExportOptions, fR as HtmlExportOptionsFormat, fS as HtmlExportOptionsMaxSegmentChars, fT as HtmlExportOptionsMaxSegmentDurationS, fU as HtmlExportOptionsSegmentOnSilenceLongerThanS, fV as PdfExportOptions, fW as PdfExportOptionsFormat, fX as PdfExportOptionsMaxSegmentChars, fY as PdfExportOptionsMaxSegmentDurationS, fZ as PdfExportOptionsSegmentOnSilenceLongerThanS, f_ as SegmentedJsonExportOptions, f$ as SegmentedJsonExportOptionsFormat, g0 as SegmentedJsonExportOptionsMaxSegmentChars, g1 as SegmentedJsonExportOptionsMaxSegmentDurationS, g2 as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, g3 as SpeechToTextCharacterResponseModel, g4 as SpeechToTextCharacterResponseModelEnd, g5 as SpeechToTextCharacterResponseModelStart, g6 as SpeechToTextChunkResponseModelAdditionalFormats, g7 as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, g8 as SpeechToTextChunkResponseModelChannelIndex, g9 as SpeechToTextChunkResponseModelEntities, ga as SpeechToTextChunkResponseModelTranscriptionId, gb as SpeechToTextWordResponseModel, gc as SpeechToTextWordResponseModelCharacters, gd as SpeechToTextWordResponseModelEnd, ge as SpeechToTextWordResponseModelSpeakerId, gf as SpeechToTextWordResponseModelStart, gg as SpeechToTextWordResponseModelType, gh as SrtExportOptions, gi as SrtExportOptionsFormat, gj as SrtExportOptionsMaxCharactersPerLine, gk as SrtExportOptionsMaxSegmentChars, gl as SrtExportOptionsMaxSegmentDurationS, gm as SrtExportOptionsSegmentOnSilenceLongerThanS, gn as TxtExportOptions, go as TxtExportOptionsFormat, gp as TxtExportOptionsMaxCharactersPerLine, gq as TxtExportOptionsMaxSegmentChars, gr as TxtExportOptionsMaxSegmentDurationS, gs as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-DjL2ncnf.js';
|
|
2
|
-
export {
|
|
1
|
+
import { T as TranscribeOptions, U as UnifiedTranscriptResponse, S as StreamingOptions, a as StreamingCallbacks, b as StreamingSession, c as StreamEvent, L as ListTranscriptsOptions, G as GladiaStreamingOptions, D as DeepgramStreamingOptions, A as AssemblyAIStreamingOptions, F as FileResponse, d as StreamingSupportedEncodingEnum, e as StreamingSupportedBitDepthEnum, f as StreamingSupportedSampleRateEnum, g as StreamingSupportedModels, h as LanguageConfig, P as PreProcessingConfig, R as RealtimeProcessingConfig, i as PostProcessingConfig, M as MessagesConfig, C as CallbackConfig, j as TranscriptionMetadataDTO, k as TranscriptionDTO, l as TranslationDTO, m as SummarizationDTO, N as NamedEntityRecognitionDTO, n as SentimentAnalysisDTO, o as ChapterizationDTO, p as PreRecordedResponse, q as UtteranceDTO, r as TranscriptionLanguageCodeEnum, s as TranslationLanguageCodeEnum, W as WordDTO, t as StreamingSupportedRegions, u as AddonErrorDTO, v as AudioToLlmDTO, w as AudioToLlmDTOError, x as AudioToLlmDTOResults, y as AudioToLlmListConfigDTO, z as AudioToLlmListDTO, B as AudioToLlmListDTOError, E as AudioToLlmResultDTO, H as CallbackConfigDto, I as CallbackMethodEnum, J as CallbackTranscriptionErrorPayload, K as CallbackTranscriptionErrorPayloadCustomMetadata, O as CallbackTranscriptionErrorPayloadEvent, Q as CallbackTranscriptionSuccessPayload, V as CallbackTranscriptionSuccessPayloadCustomMetadata, X as CallbackTranscriptionSuccessPayloadEvent, Y as ChapterizationDTOError, Z as ChapterizationDTOResults, _ as CodeSwitchingConfigDTO, $ as CustomSpellingConfigDTO, a0 as CustomSpellingConfigDTOSpellingDictionary, a1 as CustomVocabularyConfigDTO, a2 as CustomVocabularyConfigDTOVocabularyItem, a3 as CustomVocabularyEntryDTO, a4 as DiarizationConfigDTO, a5 as DiarizationDTO, a6 as DiarizationDTOError, a7 as DisplayModeDTO, a8 as DisplayModeDTOError, a9 as ErrorDTO, aa as InitTranscriptionRequest, ab as InitTranscriptionRequestCustomMetadata, ac as ModerationDTO, ad as ModerationDTOError, ae as NamedEntityRecognitionDTOError, af as NamesConsistencyDTO, ag as NamesConsistencyDTOError, ah as PreRecordedRequestParamsResponse, ai as PreRecordedResponseCustomMetadata, aj as PreRecordedResponseFile, ak as PreRecordedResponseKind, al as PreRecordedResponsePostSessionMetadata, am as PreRecordedResponseRequestParams, an as PreRecordedResponseResult, ao as PreRecordedResponseStatus, ap as SentencesDTO, aq as SentencesDTOError, ar as SentimentAnalysisDTOError, as as SpeakerReidentificationDTO, at as SpeakerReidentificationDTOError, au as StreamingRequest, av as StreamingRequestCustomMetadata, aw as StructuredDataExtractionConfigDTO, ax as StructuredDataExtractionDTO, ay as StructuredDataExtractionDTOError, az as SubtitleDTO, aA as SubtitlesConfigDTO, aB as SubtitlesFormatEnum, aC as SubtitlesStyleEnum, aD as SummarizationConfigDTO, aE as SummarizationDTOError, aF as SummaryTypesEnum, aG as TranscriptionControllerListV2KindItem, aH as TranscriptionControllerListV2Params, aI as TranscriptionControllerListV2StatusItem, aJ as TranscriptionResultDTO, aK as TranslationConfigDTO, aL as TranslationDTOError, aM as TranslationModelEnum, aN as TranslationResultDTO, aO as TranslationResultDTOError, aP as TranscriptStatus, aQ as TranscriptWord, aR as AudioIntelligenceModelStatus, aS as AutoHighlightResult, aT as AutoHighlightsResult, aU as Chapter, aV as ContentSafetyLabel, aW as ContentSafetyLabelResult, aX as ContentSafetyLabelsResult, aY as ContentSafetyLabelsResultSeverityScoreSummary, aZ as ContentSafetyLabelsResultSummary, a_ as CustomFormattingRequestBody, a$ as CustomFormattingRequestBodyCustomFormatting, b0 as CustomFormattingResponse, b1 as CustomFormattingResponseCustomFormatting, b2 as CustomFormattingResponseCustomFormattingMapping, b3 as Entity, b4 as EntityType, b5 as ListTranscriptsParams, b6 as PiiPolicy, b7 as RedactPiiAudioQuality, b8 as Sentiment, b9 as SentimentAnalysisResult$1, ba as SentimentAnalysisResultChannel, bb as SentimentAnalysisResultSpeaker, bc as SeverityScoreSummary, bd as SpeakerIdentificationRequestBody, be as SpeakerIdentificationRequestBodySpeakerIdentification, bf as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, bg as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, bh as SpeakerIdentificationResponse, bi as SpeakerIdentificationResponseSpeakerIdentification, bj as SpeakerIdentificationResponseSpeakerIdentificationMapping, bk as SpeechModel, bl as SubstitutionPolicy, bm as SummaryModel, bn as SummaryType, bo as Timestamp, bp as TopicDetectionModelResult, bq as TopicDetectionModelResultSummary, br as TopicDetectionResult, bs as TopicDetectionResultLabelsItems, bt as Transcript, bu as TranscriptAudioDuration, bv as TranscriptAudioEndAt, bw as TranscriptAudioStartFrom, bx as TranscriptAutoChapters, by as TranscriptAutoHighlightsResult, bz as TranscriptChapters, bA as TranscriptConfidence, bB as TranscriptContentSafety, bC as TranscriptContentSafetyLabels, bD as TranscriptCustomSpelling, bE as TranscriptCustomSpellingProperty, bF as TranscriptCustomTopics, bG as TranscriptDisfluencies, bH as TranscriptDomain, bI as TranscriptEntities, bJ as TranscriptEntityDetection, bK as TranscriptFilterProfanity, bL as TranscriptFormatText, bM as TranscriptIabCategories, bN as TranscriptIabCategoriesResult, bO as TranscriptLanguageCode, bP as TranscriptLanguageCodes, bQ as TranscriptLanguageConfidence, bR as TranscriptLanguageConfidenceThreshold, bS as TranscriptLanguageDetection, bT as TranscriptLanguageDetectionOptions, bU as TranscriptMultichannel, bV as TranscriptOptionalParamsLanguageDetectionOptions, bW as TranscriptOptionalParamsRedactPiiAudioOptions, bX as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, bY as TranscriptOptionalParamsRemoveAudioTags, bZ as TranscriptOptionalParamsSpeakerOptions, b_ as TranscriptOptionalParamsSpeechUnderstanding, b$ as TranscriptOptionalParamsSpeechUnderstandingRequest, c0 as TranscriptParams, c1 as TranscriptParamsDomain, c2 as TranscriptParamsLanguageCode, c3 as TranscriptParamsLanguageCodes, c4 as TranscriptParamsRedactPiiSub, c5 as TranscriptParamsRemoveAudioTags, c6 as TranscriptParamsSpeakersExpected, c7 as TranscriptParamsSpeechModel, c8 as TranscriptParamsSpeechThreshold, c9 as TranscriptParamsWebhookAuthHeaderName, ca as TranscriptParamsWebhookAuthHeaderValue, cb as TranscriptPunctuate, cc as TranscriptRedactPiiAudio, cd as TranscriptRedactPiiAudioOptions, ce as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, cf as TranscriptRedactPiiAudioQuality, cg as TranscriptRedactPiiPolicies, ch as TranscriptRemoveAudioTags, ci as TranscriptRemoveAudioTagsProperty, cj as TranscriptSentimentAnalysis, ck as TranscriptSentimentAnalysisResults, cl as TranscriptSpeakerLabels, cm as TranscriptSpeakersExpected, cn as TranscriptSpeechModel, co as TranscriptSpeechModels, cp as TranscriptSpeechThreshold, cq as TranscriptSpeechUnderstanding, cr as TranscriptSpeechUnderstandingRequest, cs as TranscriptSpeechUnderstandingResponse, ct as TranscriptSpeedBoost, cu as TranscriptSummary, cv as TranscriptSummaryModel, cw as TranscriptSummaryType, cx as TranscriptTemperature, cy as TranscriptText, cz as TranscriptThrottled, cA as TranscriptTranslatedTexts, cB as TranscriptUtterance, cC as TranscriptUtteranceChannel, cD as TranscriptUtteranceTranslatedTexts, cE as TranscriptUtterances, cF as TranscriptWebhookAuthHeaderName, cG as TranscriptWebhookStatusCode, cH as TranscriptWebhookUrl, cI as TranscriptWordChannel, cJ as TranscriptWordSpeaker, cK as TranscriptWords, cL as TranslationRequestBody, cM as TranslationRequestBodyTranslation, cN as TranslationResponse, cO as TranslationResponseTranslation, cP as StreamingUpdateConfiguration, cQ as Transcription$1, cR as EntityError, cS as Status, cT as EntityReference, cU as DiarizationProperties, cV as DiarizationSpeakersProperties, cW as LanguageIdentificationMode, cX as LanguageIdentificationProperties, cY as LanguageIdentificationPropertiesSpeechModelMapping, cZ as ProfanityFilterMode, c_ as PunctuationMode, c$ as TranscriptionCustomProperties, d0 as TranscriptionLinks, d1 as TranscriptionProperties, d2 as TranscriptTextUsageTokens, d3 as TranscriptionSegment, d4 as RealtimeSessionCreateRequestGAModel, d5 as RealtimeTranscriptionSessionCreateRequestTurnDetectionType, d6 as RealtimeTranscriptionSessionCreateRequestInputAudioFormat, d7 as AudioResponseFormat, d8 as CreateTranscription200One, d9 as CreateTranscriptionRequest, da as CreateTranscriptionRequestModel, db as CreateTranscriptionRequestStream, dc as CreateTranscriptionRequestTimestampGranularitiesItem, dd as CreateTranscriptionResponseDiarizedJson, de as CreateTranscriptionResponseDiarizedJsonTask, df as CreateTranscriptionResponseDiarizedJsonUsage, dg as CreateTranscriptionResponseJson, dh as CreateTranscriptionResponseJsonLogprobsItem, di as CreateTranscriptionResponseJsonUsage, dj as CreateTranscriptionResponseVerboseJson, dk as TranscriptTextUsageDuration, dl as TranscriptTextUsageDurationType, dm as TranscriptTextUsageTokensInputTokenDetails, dn as TranscriptTextUsageTokensType, dp as TranscriptionChunkingStrategy, dq as TranscriptionChunkingStrategyAnyOf, dr as TranscriptionDiarizedSegment, ds as TranscriptionDiarizedSegmentType, dt as TranscriptionInclude, du as TranscriptionWord, dv as VadConfig, dw as VadConfigType, dx as ListenV1Response, dy as ManageV1FilterAccessorParameter, dz as ManageV1FilterDeploymentParameter, dA as ManageV1LimitParameter, dB as ManageV1PageParameter, dC as ManageV1FilterEndpointParameter, dD as ManageV1FilterMethodParameter, dE as SharedTopics, dF as SharedIntents, dG as SharedSentiments, dH as SharedCallbackParameter, dI as SharedCallbackMethodParameter, dJ as SharedSentimentParameter, dK as SharedSummarizeParameter, dL as SharedTagParameter, dM as SharedTopicsParameter, dN as SharedCustomTopicParameter, dO as SharedCustomTopicModeParameter, dP as SharedIntentsParameter, dQ as SharedCustomIntentParameter, dR as SharedCustomIntentModeParameter, dS as SharedMipOptOutParameter, dT as ListenV1DetectEntitiesParameter, dU as ListenV1DetectLanguageParameter, dV as ListenV1DiarizeParameter, dW as ListenV1DictationParameter, dX as ListenV1EncodingParameter, dY as ListenV1FillerWordsParameter, dZ as ListenV1KeytermParameter, d_ as ListenV1KeywordsParameter, d$ as ListenV1LanguageParameter, e0 as ListenV1MeasurementsParameter, e1 as ListenV1MediaTranscribeParams, e2 as ListenV1ModelParameter, e3 as ListenV1MultichannelParameter, e4 as ListenV1NumeralsParameter, e5 as ListenV1ParagraphsParameter, e6 as ListenV1ProfanityFilterParameter, e7 as ListenV1PunctuateParameter, e8 as ListenV1RedactParameter, e9 as ListenV1RedactParameterOneOfItem, ea as ListenV1ReplaceParameter, eb as ListenV1ResponseMetadata, ec as ListenV1ResponseMetadataIntentsInfo, ed as ListenV1ResponseMetadataModelInfo, ee as ListenV1ResponseMetadataSentimentInfo, ef as ListenV1ResponseMetadataSummaryInfo, eg as ListenV1ResponseMetadataTopicsInfo, eh as ListenV1ResponseResults, ei as ListenV1ResponseResultsChannels, ej as ListenV1ResponseResultsChannelsItem, ek as ListenV1ResponseResultsChannelsItemAlternativesItem, el as ListenV1ResponseResultsChannelsItemAlternativesItemEntitiesItem, em as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphs, en as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItem, eo as ListenV1ResponseResultsChannelsItemAlternativesItemParagraphsParagraphsItemSentencesItem, ep as ListenV1ResponseResultsChannelsItemAlternativesItemSummariesItem, eq as ListenV1ResponseResultsChannelsItemAlternativesItemTopicsItem, er as ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, es as ListenV1ResponseResultsChannelsItemSearchItem, et as ListenV1ResponseResultsChannelsItemSearchItemHitsItem, eu as ListenV1ResponseResultsSummary, ev as ListenV1ResponseResultsUtterances, ew as ListenV1ResponseResultsUtterancesItem, ex as ListenV1ResponseResultsUtterancesItemWordsItem, ey as ListenV1SearchParameter, ez as ListenV1SmartFormatParameter, eA as ListenV1UttSplitParameter, eB as ListenV1UtterancesParameter, eC as ListenV1VersionParameter, eD as ManageV1EndDateTimeParameter, eE as ManageV1FilterRequestIdParameter, eF as ManageV1FilterStatusParameter, eG as ManageV1ProjectsRequestsListParams, eH as ManageV1StartDateTimeParameter, eI as SharedExtraParameter, eJ as SharedIntentsResults, eK as SharedIntentsResultsIntents, eL as SharedIntentsResultsIntentsSegmentsItem, eM as SharedIntentsResultsIntentsSegmentsItemIntentsItem, eN as SharedSentimentsAverage, eO as SharedSentimentsSegmentsItem, eP as SharedTopicsResults, eQ as SharedTopicsResultsTopics, eR as SharedTopicsResultsTopicsSegmentsItem, eS as SharedTopicsResultsTopicsSegmentsItemTopicsItem, eT as JobType, eU as AlignmentConfig, eV as TranscriptionConfig, eW as TrackingData, eX as OutputConfig, eY as OperatingPoint, eZ as AudioEventItem, e_ as AudioEventSummary, e$ as AudioEventSummaryItem, f0 as AutoChaptersResult, f1 as AutoChaptersResultError, f2 as AutoChaptersResultErrorType, f3 as Chapter$1, f4 as JobInfo, f5 as LanguageIdentificationResult, f6 as LanguageIdentificationResultAlternative, f7 as LanguageIdentificationResultError, f8 as LanguageIdentificationResultItem, f9 as LanguagePackInfo, fa as LanguagePackInfoWritingDirection, fb as OutputConfigSrtOverrides, fc as RecognitionAlternative, fd as RecognitionDisplay, fe as RecognitionDisplayDirection, ff as RecognitionMetadata, fg as RecognitionResult, fh as RecognitionResultAttachesTo, fi as RecognitionResultType, fj as RetrieveTranscriptResponse, fk as RetrieveTranscriptResponseAudioEventSummary, fl as RetrieveTranscriptResponseAudioEventSummaryChannels, fm as RetrieveTranscriptResponseTranslations, fn as SentimentAnalysisError, fo as SentimentAnalysisErrorType, fp as SentimentAnalysisResult$2, fq as SentimentAnalysisResultSentimentAnalysis, fr as SentimentChannelSummary, fs as SentimentSegment, ft as SentimentSpeakerSummary, fu as SentimentSummary, fv as SentimentSummaryDetail, fw as SpokenFormRecognitionResult, fx as SpokenFormRecognitionResultType, fy as SummarizationError, fz as SummarizationErrorType, fA as SummarizationResult, fB as TopicDetectionError, fC as TopicDetectionErrorType, fD as TopicDetectionResult$1, fE as TopicDetectionSegment, fF as TopicDetectionSegmentTopic, fG as TopicDetectionSummary, fH as TopicDetectionSummaryOverall, fI as TrackingDataDetails, fJ as TranscriptionConfigAdditionalVocabItem, fK as TranscriptionConfigDiarization, fL as TranscriptionConfigMaxDelayMode, fM as TranscriptionConfigPunctuationOverrides, fN as TranscriptionConfigSpeakerDiarizationConfig, fO as TranscriptionConfigTranscriptFilteringConfig, fP as TranscriptionConfigTranscriptFilteringConfigReplacementsItem, fQ as TranslationError, fR as TranslationErrorType, fS as TranslationSentence, fT as WrittenFormRecognitionResult, fU as WrittenFormRecognitionResultType, fV as SpeechToTextChunkResponseModel, fW as AdditionalFormatResponseModel, fX as AdditionalFormats, fY as BodySpeechToTextV1SpeechToTextPost, fZ as BodySpeechToTextV1SpeechToTextPostCloudStorageUrl, f_ as BodySpeechToTextV1SpeechToTextPostDiarizationThreshold, f$ as BodySpeechToTextV1SpeechToTextPostEntityDetection, g0 as BodySpeechToTextV1SpeechToTextPostFile, g1 as BodySpeechToTextV1SpeechToTextPostFileFormat, g2 as BodySpeechToTextV1SpeechToTextPostLanguageCode, g3 as BodySpeechToTextV1SpeechToTextPostModelId, g4 as BodySpeechToTextV1SpeechToTextPostNumSpeakers, g5 as BodySpeechToTextV1SpeechToTextPostSeed, g6 as BodySpeechToTextV1SpeechToTextPostTemperature, g7 as BodySpeechToTextV1SpeechToTextPostTimestampsGranularity, g8 as BodySpeechToTextV1SpeechToTextPostWebhookId, g9 as BodySpeechToTextV1SpeechToTextPostWebhookMetadata, ga as BodySpeechToTextV1SpeechToTextPostWebhookMetadataAnyOf, gb as DetectedEntity, gc as DocxExportOptions, gd as DocxExportOptionsFormat, ge as DocxExportOptionsMaxSegmentChars, gf as DocxExportOptionsMaxSegmentDurationS, gg as DocxExportOptionsSegmentOnSilenceLongerThanS, gh as ExportOptions, gi as HtmlExportOptions, gj as HtmlExportOptionsFormat, gk as HtmlExportOptionsMaxSegmentChars, gl as HtmlExportOptionsMaxSegmentDurationS, gm as HtmlExportOptionsSegmentOnSilenceLongerThanS, gn as PdfExportOptions, go as PdfExportOptionsFormat, gp as PdfExportOptionsMaxSegmentChars, gq as PdfExportOptionsMaxSegmentDurationS, gr as PdfExportOptionsSegmentOnSilenceLongerThanS, gs as SegmentedJsonExportOptions, gt as SegmentedJsonExportOptionsFormat, gu as SegmentedJsonExportOptionsMaxSegmentChars, gv as SegmentedJsonExportOptionsMaxSegmentDurationS, gw as SegmentedJsonExportOptionsSegmentOnSilenceLongerThanS, gx as SpeechToTextCharacterResponseModel, gy as SpeechToTextCharacterResponseModelEnd, gz as SpeechToTextCharacterResponseModelStart, gA as SpeechToTextChunkResponseModelAdditionalFormats, gB as SpeechToTextChunkResponseModelAdditionalFormatsAnyOfItem, gC as SpeechToTextChunkResponseModelChannelIndex, gD as SpeechToTextChunkResponseModelEntities, gE as SpeechToTextChunkResponseModelTranscriptionId, gF as SpeechToTextWordResponseModel, gG as SpeechToTextWordResponseModelCharacters, gH as SpeechToTextWordResponseModelEnd, gI as SpeechToTextWordResponseModelSpeakerId, gJ as SpeechToTextWordResponseModelStart, gK as SpeechToTextWordResponseModelType, gL as SrtExportOptions, gM as SrtExportOptionsFormat, gN as SrtExportOptionsMaxCharactersPerLine, gO as SrtExportOptionsMaxSegmentChars, gP as SrtExportOptionsMaxSegmentDurationS, gQ as SrtExportOptionsSegmentOnSilenceLongerThanS, gR as TxtExportOptions, gS as TxtExportOptionsFormat, gT as TxtExportOptionsMaxCharactersPerLine, gU as TxtExportOptionsMaxSegmentChars, gV as TxtExportOptionsMaxSegmentDurationS, gW as TxtExportOptionsSegmentOnSilenceLongerThanS } from './speechToTextChunkResponseModel-eq8eLKEA.js';
|
|
2
|
+
export { g_ as AssemblyAIExtendedData, hp as AssemblyAIUpdateConfiguration, hl as AudioAckEvent, hn as AudioChunk, h4 as BatchOnlyProvider, hw as BeginEvent, hk as ChapterizationEvent, h0 as DeepgramExtendedData, h1 as ElevenLabsExtendedData, hs as ElevenLabsStreamingOptions, hi as EntityEvent, hz as ErrorEvent, g$ as GladiaExtendedData, hm as LifecycleEvent, hc as ListTranscriptsResponse, hq as OpenAIStreamingOptions, h2 as ProviderExtendedDataMap, hd as ProviderRawResponseMap, ht as ProviderStreamingOptions, ho as RawWebSocketMessage, hh as SentimentEvent, h5 as SessionStatus, hr as SonioxStreamingOptions, h6 as Speaker, hf as SpeechEvent, gX as SpeechmaticsOperatingPoint, he as StreamEventType, hA as StreamingEventMessage, hC as StreamingForceEndpoint, hu as StreamingOptionsForProvider, h3 as StreamingProvider, hB as StreamingWord, hj as SummarizationEvent, hy as TerminationEvent, hv as TranscribeStreamParams, hb as TranscriptData, ha as TranscriptMetadata, gZ as TranscriptionLanguage, gY as TranscriptionModel, h9 as TranscriptionStatus, hg as TranslationEvent, hx as TurnEvent, h8 as Utterance, h7 as Word } from './speechToTextChunkResponseModel-eq8eLKEA.js';
|
|
3
3
|
import { T as TranscriptionProvider, P as ProviderCapabilities, A as AudioInput } from './provider-metadata-_gUWlRXS.js';
|
|
4
4
|
export { k as AllLanguageCodes, o as AllProviders, a as AssemblyAICapabilities, i as AssemblyAILanguageCodes, b as AzureCapabilities, B as BatchOnlyProviderType, q as BatchOnlyProviders, C as CapabilityKeys, f as CapabilityLabels, D as DeepgramCapabilities, j as DeepgramLanguageCodes, E as ElevenLabsCapabilities, G as GladiaCapabilities, h as GladiaLanguageCodes, g as LanguageCode, L as LanguageLabels, O as OpenAICapabilities, d as ProviderCapabilitiesMap, l as ProviderDisplayNames, n as ProviderDocs, m as ProviderWebsites, c as SonioxCapabilities, S as SpeechmaticsCapabilities, e as StreamingProviderType, p as StreamingProviders } from './provider-metadata-_gUWlRXS.js';
|
|
5
5
|
import { AssemblyAIRegionType, DeepgramRegionType, SpeechmaticsRegionType, SonioxModelCode, SonioxRegionType, ElevenLabsModelCode, ElevenLabsRegionType } from './constants.js';
|
|
6
6
|
export { AssemblyAIEncoding, AssemblyAIEncodingType, AssemblyAILanguage, AssemblyAILanguageType, AssemblyAISampleRate, AssemblyAISampleRateType, AssemblyAISpeechModel, AssemblyAISpeechModelType, AssemblyAIStatus, AssemblyAITranscriptionModel, AssemblyAITranscriptionModelType, AzureLocale, AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, AzureLocaleType, AzureLocales, DeepgramCallbackMethod, DeepgramEncoding, DeepgramIntentMode, DeepgramLanguage, DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, DeepgramModelCode as DeepgramModelType, DeepgramRedact, DeepgramRedactType, DeepgramRegion, DeepgramSampleRate, DeepgramStatus, DeepgramTTSContainer, DeepgramTTSEncoding, DeepgramTTSModel, DeepgramTTSSampleRate, DeepgramTopicMode, DeepgramTopicModeType, ElevenLabsLanguageCode, ElevenLabsLanguageCodes, ElevenLabsLanguageLabels, ElevenLabsLanguages, GladiaBitDepth, GladiaEncoding, GladiaLanguage, GladiaModel, GladiaRegion, GladiaSampleRate, GladiaStatus, GladiaTranslationLanguage, OpenAILanguageCodes, OpenAIModel, OpenAIRealtimeAudioFormat, OpenAIRealtimeModel, OpenAIRealtimeTranscriptionModel, OpenAIRealtimeTurnDetection, OpenAIResponseFormat, SonioxAsyncModel, SonioxAsyncModelCode, SonioxLanguage, SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, SonioxLanguageType, SonioxLanguages, SonioxModel, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, SonioxRealtimeModelCode, SonioxRegion, SpeechmaticsLanguage, SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsRegion } from './constants.js';
|
|
7
|
-
export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-
|
|
7
|
+
export { B as AssemblyAIListFilterFieldName, L as AssemblyAIListFilterSchema, E as AssemblyAIStreamingConfig, x as AssemblyAIStreamingFieldName, J as AssemblyAIStreamingSchema, H as AssemblyAIStreamingUpdateConfig, y as AssemblyAIStreamingUpdateFieldName, K as AssemblyAIStreamingUpdateSchema, C as AssemblyAITranscriptionConfig, A as AssemblyAITranscriptionFieldName, I as AssemblyAITranscriptionSchema, a as AssemblyAIZodSchemas, R as AzureListFilterFieldName, U as AzureListFilterSchema, S as AzureTranscriptionConfig, Q as AzureTranscriptionFieldName, T as AzureTranscriptionSchema, q as DeepgramListFilterFieldName, w as DeepgramListFilterSchema, p as DeepgramStreamingFieldName, t as DeepgramStreamingOnlyConfig, v as DeepgramStreamingOnlySchema, r as DeepgramTranscriptionConfig, D as DeepgramTranscriptionFieldName, u as DeepgramTranscriptionSchema, d as DeepgramZodSchemas, Z as FieldConfig, aK as FieldConfigProvider, c as FieldOverrides, F as FieldType, i as GladiaListFilterFieldName, n as GladiaListFilterSchema, k as GladiaStreamingConfig, h as GladiaStreamingFieldName, m as GladiaStreamingSchema, j as GladiaTranscriptionConfig, G as GladiaTranscriptionFieldName, l as GladiaTranscriptionSchema, g as GladiaZodSchemas, M as OpenAITranscriptionConfig, O as OpenAITranscriptionFieldName, N as OpenAITranscriptionSchema, o as OpenAIZodSchemas, P as ProviderFieldConfigs, b as SonioxApiZodSchemas, a8 as SonioxListFilterFieldName, af as SonioxListFilterSchema, aa as SonioxStreamingConfig, a6 as SonioxStreamingFieldName, ad as SonioxStreamingSchema, s as SonioxStreamingTypes, ab as SonioxStreamingUpdateConfig, a7 as SonioxStreamingUpdateFieldName, ae as SonioxStreamingUpdateSchema, s as SonioxStreamingZodSchemas, a9 as SonioxTranscriptionConfig, a5 as SonioxTranscriptionFieldName, ac as SonioxTranscriptionSchema, Y as SpeechmaticsListFilterFieldName, a4 as SpeechmaticsListFilterSchema, $ as SpeechmaticsStreamingConfig, W as SpeechmaticsStreamingFieldName, a2 as SpeechmaticsStreamingSchema, a0 as SpeechmaticsStreamingUpdateConfig, X as SpeechmaticsStreamingUpdateFieldName, a3 as SpeechmaticsStreamingUpdateSchema, _ as SpeechmaticsTranscriptionConfig, V as SpeechmaticsTranscriptionFieldName, a1 as SpeechmaticsTranscriptionSchema, ah as StreamingFieldName, ag as TranscriptionFieldName, Z as ZodFieldConfig, e as excludeFields, f as filterFields, aM as getAllFieldConfigs, au as getAssemblyAIFieldConfigs, ar as getAssemblyAIListFilterFields, as as getAssemblyAIStreamingFields, at as getAssemblyAIStreamingUpdateFields, aq as getAssemblyAITranscriptionFields, az as getAzureFieldConfigs, ay as getAzureListFilterFields, ax as getAzureTranscriptionFields, ap as getDeepgramFieldConfigs, an as getDeepgramListFilterFields, ao as getDeepgramStreamingFields, am as getDeepgramTranscriptionFields, al as getGladiaFieldConfigs, aj as getGladiaListFilterFields, ak as getGladiaStreamingFields, ai as getGladiaTranscriptionFields, aw as getOpenAIFieldConfigs, av as getOpenAITranscriptionFields, aL as getProviderFieldConfigs, aJ as getSonioxFieldConfigs, aG as getSonioxListFilterFields, aH as getSonioxStreamingFields, aI as getSonioxStreamingUpdateFields, aF as getSonioxTranscriptionFields, aE as getSpeechmaticsFieldConfigs, aB as getSpeechmaticsListFilterFields, aC as getSpeechmaticsStreamingFields, aD as getSpeechmaticsStreamingUpdateFields, aA as getSpeechmaticsTranscriptionFields, z as zodToFieldConfigs } from './field-configs-DLbrsYTk.js';
|
|
8
8
|
import { z } from 'zod';
|
|
9
9
|
import { AxiosResponse, AxiosRequestConfig } from 'axios';
|
|
10
10
|
|
|
@@ -4115,559 +4115,101 @@ declare function createGladiaAdapter(config: ProviderConfig): GladiaAdapter;
|
|
|
4115
4115
|
* Generated by orval v7.9.0 🍺
|
|
4116
4116
|
* Do not edit manually.
|
|
4117
4117
|
* AssemblyAI API
|
|
4118
|
-
* AssemblyAI API
|
|
4119
|
-
* OpenAPI spec version: 1.
|
|
4120
|
-
*/
|
|
4121
|
-
type TranscriptParamsAllOf = {
|
|
4122
|
-
/** The URL of the audio or video file to transcribe. */
|
|
4123
|
-
audio_url: string;
|
|
4124
|
-
};
|
|
4125
|
-
|
|
4126
|
-
/**
|
|
4127
|
-
* Generated by orval v7.9.0 🍺
|
|
4128
|
-
* Do not edit manually.
|
|
4129
|
-
* AssemblyAI API
|
|
4130
|
-
* AssemblyAI API
|
|
4131
|
-
* OpenAPI spec version: 1.3.4
|
|
4132
|
-
*/
|
|
4133
|
-
|
|
4134
|
-
/**
|
|
4135
|
-
* The parameters for creating a transcript
|
|
4136
|
-
*/
|
|
4137
|
-
type TranscriptParams = TranscriptParamsAllOf & TranscriptOptionalParams;
|
|
4138
|
-
|
|
4139
|
-
/**
|
|
4140
|
-
* Generated by orval v7.9.0 🍺
|
|
4141
|
-
* Do not edit manually.
|
|
4142
|
-
* AssemblyAI API
|
|
4143
|
-
* AssemblyAI API
|
|
4144
|
-
* OpenAPI spec version: 1.3.4
|
|
4145
|
-
*/
|
|
4146
|
-
/**
|
|
4147
|
-
* The date and time the transcript was completed
|
|
4148
|
-
* @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
|
|
4149
|
-
*/
|
|
4150
|
-
type TranscriptListItemCompleted = string | null;
|
|
4151
|
-
|
|
4152
|
-
/**
|
|
4153
|
-
* Generated by orval v7.9.0 🍺
|
|
4154
|
-
* Do not edit manually.
|
|
4155
|
-
* AssemblyAI API
|
|
4156
|
-
* AssemblyAI API
|
|
4157
|
-
* OpenAPI spec version: 1.3.4
|
|
4158
|
-
*/
|
|
4159
|
-
/**
|
|
4160
|
-
* Error message of why the transcript failed
|
|
4161
|
-
*/
|
|
4162
|
-
type TranscriptListItemError = string | null;
|
|
4163
|
-
|
|
4164
|
-
/**
|
|
4165
|
-
* Generated by orval v7.9.0 🍺
|
|
4166
|
-
* Do not edit manually.
|
|
4167
|
-
* AssemblyAI API
|
|
4168
|
-
* AssemblyAI API
|
|
4169
|
-
* OpenAPI spec version: 1.3.4
|
|
4170
|
-
*/
|
|
4171
|
-
|
|
4172
|
-
interface TranscriptListItem {
|
|
4173
|
-
/** The unique identifier for the transcript */
|
|
4174
|
-
id: string;
|
|
4175
|
-
/** The URL to retrieve the transcript */
|
|
4176
|
-
resource_url: string;
|
|
4177
|
-
/** The status of the transcript */
|
|
4178
|
-
status: TranscriptStatus;
|
|
4179
|
-
/**
|
|
4180
|
-
* The date and time the transcript was created
|
|
4181
|
-
* @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
|
|
4182
|
-
*/
|
|
4183
|
-
created: string;
|
|
4184
|
-
/**
|
|
4185
|
-
* The date and time the transcript was completed
|
|
4186
|
-
* @pattern ^(?:(\d{4}-\d{2}-\d{2})T(\d{2}:\d{2}:\d{2}(?:\.\d+)?))$
|
|
4187
|
-
*/
|
|
4188
|
-
completed: TranscriptListItemCompleted;
|
|
4189
|
-
/** The URL to the audio file */
|
|
4190
|
-
audio_url: string;
|
|
4191
|
-
/** Error message of why the transcript failed */
|
|
4192
|
-
error: TranscriptListItemError;
|
|
4193
|
-
}
|
|
4194
|
-
|
|
4195
|
-
/**
|
|
4196
|
-
* Bad request
|
|
4197
|
-
*/
|
|
4198
|
-
type BadRequestResponse = Error;
|
|
4199
|
-
|
|
4200
|
-
/**
|
|
4201
|
-
* Cannot access uploaded file
|
|
4202
|
-
*/
|
|
4203
|
-
type CannotAccessUploadedFileResponse = Error;
|
|
4204
|
-
|
|
4205
|
-
/**
|
|
4206
|
-
* Generated by orval v7.9.0 🍺
|
|
4207
|
-
* Do not edit manually.
|
|
4208
|
-
* AssemblyAI API
|
|
4209
|
-
* AssemblyAI API
|
|
4210
|
-
* OpenAPI spec version: 1.3.4
|
|
4211
|
-
*/
|
|
4212
|
-
interface CreateRealtimeTemporaryTokenParams {
|
|
4213
|
-
/**
|
|
4214
|
-
* The amount of time until the token expires in seconds
|
|
4215
|
-
* @minimum 60
|
|
4216
|
-
*/
|
|
4217
|
-
expires_in: number;
|
|
4218
|
-
}
|
|
4219
|
-
|
|
4220
|
-
/**
|
|
4221
|
-
* Generated by orval v7.9.0 🍺
|
|
4222
|
-
* Do not edit manually.
|
|
4223
|
-
* AssemblyAI API
|
|
4224
|
-
* AssemblyAI API
|
|
4225
|
-
* OpenAPI spec version: 1.3.4
|
|
4226
|
-
*/
|
|
4227
|
-
interface Error$2 {
|
|
4228
|
-
/** Error message */
|
|
4229
|
-
error: string;
|
|
4230
|
-
status?: "error";
|
|
4231
|
-
[key: string]: unknown;
|
|
4232
|
-
}
|
|
4233
|
-
|
|
4234
|
-
/**
|
|
4235
|
-
* Generated by orval v7.9.0 🍺
|
|
4236
|
-
* Do not edit manually.
|
|
4237
|
-
* AssemblyAI API
|
|
4238
|
-
* AssemblyAI API
|
|
4239
|
-
* OpenAPI spec version: 1.3.4
|
|
4240
|
-
*/
|
|
4241
|
-
/**
|
|
4242
|
-
* Gateway timeout
|
|
4243
|
-
*/
|
|
4244
|
-
type GatewayTimeoutResponse = void;
|
|
4245
|
-
|
|
4246
|
-
/**
|
|
4247
|
-
* Generated by orval v7.9.0 🍺
|
|
4248
|
-
* Do not edit manually.
|
|
4249
|
-
* AssemblyAI API
|
|
4250
|
-
* AssemblyAI API
|
|
4251
|
-
* OpenAPI spec version: 1.3.4
|
|
4252
|
-
*/
|
|
4253
|
-
type GetSubtitlesParams = {
|
|
4254
|
-
/**
|
|
4255
|
-
* The maximum number of characters per caption
|
|
4256
|
-
*/
|
|
4257
|
-
chars_per_caption?: number;
|
|
4258
|
-
};
|
|
4259
|
-
|
|
4260
|
-
/**
|
|
4261
|
-
* An error occurred while processing the request
|
|
4262
|
-
*/
|
|
4263
|
-
type InternalServerErrorResponse = Error;
|
|
4264
|
-
|
|
4265
|
-
/**
|
|
4266
|
-
* Generated by orval v7.9.0 🍺
|
|
4267
|
-
* Do not edit manually.
|
|
4268
|
-
* AssemblyAI API
|
|
4269
|
-
* AssemblyAI API
|
|
4270
|
-
* OpenAPI spec version: 1.3.4
|
|
4271
|
-
*/
|
|
4272
|
-
type LemurBaseParamsContextOneOf = {
|
|
4273
|
-
[key: string]: unknown;
|
|
4274
|
-
};
|
|
4275
|
-
|
|
4276
|
-
/**
|
|
4277
|
-
* Generated by orval v7.9.0 🍺
|
|
4278
|
-
* Do not edit manually.
|
|
4279
|
-
* AssemblyAI API
|
|
4280
|
-
* AssemblyAI API
|
|
4281
|
-
* OpenAPI spec version: 1.3.4
|
|
4282
|
-
*/
|
|
4283
|
-
|
|
4284
|
-
/**
|
|
4285
|
-
* Context to provide the model. This can be a string or a free-form JSON value.
|
|
4286
|
-
*/
|
|
4287
|
-
type LemurBaseParamsContext = string | LemurBaseParamsContextOneOf;
|
|
4288
|
-
|
|
4289
|
-
/**
|
|
4290
|
-
* Generated by orval v7.9.0 🍺
|
|
4291
|
-
* Do not edit manually.
|
|
4292
|
-
* AssemblyAI API
|
|
4293
|
-
* AssemblyAI API
|
|
4294
|
-
* OpenAPI spec version: 1.3.4
|
|
4295
|
-
*/
|
|
4296
|
-
/**
|
|
4297
|
-
* The model that is used for the final prompt after compression is performed.
|
|
4298
|
-
|
|
4299
|
-
*/
|
|
4300
|
-
type LemurModel = (typeof LemurModel)[keyof typeof LemurModel];
|
|
4301
|
-
declare const LemurModel: {
|
|
4302
|
-
readonly "anthropic/claude-3-5-sonnet": "anthropic/claude-3-5-sonnet";
|
|
4303
|
-
readonly "anthropic/claude-3-opus": "anthropic/claude-3-opus";
|
|
4304
|
-
readonly "anthropic/claude-3-haiku": "anthropic/claude-3-haiku";
|
|
4305
|
-
};
|
|
4306
|
-
|
|
4307
|
-
/**
|
|
4308
|
-
* Generated by orval v7.9.0 🍺
|
|
4309
|
-
* Do not edit manually.
|
|
4310
|
-
* AssemblyAI API
|
|
4311
|
-
* AssemblyAI API
|
|
4312
|
-
* OpenAPI spec version: 1.3.4
|
|
4313
|
-
*/
|
|
4314
|
-
|
|
4315
|
-
/**
|
|
4316
|
-
* The model that is used for the final prompt after compression is performed.
|
|
4317
|
-
|
|
4318
|
-
*/
|
|
4319
|
-
type LemurBaseParamsFinalModel = LemurModel | string;
|
|
4320
|
-
|
|
4321
|
-
/**
|
|
4322
|
-
* Generated by orval v7.9.0 🍺
|
|
4323
|
-
* Do not edit manually.
|
|
4324
|
-
* AssemblyAI API
|
|
4325
|
-
* AssemblyAI API
|
|
4326
|
-
* OpenAPI spec version: 1.3.4
|
|
4327
|
-
*/
|
|
4328
|
-
|
|
4329
|
-
interface LemurBaseParams {
|
|
4330
|
-
/** A list of completed transcripts with text. Up to a maximum of 100 hours of audio.
|
|
4331
|
-
Use either transcript_ids or input_text as input into LeMUR.
|
|
4332
|
-
*/
|
|
4333
|
-
transcript_ids?: string[];
|
|
4334
|
-
/** Custom formatted transcript data. Maximum size is the context limit of the selected model.
|
|
4335
|
-
Use either transcript_ids or input_text as input into LeMUR.
|
|
4336
|
-
*/
|
|
4337
|
-
input_text?: string;
|
|
4338
|
-
/** Context to provide the model. This can be a string or a free-form JSON value. */
|
|
4339
|
-
context?: LemurBaseParamsContext;
|
|
4340
|
-
/** The model that is used for the final prompt after compression is performed.
|
|
4341
|
-
*/
|
|
4342
|
-
final_model: LemurBaseParamsFinalModel;
|
|
4343
|
-
/** Max output size in tokens. */
|
|
4344
|
-
max_output_size?: number;
|
|
4345
|
-
/**
|
|
4346
|
-
* The temperature to use for the model.
|
|
4347
|
-
Higher values result in answers that are more creative, lower values are more conservative.
|
|
4348
|
-
Can be any value between 0.0 and 1.0 inclusive.
|
|
4349
|
-
|
|
4350
|
-
* @minimum 0
|
|
4351
|
-
* @maximum 1
|
|
4352
|
-
*/
|
|
4353
|
-
temperature?: number;
|
|
4354
|
-
}
|
|
4355
|
-
|
|
4356
|
-
/**
|
|
4357
|
-
* Generated by orval v7.9.0 🍺
|
|
4358
|
-
* Do not edit manually.
|
|
4359
|
-
* AssemblyAI API
|
|
4360
|
-
* AssemblyAI API
|
|
4361
|
-
* OpenAPI spec version: 1.3.4
|
|
4362
|
-
*/
|
|
4363
|
-
type LemurActionItemsParamsAllOf = {
|
|
4364
|
-
/** How you want the action items to be returned. This can be any text.
|
|
4365
|
-
Defaults to "Bullet Points".
|
|
4366
|
-
*/
|
|
4367
|
-
answer_format?: string;
|
|
4368
|
-
};
|
|
4369
|
-
|
|
4370
|
-
/**
|
|
4371
|
-
* Generated by orval v7.9.0 🍺
|
|
4372
|
-
* Do not edit manually.
|
|
4373
|
-
* AssemblyAI API
|
|
4374
|
-
* AssemblyAI API
|
|
4375
|
-
* OpenAPI spec version: 1.3.4
|
|
4376
|
-
*/
|
|
4377
|
-
|
|
4378
|
-
type LemurActionItemsParams = LemurBaseParams & LemurActionItemsParamsAllOf;
|
|
4379
|
-
|
|
4380
|
-
/**
|
|
4381
|
-
* Generated by orval v7.9.0 🍺
|
|
4382
|
-
* Do not edit manually.
|
|
4383
|
-
* AssemblyAI API
|
|
4384
|
-
* AssemblyAI API
|
|
4385
|
-
* OpenAPI spec version: 1.3.4
|
|
4386
|
-
*/
|
|
4387
|
-
type LemurStringResponseAllOf = {
|
|
4388
|
-
/** The response generated by LeMUR. */
|
|
4389
|
-
response: string;
|
|
4390
|
-
};
|
|
4391
|
-
|
|
4392
|
-
/**
|
|
4393
|
-
* Generated by orval v7.9.0 🍺
|
|
4394
|
-
* Do not edit manually.
|
|
4395
|
-
* AssemblyAI API
|
|
4396
|
-
* AssemblyAI API
|
|
4397
|
-
* OpenAPI spec version: 1.3.4
|
|
4398
|
-
*/
|
|
4399
|
-
/**
|
|
4400
|
-
* The usage numbers for the LeMUR request
|
|
4401
|
-
*/
|
|
4402
|
-
interface LemurUsage {
|
|
4403
|
-
/**
|
|
4404
|
-
* The number of input tokens used by the model
|
|
4405
|
-
* @minimum 0
|
|
4406
|
-
*/
|
|
4407
|
-
input_tokens: number;
|
|
4408
|
-
/**
|
|
4409
|
-
* The number of output tokens generated by the model
|
|
4410
|
-
* @minimum 0
|
|
4411
|
-
*/
|
|
4412
|
-
output_tokens: number;
|
|
4413
|
-
}
|
|
4414
|
-
|
|
4415
|
-
/**
|
|
4416
|
-
* Generated by orval v7.9.0 🍺
|
|
4417
|
-
* Do not edit manually.
|
|
4418
|
-
* AssemblyAI API
|
|
4419
|
-
* AssemblyAI API
|
|
4420
|
-
* OpenAPI spec version: 1.3.4
|
|
4421
|
-
*/
|
|
4422
|
-
|
|
4423
|
-
interface LemurBaseResponse {
|
|
4424
|
-
/** The ID of the LeMUR request */
|
|
4425
|
-
request_id: string;
|
|
4426
|
-
/** The usage numbers for the LeMUR request */
|
|
4427
|
-
usage: LemurUsage;
|
|
4428
|
-
}
|
|
4429
|
-
|
|
4430
|
-
/**
|
|
4431
|
-
* Generated by orval v7.9.0 🍺
|
|
4432
|
-
* Do not edit manually.
|
|
4433
|
-
* AssemblyAI API
|
|
4434
|
-
* AssemblyAI API
|
|
4435
|
-
* OpenAPI spec version: 1.3.4
|
|
4436
|
-
*/
|
|
4437
|
-
|
|
4438
|
-
type LemurStringResponse = LemurStringResponseAllOf & LemurBaseResponse;
|
|
4439
|
-
|
|
4440
|
-
/**
|
|
4441
|
-
* Generated by orval v7.9.0 🍺
|
|
4442
|
-
* Do not edit manually.
|
|
4443
|
-
* AssemblyAI API
|
|
4444
|
-
* AssemblyAI API
|
|
4445
|
-
* OpenAPI spec version: 1.3.4
|
|
4446
|
-
*/
|
|
4447
|
-
|
|
4448
|
-
type LemurActionItemsResponse = LemurStringResponse;
|
|
4449
|
-
|
|
4450
|
-
/**
|
|
4451
|
-
* Generated by orval v7.9.0 🍺
|
|
4452
|
-
* Do not edit manually.
|
|
4453
|
-
* AssemblyAI API
|
|
4454
|
-
* AssemblyAI API
|
|
4455
|
-
* OpenAPI spec version: 1.3.4
|
|
4456
|
-
*/
|
|
4457
|
-
type LemurQuestionContextOneOf = {
|
|
4458
|
-
[key: string]: unknown;
|
|
4459
|
-
};
|
|
4460
|
-
|
|
4461
|
-
/**
|
|
4462
|
-
* Generated by orval v7.9.0 🍺
|
|
4463
|
-
* Do not edit manually.
|
|
4464
|
-
* AssemblyAI API
|
|
4465
|
-
* AssemblyAI API
|
|
4466
|
-
* OpenAPI spec version: 1.3.4
|
|
4467
|
-
*/
|
|
4468
|
-
|
|
4469
|
-
/**
|
|
4470
|
-
* Any context about the transcripts you wish to provide. This can be a string or any object.
|
|
4471
|
-
*/
|
|
4472
|
-
type LemurQuestionContext = string | LemurQuestionContextOneOf;
|
|
4473
|
-
|
|
4474
|
-
/**
|
|
4475
|
-
* Generated by orval v7.9.0 🍺
|
|
4476
|
-
* Do not edit manually.
|
|
4477
|
-
* AssemblyAI API
|
|
4478
|
-
* AssemblyAI API
|
|
4479
|
-
* OpenAPI spec version: 1.3.4
|
|
4480
|
-
*/
|
|
4481
|
-
|
|
4482
|
-
interface LemurQuestion {
|
|
4483
|
-
/** The question you wish to ask. For more complex questions use default model. */
|
|
4484
|
-
question: string;
|
|
4485
|
-
/** Any context about the transcripts you wish to provide. This can be a string or any object. */
|
|
4486
|
-
context?: LemurQuestionContext;
|
|
4487
|
-
/** How you want the answer to be returned. This can be any text. Can't be used with answer_options. Examples: "short sentence", "bullet points"
|
|
4488
|
-
*/
|
|
4489
|
-
answer_format?: string;
|
|
4490
|
-
/** What discrete options to return. Useful for precise responses. Can't be used with answer_format. Example: ["Yes", "No"]
|
|
4491
|
-
*/
|
|
4492
|
-
answer_options?: string[];
|
|
4493
|
-
}
|
|
4494
|
-
|
|
4495
|
-
/**
|
|
4496
|
-
* Generated by orval v7.9.0 🍺
|
|
4497
|
-
* Do not edit manually.
|
|
4498
|
-
* AssemblyAI API
|
|
4499
|
-
* AssemblyAI API
|
|
4500
|
-
* OpenAPI spec version: 1.3.4
|
|
4501
|
-
*/
|
|
4502
|
-
/**
|
|
4503
|
-
* An answer generated by LeMUR and its question
|
|
4504
|
-
*/
|
|
4505
|
-
interface LemurQuestionAnswer {
|
|
4506
|
-
/** The question for LeMUR to answer */
|
|
4507
|
-
question: string;
|
|
4508
|
-
/** The answer generated by LeMUR */
|
|
4509
|
-
answer: string;
|
|
4510
|
-
}
|
|
4511
|
-
|
|
4512
|
-
/**
|
|
4513
|
-
* Generated by orval v7.9.0 🍺
|
|
4514
|
-
* Do not edit manually.
|
|
4515
|
-
* AssemblyAI API
|
|
4516
|
-
* AssemblyAI API
|
|
4517
|
-
* OpenAPI spec version: 1.3.4
|
|
4518
|
-
*/
|
|
4519
|
-
|
|
4520
|
-
type LemurQuestionAnswerParamsAllOf = {
|
|
4521
|
-
/** A list of questions to ask */
|
|
4522
|
-
questions: LemurQuestion[];
|
|
4523
|
-
};
|
|
4524
|
-
|
|
4525
|
-
/**
|
|
4526
|
-
* Generated by orval v7.9.0 🍺
|
|
4527
|
-
* Do not edit manually.
|
|
4528
|
-
* AssemblyAI API
|
|
4529
|
-
* AssemblyAI API
|
|
4530
|
-
* OpenAPI spec version: 1.3.4
|
|
4531
|
-
*/
|
|
4532
|
-
|
|
4533
|
-
type LemurQuestionAnswerParams = LemurBaseParams & LemurQuestionAnswerParamsAllOf;
|
|
4534
|
-
|
|
4535
|
-
/**
|
|
4536
|
-
* Generated by orval v7.9.0 🍺
|
|
4537
|
-
* Do not edit manually.
|
|
4538
|
-
* AssemblyAI API
|
|
4539
|
-
* AssemblyAI API
|
|
4540
|
-
* OpenAPI spec version: 1.3.4
|
|
4118
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4119
|
+
* OpenAPI spec version: 1.0.0
|
|
4541
4120
|
*/
|
|
4542
|
-
|
|
4543
|
-
type LemurQuestionAnswerResponseAllOf = {
|
|
4544
|
-
/** The answers generated by LeMUR and their questions */
|
|
4545
|
-
response: LemurQuestionAnswer[];
|
|
4546
|
-
};
|
|
4547
|
-
|
|
4548
4121
|
/**
|
|
4549
|
-
*
|
|
4550
|
-
* Do not edit manually.
|
|
4551
|
-
* AssemblyAI API
|
|
4552
|
-
* AssemblyAI API
|
|
4553
|
-
* OpenAPI spec version: 1.3.4
|
|
4122
|
+
* Error message of why the transcript failed
|
|
4554
4123
|
*/
|
|
4555
|
-
|
|
4556
|
-
type LemurQuestionAnswerResponse = LemurBaseResponse & LemurQuestionAnswerResponseAllOf;
|
|
4124
|
+
type TranscriptListItemError = string | null;
|
|
4557
4125
|
|
|
4558
4126
|
/**
|
|
4559
4127
|
* Generated by orval v7.9.0 🍺
|
|
4560
4128
|
* Do not edit manually.
|
|
4561
4129
|
* AssemblyAI API
|
|
4562
|
-
* AssemblyAI API
|
|
4563
|
-
* OpenAPI spec version: 1.
|
|
4130
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4131
|
+
* OpenAPI spec version: 1.0.0
|
|
4564
4132
|
*/
|
|
4565
4133
|
|
|
4566
|
-
|
|
4134
|
+
interface TranscriptListItem {
|
|
4135
|
+
/** The unique identifier for the transcript */
|
|
4136
|
+
id: string;
|
|
4137
|
+
/** The URL to retrieve the transcript */
|
|
4138
|
+
resource_url: string;
|
|
4139
|
+
/** The status of the transcript */
|
|
4140
|
+
status: TranscriptStatus;
|
|
4141
|
+
/** The date and time the transcript was created */
|
|
4142
|
+
created: string;
|
|
4143
|
+
/** The date and time the transcript was completed */
|
|
4144
|
+
completed?: string;
|
|
4145
|
+
/** The URL to the audio file */
|
|
4146
|
+
audio_url: string;
|
|
4147
|
+
/** Error message of why the transcript failed */
|
|
4148
|
+
error: TranscriptListItemError;
|
|
4149
|
+
}
|
|
4567
4150
|
|
|
4568
4151
|
/**
|
|
4569
4152
|
* Generated by orval v7.9.0 🍺
|
|
4570
4153
|
* Do not edit manually.
|
|
4571
4154
|
* AssemblyAI API
|
|
4572
|
-
* AssemblyAI API
|
|
4573
|
-
* OpenAPI spec version: 1.
|
|
4155
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4156
|
+
* OpenAPI spec version: 1.0.0
|
|
4574
4157
|
*/
|
|
4575
|
-
type LemurSummaryParamsAllOf = {
|
|
4576
|
-
/** How you want the summary to be returned. This can be any text. Examples: "TLDR", "bullet points"
|
|
4577
|
-
*/
|
|
4578
|
-
answer_format?: string;
|
|
4579
|
-
};
|
|
4580
|
-
|
|
4581
4158
|
/**
|
|
4582
|
-
*
|
|
4583
|
-
* Do not edit manually.
|
|
4584
|
-
* AssemblyAI API
|
|
4585
|
-
* AssemblyAI API
|
|
4586
|
-
* OpenAPI spec version: 1.3.4
|
|
4159
|
+
* Additional error details if available
|
|
4587
4160
|
*/
|
|
4588
|
-
|
|
4589
|
-
|
|
4161
|
+
interface ErrorDetails {
|
|
4162
|
+
[key: string]: unknown;
|
|
4163
|
+
}
|
|
4590
4164
|
|
|
4591
4165
|
/**
|
|
4592
4166
|
* Generated by orval v7.9.0 🍺
|
|
4593
4167
|
* Do not edit manually.
|
|
4594
4168
|
* AssemblyAI API
|
|
4595
|
-
* AssemblyAI API
|
|
4596
|
-
* OpenAPI spec version: 1.
|
|
4169
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4170
|
+
* OpenAPI spec version: 1.0.0
|
|
4597
4171
|
*/
|
|
4598
4172
|
|
|
4599
|
-
|
|
4173
|
+
interface Error$2 {
|
|
4174
|
+
/** Error message describing what went wrong */
|
|
4175
|
+
error: string;
|
|
4176
|
+
/** Error code for programmatic handling */
|
|
4177
|
+
code?: string;
|
|
4178
|
+
/** Additional error details if available */
|
|
4179
|
+
details?: ErrorDetails;
|
|
4180
|
+
}
|
|
4600
4181
|
|
|
4601
4182
|
/**
|
|
4602
4183
|
* Generated by orval v7.9.0 🍺
|
|
4603
4184
|
* Do not edit manually.
|
|
4604
4185
|
* AssemblyAI API
|
|
4605
|
-
* AssemblyAI API
|
|
4606
|
-
* OpenAPI spec version: 1.
|
|
4186
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4187
|
+
* OpenAPI spec version: 1.0.0
|
|
4607
4188
|
*/
|
|
4608
|
-
type
|
|
4609
|
-
|
|
4610
|
-
prompt: string;
|
|
4189
|
+
type GetSubtitles200 = {
|
|
4190
|
+
[key: string]: unknown;
|
|
4611
4191
|
};
|
|
4612
4192
|
|
|
4613
4193
|
/**
|
|
4614
4194
|
* Generated by orval v7.9.0 🍺
|
|
4615
4195
|
* Do not edit manually.
|
|
4616
4196
|
* AssemblyAI API
|
|
4617
|
-
* AssemblyAI API
|
|
4618
|
-
* OpenAPI spec version: 1.
|
|
4619
|
-
*/
|
|
4620
|
-
|
|
4621
|
-
type LemurTaskParams = LemurTaskParamsAllOf & LemurBaseParams;
|
|
4622
|
-
|
|
4623
|
-
/**
|
|
4624
|
-
* Generated by orval v7.9.0 🍺
|
|
4625
|
-
* Do not edit manually.
|
|
4626
|
-
* AssemblyAI API
|
|
4627
|
-
* AssemblyAI API
|
|
4628
|
-
* OpenAPI spec version: 1.3.4
|
|
4629
|
-
*/
|
|
4630
|
-
|
|
4631
|
-
type LemurTaskResponse = LemurStringResponse;
|
|
4632
|
-
|
|
4633
|
-
/**
|
|
4634
|
-
* Generated by orval v7.9.0 🍺
|
|
4635
|
-
* Do not edit manually.
|
|
4636
|
-
* AssemblyAI API
|
|
4637
|
-
* AssemblyAI API
|
|
4638
|
-
* OpenAPI spec version: 1.3.4
|
|
4197
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4198
|
+
* OpenAPI spec version: 1.0.0
|
|
4639
4199
|
*/
|
|
4640
|
-
|
|
4641
|
-
interface ListTranscriptParams {
|
|
4200
|
+
type GetSubtitlesParams = {
|
|
4642
4201
|
/**
|
|
4643
|
-
*
|
|
4644
|
-
* @minimum 1
|
|
4645
|
-
* @maximum 200
|
|
4202
|
+
* The maximum number of characters per caption
|
|
4646
4203
|
*/
|
|
4647
|
-
|
|
4648
|
-
|
|
4649
|
-
status?: TranscriptStatus;
|
|
4650
|
-
/** Only get transcripts created on this date */
|
|
4651
|
-
created_on?: string;
|
|
4652
|
-
/** Get transcripts that were created before this transcript ID */
|
|
4653
|
-
before_id?: string;
|
|
4654
|
-
/** Get transcripts that were created after this transcript ID */
|
|
4655
|
-
after_id?: string;
|
|
4656
|
-
/** Only get throttled transcripts, overrides the status filter */
|
|
4657
|
-
throttled_only?: boolean;
|
|
4658
|
-
}
|
|
4659
|
-
|
|
4660
|
-
/**
|
|
4661
|
-
* Not found
|
|
4662
|
-
*/
|
|
4663
|
-
type NotFoundResponse = Error;
|
|
4204
|
+
chars_per_caption?: number;
|
|
4205
|
+
};
|
|
4664
4206
|
|
|
4665
4207
|
/**
|
|
4666
4208
|
* Generated by orval v7.9.0 🍺
|
|
4667
4209
|
* Do not edit manually.
|
|
4668
4210
|
* AssemblyAI API
|
|
4669
|
-
* AssemblyAI API
|
|
4670
|
-
* OpenAPI spec version: 1.
|
|
4211
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4212
|
+
* OpenAPI spec version: 1.0.0
|
|
4671
4213
|
*/
|
|
4672
4214
|
/**
|
|
4673
4215
|
* The URL to the next page of transcripts. The previous URL always points to a page with older transcripts.
|
|
@@ -4678,8 +4220,8 @@ type PageDetailsPrevUrl = string | null;
|
|
|
4678
4220
|
* Generated by orval v7.9.0 🍺
|
|
4679
4221
|
* Do not edit manually.
|
|
4680
4222
|
* AssemblyAI API
|
|
4681
|
-
* AssemblyAI API
|
|
4682
|
-
* OpenAPI spec version: 1.
|
|
4223
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4224
|
+
* OpenAPI spec version: 1.0.0
|
|
4683
4225
|
*/
|
|
4684
4226
|
/**
|
|
4685
4227
|
* The URL to the next page of transcripts. The next URL always points to a page with newer transcripts.
|
|
@@ -4690,8 +4232,8 @@ type PageDetailsNextUrl = string | null;
|
|
|
4690
4232
|
* Generated by orval v7.9.0 🍺
|
|
4691
4233
|
* Do not edit manually.
|
|
4692
4234
|
* AssemblyAI API
|
|
4693
|
-
* AssemblyAI API
|
|
4694
|
-
* OpenAPI spec version: 1.
|
|
4235
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4236
|
+
* OpenAPI spec version: 1.0.0
|
|
4695
4237
|
*/
|
|
4696
4238
|
|
|
4697
4239
|
/**
|
|
@@ -4714,8 +4256,8 @@ interface PageDetails {
|
|
|
4714
4256
|
* Generated by orval v7.9.0 🍺
|
|
4715
4257
|
* Do not edit manually.
|
|
4716
4258
|
* AssemblyAI API
|
|
4717
|
-
* AssemblyAI API
|
|
4718
|
-
* OpenAPI spec version: 1.
|
|
4259
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4260
|
+
* OpenAPI spec version: 1.0.0
|
|
4719
4261
|
*/
|
|
4720
4262
|
|
|
4721
4263
|
interface TranscriptParagraph {
|
|
@@ -4725,11 +4267,7 @@ interface TranscriptParagraph {
|
|
|
4725
4267
|
start: number;
|
|
4726
4268
|
/** The ending time, in milliseconds, of the paragraph */
|
|
4727
4269
|
end: number;
|
|
4728
|
-
/**
|
|
4729
|
-
* The confidence score for the transcript of this paragraph
|
|
4730
|
-
* @minimum 0
|
|
4731
|
-
* @maximum 1
|
|
4732
|
-
*/
|
|
4270
|
+
/** The confidence score for the transcript of this paragraph */
|
|
4733
4271
|
confidence: number;
|
|
4734
4272
|
/** An array of words in the paragraph */
|
|
4735
4273
|
words: TranscriptWord[];
|
|
@@ -4739,18 +4277,14 @@ interface TranscriptParagraph {
|
|
|
4739
4277
|
* Generated by orval v7.9.0 🍺
|
|
4740
4278
|
* Do not edit manually.
|
|
4741
4279
|
* AssemblyAI API
|
|
4742
|
-
* AssemblyAI API
|
|
4743
|
-
* OpenAPI spec version: 1.
|
|
4280
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4281
|
+
* OpenAPI spec version: 1.0.0
|
|
4744
4282
|
*/
|
|
4745
4283
|
|
|
4746
4284
|
interface ParagraphsResponse {
|
|
4747
4285
|
/** The unique identifier of your transcript */
|
|
4748
4286
|
id: string;
|
|
4749
|
-
/**
|
|
4750
|
-
* The confidence score for the transcript
|
|
4751
|
-
* @minimum 0
|
|
4752
|
-
* @maximum 1
|
|
4753
|
-
*/
|
|
4287
|
+
/** The confidence score for the transcript */
|
|
4754
4288
|
confidence: number;
|
|
4755
4289
|
/** The duration of the audio file in seconds */
|
|
4756
4290
|
audio_duration: number;
|
|
@@ -4762,36 +4296,38 @@ interface ParagraphsResponse {
|
|
|
4762
4296
|
* Generated by orval v7.9.0 🍺
|
|
4763
4297
|
* Do not edit manually.
|
|
4764
4298
|
* AssemblyAI API
|
|
4765
|
-
* AssemblyAI API
|
|
4766
|
-
* OpenAPI spec version: 1.
|
|
4299
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4300
|
+
* OpenAPI spec version: 1.0.0
|
|
4767
4301
|
*/
|
|
4768
|
-
|
|
4769
|
-
|
|
4770
|
-
|
|
4771
|
-
|
|
4772
|
-
|
|
4773
|
-
|
|
4774
|
-
|
|
4775
|
-
}
|
|
4302
|
+
/**
|
|
4303
|
+
* The status of the redacted audio
|
|
4304
|
+
*/
|
|
4305
|
+
type RedactedAudioStatus = (typeof RedactedAudioStatus)[keyof typeof RedactedAudioStatus];
|
|
4306
|
+
declare const RedactedAudioStatus: {
|
|
4307
|
+
readonly redacted_audio_ready: "redacted_audio_ready";
|
|
4308
|
+
};
|
|
4776
4309
|
|
|
4777
4310
|
/**
|
|
4778
4311
|
* Generated by orval v7.9.0 🍺
|
|
4779
4312
|
* Do not edit manually.
|
|
4780
4313
|
* AssemblyAI API
|
|
4781
|
-
* AssemblyAI API
|
|
4782
|
-
* OpenAPI spec version: 1.
|
|
4314
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4315
|
+
* OpenAPI spec version: 1.0.0
|
|
4783
4316
|
*/
|
|
4784
|
-
|
|
4785
|
-
|
|
4786
|
-
|
|
4317
|
+
|
|
4318
|
+
interface RedactedAudioResponse {
|
|
4319
|
+
/** The status of the redacted audio */
|
|
4320
|
+
status: RedactedAudioStatus;
|
|
4321
|
+
/** The URL of the redacted audio file */
|
|
4322
|
+
redacted_audio_url: string;
|
|
4787
4323
|
}
|
|
4788
4324
|
|
|
4789
4325
|
/**
|
|
4790
4326
|
* Generated by orval v7.9.0 🍺
|
|
4791
4327
|
* Do not edit manually.
|
|
4792
4328
|
* AssemblyAI API
|
|
4793
|
-
* AssemblyAI API
|
|
4794
|
-
* OpenAPI spec version: 1.
|
|
4329
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4330
|
+
* OpenAPI spec version: 1.0.0
|
|
4795
4331
|
*/
|
|
4796
4332
|
/**
|
|
4797
4333
|
* The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially.
|
|
@@ -4802,11 +4338,11 @@ type TranscriptSentenceChannel = string | null;
|
|
|
4802
4338
|
* Generated by orval v7.9.0 🍺
|
|
4803
4339
|
* Do not edit manually.
|
|
4804
4340
|
* AssemblyAI API
|
|
4805
|
-
* AssemblyAI API
|
|
4806
|
-
* OpenAPI spec version: 1.
|
|
4341
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4342
|
+
* OpenAPI spec version: 1.0.0
|
|
4807
4343
|
*/
|
|
4808
4344
|
/**
|
|
4809
|
-
* The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/
|
|
4345
|
+
* The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null
|
|
4810
4346
|
*/
|
|
4811
4347
|
type TranscriptSentenceSpeaker = string | null;
|
|
4812
4348
|
|
|
@@ -4814,8 +4350,8 @@ type TranscriptSentenceSpeaker = string | null;
|
|
|
4814
4350
|
* Generated by orval v7.9.0 🍺
|
|
4815
4351
|
* Do not edit manually.
|
|
4816
4352
|
* AssemblyAI API
|
|
4817
|
-
* AssemblyAI API
|
|
4818
|
-
* OpenAPI spec version: 1.
|
|
4353
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4354
|
+
* OpenAPI spec version: 1.0.0
|
|
4819
4355
|
*/
|
|
4820
4356
|
|
|
4821
4357
|
interface TranscriptSentence {
|
|
@@ -4825,17 +4361,13 @@ interface TranscriptSentence {
|
|
|
4825
4361
|
start: number;
|
|
4826
4362
|
/** The ending time, in milliseconds, for the sentence */
|
|
4827
4363
|
end: number;
|
|
4828
|
-
/**
|
|
4829
|
-
* The confidence score for the transcript of this sentence
|
|
4830
|
-
* @minimum 0
|
|
4831
|
-
* @maximum 1
|
|
4832
|
-
*/
|
|
4364
|
+
/** The confidence score for the transcript of this sentence */
|
|
4833
4365
|
confidence: number;
|
|
4834
4366
|
/** An array of words in the sentence */
|
|
4835
4367
|
words: TranscriptWord[];
|
|
4836
4368
|
/** The channel of the sentence. The left and right channels are channels 1 and 2. Additional channels increment the channel number sequentially. */
|
|
4837
4369
|
channel?: TranscriptSentenceChannel;
|
|
4838
|
-
/** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/
|
|
4370
|
+
/** The speaker of the sentence if [Speaker Diarization](https://www.assemblyai.com/docs/pre-recorded-audio/speaker-diarization) is enabled, else null */
|
|
4839
4371
|
speaker: TranscriptSentenceSpeaker;
|
|
4840
4372
|
}
|
|
4841
4373
|
|
|
@@ -4843,18 +4375,14 @@ interface TranscriptSentence {
|
|
|
4843
4375
|
* Generated by orval v7.9.0 🍺
|
|
4844
4376
|
* Do not edit manually.
|
|
4845
4377
|
* AssemblyAI API
|
|
4846
|
-
* AssemblyAI API
|
|
4847
|
-
* OpenAPI spec version: 1.
|
|
4378
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4379
|
+
* OpenAPI spec version: 1.0.0
|
|
4848
4380
|
*/
|
|
4849
4381
|
|
|
4850
4382
|
interface SentencesResponse {
|
|
4851
4383
|
/** The unique identifier for the transcript */
|
|
4852
4384
|
id: string;
|
|
4853
|
-
/**
|
|
4854
|
-
* The confidence score for the transcript
|
|
4855
|
-
* @minimum 0
|
|
4856
|
-
* @maximum 1
|
|
4857
|
-
*/
|
|
4385
|
+
/** The confidence score for the transcript */
|
|
4858
4386
|
confidence: number;
|
|
4859
4387
|
/** The duration of the audio file in seconds */
|
|
4860
4388
|
audio_duration: number;
|
|
@@ -4866,20 +4394,8 @@ interface SentencesResponse {
|
|
|
4866
4394
|
* Generated by orval v7.9.0 🍺
|
|
4867
4395
|
* Do not edit manually.
|
|
4868
4396
|
* AssemblyAI API
|
|
4869
|
-
* AssemblyAI API
|
|
4870
|
-
* OpenAPI spec version: 1.
|
|
4871
|
-
*/
|
|
4872
|
-
/**
|
|
4873
|
-
* Service unavailable
|
|
4874
|
-
*/
|
|
4875
|
-
type ServiceUnavailableResponse = void;
|
|
4876
|
-
|
|
4877
|
-
/**
|
|
4878
|
-
* Generated by orval v7.9.0 🍺
|
|
4879
|
-
* Do not edit manually.
|
|
4880
|
-
* AssemblyAI API
|
|
4881
|
-
* AssemblyAI API
|
|
4882
|
-
* OpenAPI spec version: 1.3.4
|
|
4397
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4398
|
+
* OpenAPI spec version: 1.0.0
|
|
4883
4399
|
*/
|
|
4884
4400
|
/**
|
|
4885
4401
|
* Format of the subtitles
|
|
@@ -4890,17 +4406,12 @@ declare const SubtitleFormat: {
|
|
|
4890
4406
|
readonly vtt: "vtt";
|
|
4891
4407
|
};
|
|
4892
4408
|
|
|
4893
|
-
/**
|
|
4894
|
-
* Too many requests
|
|
4895
|
-
*/
|
|
4896
|
-
type TooManyRequestsResponse = Error;
|
|
4897
|
-
|
|
4898
4409
|
/**
|
|
4899
4410
|
* Generated by orval v7.9.0 🍺
|
|
4900
4411
|
* Do not edit manually.
|
|
4901
4412
|
* AssemblyAI API
|
|
4902
|
-
* AssemblyAI API
|
|
4903
|
-
* OpenAPI spec version: 1.
|
|
4413
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4414
|
+
* OpenAPI spec version: 1.0.0
|
|
4904
4415
|
*/
|
|
4905
4416
|
|
|
4906
4417
|
/**
|
|
@@ -4913,17 +4424,12 @@ interface TranscriptList {
|
|
|
4913
4424
|
transcripts: TranscriptListItem[];
|
|
4914
4425
|
}
|
|
4915
4426
|
|
|
4916
|
-
/**
|
|
4917
|
-
* Unauthorized
|
|
4918
|
-
*/
|
|
4919
|
-
type UnauthorizedResponse = Error;
|
|
4920
|
-
|
|
4921
4427
|
/**
|
|
4922
4428
|
* Generated by orval v7.9.0 🍺
|
|
4923
4429
|
* Do not edit manually.
|
|
4924
4430
|
* AssemblyAI API
|
|
4925
|
-
* AssemblyAI API
|
|
4926
|
-
* OpenAPI spec version: 1.
|
|
4431
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4432
|
+
* OpenAPI spec version: 1.0.0
|
|
4927
4433
|
*/
|
|
4928
4434
|
interface UploadedFile {
|
|
4929
4435
|
/** A URL that points to your audio file, accessible only by AssemblyAI's servers
|
|
@@ -4935,8 +4441,8 @@ interface UploadedFile {
|
|
|
4935
4441
|
* Generated by orval v7.9.0 🍺
|
|
4936
4442
|
* Do not edit manually.
|
|
4937
4443
|
* AssemblyAI API
|
|
4938
|
-
* AssemblyAI API
|
|
4939
|
-
* OpenAPI spec version: 1.
|
|
4444
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4445
|
+
* OpenAPI spec version: 1.0.0
|
|
4940
4446
|
*/
|
|
4941
4447
|
/**
|
|
4942
4448
|
* An array of timestamps structured as [`start_time`, `end_time`] in milliseconds
|
|
@@ -4947,8 +4453,8 @@ type WordSearchTimestamp = number[];
|
|
|
4947
4453
|
* Generated by orval v7.9.0 🍺
|
|
4948
4454
|
* Do not edit manually.
|
|
4949
4455
|
* AssemblyAI API
|
|
4950
|
-
* AssemblyAI API
|
|
4951
|
-
* OpenAPI spec version: 1.
|
|
4456
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4457
|
+
* OpenAPI spec version: 1.0.0
|
|
4952
4458
|
*/
|
|
4953
4459
|
|
|
4954
4460
|
interface WordSearchMatch {
|
|
@@ -4966,8 +4472,8 @@ interface WordSearchMatch {
|
|
|
4966
4472
|
* Generated by orval v7.9.0 🍺
|
|
4967
4473
|
* Do not edit manually.
|
|
4968
4474
|
* AssemblyAI API
|
|
4969
|
-
* AssemblyAI API
|
|
4970
|
-
* OpenAPI spec version: 1.
|
|
4475
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4476
|
+
* OpenAPI spec version: 1.0.0
|
|
4971
4477
|
*/
|
|
4972
4478
|
type WordSearchParams = {
|
|
4973
4479
|
/**
|
|
@@ -4980,8 +4486,8 @@ type WordSearchParams = {
|
|
|
4980
4486
|
* Generated by orval v7.9.0 🍺
|
|
4981
4487
|
* Do not edit manually.
|
|
4982
4488
|
* AssemblyAI API
|
|
4983
|
-
* AssemblyAI API
|
|
4984
|
-
* OpenAPI spec version: 1.
|
|
4489
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4490
|
+
* OpenAPI spec version: 1.0.0
|
|
4985
4491
|
*/
|
|
4986
4492
|
|
|
4987
4493
|
interface WordSearchResponse {
|
|
@@ -4997,93 +4503,66 @@ interface WordSearchResponse {
|
|
|
4997
4503
|
* Generated by orval v7.9.0 🍺
|
|
4998
4504
|
* Do not edit manually.
|
|
4999
4505
|
* AssemblyAI API
|
|
5000
|
-
* AssemblyAI API
|
|
5001
|
-
* OpenAPI spec version: 1.
|
|
4506
|
+
* AssemblyAI Speech-to-Text API - Batch transcription endpoints. Filtered from the official AssemblyAI docs spec.
|
|
4507
|
+
* OpenAPI spec version: 1.0.0
|
|
5002
4508
|
*/
|
|
5003
4509
|
|
|
5004
4510
|
declare const index$6_AudioIntelligenceModelStatus: typeof AudioIntelligenceModelStatus;
|
|
5005
4511
|
declare const index$6_AutoHighlightResult: typeof AutoHighlightResult;
|
|
5006
4512
|
declare const index$6_AutoHighlightsResult: typeof AutoHighlightsResult;
|
|
5007
|
-
type index$6_BadRequestResponse = BadRequestResponse;
|
|
5008
|
-
type index$6_CannotAccessUploadedFileResponse = CannotAccessUploadedFileResponse;
|
|
5009
4513
|
declare const index$6_Chapter: typeof Chapter;
|
|
5010
4514
|
declare const index$6_ContentSafetyLabel: typeof ContentSafetyLabel;
|
|
5011
4515
|
declare const index$6_ContentSafetyLabelResult: typeof ContentSafetyLabelResult;
|
|
5012
4516
|
declare const index$6_ContentSafetyLabelsResult: typeof ContentSafetyLabelsResult;
|
|
5013
4517
|
declare const index$6_ContentSafetyLabelsResultSeverityScoreSummary: typeof ContentSafetyLabelsResultSeverityScoreSummary;
|
|
5014
4518
|
declare const index$6_ContentSafetyLabelsResultSummary: typeof ContentSafetyLabelsResultSummary;
|
|
5015
|
-
|
|
4519
|
+
declare const index$6_CustomFormattingRequestBody: typeof CustomFormattingRequestBody;
|
|
4520
|
+
declare const index$6_CustomFormattingRequestBodyCustomFormatting: typeof CustomFormattingRequestBodyCustomFormatting;
|
|
4521
|
+
declare const index$6_CustomFormattingResponse: typeof CustomFormattingResponse;
|
|
4522
|
+
declare const index$6_CustomFormattingResponseCustomFormatting: typeof CustomFormattingResponseCustomFormatting;
|
|
4523
|
+
declare const index$6_CustomFormattingResponseCustomFormattingMapping: typeof CustomFormattingResponseCustomFormattingMapping;
|
|
5016
4524
|
declare const index$6_Entity: typeof Entity;
|
|
5017
4525
|
declare const index$6_EntityType: typeof EntityType;
|
|
5018
|
-
type index$
|
|
4526
|
+
type index$6_ErrorDetails = ErrorDetails;
|
|
4527
|
+
type index$6_GetSubtitles200 = GetSubtitles200;
|
|
5019
4528
|
type index$6_GetSubtitlesParams = GetSubtitlesParams;
|
|
5020
|
-
type index$6_InternalServerErrorResponse = InternalServerErrorResponse;
|
|
5021
|
-
type index$6_LemurActionItemsParams = LemurActionItemsParams;
|
|
5022
|
-
type index$6_LemurActionItemsParamsAllOf = LemurActionItemsParamsAllOf;
|
|
5023
|
-
type index$6_LemurActionItemsResponse = LemurActionItemsResponse;
|
|
5024
|
-
type index$6_LemurBaseParams = LemurBaseParams;
|
|
5025
|
-
type index$6_LemurBaseParamsContext = LemurBaseParamsContext;
|
|
5026
|
-
type index$6_LemurBaseParamsContextOneOf = LemurBaseParamsContextOneOf;
|
|
5027
|
-
type index$6_LemurBaseParamsFinalModel = LemurBaseParamsFinalModel;
|
|
5028
|
-
type index$6_LemurBaseResponse = LemurBaseResponse;
|
|
5029
|
-
declare const index$6_LemurModel: typeof LemurModel;
|
|
5030
|
-
type index$6_LemurQuestion = LemurQuestion;
|
|
5031
|
-
type index$6_LemurQuestionAnswer = LemurQuestionAnswer;
|
|
5032
|
-
type index$6_LemurQuestionAnswerParams = LemurQuestionAnswerParams;
|
|
5033
|
-
type index$6_LemurQuestionAnswerParamsAllOf = LemurQuestionAnswerParamsAllOf;
|
|
5034
|
-
type index$6_LemurQuestionAnswerResponse = LemurQuestionAnswerResponse;
|
|
5035
|
-
type index$6_LemurQuestionAnswerResponseAllOf = LemurQuestionAnswerResponseAllOf;
|
|
5036
|
-
type index$6_LemurQuestionContext = LemurQuestionContext;
|
|
5037
|
-
type index$6_LemurQuestionContextOneOf = LemurQuestionContextOneOf;
|
|
5038
|
-
type index$6_LemurResponse = LemurResponse;
|
|
5039
|
-
type index$6_LemurStringResponse = LemurStringResponse;
|
|
5040
|
-
type index$6_LemurStringResponseAllOf = LemurStringResponseAllOf;
|
|
5041
|
-
type index$6_LemurSummaryParams = LemurSummaryParams;
|
|
5042
|
-
type index$6_LemurSummaryParamsAllOf = LemurSummaryParamsAllOf;
|
|
5043
|
-
type index$6_LemurSummaryResponse = LemurSummaryResponse;
|
|
5044
|
-
type index$6_LemurTaskParams = LemurTaskParams;
|
|
5045
|
-
type index$6_LemurTaskParamsAllOf = LemurTaskParamsAllOf;
|
|
5046
|
-
type index$6_LemurTaskResponse = LemurTaskResponse;
|
|
5047
|
-
type index$6_LemurUsage = LemurUsage;
|
|
5048
|
-
type index$6_ListTranscriptParams = ListTranscriptParams;
|
|
5049
4529
|
declare const index$6_ListTranscriptsParams: typeof ListTranscriptsParams;
|
|
5050
|
-
type index$6_NotFoundResponse = NotFoundResponse;
|
|
5051
4530
|
type index$6_PageDetails = PageDetails;
|
|
5052
4531
|
type index$6_PageDetailsNextUrl = PageDetailsNextUrl;
|
|
5053
4532
|
type index$6_PageDetailsPrevUrl = PageDetailsPrevUrl;
|
|
5054
4533
|
type index$6_ParagraphsResponse = ParagraphsResponse;
|
|
5055
4534
|
declare const index$6_PiiPolicy: typeof PiiPolicy;
|
|
5056
|
-
type index$6_PurgeLemurRequestDataResponse = PurgeLemurRequestDataResponse;
|
|
5057
|
-
type index$6_RealtimeTemporaryTokenResponse = RealtimeTemporaryTokenResponse;
|
|
5058
4535
|
declare const index$6_RedactPiiAudioQuality: typeof RedactPiiAudioQuality;
|
|
5059
|
-
|
|
5060
|
-
declare const index$6_RedactedAudioResponse: typeof RedactedAudioResponse;
|
|
4536
|
+
type index$6_RedactedAudioResponse = RedactedAudioResponse;
|
|
5061
4537
|
declare const index$6_RedactedAudioStatus: typeof RedactedAudioStatus;
|
|
5062
4538
|
type index$6_SentencesResponse = SentencesResponse;
|
|
5063
4539
|
declare const index$6_Sentiment: typeof Sentiment;
|
|
5064
4540
|
declare const index$6_SentimentAnalysisResultChannel: typeof SentimentAnalysisResultChannel;
|
|
5065
4541
|
declare const index$6_SentimentAnalysisResultSpeaker: typeof SentimentAnalysisResultSpeaker;
|
|
5066
|
-
type index$6_ServiceUnavailableResponse = ServiceUnavailableResponse;
|
|
5067
4542
|
declare const index$6_SeverityScoreSummary: typeof SeverityScoreSummary;
|
|
4543
|
+
declare const index$6_SpeakerIdentificationRequestBody: typeof SpeakerIdentificationRequestBody;
|
|
4544
|
+
declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentification: typeof SpeakerIdentificationRequestBodySpeakerIdentification;
|
|
4545
|
+
declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType;
|
|
4546
|
+
declare const index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems: typeof SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems;
|
|
4547
|
+
declare const index$6_SpeakerIdentificationResponse: typeof SpeakerIdentificationResponse;
|
|
4548
|
+
declare const index$6_SpeakerIdentificationResponseSpeakerIdentification: typeof SpeakerIdentificationResponseSpeakerIdentification;
|
|
4549
|
+
declare const index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping: typeof SpeakerIdentificationResponseSpeakerIdentificationMapping;
|
|
5068
4550
|
declare const index$6_SpeechModel: typeof SpeechModel;
|
|
5069
4551
|
declare const index$6_SubstitutionPolicy: typeof SubstitutionPolicy;
|
|
5070
4552
|
declare const index$6_SubtitleFormat: typeof SubtitleFormat;
|
|
5071
4553
|
declare const index$6_SummaryModel: typeof SummaryModel;
|
|
5072
4554
|
declare const index$6_SummaryType: typeof SummaryType;
|
|
5073
4555
|
declare const index$6_Timestamp: typeof Timestamp;
|
|
5074
|
-
type index$6_TooManyRequestsResponse = TooManyRequestsResponse;
|
|
5075
4556
|
declare const index$6_TopicDetectionModelResult: typeof TopicDetectionModelResult;
|
|
5076
4557
|
declare const index$6_TopicDetectionModelResultSummary: typeof TopicDetectionModelResultSummary;
|
|
5077
4558
|
declare const index$6_TopicDetectionResult: typeof TopicDetectionResult;
|
|
5078
|
-
declare const index$
|
|
4559
|
+
declare const index$6_TopicDetectionResultLabelsItems: typeof TopicDetectionResultLabelsItems;
|
|
5079
4560
|
declare const index$6_Transcript: typeof Transcript;
|
|
5080
4561
|
declare const index$6_TranscriptAudioDuration: typeof TranscriptAudioDuration;
|
|
5081
4562
|
declare const index$6_TranscriptAudioEndAt: typeof TranscriptAudioEndAt;
|
|
5082
4563
|
declare const index$6_TranscriptAudioStartFrom: typeof TranscriptAudioStartFrom;
|
|
5083
4564
|
declare const index$6_TranscriptAutoChapters: typeof TranscriptAutoChapters;
|
|
5084
4565
|
declare const index$6_TranscriptAutoHighlightsResult: typeof TranscriptAutoHighlightsResult;
|
|
5085
|
-
declare const index$6_TranscriptBoostParam: typeof TranscriptBoostParam;
|
|
5086
|
-
declare const index$6_TranscriptBoostParamProperty: typeof TranscriptBoostParamProperty;
|
|
5087
4566
|
declare const index$6_TranscriptChapters: typeof TranscriptChapters;
|
|
5088
4567
|
declare const index$6_TranscriptConfidence: typeof TranscriptConfidence;
|
|
5089
4568
|
declare const index$6_TranscriptContentSafety: typeof TranscriptContentSafety;
|
|
@@ -5092,6 +4571,7 @@ declare const index$6_TranscriptCustomSpelling: typeof TranscriptCustomSpelling;
|
|
|
5092
4571
|
declare const index$6_TranscriptCustomSpellingProperty: typeof TranscriptCustomSpellingProperty;
|
|
5093
4572
|
declare const index$6_TranscriptCustomTopics: typeof TranscriptCustomTopics;
|
|
5094
4573
|
declare const index$6_TranscriptDisfluencies: typeof TranscriptDisfluencies;
|
|
4574
|
+
declare const index$6_TranscriptDomain: typeof TranscriptDomain;
|
|
5095
4575
|
declare const index$6_TranscriptEntities: typeof TranscriptEntities;
|
|
5096
4576
|
declare const index$6_TranscriptEntityDetection: typeof TranscriptEntityDetection;
|
|
5097
4577
|
declare const index$6_TranscriptFilterProfanity: typeof TranscriptFilterProfanity;
|
|
@@ -5099,33 +4579,42 @@ declare const index$6_TranscriptFormatText: typeof TranscriptFormatText;
|
|
|
5099
4579
|
declare const index$6_TranscriptIabCategories: typeof TranscriptIabCategories;
|
|
5100
4580
|
declare const index$6_TranscriptIabCategoriesResult: typeof TranscriptIabCategoriesResult;
|
|
5101
4581
|
declare const index$6_TranscriptLanguageCode: typeof TranscriptLanguageCode;
|
|
5102
|
-
declare const index$
|
|
4582
|
+
declare const index$6_TranscriptLanguageCodes: typeof TranscriptLanguageCodes;
|
|
5103
4583
|
declare const index$6_TranscriptLanguageConfidence: typeof TranscriptLanguageConfidence;
|
|
5104
4584
|
declare const index$6_TranscriptLanguageConfidenceThreshold: typeof TranscriptLanguageConfidenceThreshold;
|
|
5105
4585
|
declare const index$6_TranscriptLanguageDetection: typeof TranscriptLanguageDetection;
|
|
4586
|
+
declare const index$6_TranscriptLanguageDetectionOptions: typeof TranscriptLanguageDetectionOptions;
|
|
5106
4587
|
type index$6_TranscriptList = TranscriptList;
|
|
5107
4588
|
type index$6_TranscriptListItem = TranscriptListItem;
|
|
5108
|
-
type index$6_TranscriptListItemCompleted = TranscriptListItemCompleted;
|
|
5109
4589
|
type index$6_TranscriptListItemError = TranscriptListItemError;
|
|
5110
4590
|
declare const index$6_TranscriptMultichannel: typeof TranscriptMultichannel;
|
|
5111
|
-
declare const index$
|
|
5112
|
-
declare const index$
|
|
5113
|
-
declare const index$
|
|
5114
|
-
declare const index$
|
|
5115
|
-
declare const index$
|
|
5116
|
-
declare const index$
|
|
5117
|
-
declare const index$
|
|
5118
|
-
declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderName: typeof TranscriptOptionalParamsWebhookAuthHeaderName;
|
|
5119
|
-
declare const index$6_TranscriptOptionalParamsWebhookAuthHeaderValue: typeof TranscriptOptionalParamsWebhookAuthHeaderValue;
|
|
4591
|
+
declare const index$6_TranscriptOptionalParamsLanguageDetectionOptions: typeof TranscriptOptionalParamsLanguageDetectionOptions;
|
|
4592
|
+
declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptions: typeof TranscriptOptionalParamsRedactPiiAudioOptions;
|
|
4593
|
+
declare const index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod;
|
|
4594
|
+
declare const index$6_TranscriptOptionalParamsRemoveAudioTags: typeof TranscriptOptionalParamsRemoveAudioTags;
|
|
4595
|
+
declare const index$6_TranscriptOptionalParamsSpeakerOptions: typeof TranscriptOptionalParamsSpeakerOptions;
|
|
4596
|
+
declare const index$6_TranscriptOptionalParamsSpeechUnderstanding: typeof TranscriptOptionalParamsSpeechUnderstanding;
|
|
4597
|
+
declare const index$6_TranscriptOptionalParamsSpeechUnderstandingRequest: typeof TranscriptOptionalParamsSpeechUnderstandingRequest;
|
|
5120
4598
|
type index$6_TranscriptParagraph = TranscriptParagraph;
|
|
5121
|
-
|
|
5122
|
-
|
|
4599
|
+
declare const index$6_TranscriptParams: typeof TranscriptParams;
|
|
4600
|
+
declare const index$6_TranscriptParamsDomain: typeof TranscriptParamsDomain;
|
|
4601
|
+
declare const index$6_TranscriptParamsLanguageCode: typeof TranscriptParamsLanguageCode;
|
|
4602
|
+
declare const index$6_TranscriptParamsLanguageCodes: typeof TranscriptParamsLanguageCodes;
|
|
4603
|
+
declare const index$6_TranscriptParamsRedactPiiSub: typeof TranscriptParamsRedactPiiSub;
|
|
4604
|
+
declare const index$6_TranscriptParamsRemoveAudioTags: typeof TranscriptParamsRemoveAudioTags;
|
|
4605
|
+
declare const index$6_TranscriptParamsSpeakersExpected: typeof TranscriptParamsSpeakersExpected;
|
|
4606
|
+
declare const index$6_TranscriptParamsSpeechModel: typeof TranscriptParamsSpeechModel;
|
|
4607
|
+
declare const index$6_TranscriptParamsSpeechThreshold: typeof TranscriptParamsSpeechThreshold;
|
|
4608
|
+
declare const index$6_TranscriptParamsWebhookAuthHeaderName: typeof TranscriptParamsWebhookAuthHeaderName;
|
|
4609
|
+
declare const index$6_TranscriptParamsWebhookAuthHeaderValue: typeof TranscriptParamsWebhookAuthHeaderValue;
|
|
5123
4610
|
declare const index$6_TranscriptPunctuate: typeof TranscriptPunctuate;
|
|
5124
|
-
declare const index$6_TranscriptReadyNotification: typeof TranscriptReadyNotification;
|
|
5125
|
-
declare const index$6_TranscriptReadyStatus: typeof TranscriptReadyStatus;
|
|
5126
4611
|
declare const index$6_TranscriptRedactPiiAudio: typeof TranscriptRedactPiiAudio;
|
|
4612
|
+
declare const index$6_TranscriptRedactPiiAudioOptions: typeof TranscriptRedactPiiAudioOptions;
|
|
4613
|
+
declare const index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod: typeof TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod;
|
|
5127
4614
|
declare const index$6_TranscriptRedactPiiAudioQuality: typeof TranscriptRedactPiiAudioQuality;
|
|
5128
4615
|
declare const index$6_TranscriptRedactPiiPolicies: typeof TranscriptRedactPiiPolicies;
|
|
4616
|
+
declare const index$6_TranscriptRemoveAudioTags: typeof TranscriptRemoveAudioTags;
|
|
4617
|
+
declare const index$6_TranscriptRemoveAudioTagsProperty: typeof TranscriptRemoveAudioTagsProperty;
|
|
5129
4618
|
type index$6_TranscriptSentence = TranscriptSentence;
|
|
5130
4619
|
type index$6_TranscriptSentenceChannel = TranscriptSentenceChannel;
|
|
5131
4620
|
type index$6_TranscriptSentenceSpeaker = TranscriptSentenceSpeaker;
|
|
@@ -5134,33 +4623,42 @@ declare const index$6_TranscriptSentimentAnalysisResults: typeof TranscriptSenti
|
|
|
5134
4623
|
declare const index$6_TranscriptSpeakerLabels: typeof TranscriptSpeakerLabels;
|
|
5135
4624
|
declare const index$6_TranscriptSpeakersExpected: typeof TranscriptSpeakersExpected;
|
|
5136
4625
|
declare const index$6_TranscriptSpeechModel: typeof TranscriptSpeechModel;
|
|
4626
|
+
declare const index$6_TranscriptSpeechModels: typeof TranscriptSpeechModels;
|
|
5137
4627
|
declare const index$6_TranscriptSpeechThreshold: typeof TranscriptSpeechThreshold;
|
|
4628
|
+
declare const index$6_TranscriptSpeechUnderstanding: typeof TranscriptSpeechUnderstanding;
|
|
4629
|
+
declare const index$6_TranscriptSpeechUnderstandingRequest: typeof TranscriptSpeechUnderstandingRequest;
|
|
4630
|
+
declare const index$6_TranscriptSpeechUnderstandingResponse: typeof TranscriptSpeechUnderstandingResponse;
|
|
5138
4631
|
declare const index$6_TranscriptSpeedBoost: typeof TranscriptSpeedBoost;
|
|
5139
4632
|
declare const index$6_TranscriptStatus: typeof TranscriptStatus;
|
|
5140
4633
|
declare const index$6_TranscriptSummary: typeof TranscriptSummary;
|
|
5141
4634
|
declare const index$6_TranscriptSummaryModel: typeof TranscriptSummaryModel;
|
|
5142
4635
|
declare const index$6_TranscriptSummaryType: typeof TranscriptSummaryType;
|
|
4636
|
+
declare const index$6_TranscriptTemperature: typeof TranscriptTemperature;
|
|
5143
4637
|
declare const index$6_TranscriptText: typeof TranscriptText;
|
|
5144
4638
|
declare const index$6_TranscriptThrottled: typeof TranscriptThrottled;
|
|
4639
|
+
declare const index$6_TranscriptTranslatedTexts: typeof TranscriptTranslatedTexts;
|
|
5145
4640
|
declare const index$6_TranscriptUtterance: typeof TranscriptUtterance;
|
|
5146
4641
|
declare const index$6_TranscriptUtteranceChannel: typeof TranscriptUtteranceChannel;
|
|
4642
|
+
declare const index$6_TranscriptUtteranceTranslatedTexts: typeof TranscriptUtteranceTranslatedTexts;
|
|
5147
4643
|
declare const index$6_TranscriptUtterances: typeof TranscriptUtterances;
|
|
5148
4644
|
declare const index$6_TranscriptWebhookAuthHeaderName: typeof TranscriptWebhookAuthHeaderName;
|
|
5149
|
-
declare const index$6_TranscriptWebhookNotification: typeof TranscriptWebhookNotification;
|
|
5150
4645
|
declare const index$6_TranscriptWebhookStatusCode: typeof TranscriptWebhookStatusCode;
|
|
5151
4646
|
declare const index$6_TranscriptWebhookUrl: typeof TranscriptWebhookUrl;
|
|
5152
4647
|
declare const index$6_TranscriptWord: typeof TranscriptWord;
|
|
5153
4648
|
declare const index$6_TranscriptWordChannel: typeof TranscriptWordChannel;
|
|
5154
4649
|
declare const index$6_TranscriptWordSpeaker: typeof TranscriptWordSpeaker;
|
|
5155
4650
|
declare const index$6_TranscriptWords: typeof TranscriptWords;
|
|
5156
|
-
|
|
4651
|
+
declare const index$6_TranslationRequestBody: typeof TranslationRequestBody;
|
|
4652
|
+
declare const index$6_TranslationRequestBodyTranslation: typeof TranslationRequestBodyTranslation;
|
|
4653
|
+
declare const index$6_TranslationResponse: typeof TranslationResponse;
|
|
4654
|
+
declare const index$6_TranslationResponseTranslation: typeof TranslationResponseTranslation;
|
|
5157
4655
|
type index$6_UploadedFile = UploadedFile;
|
|
5158
4656
|
type index$6_WordSearchMatch = WordSearchMatch;
|
|
5159
4657
|
type index$6_WordSearchParams = WordSearchParams;
|
|
5160
4658
|
type index$6_WordSearchResponse = WordSearchResponse;
|
|
5161
4659
|
type index$6_WordSearchTimestamp = WordSearchTimestamp;
|
|
5162
4660
|
declare namespace index$6 {
|
|
5163
|
-
export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, type index$6_BadRequestResponse as BadRequestResponse, type index$6_CannotAccessUploadedFileResponse as CannotAccessUploadedFileResponse, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, type index$6_CreateRealtimeTemporaryTokenParams as CreateRealtimeTemporaryTokenParams, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_GatewayTimeoutResponse as GatewayTimeoutResponse, type index$6_GetSubtitlesParams as GetSubtitlesParams, type index$6_InternalServerErrorResponse as InternalServerErrorResponse, type index$6_LemurActionItemsParams as LemurActionItemsParams, type index$6_LemurActionItemsParamsAllOf as LemurActionItemsParamsAllOf, type index$6_LemurActionItemsResponse as LemurActionItemsResponse, type index$6_LemurBaseParams as LemurBaseParams, type index$6_LemurBaseParamsContext as LemurBaseParamsContext, type index$6_LemurBaseParamsContextOneOf as LemurBaseParamsContextOneOf, type index$6_LemurBaseParamsFinalModel as LemurBaseParamsFinalModel, type index$6_LemurBaseResponse as LemurBaseResponse, index$6_LemurModel as LemurModel, type index$6_LemurQuestion as LemurQuestion, type index$6_LemurQuestionAnswer as LemurQuestionAnswer, type index$6_LemurQuestionAnswerParams as LemurQuestionAnswerParams, type index$6_LemurQuestionAnswerParamsAllOf as LemurQuestionAnswerParamsAllOf, type index$6_LemurQuestionAnswerResponse as LemurQuestionAnswerResponse, type index$6_LemurQuestionAnswerResponseAllOf as LemurQuestionAnswerResponseAllOf, type index$6_LemurQuestionContext as LemurQuestionContext, type index$6_LemurQuestionContextOneOf as LemurQuestionContextOneOf, type index$6_LemurResponse as LemurResponse, type index$6_LemurStringResponse as LemurStringResponse, type index$6_LemurStringResponseAllOf as LemurStringResponseAllOf, type index$6_LemurSummaryParams as LemurSummaryParams, type index$6_LemurSummaryParamsAllOf as LemurSummaryParamsAllOf, type index$6_LemurSummaryResponse as LemurSummaryResponse, type index$6_LemurTaskParams as LemurTaskParams, type index$6_LemurTaskParamsAllOf as LemurTaskParamsAllOf, type index$6_LemurTaskResponse as LemurTaskResponse, type index$6_LemurUsage as LemurUsage, type index$6_ListTranscriptParams as ListTranscriptParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_NotFoundResponse as NotFoundResponse, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, type index$6_PurgeLemurRequestDataResponse as PurgeLemurRequestDataResponse, type index$6_RealtimeTemporaryTokenResponse as RealtimeTemporaryTokenResponse, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, index$6_RedactedAudioNotification as RedactedAudioNotification, index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, type index$6_ServiceUnavailableResponse as ServiceUnavailableResponse, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, type index$6_TooManyRequestsResponse as TooManyRequestsResponse, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItem as TopicDetectionResultLabelsItem, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptBoostParam as TranscriptBoostParam, index$6_TranscriptBoostParamProperty as TranscriptBoostParamProperty, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodeProperty as TranscriptLanguageCodeProperty, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemCompleted as TranscriptListItemCompleted, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParams as TranscriptOptionalParams, index$6_TranscriptOptionalParamsLanguageCode as TranscriptOptionalParamsLanguageCode, index$6_TranscriptOptionalParamsLanguageCodeOneOf as TranscriptOptionalParamsLanguageCodeOneOf, index$6_TranscriptOptionalParamsRedactPiiSub as TranscriptOptionalParamsRedactPiiSub, index$6_TranscriptOptionalParamsSpeakersExpected as TranscriptOptionalParamsSpeakersExpected, index$6_TranscriptOptionalParamsSpeechModel as TranscriptOptionalParamsSpeechModel, index$6_TranscriptOptionalParamsSpeechThreshold as TranscriptOptionalParamsSpeechThreshold, index$6_TranscriptOptionalParamsWebhookAuthHeaderName as TranscriptOptionalParamsWebhookAuthHeaderName, index$6_TranscriptOptionalParamsWebhookAuthHeaderValue as TranscriptOptionalParamsWebhookAuthHeaderValue, type index$6_TranscriptParagraph as TranscriptParagraph, type index$6_TranscriptParams as TranscriptParams, type index$6_TranscriptParamsAllOf as TranscriptParamsAllOf, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptReadyNotification as TranscriptReadyNotification, index$6_TranscriptReadyStatus as TranscriptReadyStatus, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookNotification as TranscriptWebhookNotification, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, type index$6_UnauthorizedResponse as UnauthorizedResponse, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
|
|
4661
|
+
export { index$6_AudioIntelligenceModelStatus as AudioIntelligenceModelStatus, index$6_AutoHighlightResult as AutoHighlightResult, index$6_AutoHighlightsResult as AutoHighlightsResult, index$6_Chapter as Chapter, index$6_ContentSafetyLabel as ContentSafetyLabel, index$6_ContentSafetyLabelResult as ContentSafetyLabelResult, index$6_ContentSafetyLabelsResult as ContentSafetyLabelsResult, index$6_ContentSafetyLabelsResultSeverityScoreSummary as ContentSafetyLabelsResultSeverityScoreSummary, index$6_ContentSafetyLabelsResultSummary as ContentSafetyLabelsResultSummary, index$6_CustomFormattingRequestBody as CustomFormattingRequestBody, index$6_CustomFormattingRequestBodyCustomFormatting as CustomFormattingRequestBodyCustomFormatting, index$6_CustomFormattingResponse as CustomFormattingResponse, index$6_CustomFormattingResponseCustomFormatting as CustomFormattingResponseCustomFormatting, index$6_CustomFormattingResponseCustomFormattingMapping as CustomFormattingResponseCustomFormattingMapping, index$6_Entity as Entity, index$6_EntityType as EntityType, type Error$2 as Error, type index$6_ErrorDetails as ErrorDetails, type index$6_GetSubtitles200 as GetSubtitles200, type index$6_GetSubtitlesParams as GetSubtitlesParams, index$6_ListTranscriptsParams as ListTranscriptsParams, type index$6_PageDetails as PageDetails, type index$6_PageDetailsNextUrl as PageDetailsNextUrl, type index$6_PageDetailsPrevUrl as PageDetailsPrevUrl, type index$6_ParagraphsResponse as ParagraphsResponse, index$6_PiiPolicy as PiiPolicy, index$6_RedactPiiAudioQuality as RedactPiiAudioQuality, type index$6_RedactedAudioResponse as RedactedAudioResponse, index$6_RedactedAudioStatus as RedactedAudioStatus, type index$6_SentencesResponse as SentencesResponse, index$6_Sentiment as Sentiment, SentimentAnalysisResult$1 as SentimentAnalysisResult, index$6_SentimentAnalysisResultChannel as SentimentAnalysisResultChannel, index$6_SentimentAnalysisResultSpeaker as SentimentAnalysisResultSpeaker, index$6_SeverityScoreSummary as SeverityScoreSummary, index$6_SpeakerIdentificationRequestBody as SpeakerIdentificationRequestBody, index$6_SpeakerIdentificationRequestBodySpeakerIdentification as SpeakerIdentificationRequestBodySpeakerIdentification, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakerType, index$6_SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems as SpeakerIdentificationRequestBodySpeakerIdentificationSpeakersItems, index$6_SpeakerIdentificationResponse as SpeakerIdentificationResponse, index$6_SpeakerIdentificationResponseSpeakerIdentification as SpeakerIdentificationResponseSpeakerIdentification, index$6_SpeakerIdentificationResponseSpeakerIdentificationMapping as SpeakerIdentificationResponseSpeakerIdentificationMapping, index$6_SpeechModel as SpeechModel, index$6_SubstitutionPolicy as SubstitutionPolicy, index$6_SubtitleFormat as SubtitleFormat, index$6_SummaryModel as SummaryModel, index$6_SummaryType as SummaryType, index$6_Timestamp as Timestamp, index$6_TopicDetectionModelResult as TopicDetectionModelResult, index$6_TopicDetectionModelResultSummary as TopicDetectionModelResultSummary, index$6_TopicDetectionResult as TopicDetectionResult, index$6_TopicDetectionResultLabelsItems as TopicDetectionResultLabelsItems, index$6_Transcript as Transcript, index$6_TranscriptAudioDuration as TranscriptAudioDuration, index$6_TranscriptAudioEndAt as TranscriptAudioEndAt, index$6_TranscriptAudioStartFrom as TranscriptAudioStartFrom, index$6_TranscriptAutoChapters as TranscriptAutoChapters, index$6_TranscriptAutoHighlightsResult as TranscriptAutoHighlightsResult, index$6_TranscriptChapters as TranscriptChapters, index$6_TranscriptConfidence as TranscriptConfidence, index$6_TranscriptContentSafety as TranscriptContentSafety, index$6_TranscriptContentSafetyLabels as TranscriptContentSafetyLabels, index$6_TranscriptCustomSpelling as TranscriptCustomSpelling, index$6_TranscriptCustomSpellingProperty as TranscriptCustomSpellingProperty, index$6_TranscriptCustomTopics as TranscriptCustomTopics, index$6_TranscriptDisfluencies as TranscriptDisfluencies, index$6_TranscriptDomain as TranscriptDomain, index$6_TranscriptEntities as TranscriptEntities, index$6_TranscriptEntityDetection as TranscriptEntityDetection, index$6_TranscriptFilterProfanity as TranscriptFilterProfanity, index$6_TranscriptFormatText as TranscriptFormatText, index$6_TranscriptIabCategories as TranscriptIabCategories, index$6_TranscriptIabCategoriesResult as TranscriptIabCategoriesResult, index$6_TranscriptLanguageCode as TranscriptLanguageCode, index$6_TranscriptLanguageCodes as TranscriptLanguageCodes, index$6_TranscriptLanguageConfidence as TranscriptLanguageConfidence, index$6_TranscriptLanguageConfidenceThreshold as TranscriptLanguageConfidenceThreshold, index$6_TranscriptLanguageDetection as TranscriptLanguageDetection, index$6_TranscriptLanguageDetectionOptions as TranscriptLanguageDetectionOptions, type index$6_TranscriptList as TranscriptList, type index$6_TranscriptListItem as TranscriptListItem, type index$6_TranscriptListItemError as TranscriptListItemError, index$6_TranscriptMultichannel as TranscriptMultichannel, index$6_TranscriptOptionalParamsLanguageDetectionOptions as TranscriptOptionalParamsLanguageDetectionOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptions as TranscriptOptionalParamsRedactPiiAudioOptions, index$6_TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptOptionalParamsRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptOptionalParamsRemoveAudioTags as TranscriptOptionalParamsRemoveAudioTags, index$6_TranscriptOptionalParamsSpeakerOptions as TranscriptOptionalParamsSpeakerOptions, index$6_TranscriptOptionalParamsSpeechUnderstanding as TranscriptOptionalParamsSpeechUnderstanding, index$6_TranscriptOptionalParamsSpeechUnderstandingRequest as TranscriptOptionalParamsSpeechUnderstandingRequest, type index$6_TranscriptParagraph as TranscriptParagraph, index$6_TranscriptParams as TranscriptParams, index$6_TranscriptParamsDomain as TranscriptParamsDomain, index$6_TranscriptParamsLanguageCode as TranscriptParamsLanguageCode, index$6_TranscriptParamsLanguageCodes as TranscriptParamsLanguageCodes, index$6_TranscriptParamsRedactPiiSub as TranscriptParamsRedactPiiSub, index$6_TranscriptParamsRemoveAudioTags as TranscriptParamsRemoveAudioTags, index$6_TranscriptParamsSpeakersExpected as TranscriptParamsSpeakersExpected, index$6_TranscriptParamsSpeechModel as TranscriptParamsSpeechModel, index$6_TranscriptParamsSpeechThreshold as TranscriptParamsSpeechThreshold, index$6_TranscriptParamsWebhookAuthHeaderName as TranscriptParamsWebhookAuthHeaderName, index$6_TranscriptParamsWebhookAuthHeaderValue as TranscriptParamsWebhookAuthHeaderValue, index$6_TranscriptPunctuate as TranscriptPunctuate, index$6_TranscriptRedactPiiAudio as TranscriptRedactPiiAudio, index$6_TranscriptRedactPiiAudioOptions as TranscriptRedactPiiAudioOptions, index$6_TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod as TranscriptRedactPiiAudioOptionsOverrideAudioRedactionMethod, index$6_TranscriptRedactPiiAudioQuality as TranscriptRedactPiiAudioQuality, index$6_TranscriptRedactPiiPolicies as TranscriptRedactPiiPolicies, index$6_TranscriptRemoveAudioTags as TranscriptRemoveAudioTags, index$6_TranscriptRemoveAudioTagsProperty as TranscriptRemoveAudioTagsProperty, type index$6_TranscriptSentence as TranscriptSentence, type index$6_TranscriptSentenceChannel as TranscriptSentenceChannel, type index$6_TranscriptSentenceSpeaker as TranscriptSentenceSpeaker, index$6_TranscriptSentimentAnalysis as TranscriptSentimentAnalysis, index$6_TranscriptSentimentAnalysisResults as TranscriptSentimentAnalysisResults, index$6_TranscriptSpeakerLabels as TranscriptSpeakerLabels, index$6_TranscriptSpeakersExpected as TranscriptSpeakersExpected, index$6_TranscriptSpeechModel as TranscriptSpeechModel, index$6_TranscriptSpeechModels as TranscriptSpeechModels, index$6_TranscriptSpeechThreshold as TranscriptSpeechThreshold, index$6_TranscriptSpeechUnderstanding as TranscriptSpeechUnderstanding, index$6_TranscriptSpeechUnderstandingRequest as TranscriptSpeechUnderstandingRequest, index$6_TranscriptSpeechUnderstandingResponse as TranscriptSpeechUnderstandingResponse, index$6_TranscriptSpeedBoost as TranscriptSpeedBoost, index$6_TranscriptStatus as TranscriptStatus, index$6_TranscriptSummary as TranscriptSummary, index$6_TranscriptSummaryModel as TranscriptSummaryModel, index$6_TranscriptSummaryType as TranscriptSummaryType, index$6_TranscriptTemperature as TranscriptTemperature, index$6_TranscriptText as TranscriptText, index$6_TranscriptThrottled as TranscriptThrottled, index$6_TranscriptTranslatedTexts as TranscriptTranslatedTexts, index$6_TranscriptUtterance as TranscriptUtterance, index$6_TranscriptUtteranceChannel as TranscriptUtteranceChannel, index$6_TranscriptUtteranceTranslatedTexts as TranscriptUtteranceTranslatedTexts, index$6_TranscriptUtterances as TranscriptUtterances, index$6_TranscriptWebhookAuthHeaderName as TranscriptWebhookAuthHeaderName, index$6_TranscriptWebhookStatusCode as TranscriptWebhookStatusCode, index$6_TranscriptWebhookUrl as TranscriptWebhookUrl, index$6_TranscriptWord as TranscriptWord, index$6_TranscriptWordChannel as TranscriptWordChannel, index$6_TranscriptWordSpeaker as TranscriptWordSpeaker, index$6_TranscriptWords as TranscriptWords, index$6_TranslationRequestBody as TranslationRequestBody, index$6_TranslationRequestBodyTranslation as TranslationRequestBodyTranslation, index$6_TranslationResponse as TranslationResponse, index$6_TranslationResponseTranslation as TranslationResponseTranslation, type index$6_UploadedFile as UploadedFile, type index$6_WordSearchMatch as WordSearchMatch, type index$6_WordSearchParams as WordSearchParams, type index$6_WordSearchResponse as WordSearchResponse, type index$6_WordSearchTimestamp as WordSearchTimestamp };
|
|
5164
4662
|
}
|
|
5165
4663
|
|
|
5166
4664
|
/**
|
|
@@ -5177,6 +4675,15 @@ Transcripts are sorted from newest to oldest and can be retrieved for the last 9
|
|
|
5177
4675
|
|
|
5178
4676
|
If you need to retrieve transcripts from more than 90 days ago please reach out to our Support team at support@assemblyai.com.
|
|
5179
4677
|
|
|
4678
|
+
**Pagination**
|
|
4679
|
+
|
|
4680
|
+
This endpoint returns paginated results. The response includes a `page_details` object with the following properties:
|
|
4681
|
+
- `page_details.limit` - Maximum number of transcripts per page.
|
|
4682
|
+
- `page_details.result_count` - Total number of transcripts returned on the current page.
|
|
4683
|
+
- `page_details.current_url` - URL to the current page.
|
|
4684
|
+
- `page_details.prev_url` - URL to the previous page of older transcripts.
|
|
4685
|
+
- `page_details.next_url` - URL to the next page of newer transcripts.
|
|
4686
|
+
|
|
5180
4687
|
* @summary List transcripts
|
|
5181
4688
|
*/
|
|
5182
4689
|
declare const listTranscripts: <TData = AxiosResponse<TranscriptList>>(params?: ListTranscriptsParams, options?: AxiosRequestConfig) => Promise<TData>;
|
|
@@ -5195,14 +4702,6 @@ Remove the data from the transcript and mark it as deleted.
|
|
|
5195
4702
|
* @summary Delete transcript
|
|
5196
4703
|
*/
|
|
5197
4704
|
declare const deleteTranscript: <TData = AxiosResponse<Transcript>>(transcriptId: string, options?: AxiosRequestConfig) => Promise<TData>;
|
|
5198
|
-
/**
|
|
5199
|
-
* <Warning>Streaming Speech-to-Text is currently not available on the EU endpoint.</Warning>
|
|
5200
|
-
<Note>Any usage associated with a temporary token will be attributed to the API key that generated it.</Note>
|
|
5201
|
-
Create a temporary authentication token for Streaming Speech-to-Text
|
|
5202
|
-
|
|
5203
|
-
* @summary Create temporary authentication token for Streaming STT
|
|
5204
|
-
*/
|
|
5205
|
-
declare const createTemporaryToken: <TData = AxiosResponse<RealtimeTemporaryTokenResponse>>(createRealtimeTemporaryTokenParams: CreateRealtimeTemporaryTokenParams, options?: AxiosRequestConfig) => Promise<TData>;
|
|
5206
4705
|
|
|
5207
4706
|
/**
|
|
5208
4707
|
* AssemblyAI transcription provider adapter
|
|
@@ -15349,6 +14848,18 @@ type ManageV1FilterVersionParameter = boolean;
|
|
|
15349
14848
|
/**
|
|
15350
14849
|
* ManageV1GroupingParameter type definition
|
|
15351
14850
|
*/
|
|
14851
|
+
/**
|
|
14852
|
+
* ManageV1GroupingParameter type definition
|
|
14853
|
+
*/
|
|
14854
|
+
/**
|
|
14855
|
+
* ManageV1GroupingParameter type definition
|
|
14856
|
+
*/
|
|
14857
|
+
/**
|
|
14858
|
+
* ManageV1GroupingParameter type definition
|
|
14859
|
+
*/
|
|
14860
|
+
/**
|
|
14861
|
+
* ManageV1GroupingParameter type definition
|
|
14862
|
+
*/
|
|
15352
14863
|
type ManageV1GroupingParameter = typeof ManageV1GroupingParameter[keyof typeof ManageV1GroupingParameter];
|
|
15353
14864
|
declare const ManageV1GroupingParameter: {
|
|
15354
14865
|
readonly accessor: "accessor";
|
|
@@ -15421,6 +14932,18 @@ type ManageV1ModelsListParams = {
|
|
|
15421
14932
|
/**
|
|
15422
14933
|
* ManageV1ProjectKeysStatusParameter type definition
|
|
15423
14934
|
*/
|
|
14935
|
+
/**
|
|
14936
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14937
|
+
*/
|
|
14938
|
+
/**
|
|
14939
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14940
|
+
*/
|
|
14941
|
+
/**
|
|
14942
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14943
|
+
*/
|
|
14944
|
+
/**
|
|
14945
|
+
* ManageV1ProjectKeysStatusParameter type definition
|
|
14946
|
+
*/
|
|
15424
14947
|
type ManageV1ProjectKeysStatusParameter = typeof ManageV1ProjectKeysStatusParameter[keyof typeof ManageV1ProjectKeysStatusParameter];
|
|
15425
14948
|
declare const ManageV1ProjectKeysStatusParameter: {
|
|
15426
14949
|
readonly active: "active";
|
|
@@ -16306,6 +15829,18 @@ type SelfHostedV1ParamsDistributionScopesParameter = SelfHostedV1ParamsDistribut
|
|
|
16306
15829
|
/**
|
|
16307
15830
|
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
16308
15831
|
*/
|
|
15832
|
+
/**
|
|
15833
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15834
|
+
*/
|
|
15835
|
+
/**
|
|
15836
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15837
|
+
*/
|
|
15838
|
+
/**
|
|
15839
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15840
|
+
*/
|
|
15841
|
+
/**
|
|
15842
|
+
* SelfHostedV1ParamsDistributionProviderParameter type definition
|
|
15843
|
+
*/
|
|
16309
15844
|
type SelfHostedV1ParamsDistributionProviderParameter = typeof SelfHostedV1ParamsDistributionProviderParameter[keyof typeof SelfHostedV1ParamsDistributionProviderParameter];
|
|
16310
15845
|
declare const SelfHostedV1ParamsDistributionProviderParameter: {
|
|
16311
15846
|
readonly quay: "quay";
|
|
@@ -19533,8 +19068,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19533
19068
|
}, "strip", z.ZodTypeAny, {
|
|
19534
19069
|
language: string;
|
|
19535
19070
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19536
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19537
19071
|
domain?: string | undefined;
|
|
19072
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19538
19073
|
output_locale?: string | undefined;
|
|
19539
19074
|
additional_vocab?: {
|
|
19540
19075
|
content: string;
|
|
@@ -19561,8 +19096,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19561
19096
|
}, {
|
|
19562
19097
|
language: string;
|
|
19563
19098
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19564
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19565
19099
|
domain?: string | undefined;
|
|
19100
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19566
19101
|
output_locale?: string | undefined;
|
|
19567
19102
|
additional_vocab?: {
|
|
19568
19103
|
content: string;
|
|
@@ -19720,8 +19255,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19720
19255
|
transcription_config?: {
|
|
19721
19256
|
language: string;
|
|
19722
19257
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19723
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19724
19258
|
domain?: string | undefined;
|
|
19259
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19725
19260
|
output_locale?: string | undefined;
|
|
19726
19261
|
additional_vocab?: {
|
|
19727
19262
|
content: string;
|
|
@@ -19801,8 +19336,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19801
19336
|
transcription_config?: {
|
|
19802
19337
|
language: string;
|
|
19803
19338
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19804
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19805
19339
|
domain?: string | undefined;
|
|
19340
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19806
19341
|
output_locale?: string | undefined;
|
|
19807
19342
|
additional_vocab?: {
|
|
19808
19343
|
content: string;
|
|
@@ -19900,8 +19435,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19900
19435
|
transcription_config?: {
|
|
19901
19436
|
language: string;
|
|
19902
19437
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19903
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19904
19438
|
domain?: string | undefined;
|
|
19439
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19905
19440
|
output_locale?: string | undefined;
|
|
19906
19441
|
additional_vocab?: {
|
|
19907
19442
|
content: string;
|
|
@@ -19994,8 +19529,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
19994
19529
|
transcription_config?: {
|
|
19995
19530
|
language: string;
|
|
19996
19531
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
19997
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
19998
19532
|
domain?: string | undefined;
|
|
19533
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
19999
19534
|
output_locale?: string | undefined;
|
|
20000
19535
|
additional_vocab?: {
|
|
20001
19536
|
content: string;
|
|
@@ -20090,8 +19625,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
20090
19625
|
transcription_config?: {
|
|
20091
19626
|
language: string;
|
|
20092
19627
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20093
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20094
19628
|
domain?: string | undefined;
|
|
19629
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20095
19630
|
output_locale?: string | undefined;
|
|
20096
19631
|
additional_vocab?: {
|
|
20097
19632
|
content: string;
|
|
@@ -20186,8 +19721,8 @@ declare const getJobsResponse: z.ZodObject<{
|
|
|
20186
19721
|
transcription_config?: {
|
|
20187
19722
|
language: string;
|
|
20188
19723
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20189
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20190
19724
|
domain?: string | undefined;
|
|
19725
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20191
19726
|
output_locale?: string | undefined;
|
|
20192
19727
|
additional_vocab?: {
|
|
20193
19728
|
content: string;
|
|
@@ -20368,8 +19903,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20368
19903
|
}, "strip", z.ZodTypeAny, {
|
|
20369
19904
|
language: string;
|
|
20370
19905
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20371
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20372
19906
|
domain?: string | undefined;
|
|
19907
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20373
19908
|
output_locale?: string | undefined;
|
|
20374
19909
|
additional_vocab?: {
|
|
20375
19910
|
content: string;
|
|
@@ -20396,8 +19931,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20396
19931
|
}, {
|
|
20397
19932
|
language: string;
|
|
20398
19933
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20399
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20400
19934
|
domain?: string | undefined;
|
|
19935
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20401
19936
|
output_locale?: string | undefined;
|
|
20402
19937
|
additional_vocab?: {
|
|
20403
19938
|
content: string;
|
|
@@ -20555,8 +20090,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20555
20090
|
transcription_config?: {
|
|
20556
20091
|
language: string;
|
|
20557
20092
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20558
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20559
20093
|
domain?: string | undefined;
|
|
20094
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20560
20095
|
output_locale?: string | undefined;
|
|
20561
20096
|
additional_vocab?: {
|
|
20562
20097
|
content: string;
|
|
@@ -20636,8 +20171,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20636
20171
|
transcription_config?: {
|
|
20637
20172
|
language: string;
|
|
20638
20173
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20639
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20640
20174
|
domain?: string | undefined;
|
|
20175
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20641
20176
|
output_locale?: string | undefined;
|
|
20642
20177
|
additional_vocab?: {
|
|
20643
20178
|
content: string;
|
|
@@ -20735,8 +20270,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20735
20270
|
transcription_config?: {
|
|
20736
20271
|
language: string;
|
|
20737
20272
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20738
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20739
20273
|
domain?: string | undefined;
|
|
20274
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20740
20275
|
output_locale?: string | undefined;
|
|
20741
20276
|
additional_vocab?: {
|
|
20742
20277
|
content: string;
|
|
@@ -20829,8 +20364,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20829
20364
|
transcription_config?: {
|
|
20830
20365
|
language: string;
|
|
20831
20366
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20832
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20833
20367
|
domain?: string | undefined;
|
|
20368
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20834
20369
|
output_locale?: string | undefined;
|
|
20835
20370
|
additional_vocab?: {
|
|
20836
20371
|
content: string;
|
|
@@ -20925,8 +20460,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
20925
20460
|
transcription_config?: {
|
|
20926
20461
|
language: string;
|
|
20927
20462
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
20928
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
20929
20463
|
domain?: string | undefined;
|
|
20464
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
20930
20465
|
output_locale?: string | undefined;
|
|
20931
20466
|
additional_vocab?: {
|
|
20932
20467
|
content: string;
|
|
@@ -21021,8 +20556,8 @@ declare const getJobsJobidResponse: z.ZodObject<{
|
|
|
21021
20556
|
transcription_config?: {
|
|
21022
20557
|
language: string;
|
|
21023
20558
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21024
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21025
20559
|
domain?: string | undefined;
|
|
20560
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21026
20561
|
output_locale?: string | undefined;
|
|
21027
20562
|
additional_vocab?: {
|
|
21028
20563
|
content: string;
|
|
@@ -21210,8 +20745,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21210
20745
|
}, "strip", z.ZodTypeAny, {
|
|
21211
20746
|
language: string;
|
|
21212
20747
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21213
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21214
20748
|
domain?: string | undefined;
|
|
20749
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21215
20750
|
output_locale?: string | undefined;
|
|
21216
20751
|
additional_vocab?: {
|
|
21217
20752
|
content: string;
|
|
@@ -21238,8 +20773,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21238
20773
|
}, {
|
|
21239
20774
|
language: string;
|
|
21240
20775
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21241
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21242
20776
|
domain?: string | undefined;
|
|
20777
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21243
20778
|
output_locale?: string | undefined;
|
|
21244
20779
|
additional_vocab?: {
|
|
21245
20780
|
content: string;
|
|
@@ -21397,8 +20932,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21397
20932
|
transcription_config?: {
|
|
21398
20933
|
language: string;
|
|
21399
20934
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21400
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21401
20935
|
domain?: string | undefined;
|
|
20936
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21402
20937
|
output_locale?: string | undefined;
|
|
21403
20938
|
additional_vocab?: {
|
|
21404
20939
|
content: string;
|
|
@@ -21478,8 +21013,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21478
21013
|
transcription_config?: {
|
|
21479
21014
|
language: string;
|
|
21480
21015
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21481
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21482
21016
|
domain?: string | undefined;
|
|
21017
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21483
21018
|
output_locale?: string | undefined;
|
|
21484
21019
|
additional_vocab?: {
|
|
21485
21020
|
content: string;
|
|
@@ -21577,8 +21112,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21577
21112
|
transcription_config?: {
|
|
21578
21113
|
language: string;
|
|
21579
21114
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21580
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21581
21115
|
domain?: string | undefined;
|
|
21116
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21582
21117
|
output_locale?: string | undefined;
|
|
21583
21118
|
additional_vocab?: {
|
|
21584
21119
|
content: string;
|
|
@@ -21671,8 +21206,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21671
21206
|
transcription_config?: {
|
|
21672
21207
|
language: string;
|
|
21673
21208
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21674
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21675
21209
|
domain?: string | undefined;
|
|
21210
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21676
21211
|
output_locale?: string | undefined;
|
|
21677
21212
|
additional_vocab?: {
|
|
21678
21213
|
content: string;
|
|
@@ -21767,8 +21302,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21767
21302
|
transcription_config?: {
|
|
21768
21303
|
language: string;
|
|
21769
21304
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21770
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21771
21305
|
domain?: string | undefined;
|
|
21306
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21772
21307
|
output_locale?: string | undefined;
|
|
21773
21308
|
additional_vocab?: {
|
|
21774
21309
|
content: string;
|
|
@@ -21863,8 +21398,8 @@ declare const deleteJobsJobidResponse: z.ZodObject<{
|
|
|
21863
21398
|
transcription_config?: {
|
|
21864
21399
|
language: string;
|
|
21865
21400
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
21866
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
21867
21401
|
domain?: string | undefined;
|
|
21402
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
21868
21403
|
output_locale?: string | undefined;
|
|
21869
21404
|
additional_vocab?: {
|
|
21870
21405
|
content: string;
|
|
@@ -22090,8 +21625,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22090
21625
|
}, "strip", z.ZodTypeAny, {
|
|
22091
21626
|
language: string;
|
|
22092
21627
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22093
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22094
21628
|
domain?: string | undefined;
|
|
21629
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22095
21630
|
output_locale?: string | undefined;
|
|
22096
21631
|
additional_vocab?: {
|
|
22097
21632
|
content: string;
|
|
@@ -22118,8 +21653,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22118
21653
|
}, {
|
|
22119
21654
|
language: string;
|
|
22120
21655
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22121
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22122
21656
|
domain?: string | undefined;
|
|
21657
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22123
21658
|
output_locale?: string | undefined;
|
|
22124
21659
|
additional_vocab?: {
|
|
22125
21660
|
content: string;
|
|
@@ -22306,8 +21841,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22306
21841
|
transcription_config?: {
|
|
22307
21842
|
language: string;
|
|
22308
21843
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22309
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22310
21844
|
domain?: string | undefined;
|
|
21845
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22311
21846
|
output_locale?: string | undefined;
|
|
22312
21847
|
additional_vocab?: {
|
|
22313
21848
|
content: string;
|
|
@@ -22387,8 +21922,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
22387
21922
|
transcription_config?: {
|
|
22388
21923
|
language: string;
|
|
22389
21924
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
22390
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
22391
21925
|
domain?: string | undefined;
|
|
21926
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
22392
21927
|
output_locale?: string | undefined;
|
|
22393
21928
|
additional_vocab?: {
|
|
22394
21929
|
content: string;
|
|
@@ -23129,8 +22664,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
23129
22664
|
transcription_config?: {
|
|
23130
22665
|
language: string;
|
|
23131
22666
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
23132
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
23133
22667
|
domain?: string | undefined;
|
|
22668
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
23134
22669
|
output_locale?: string | undefined;
|
|
23135
22670
|
additional_vocab?: {
|
|
23136
22671
|
content: string;
|
|
@@ -23353,8 +22888,8 @@ declare const getJobsJobidTranscriptResponse: z.ZodObject<{
|
|
|
23353
22888
|
transcription_config?: {
|
|
23354
22889
|
language: string;
|
|
23355
22890
|
diarization?: "none" | "speaker" | "channel" | undefined;
|
|
23356
|
-
operating_point?: "standard" | "enhanced" | undefined;
|
|
23357
22891
|
domain?: string | undefined;
|
|
22892
|
+
operating_point?: "standard" | "enhanced" | undefined;
|
|
23358
22893
|
output_locale?: string | undefined;
|
|
23359
22894
|
additional_vocab?: {
|
|
23360
22895
|
content: string;
|
|
@@ -24935,4 +24470,4 @@ declare namespace elevenLabsSpeechToTextAPI_zod {
|
|
|
24935
24470
|
export { elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdHeader as deleteTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdParams as deleteTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_deleteTranscriptByIdResponse as deleteTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdHeader as getTranscriptByIdHeader, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdParams as getTranscriptByIdParams, elevenLabsSpeechToTextAPI_zod_getTranscriptByIdResponse as getTranscriptByIdResponse, elevenLabsSpeechToTextAPI_zod_speechToTextBody as speechToTextBody, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeSpeakersDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultFour, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo as speechToTextBodyAdditionalFormatsItemIncludeTimestampsDefaultTwo, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree as speechToTextBodyAdditionalFormatsItemMaxCharactersPerLineDefaultThree, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentCharsDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive as speechToTextBodyAdditionalFormatsItemMaxSegmentDurationSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive as speechToTextBodyAdditionalFormatsItemSegmentOnSilenceLongerThanSDefaultOnefive, elevenLabsSpeechToTextAPI_zod_speechToTextBodyAdditionalFormatsMax as speechToTextBodyAdditionalFormatsMax, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMaxOne as speechToTextBodyDiarizationThresholdMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizationThresholdMinOne as speechToTextBodyDiarizationThresholdMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyDiarizeDefault as speechToTextBodyDiarizeDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyFileFormatDefault as speechToTextBodyFileFormatDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyKeytermsDefault as speechToTextBodyKeytermsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNoVerbatimDefault as speechToTextBodyNoVerbatimDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyNumSpeakersMaxOne as speechToTextBodyNumSpeakersMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMaxOne as speechToTextBodySeedMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodySeedMinOne as speechToTextBodySeedMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTagAudioEventsDefault as speechToTextBodyTagAudioEventsDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMaxOne as speechToTextBodyTemperatureMaxOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTemperatureMinOne as speechToTextBodyTemperatureMinOne, elevenLabsSpeechToTextAPI_zod_speechToTextBodyTimestampsGranularityDefault as speechToTextBodyTimestampsGranularityDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyUseMultiChannelDefault as speechToTextBodyUseMultiChannelDefault, elevenLabsSpeechToTextAPI_zod_speechToTextBodyWebhookDefault as speechToTextBodyWebhookDefault, elevenLabsSpeechToTextAPI_zod_speechToTextHeader as speechToTextHeader, elevenLabsSpeechToTextAPI_zod_speechToTextQueryEnableLoggingDefault as speechToTextQueryEnableLoggingDefault, elevenLabsSpeechToTextAPI_zod_speechToTextQueryParams as speechToTextQueryParams, elevenLabsSpeechToTextAPI_zod_speechToTextResponse as speechToTextResponse };
|
|
24936
24471
|
}
|
|
24937
24472
|
|
|
24938
|
-
export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult,
|
|
24473
|
+
export { AssemblyAIAdapter, Chapter as AssemblyAIChapter, type AssemblyAIConfig, ContentSafetyLabelsResult as AssemblyAIContentSafetyResult, Entity as AssemblyAIEntity, AutoHighlightsResult as AssemblyAIHighlightsResult, TranscriptParams as AssemblyAIOptions, SentimentAnalysisResult$1 as AssemblyAISentimentResult, AssemblyAIStreamingOptions, TopicDetectionModelResult as AssemblyAITopicsResult, index$6 as AssemblyAITypes, type AudioChunkAckMessage, AudioInput, AudioResponseFormat, AzureSTTAdapter, Status as AzureStatus, index$5 as AzureTypes, BaseAdapter, CreateTranscriptionRequest, CreateTranscriptionRequestModel, CreateTranscriptionResponseDiarizedJson, CreateTranscriptionResponseVerboseJson, DeepgramAdapter, type DeepgramConfig, ListenV1ResponseMetadata as DeepgramMetadata, ListenV1MediaTranscribeParams as DeepgramOptions, DeepgramStreamingOptions, index$3 as DeepgramTypes, ElevenLabsAdapter, type ElevenLabsConfig, index as ElevenLabsTypes, elevenLabsSpeechToTextAPI_zod as ElevenLabsZodSchemas, type EndRecordingMessage, type EndSessionMessage, type GetProjectRequestV1Response, GladiaAdapter, AudioToLlmListConfigDTO as GladiaAudioToLlmConfig, AudioToLlmListDTO as GladiaAudioToLlmResult, ChapterizationDTO as GladiaChapters, CodeSwitchingConfigDTO as GladiaCodeSwitchingConfig, NamedEntityRecognitionDTO as GladiaEntities, ModerationDTO as GladiaModeration, InitTranscriptionRequest as GladiaOptions, SentimentAnalysisDTO as GladiaSentiment, SpeakerReidentificationDTO as GladiaSpeakerReidentification, GladiaStreamingOptions, StreamingRequest as GladiaStreamingRequest, StructuredDataExtractionDTO as GladiaStructuredData, TranslationDTO as GladiaTranslation, index$7 as GladiaTypes, InitTranscriptionRequest, LanguageIdentificationMode, LanguageIdentificationProperties, type ListProjectRequestsV1Response, type ListTranscriptionResponseItemsItem, ListTranscriptsOptions, ListTranscriptsParams, ListenV1EncodingParameter, ListenV1LanguageParameter, ListenV1MediaTranscribeParams, ListenV1ModelParameter, ListenV1Response, ListenV1ResponseResultsChannelsItemAlternativesItem, ListenV1ResponseResultsChannelsItemAlternativesItemWordsItem, ListenV1ResponseResultsUtterancesItem, ListenV1VersionParameter, ManageV1FilterEndpointParameter, ManageV1FilterStatusParameter, ManageV1ProjectsRequestsListParams, type NamedEntityRecognitionMessage, streamingTypes as OpenAIStreamingTypes, index$4 as OpenAITypes, OpenAIWhisperAdapter, CreateTranscriptionRequest as OpenAIWhisperOptions, type PaginatedTranscriptions, type PostChapterizationMessage, type PostFinalTranscriptMessage, type PostSummarizationMessage, type PostTranscriptMessage, PreRecordedResponse, ProfanityFilterMode, type ProjectRequestResponse, ProviderCapabilities, type ProviderConfig, PunctuationMode, type SentimentAnalysisMessage, SonioxAdapter, type SonioxConfig, SonioxModelCode, index$1 as SonioxTypes, SpeakV1ContainerParameter, SpeakV1EncodingParameter, SpeakV1SampleRateParameter, type SpeechEndMessage, SpeechModel, type SpeechStartMessage, SpeechmaticsAdapter, type SpeechmaticsConfig, index$2 as SpeechmaticsTypes, speechmaticsASRRESTAPI_zod as SpeechmaticsZodSchemas, type StartRecordingMessage, type StartSessionMessage, type StopRecordingAckMessage, StreamEvent, StreamingCallbacks, StreamingOptions, StreamingRequest, type StreamingResponse, StreamingSession, StreamingSupportedBitDepthEnum, StreamingSupportedEncodingEnum, StreamingSupportedModels, StreamingSupportedSampleRateEnum, StreamingUpdateConfiguration, TranscribeOptions, Transcript, type TranscriptListItem, type TranscriptMessage, TranscriptParams, TranscriptStatus, TranscriptUtterance, TranscriptWord, Transcription$1 as Transcription, type TranscriptionAdapter, TranscriptionControllerListV2Params, TranscriptionControllerListV2StatusItem, TranscriptionDTO, TranscriptionLanguageCodeEnum, TranscriptionProperties, TranscriptionProvider, type TranscriptionsListParams, type TranslationMessage, UnifiedTranscriptResponse, UtteranceDTO, VoiceRouter, type VoiceRouterConfig, type WebHook, type WebHookEvents, WordDTO, createAssemblyAIAdapter, createAzureSTTAdapter, createDeepgramAdapter, createElevenLabsAdapter, createGladiaAdapter, createOpenAIWhisperAdapter, createSonioxAdapter, createSpeechmaticsAdapter, createTranscript, createTranscription, createVoiceRouter, deleteTranscript as deleteTranscriptAPI, getTranscript as getTranscriptAPI, listTranscripts as listTranscriptsAPI, preRecordedControllerDeletePreRecordedJobV2, preRecordedControllerGetAudioV2, preRecordedControllerGetPreRecordedJobV2, preRecordedControllerInitPreRecordedJobV2, streamingControllerDeleteStreamingJobV2, streamingControllerGetAudioV2, streamingControllerInitStreamingSessionV2, transcriptionControllerListV2, transcriptionsCreate, transcriptionsDelete, transcriptionsGet, transcriptionsList, transcriptionsListFiles, webHooksCreate, webHooksDelete, webHooksList };
|