@mux/ai 0.5.1 → 0.6.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -0
- package/dist/{index-D4eU6RMH.d.ts → index-B6ro59tn.d.ts} +175 -31
- package/dist/{index-B5bQ-IQk.d.ts → index-DAlX4SNJ.d.ts} +1 -1
- package/dist/index.d.ts +5 -5
- package/dist/index.js +666 -28293
- package/dist/index.js.map +1 -1
- package/dist/primitives/index.d.ts +2 -2
- package/dist/primitives/index.js +72 -28068
- package/dist/primitives/index.js.map +1 -1
- package/dist/{types-KcVfWtUl.d.ts → types-CqjLMB84.d.ts} +14 -7
- package/dist/workflows/index.d.ts +2 -2
- package/dist/workflows/index.js +793 -28436
- package/dist/workflows/index.js.map +1 -1
- package/package.json +8 -5
package/README.md
CHANGED
|
@@ -86,6 +86,7 @@ S3_SECRET_ACCESS_KEY=your-secret-key
|
|
|
86
86
|
| [`getSummaryAndTags`](./docs/WORKFLOWS.md#video-summarization)<br/>[API](./docs/API.md#getsummaryandtagsassetid-options) · [Source](./src/workflows/summarization.ts) | Generate titles, descriptions, and tags for an asset | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video (required), Captions (optional) | None |
|
|
87
87
|
| [`getModerationScores`](./docs/WORKFLOWS.md#content-moderation)<br/>[API](./docs/API.md#getmoderationscoresassetid-options) · [Source](./src/workflows/moderation.ts) | Detect inappropriate (sexual or violent) content in an asset | OpenAI, Hive | `omni-moderation-latest` (OpenAI) or Hive visual moderation task | Video (required) | None |
|
|
88
88
|
| [`hasBurnedInCaptions`](./docs/WORKFLOWS.md#burned-in-caption-detection)<br/>[API](./docs/API.md#hasburnedincaptionsassetid-options) · [Source](./src/workflows/burned-in-captions.ts) | Detect burned-in captions (hardcoded subtitles) in an asset | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video (required) | None |
|
|
89
|
+
| [`askQuestions`](./docs/WORKFLOWS.md#ask-questions)<br/>[API](./docs/API.md#askquestionsassetid-questions-options) · [Source](./src/workflows/ask-questions.ts) | Answer yes/no questions about an asset's content | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video (required), Captions (optional) | None |
|
|
89
90
|
| [`generateChapters`](./docs/WORKFLOWS.md#chapter-generation)<br/>[API](./docs/API.md#generatechaptersassetid-languagecode-options) · [Source](./src/workflows/chapters.ts) | Generate chapter markers for an asset using the transcript | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video or audio-only, Captions/Transcripts (required) | None |
|
|
90
91
|
| [`generateEmbeddings`](./docs/WORKFLOWS.md#embeddings)<br/>[API](./docs/API.md#generateembeddingsassetid-options) · [Source](./src/workflows/embeddings.ts) | Generate vector embeddings for an asset's transcript chunks | OpenAI, Google | `text-embedding-3-small` (OpenAI), `gemini-embedding-001` (Google) | Video or audio-only, Captions/Transcripts (required) | None |
|
|
91
92
|
| [`translateCaptions`](./docs/WORKFLOWS.md#caption-translation)<br/>[API](./docs/API.md#translatecaptionsassetid-fromlanguagecode-tolanguagecode-options) · [Source](./src/workflows/translate-captions.ts) | Translate an asset's captions into different languages | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video or audio-only, Captions/Transcripts (required) | AWS S3 (if `uploadToMux=true`) |
|
|
@@ -2,7 +2,7 @@ import { z } from 'zod';
|
|
|
2
2
|
import { createAnthropic } from '@ai-sdk/anthropic';
|
|
3
3
|
import { createGoogleGenerativeAI } from '@ai-sdk/google';
|
|
4
4
|
import { createOpenAI } from '@ai-sdk/openai';
|
|
5
|
-
import {
|
|
5
|
+
import { M as MuxAIOptions, I as ImageSubmissionMode, k as TokenUsage, C as ChunkingStrategy, j as VideoEmbeddingsResult, T as ToneType } from './types-CqjLMB84.js';
|
|
6
6
|
|
|
7
7
|
interface ImageDownloadOptions {
|
|
8
8
|
/** Request timeout in milliseconds (default: 10000) */
|
|
@@ -17,6 +17,155 @@ interface ImageDownloadOptions {
|
|
|
17
17
|
exponentialBackoff?: boolean;
|
|
18
18
|
}
|
|
19
19
|
|
|
20
|
+
type SupportedProvider = "openai" | "anthropic" | "google";
|
|
21
|
+
type SupportedEmbeddingProvider = "openai" | "google";
|
|
22
|
+
type OpenAIModelId = Parameters<ReturnType<typeof createOpenAI>["chat"]>[0];
|
|
23
|
+
type AnthropicModelId = Parameters<ReturnType<typeof createAnthropic>["chat"]>[0];
|
|
24
|
+
type GoogleModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["chat"]>[0];
|
|
25
|
+
type OpenAIEmbeddingModelId = Parameters<ReturnType<typeof createOpenAI>["embedding"]>[0];
|
|
26
|
+
type GoogleEmbeddingModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["textEmbeddingModel"]>[0];
|
|
27
|
+
interface ModelIdByProvider {
|
|
28
|
+
openai: OpenAIModelId;
|
|
29
|
+
anthropic: AnthropicModelId;
|
|
30
|
+
google: GoogleModelId;
|
|
31
|
+
}
|
|
32
|
+
interface EmbeddingModelIdByProvider {
|
|
33
|
+
openai: OpenAIEmbeddingModelId;
|
|
34
|
+
google: GoogleEmbeddingModelId;
|
|
35
|
+
}
|
|
36
|
+
|
|
37
|
+
/** A single yes/no question to be answered about video content. */
|
|
38
|
+
interface Question {
|
|
39
|
+
/** The question text */
|
|
40
|
+
question: string;
|
|
41
|
+
}
|
|
42
|
+
/** A single answer to a question. */
|
|
43
|
+
interface QuestionAnswer {
|
|
44
|
+
/** The original question */
|
|
45
|
+
question: string;
|
|
46
|
+
/** Answer selected from the allowed options */
|
|
47
|
+
answer: string;
|
|
48
|
+
/** Confidence score between 0 and 1 */
|
|
49
|
+
confidence: number;
|
|
50
|
+
/** Reasoning explaining the answer based on observable evidence */
|
|
51
|
+
reasoning: string;
|
|
52
|
+
}
|
|
53
|
+
/** Configuration options for askQuestions workflow. */
|
|
54
|
+
interface AskQuestionsOptions extends MuxAIOptions {
|
|
55
|
+
/** AI provider to run (defaults to 'openai'). */
|
|
56
|
+
provider?: SupportedProvider;
|
|
57
|
+
/** Provider-specific chat model identifier. */
|
|
58
|
+
model?: ModelIdByProvider[SupportedProvider];
|
|
59
|
+
/** Allowed answers for each question (defaults to ["yes", "no"]). */
|
|
60
|
+
answerOptions?: string[];
|
|
61
|
+
/** Fetch transcript alongside storyboard (defaults to true). */
|
|
62
|
+
includeTranscript?: boolean;
|
|
63
|
+
/** Strip timestamps/markup from transcripts (defaults to true). */
|
|
64
|
+
cleanTranscript?: boolean;
|
|
65
|
+
/** How storyboard should be delivered to the provider (defaults to 'url'). */
|
|
66
|
+
imageSubmissionMode?: ImageSubmissionMode;
|
|
67
|
+
/** Fine-tune storyboard downloads when imageSubmissionMode === 'base64'. */
|
|
68
|
+
imageDownloadOptions?: ImageDownloadOptions;
|
|
69
|
+
/** Storyboard width in pixels (defaults to 640). */
|
|
70
|
+
storyboardWidth?: number;
|
|
71
|
+
}
|
|
72
|
+
/** Structured return payload for askQuestions workflow. */
|
|
73
|
+
interface AskQuestionsResult {
|
|
74
|
+
/** Asset ID passed into the workflow. */
|
|
75
|
+
assetId: string;
|
|
76
|
+
/** Array of answers for each question. */
|
|
77
|
+
answers: QuestionAnswer[];
|
|
78
|
+
/** Storyboard image URL that was analyzed. */
|
|
79
|
+
storyboardUrl: string;
|
|
80
|
+
/** Token usage from the AI provider (for efficiency/cost analysis). */
|
|
81
|
+
usage?: TokenUsage;
|
|
82
|
+
/** Raw transcript text used for analysis (when includeTranscript is true). */
|
|
83
|
+
transcriptText?: string;
|
|
84
|
+
}
|
|
85
|
+
/** Zod schema for a single answer. */
|
|
86
|
+
declare const questionAnswerSchema: z.ZodObject<{
|
|
87
|
+
question: z.ZodString;
|
|
88
|
+
answer: z.ZodString;
|
|
89
|
+
confidence: z.ZodNumber;
|
|
90
|
+
reasoning: z.ZodString;
|
|
91
|
+
}, "strip", z.ZodTypeAny, {
|
|
92
|
+
question: string;
|
|
93
|
+
answer: string;
|
|
94
|
+
confidence: number;
|
|
95
|
+
reasoning: string;
|
|
96
|
+
}, {
|
|
97
|
+
question: string;
|
|
98
|
+
answer: string;
|
|
99
|
+
confidence: number;
|
|
100
|
+
reasoning: string;
|
|
101
|
+
}>;
|
|
102
|
+
type QuestionAnswerType = z.infer<typeof questionAnswerSchema>;
|
|
103
|
+
declare function createAskQuestionsSchema(allowedAnswers: [string, ...string[]]): z.ZodObject<{
|
|
104
|
+
answers: z.ZodArray<z.ZodObject<{
|
|
105
|
+
question: z.ZodString;
|
|
106
|
+
confidence: z.ZodNumber;
|
|
107
|
+
reasoning: z.ZodString;
|
|
108
|
+
} & {
|
|
109
|
+
answer: z.ZodEnum<[string, ...string[]]>;
|
|
110
|
+
}, "strip", z.ZodTypeAny, {
|
|
111
|
+
question: string;
|
|
112
|
+
answer: string;
|
|
113
|
+
confidence: number;
|
|
114
|
+
reasoning: string;
|
|
115
|
+
}, {
|
|
116
|
+
question: string;
|
|
117
|
+
answer: string;
|
|
118
|
+
confidence: number;
|
|
119
|
+
reasoning: string;
|
|
120
|
+
}>, "many">;
|
|
121
|
+
}, "strip", z.ZodTypeAny, {
|
|
122
|
+
answers: {
|
|
123
|
+
question: string;
|
|
124
|
+
answer: string;
|
|
125
|
+
confidence: number;
|
|
126
|
+
reasoning: string;
|
|
127
|
+
}[];
|
|
128
|
+
}, {
|
|
129
|
+
answers: {
|
|
130
|
+
question: string;
|
|
131
|
+
answer: string;
|
|
132
|
+
confidence: number;
|
|
133
|
+
reasoning: string;
|
|
134
|
+
}[];
|
|
135
|
+
}>;
|
|
136
|
+
type AskQuestionsSchema = ReturnType<typeof createAskQuestionsSchema>;
|
|
137
|
+
type AskQuestionsType = z.infer<AskQuestionsSchema>;
|
|
138
|
+
/**
|
|
139
|
+
* Answer questions about a Mux video asset by analyzing storyboard frames and transcript.
|
|
140
|
+
* Defaults to yes/no answers unless `answerOptions` are provided.
|
|
141
|
+
*
|
|
142
|
+
* This workflow takes a list of questions and returns structured answers with confidence
|
|
143
|
+
* scores and reasoning for each question. All questions are processed in a single LLM call for
|
|
144
|
+
* efficiency.
|
|
145
|
+
*
|
|
146
|
+
* @param assetId - The Mux asset ID to analyze
|
|
147
|
+
* @param questions - Array of questions to answer (each must have a 'question' field)
|
|
148
|
+
* @param options - Configuration options for the workflow
|
|
149
|
+
* @returns Structured answers with confidence scores and reasoning
|
|
150
|
+
*
|
|
151
|
+
* @example
|
|
152
|
+
* ```typescript
|
|
153
|
+
* const result = await askQuestions("abc123", [
|
|
154
|
+
* { question: "Does this video contain cooking?" },
|
|
155
|
+
* { question: "Are there people visible in the video?" },
|
|
156
|
+
* ]);
|
|
157
|
+
*
|
|
158
|
+
* console.log(result.answers[0]);
|
|
159
|
+
* // {
|
|
160
|
+
* // question: "Does this video contain cooking?",
|
|
161
|
+
* // answer: "yes",
|
|
162
|
+
* // confidence: 0.95,
|
|
163
|
+
* // reasoning: "A chef prepares ingredients and cooks in a kitchen throughout the video."
|
|
164
|
+
* // }
|
|
165
|
+
* ```
|
|
166
|
+
*/
|
|
167
|
+
declare function askQuestions(assetId: string, questions: Question[], options?: AskQuestionsOptions): Promise<AskQuestionsResult>;
|
|
168
|
+
|
|
20
169
|
/**
|
|
21
170
|
* A single section of a prompt, rendered as an XML-like tag.
|
|
22
171
|
*/
|
|
@@ -39,23 +188,6 @@ type SectionOverride = string | PromptSection | undefined;
|
|
|
39
188
|
*/
|
|
40
189
|
type PromptOverrides<TSections extends string> = Partial<Record<TSections, SectionOverride>>;
|
|
41
190
|
|
|
42
|
-
type SupportedProvider = "openai" | "anthropic" | "google";
|
|
43
|
-
type SupportedEmbeddingProvider = "openai" | "google";
|
|
44
|
-
type OpenAIModelId = Parameters<ReturnType<typeof createOpenAI>["chat"]>[0];
|
|
45
|
-
type AnthropicModelId = Parameters<ReturnType<typeof createAnthropic>["chat"]>[0];
|
|
46
|
-
type GoogleModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["chat"]>[0];
|
|
47
|
-
type OpenAIEmbeddingModelId = Parameters<ReturnType<typeof createOpenAI>["embedding"]>[0];
|
|
48
|
-
type GoogleEmbeddingModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["textEmbeddingModel"]>[0];
|
|
49
|
-
interface ModelIdByProvider {
|
|
50
|
-
openai: OpenAIModelId;
|
|
51
|
-
anthropic: AnthropicModelId;
|
|
52
|
-
google: GoogleModelId;
|
|
53
|
-
}
|
|
54
|
-
interface EmbeddingModelIdByProvider {
|
|
55
|
-
openai: OpenAIEmbeddingModelId;
|
|
56
|
-
google: GoogleEmbeddingModelId;
|
|
57
|
-
}
|
|
58
|
-
|
|
59
191
|
/** Structured payload returned from `hasBurnedInCaptions`. */
|
|
60
192
|
interface BurnedInCaptionsResult {
|
|
61
193
|
assetId: string;
|
|
@@ -108,12 +240,12 @@ declare const burnedInCaptionsSchema: z.ZodObject<{
|
|
|
108
240
|
confidence: z.ZodNumber;
|
|
109
241
|
detectedLanguage: z.ZodNullable<z.ZodString>;
|
|
110
242
|
}, "strip", z.ZodTypeAny, {
|
|
111
|
-
hasBurnedInCaptions: boolean;
|
|
112
243
|
confidence: number;
|
|
244
|
+
hasBurnedInCaptions: boolean;
|
|
113
245
|
detectedLanguage: string | null;
|
|
114
246
|
}, {
|
|
115
|
-
hasBurnedInCaptions: boolean;
|
|
116
247
|
confidence: number;
|
|
248
|
+
hasBurnedInCaptions: boolean;
|
|
117
249
|
detectedLanguage: string | null;
|
|
118
250
|
}>;
|
|
119
251
|
/** Inferred shape returned from the burned-in captions schema. */
|
|
@@ -124,11 +256,11 @@ declare const chapterSchema: z.ZodObject<{
|
|
|
124
256
|
startTime: z.ZodNumber;
|
|
125
257
|
title: z.ZodString;
|
|
126
258
|
}, "strip", z.ZodTypeAny, {
|
|
127
|
-
startTime: number;
|
|
128
259
|
title: string;
|
|
129
|
-
}, {
|
|
130
260
|
startTime: number;
|
|
261
|
+
}, {
|
|
131
262
|
title: string;
|
|
263
|
+
startTime: number;
|
|
132
264
|
}>;
|
|
133
265
|
type Chapter = z.infer<typeof chapterSchema>;
|
|
134
266
|
declare const chaptersSchema: z.ZodObject<{
|
|
@@ -136,21 +268,21 @@ declare const chaptersSchema: z.ZodObject<{
|
|
|
136
268
|
startTime: z.ZodNumber;
|
|
137
269
|
title: z.ZodString;
|
|
138
270
|
}, "strip", z.ZodTypeAny, {
|
|
139
|
-
startTime: number;
|
|
140
271
|
title: string;
|
|
141
|
-
}, {
|
|
142
272
|
startTime: number;
|
|
273
|
+
}, {
|
|
143
274
|
title: string;
|
|
275
|
+
startTime: number;
|
|
144
276
|
}>, "many">;
|
|
145
277
|
}, "strip", z.ZodTypeAny, {
|
|
146
278
|
chapters: {
|
|
147
|
-
startTime: number;
|
|
148
279
|
title: string;
|
|
280
|
+
startTime: number;
|
|
149
281
|
}[];
|
|
150
282
|
}, {
|
|
151
283
|
chapters: {
|
|
152
|
-
startTime: number;
|
|
153
284
|
title: string;
|
|
285
|
+
startTime: number;
|
|
154
286
|
}[];
|
|
155
287
|
}>;
|
|
156
288
|
type ChaptersType = z.infer<typeof chaptersSchema>;
|
|
@@ -246,6 +378,8 @@ interface ModerationResult {
|
|
|
246
378
|
/** Convenience flag so callers can understand why `thumbnailScores` may contain a transcript entry. */
|
|
247
379
|
isAudioOnly: boolean;
|
|
248
380
|
thumbnailScores: ThumbnailModerationScore[];
|
|
381
|
+
/** Workflow usage metadata (asset duration, thumbnails, etc.). */
|
|
382
|
+
usage?: TokenUsage;
|
|
249
383
|
maxScores: {
|
|
250
384
|
sexual: number;
|
|
251
385
|
violence: number;
|
|
@@ -318,14 +452,14 @@ declare const summarySchema: z.ZodObject<{
|
|
|
318
452
|
keywords: z.ZodArray<z.ZodString, "many">;
|
|
319
453
|
title: z.ZodString;
|
|
320
454
|
description: z.ZodString;
|
|
321
|
-
}, "
|
|
455
|
+
}, "strict", z.ZodTypeAny, {
|
|
322
456
|
title: string;
|
|
323
|
-
keywords: string[];
|
|
324
457
|
description: string;
|
|
458
|
+
keywords: string[];
|
|
325
459
|
}, {
|
|
326
460
|
title: string;
|
|
327
|
-
keywords: string[];
|
|
328
461
|
description: string;
|
|
462
|
+
keywords: string[];
|
|
329
463
|
}>;
|
|
330
464
|
type SummaryType = z.infer<typeof summarySchema>;
|
|
331
465
|
/** Structured return payload for `getSummaryAndTags`. */
|
|
@@ -497,6 +631,8 @@ interface AudioTranslationResult {
|
|
|
497
631
|
dubbingId: string;
|
|
498
632
|
uploadedTrackId?: string;
|
|
499
633
|
presignedUrl?: string;
|
|
634
|
+
/** Workflow usage metadata (asset duration, thumbnails, etc.). */
|
|
635
|
+
usage?: TokenUsage;
|
|
500
636
|
}
|
|
501
637
|
/** Configuration accepted by `translateAudio`. */
|
|
502
638
|
interface AudioTranslationOptions extends MuxAIOptions {
|
|
@@ -574,6 +710,9 @@ declare const translationSchema: z.ZodObject<{
|
|
|
574
710
|
type TranslationPayload = z.infer<typeof translationSchema>;
|
|
575
711
|
declare function translateCaptions<P extends SupportedProvider = SupportedProvider>(assetId: string, fromLanguageCode: string, toLanguageCode: string, options: TranslationOptions<P>): Promise<TranslationResult>;
|
|
576
712
|
|
|
713
|
+
type index_AskQuestionsOptions = AskQuestionsOptions;
|
|
714
|
+
type index_AskQuestionsResult = AskQuestionsResult;
|
|
715
|
+
type index_AskQuestionsType = AskQuestionsType;
|
|
577
716
|
type index_AudioTranslationOptions = AudioTranslationOptions;
|
|
578
717
|
type index_AudioTranslationResult = AudioTranslationResult;
|
|
579
718
|
type index_BurnedInCaptionsAnalysis = BurnedInCaptionsAnalysis;
|
|
@@ -595,6 +734,9 @@ type index_HiveModerationSource = HiveModerationSource;
|
|
|
595
734
|
type index_ModerationOptions = ModerationOptions;
|
|
596
735
|
type index_ModerationProvider = ModerationProvider;
|
|
597
736
|
type index_ModerationResult = ModerationResult;
|
|
737
|
+
type index_Question = Question;
|
|
738
|
+
type index_QuestionAnswer = QuestionAnswer;
|
|
739
|
+
type index_QuestionAnswerType = QuestionAnswerType;
|
|
598
740
|
declare const index_SUMMARY_KEYWORD_LIMIT: typeof SUMMARY_KEYWORD_LIMIT;
|
|
599
741
|
type index_SummarizationOptions = SummarizationOptions;
|
|
600
742
|
type index_SummarizationPromptOverrides = SummarizationPromptOverrides;
|
|
@@ -605,6 +747,7 @@ type index_ThumbnailModerationScore = ThumbnailModerationScore;
|
|
|
605
747
|
type index_TranslationOptions<P extends SupportedProvider = SupportedProvider> = TranslationOptions<P>;
|
|
606
748
|
type index_TranslationPayload = TranslationPayload;
|
|
607
749
|
type index_TranslationResult = TranslationResult;
|
|
750
|
+
declare const index_askQuestions: typeof askQuestions;
|
|
608
751
|
declare const index_burnedInCaptionsSchema: typeof burnedInCaptionsSchema;
|
|
609
752
|
declare const index_chapterSchema: typeof chapterSchema;
|
|
610
753
|
declare const index_chaptersSchema: typeof chaptersSchema;
|
|
@@ -614,12 +757,13 @@ declare const index_generateVideoEmbeddings: typeof generateVideoEmbeddings;
|
|
|
614
757
|
declare const index_getModerationScores: typeof getModerationScores;
|
|
615
758
|
declare const index_getSummaryAndTags: typeof getSummaryAndTags;
|
|
616
759
|
declare const index_hasBurnedInCaptions: typeof hasBurnedInCaptions;
|
|
760
|
+
declare const index_questionAnswerSchema: typeof questionAnswerSchema;
|
|
617
761
|
declare const index_summarySchema: typeof summarySchema;
|
|
618
762
|
declare const index_translateAudio: typeof translateAudio;
|
|
619
763
|
declare const index_translateCaptions: typeof translateCaptions;
|
|
620
764
|
declare const index_translationSchema: typeof translationSchema;
|
|
621
765
|
declare namespace index {
|
|
622
|
-
export { type index_AudioTranslationOptions as AudioTranslationOptions, type index_AudioTranslationResult as AudioTranslationResult, type index_BurnedInCaptionsAnalysis as BurnedInCaptionsAnalysis, type index_BurnedInCaptionsOptions as BurnedInCaptionsOptions, type index_BurnedInCaptionsPromptOverrides as BurnedInCaptionsPromptOverrides, type index_BurnedInCaptionsPromptSections as BurnedInCaptionsPromptSections, type index_BurnedInCaptionsResult as BurnedInCaptionsResult, type index_Chapter as Chapter, type index_ChapterSystemPromptSections as ChapterSystemPromptSections, type index_ChaptersOptions as ChaptersOptions, type index_ChaptersPromptOverrides as ChaptersPromptOverrides, type index_ChaptersPromptSections as ChaptersPromptSections, type index_ChaptersResult as ChaptersResult, type index_ChaptersType as ChaptersType, type index_EmbeddingsOptions as EmbeddingsOptions, type index_EmbeddingsResult as EmbeddingsResult, type index_HiveModerationOutput as HiveModerationOutput, type index_HiveModerationSource as HiveModerationSource, type index_ModerationOptions as ModerationOptions, type index_ModerationProvider as ModerationProvider, type index_ModerationResult as ModerationResult, index_SUMMARY_KEYWORD_LIMIT as SUMMARY_KEYWORD_LIMIT, type index_SummarizationOptions as SummarizationOptions, type index_SummarizationPromptOverrides as SummarizationPromptOverrides, type index_SummarizationPromptSections as SummarizationPromptSections, type index_SummaryAndTagsResult as SummaryAndTagsResult, type index_SummaryType as SummaryType, type index_ThumbnailModerationScore as ThumbnailModerationScore, type index_TranslationOptions as TranslationOptions, type index_TranslationPayload as TranslationPayload, type index_TranslationResult as TranslationResult, index_burnedInCaptionsSchema as burnedInCaptionsSchema, index_chapterSchema as chapterSchema, index_chaptersSchema as chaptersSchema, index_generateChapters as generateChapters, index_generateEmbeddings as generateEmbeddings, index_generateVideoEmbeddings as generateVideoEmbeddings, index_getModerationScores as getModerationScores, index_getSummaryAndTags as getSummaryAndTags, index_hasBurnedInCaptions as hasBurnedInCaptions, index_summarySchema as summarySchema, index_translateAudio as translateAudio, index_translateCaptions as translateCaptions, index_translationSchema as translationSchema };
|
|
766
|
+
export { type index_AskQuestionsOptions as AskQuestionsOptions, type index_AskQuestionsResult as AskQuestionsResult, type index_AskQuestionsType as AskQuestionsType, type index_AudioTranslationOptions as AudioTranslationOptions, type index_AudioTranslationResult as AudioTranslationResult, type index_BurnedInCaptionsAnalysis as BurnedInCaptionsAnalysis, type index_BurnedInCaptionsOptions as BurnedInCaptionsOptions, type index_BurnedInCaptionsPromptOverrides as BurnedInCaptionsPromptOverrides, type index_BurnedInCaptionsPromptSections as BurnedInCaptionsPromptSections, type index_BurnedInCaptionsResult as BurnedInCaptionsResult, type index_Chapter as Chapter, type index_ChapterSystemPromptSections as ChapterSystemPromptSections, type index_ChaptersOptions as ChaptersOptions, type index_ChaptersPromptOverrides as ChaptersPromptOverrides, type index_ChaptersPromptSections as ChaptersPromptSections, type index_ChaptersResult as ChaptersResult, type index_ChaptersType as ChaptersType, type index_EmbeddingsOptions as EmbeddingsOptions, type index_EmbeddingsResult as EmbeddingsResult, type index_HiveModerationOutput as HiveModerationOutput, type index_HiveModerationSource as HiveModerationSource, type index_ModerationOptions as ModerationOptions, type index_ModerationProvider as ModerationProvider, type index_ModerationResult as ModerationResult, type index_Question as Question, type index_QuestionAnswer as QuestionAnswer, type index_QuestionAnswerType as QuestionAnswerType, index_SUMMARY_KEYWORD_LIMIT as SUMMARY_KEYWORD_LIMIT, type index_SummarizationOptions as SummarizationOptions, type index_SummarizationPromptOverrides as SummarizationPromptOverrides, type index_SummarizationPromptSections as SummarizationPromptSections, type index_SummaryAndTagsResult as SummaryAndTagsResult, type index_SummaryType as SummaryType, type index_ThumbnailModerationScore as ThumbnailModerationScore, type index_TranslationOptions as TranslationOptions, type index_TranslationPayload as TranslationPayload, type index_TranslationResult as TranslationResult, index_askQuestions as askQuestions, index_burnedInCaptionsSchema as burnedInCaptionsSchema, index_chapterSchema as chapterSchema, index_chaptersSchema as chaptersSchema, index_generateChapters as generateChapters, index_generateEmbeddings as generateEmbeddings, index_generateVideoEmbeddings as generateVideoEmbeddings, index_getModerationScores as getModerationScores, index_getSummaryAndTags as getSummaryAndTags, index_hasBurnedInCaptions as hasBurnedInCaptions, index_questionAnswerSchema as questionAnswerSchema, index_summarySchema as summarySchema, index_translateAudio as translateAudio, index_translateCaptions as translateCaptions, index_translationSchema as translationSchema };
|
|
623
767
|
}
|
|
624
768
|
|
|
625
|
-
export { type
|
|
769
|
+
export { type AskQuestionsOptions as A, type BurnedInCaptionsResult as B, type Chapter as C, type HiveModerationOutput as D, type EmbeddingsOptions as E, type ModerationOptions as F, getModerationScores as G, type HiveModerationSource as H, summarySchema as I, type SummaryType as J, type SummaryAndTagsResult as K, type SummarizationPromptSections as L, type ModerationResult as M, type SummarizationPromptOverrides as N, type SummarizationOptions as O, getSummaryAndTags as P, type Question as Q, type AudioTranslationResult as R, SUMMARY_KEYWORD_LIMIT as S, type ThumbnailModerationScore as T, type AudioTranslationOptions as U, translateAudio as V, type TranslationResult as W, type TranslationOptions as X, translationSchema as Y, type TranslationPayload as Z, translateCaptions as _, type QuestionAnswer as a, type AskQuestionsResult as b, type QuestionAnswerType as c, type AskQuestionsType as d, askQuestions as e, type BurnedInCaptionsPromptSections as f, type BurnedInCaptionsPromptOverrides as g, type BurnedInCaptionsOptions as h, index as i, burnedInCaptionsSchema as j, type BurnedInCaptionsAnalysis as k, hasBurnedInCaptions as l, chapterSchema as m, chaptersSchema as n, type ChaptersType as o, type ChaptersResult as p, questionAnswerSchema as q, type ChaptersPromptSections as r, type ChaptersPromptOverrides as s, type ChaptersOptions as t, type ChapterSystemPromptSections as u, generateChapters as v, type EmbeddingsResult as w, generateEmbeddings as x, generateVideoEmbeddings as y, type ModerationProvider as z };
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { b as WorkflowCredentialsInput, A as AssetTextTrack, c as MuxAsset, h as TextChunk, C as ChunkingStrategy } from './types-
|
|
1
|
+
import { b as WorkflowCredentialsInput, A as AssetTextTrack, c as MuxAsset, h as TextChunk, C as ChunkingStrategy } from './types-CqjLMB84.js';
|
|
2
2
|
|
|
3
3
|
declare const DEFAULT_STORYBOARD_WIDTH = 640;
|
|
4
4
|
/**
|
package/dist/index.d.ts
CHANGED
|
@@ -1,14 +1,14 @@
|
|
|
1
|
-
import { W as WorkflowCredentials } from './types-
|
|
2
|
-
export { A as AssetTextTrack, i as ChunkEmbedding, C as ChunkingStrategy, E as Encrypted, a as EncryptedPayload, I as ImageSubmissionMode, M as MuxAIOptions, c as MuxAsset, f as PlaybackAsset, P as PlaybackPolicy, h as TextChunk, g as TokenChunkingConfig, k as TokenUsage, T as ToneType, V as VTTChunkingConfig, j as VideoEmbeddingsResult, b as WorkflowCredentialsInput, d as decryptFromWorkflow, e as encryptForWorkflow } from './types-
|
|
3
|
-
export { i as primitives } from './index-
|
|
4
|
-
export { i as workflows } from './index-
|
|
1
|
+
import { W as WorkflowCredentials } from './types-CqjLMB84.js';
|
|
2
|
+
export { A as AssetTextTrack, i as ChunkEmbedding, C as ChunkingStrategy, E as Encrypted, a as EncryptedPayload, I as ImageSubmissionMode, M as MuxAIOptions, c as MuxAsset, f as PlaybackAsset, P as PlaybackPolicy, h as TextChunk, g as TokenChunkingConfig, k as TokenUsage, T as ToneType, U as UsageMetadata, V as VTTChunkingConfig, j as VideoEmbeddingsResult, b as WorkflowCredentialsInput, d as decryptFromWorkflow, e as encryptForWorkflow } from './types-CqjLMB84.js';
|
|
3
|
+
export { i as primitives } from './index-DAlX4SNJ.js';
|
|
4
|
+
export { i as workflows } from './index-B6ro59tn.js';
|
|
5
5
|
import '@mux/mux-node';
|
|
6
6
|
import 'zod';
|
|
7
7
|
import '@ai-sdk/anthropic';
|
|
8
8
|
import '@ai-sdk/google';
|
|
9
9
|
import '@ai-sdk/openai';
|
|
10
10
|
|
|
11
|
-
var version = "0.
|
|
11
|
+
var version = "0.6.0";
|
|
12
12
|
|
|
13
13
|
/**
|
|
14
14
|
* A function that returns workflow credentials, either synchronously or asynchronously.
|