@mux/ai 0.5.1 → 0.6.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -86,6 +86,7 @@ S3_SECRET_ACCESS_KEY=your-secret-key
86
86
  | [`getSummaryAndTags`](./docs/WORKFLOWS.md#video-summarization)<br/>[API](./docs/API.md#getsummaryandtagsassetid-options) · [Source](./src/workflows/summarization.ts) | Generate titles, descriptions, and tags for an asset | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video (required), Captions (optional) | None |
87
87
  | [`getModerationScores`](./docs/WORKFLOWS.md#content-moderation)<br/>[API](./docs/API.md#getmoderationscoresassetid-options) · [Source](./src/workflows/moderation.ts) | Detect inappropriate (sexual or violent) content in an asset | OpenAI, Hive | `omni-moderation-latest` (OpenAI) or Hive visual moderation task | Video (required) | None |
88
88
  | [`hasBurnedInCaptions`](./docs/WORKFLOWS.md#burned-in-caption-detection)<br/>[API](./docs/API.md#hasburnedincaptionsassetid-options) · [Source](./src/workflows/burned-in-captions.ts) | Detect burned-in captions (hardcoded subtitles) in an asset | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video (required) | None |
89
+ | [`askQuestions`](./docs/WORKFLOWS.md#ask-questions)<br/>[API](./docs/API.md#askquestionsassetid-questions-options) · [Source](./src/workflows/ask-questions.ts) | Answer yes/no questions about an asset's content | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video (required), Captions (optional) | None |
89
90
  | [`generateChapters`](./docs/WORKFLOWS.md#chapter-generation)<br/>[API](./docs/API.md#generatechaptersassetid-languagecode-options) · [Source](./src/workflows/chapters.ts) | Generate chapter markers for an asset using the transcript | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video or audio-only, Captions/Transcripts (required) | None |
90
91
  | [`generateEmbeddings`](./docs/WORKFLOWS.md#embeddings)<br/>[API](./docs/API.md#generateembeddingsassetid-options) · [Source](./src/workflows/embeddings.ts) | Generate vector embeddings for an asset's transcript chunks | OpenAI, Google | `text-embedding-3-small` (OpenAI), `gemini-embedding-001` (Google) | Video or audio-only, Captions/Transcripts (required) | None |
91
92
  | [`translateCaptions`](./docs/WORKFLOWS.md#caption-translation)<br/>[API](./docs/API.md#translatecaptionsassetid-fromlanguagecode-tolanguagecode-options) · [Source](./src/workflows/translate-captions.ts) | Translate an asset's captions into different languages | OpenAI, Anthropic, Google | `gpt-5.1` (OpenAI), `claude-sonnet-4-5` (Anthropic), `gemini-3-flash-preview` (Google) | Video or audio-only, Captions/Transcripts (required) | AWS S3 (if `uploadToMux=true`) |
@@ -2,7 +2,7 @@ import { z } from 'zod';
2
2
  import { createAnthropic } from '@ai-sdk/anthropic';
3
3
  import { createGoogleGenerativeAI } from '@ai-sdk/google';
4
4
  import { createOpenAI } from '@ai-sdk/openai';
5
- import { k as TokenUsage, M as MuxAIOptions, I as ImageSubmissionMode, C as ChunkingStrategy, j as VideoEmbeddingsResult, T as ToneType } from './types-KcVfWtUl.js';
5
+ import { M as MuxAIOptions, I as ImageSubmissionMode, k as TokenUsage, C as ChunkingStrategy, j as VideoEmbeddingsResult, T as ToneType } from './types-CqjLMB84.js';
6
6
 
7
7
  interface ImageDownloadOptions {
8
8
  /** Request timeout in milliseconds (default: 10000) */
@@ -17,6 +17,155 @@ interface ImageDownloadOptions {
17
17
  exponentialBackoff?: boolean;
18
18
  }
19
19
 
20
+ type SupportedProvider = "openai" | "anthropic" | "google";
21
+ type SupportedEmbeddingProvider = "openai" | "google";
22
+ type OpenAIModelId = Parameters<ReturnType<typeof createOpenAI>["chat"]>[0];
23
+ type AnthropicModelId = Parameters<ReturnType<typeof createAnthropic>["chat"]>[0];
24
+ type GoogleModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["chat"]>[0];
25
+ type OpenAIEmbeddingModelId = Parameters<ReturnType<typeof createOpenAI>["embedding"]>[0];
26
+ type GoogleEmbeddingModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["textEmbeddingModel"]>[0];
27
+ interface ModelIdByProvider {
28
+ openai: OpenAIModelId;
29
+ anthropic: AnthropicModelId;
30
+ google: GoogleModelId;
31
+ }
32
+ interface EmbeddingModelIdByProvider {
33
+ openai: OpenAIEmbeddingModelId;
34
+ google: GoogleEmbeddingModelId;
35
+ }
36
+
37
+ /** A single yes/no question to be answered about video content. */
38
+ interface Question {
39
+ /** The question text */
40
+ question: string;
41
+ }
42
+ /** A single answer to a question. */
43
+ interface QuestionAnswer {
44
+ /** The original question */
45
+ question: string;
46
+ /** Answer selected from the allowed options */
47
+ answer: string;
48
+ /** Confidence score between 0 and 1 */
49
+ confidence: number;
50
+ /** Reasoning explaining the answer based on observable evidence */
51
+ reasoning: string;
52
+ }
53
+ /** Configuration options for askQuestions workflow. */
54
+ interface AskQuestionsOptions extends MuxAIOptions {
55
+ /** AI provider to run (defaults to 'openai'). */
56
+ provider?: SupportedProvider;
57
+ /** Provider-specific chat model identifier. */
58
+ model?: ModelIdByProvider[SupportedProvider];
59
+ /** Allowed answers for each question (defaults to ["yes", "no"]). */
60
+ answerOptions?: string[];
61
+ /** Fetch transcript alongside storyboard (defaults to true). */
62
+ includeTranscript?: boolean;
63
+ /** Strip timestamps/markup from transcripts (defaults to true). */
64
+ cleanTranscript?: boolean;
65
+ /** How storyboard should be delivered to the provider (defaults to 'url'). */
66
+ imageSubmissionMode?: ImageSubmissionMode;
67
+ /** Fine-tune storyboard downloads when imageSubmissionMode === 'base64'. */
68
+ imageDownloadOptions?: ImageDownloadOptions;
69
+ /** Storyboard width in pixels (defaults to 640). */
70
+ storyboardWidth?: number;
71
+ }
72
+ /** Structured return payload for askQuestions workflow. */
73
+ interface AskQuestionsResult {
74
+ /** Asset ID passed into the workflow. */
75
+ assetId: string;
76
+ /** Array of answers for each question. */
77
+ answers: QuestionAnswer[];
78
+ /** Storyboard image URL that was analyzed. */
79
+ storyboardUrl: string;
80
+ /** Token usage from the AI provider (for efficiency/cost analysis). */
81
+ usage?: TokenUsage;
82
+ /** Raw transcript text used for analysis (when includeTranscript is true). */
83
+ transcriptText?: string;
84
+ }
85
+ /** Zod schema for a single answer. */
86
+ declare const questionAnswerSchema: z.ZodObject<{
87
+ question: z.ZodString;
88
+ answer: z.ZodString;
89
+ confidence: z.ZodNumber;
90
+ reasoning: z.ZodString;
91
+ }, "strip", z.ZodTypeAny, {
92
+ question: string;
93
+ answer: string;
94
+ confidence: number;
95
+ reasoning: string;
96
+ }, {
97
+ question: string;
98
+ answer: string;
99
+ confidence: number;
100
+ reasoning: string;
101
+ }>;
102
+ type QuestionAnswerType = z.infer<typeof questionAnswerSchema>;
103
+ declare function createAskQuestionsSchema(allowedAnswers: [string, ...string[]]): z.ZodObject<{
104
+ answers: z.ZodArray<z.ZodObject<{
105
+ question: z.ZodString;
106
+ confidence: z.ZodNumber;
107
+ reasoning: z.ZodString;
108
+ } & {
109
+ answer: z.ZodEnum<[string, ...string[]]>;
110
+ }, "strip", z.ZodTypeAny, {
111
+ question: string;
112
+ answer: string;
113
+ confidence: number;
114
+ reasoning: string;
115
+ }, {
116
+ question: string;
117
+ answer: string;
118
+ confidence: number;
119
+ reasoning: string;
120
+ }>, "many">;
121
+ }, "strip", z.ZodTypeAny, {
122
+ answers: {
123
+ question: string;
124
+ answer: string;
125
+ confidence: number;
126
+ reasoning: string;
127
+ }[];
128
+ }, {
129
+ answers: {
130
+ question: string;
131
+ answer: string;
132
+ confidence: number;
133
+ reasoning: string;
134
+ }[];
135
+ }>;
136
+ type AskQuestionsSchema = ReturnType<typeof createAskQuestionsSchema>;
137
+ type AskQuestionsType = z.infer<AskQuestionsSchema>;
138
+ /**
139
+ * Answer questions about a Mux video asset by analyzing storyboard frames and transcript.
140
+ * Defaults to yes/no answers unless `answerOptions` are provided.
141
+ *
142
+ * This workflow takes a list of questions and returns structured answers with confidence
143
+ * scores and reasoning for each question. All questions are processed in a single LLM call for
144
+ * efficiency.
145
+ *
146
+ * @param assetId - The Mux asset ID to analyze
147
+ * @param questions - Array of questions to answer (each must have a 'question' field)
148
+ * @param options - Configuration options for the workflow
149
+ * @returns Structured answers with confidence scores and reasoning
150
+ *
151
+ * @example
152
+ * ```typescript
153
+ * const result = await askQuestions("abc123", [
154
+ * { question: "Does this video contain cooking?" },
155
+ * { question: "Are there people visible in the video?" },
156
+ * ]);
157
+ *
158
+ * console.log(result.answers[0]);
159
+ * // {
160
+ * // question: "Does this video contain cooking?",
161
+ * // answer: "yes",
162
+ * // confidence: 0.95,
163
+ * // reasoning: "A chef prepares ingredients and cooks in a kitchen throughout the video."
164
+ * // }
165
+ * ```
166
+ */
167
+ declare function askQuestions(assetId: string, questions: Question[], options?: AskQuestionsOptions): Promise<AskQuestionsResult>;
168
+
20
169
  /**
21
170
  * A single section of a prompt, rendered as an XML-like tag.
22
171
  */
@@ -39,23 +188,6 @@ type SectionOverride = string | PromptSection | undefined;
39
188
  */
40
189
  type PromptOverrides<TSections extends string> = Partial<Record<TSections, SectionOverride>>;
41
190
 
42
- type SupportedProvider = "openai" | "anthropic" | "google";
43
- type SupportedEmbeddingProvider = "openai" | "google";
44
- type OpenAIModelId = Parameters<ReturnType<typeof createOpenAI>["chat"]>[0];
45
- type AnthropicModelId = Parameters<ReturnType<typeof createAnthropic>["chat"]>[0];
46
- type GoogleModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["chat"]>[0];
47
- type OpenAIEmbeddingModelId = Parameters<ReturnType<typeof createOpenAI>["embedding"]>[0];
48
- type GoogleEmbeddingModelId = Parameters<ReturnType<typeof createGoogleGenerativeAI>["textEmbeddingModel"]>[0];
49
- interface ModelIdByProvider {
50
- openai: OpenAIModelId;
51
- anthropic: AnthropicModelId;
52
- google: GoogleModelId;
53
- }
54
- interface EmbeddingModelIdByProvider {
55
- openai: OpenAIEmbeddingModelId;
56
- google: GoogleEmbeddingModelId;
57
- }
58
-
59
191
  /** Structured payload returned from `hasBurnedInCaptions`. */
60
192
  interface BurnedInCaptionsResult {
61
193
  assetId: string;
@@ -108,12 +240,12 @@ declare const burnedInCaptionsSchema: z.ZodObject<{
108
240
  confidence: z.ZodNumber;
109
241
  detectedLanguage: z.ZodNullable<z.ZodString>;
110
242
  }, "strip", z.ZodTypeAny, {
111
- hasBurnedInCaptions: boolean;
112
243
  confidence: number;
244
+ hasBurnedInCaptions: boolean;
113
245
  detectedLanguage: string | null;
114
246
  }, {
115
- hasBurnedInCaptions: boolean;
116
247
  confidence: number;
248
+ hasBurnedInCaptions: boolean;
117
249
  detectedLanguage: string | null;
118
250
  }>;
119
251
  /** Inferred shape returned from the burned-in captions schema. */
@@ -124,11 +256,11 @@ declare const chapterSchema: z.ZodObject<{
124
256
  startTime: z.ZodNumber;
125
257
  title: z.ZodString;
126
258
  }, "strip", z.ZodTypeAny, {
127
- startTime: number;
128
259
  title: string;
129
- }, {
130
260
  startTime: number;
261
+ }, {
131
262
  title: string;
263
+ startTime: number;
132
264
  }>;
133
265
  type Chapter = z.infer<typeof chapterSchema>;
134
266
  declare const chaptersSchema: z.ZodObject<{
@@ -136,21 +268,21 @@ declare const chaptersSchema: z.ZodObject<{
136
268
  startTime: z.ZodNumber;
137
269
  title: z.ZodString;
138
270
  }, "strip", z.ZodTypeAny, {
139
- startTime: number;
140
271
  title: string;
141
- }, {
142
272
  startTime: number;
273
+ }, {
143
274
  title: string;
275
+ startTime: number;
144
276
  }>, "many">;
145
277
  }, "strip", z.ZodTypeAny, {
146
278
  chapters: {
147
- startTime: number;
148
279
  title: string;
280
+ startTime: number;
149
281
  }[];
150
282
  }, {
151
283
  chapters: {
152
- startTime: number;
153
284
  title: string;
285
+ startTime: number;
154
286
  }[];
155
287
  }>;
156
288
  type ChaptersType = z.infer<typeof chaptersSchema>;
@@ -246,6 +378,8 @@ interface ModerationResult {
246
378
  /** Convenience flag so callers can understand why `thumbnailScores` may contain a transcript entry. */
247
379
  isAudioOnly: boolean;
248
380
  thumbnailScores: ThumbnailModerationScore[];
381
+ /** Workflow usage metadata (asset duration, thumbnails, etc.). */
382
+ usage?: TokenUsage;
249
383
  maxScores: {
250
384
  sexual: number;
251
385
  violence: number;
@@ -318,14 +452,14 @@ declare const summarySchema: z.ZodObject<{
318
452
  keywords: z.ZodArray<z.ZodString, "many">;
319
453
  title: z.ZodString;
320
454
  description: z.ZodString;
321
- }, "strip", z.ZodTypeAny, {
455
+ }, "strict", z.ZodTypeAny, {
322
456
  title: string;
323
- keywords: string[];
324
457
  description: string;
458
+ keywords: string[];
325
459
  }, {
326
460
  title: string;
327
- keywords: string[];
328
461
  description: string;
462
+ keywords: string[];
329
463
  }>;
330
464
  type SummaryType = z.infer<typeof summarySchema>;
331
465
  /** Structured return payload for `getSummaryAndTags`. */
@@ -497,6 +631,8 @@ interface AudioTranslationResult {
497
631
  dubbingId: string;
498
632
  uploadedTrackId?: string;
499
633
  presignedUrl?: string;
634
+ /** Workflow usage metadata (asset duration, thumbnails, etc.). */
635
+ usage?: TokenUsage;
500
636
  }
501
637
  /** Configuration accepted by `translateAudio`. */
502
638
  interface AudioTranslationOptions extends MuxAIOptions {
@@ -574,6 +710,9 @@ declare const translationSchema: z.ZodObject<{
574
710
  type TranslationPayload = z.infer<typeof translationSchema>;
575
711
  declare function translateCaptions<P extends SupportedProvider = SupportedProvider>(assetId: string, fromLanguageCode: string, toLanguageCode: string, options: TranslationOptions<P>): Promise<TranslationResult>;
576
712
 
713
+ type index_AskQuestionsOptions = AskQuestionsOptions;
714
+ type index_AskQuestionsResult = AskQuestionsResult;
715
+ type index_AskQuestionsType = AskQuestionsType;
577
716
  type index_AudioTranslationOptions = AudioTranslationOptions;
578
717
  type index_AudioTranslationResult = AudioTranslationResult;
579
718
  type index_BurnedInCaptionsAnalysis = BurnedInCaptionsAnalysis;
@@ -595,6 +734,9 @@ type index_HiveModerationSource = HiveModerationSource;
595
734
  type index_ModerationOptions = ModerationOptions;
596
735
  type index_ModerationProvider = ModerationProvider;
597
736
  type index_ModerationResult = ModerationResult;
737
+ type index_Question = Question;
738
+ type index_QuestionAnswer = QuestionAnswer;
739
+ type index_QuestionAnswerType = QuestionAnswerType;
598
740
  declare const index_SUMMARY_KEYWORD_LIMIT: typeof SUMMARY_KEYWORD_LIMIT;
599
741
  type index_SummarizationOptions = SummarizationOptions;
600
742
  type index_SummarizationPromptOverrides = SummarizationPromptOverrides;
@@ -605,6 +747,7 @@ type index_ThumbnailModerationScore = ThumbnailModerationScore;
605
747
  type index_TranslationOptions<P extends SupportedProvider = SupportedProvider> = TranslationOptions<P>;
606
748
  type index_TranslationPayload = TranslationPayload;
607
749
  type index_TranslationResult = TranslationResult;
750
+ declare const index_askQuestions: typeof askQuestions;
608
751
  declare const index_burnedInCaptionsSchema: typeof burnedInCaptionsSchema;
609
752
  declare const index_chapterSchema: typeof chapterSchema;
610
753
  declare const index_chaptersSchema: typeof chaptersSchema;
@@ -614,12 +757,13 @@ declare const index_generateVideoEmbeddings: typeof generateVideoEmbeddings;
614
757
  declare const index_getModerationScores: typeof getModerationScores;
615
758
  declare const index_getSummaryAndTags: typeof getSummaryAndTags;
616
759
  declare const index_hasBurnedInCaptions: typeof hasBurnedInCaptions;
760
+ declare const index_questionAnswerSchema: typeof questionAnswerSchema;
617
761
  declare const index_summarySchema: typeof summarySchema;
618
762
  declare const index_translateAudio: typeof translateAudio;
619
763
  declare const index_translateCaptions: typeof translateCaptions;
620
764
  declare const index_translationSchema: typeof translationSchema;
621
765
  declare namespace index {
622
- export { type index_AudioTranslationOptions as AudioTranslationOptions, type index_AudioTranslationResult as AudioTranslationResult, type index_BurnedInCaptionsAnalysis as BurnedInCaptionsAnalysis, type index_BurnedInCaptionsOptions as BurnedInCaptionsOptions, type index_BurnedInCaptionsPromptOverrides as BurnedInCaptionsPromptOverrides, type index_BurnedInCaptionsPromptSections as BurnedInCaptionsPromptSections, type index_BurnedInCaptionsResult as BurnedInCaptionsResult, type index_Chapter as Chapter, type index_ChapterSystemPromptSections as ChapterSystemPromptSections, type index_ChaptersOptions as ChaptersOptions, type index_ChaptersPromptOverrides as ChaptersPromptOverrides, type index_ChaptersPromptSections as ChaptersPromptSections, type index_ChaptersResult as ChaptersResult, type index_ChaptersType as ChaptersType, type index_EmbeddingsOptions as EmbeddingsOptions, type index_EmbeddingsResult as EmbeddingsResult, type index_HiveModerationOutput as HiveModerationOutput, type index_HiveModerationSource as HiveModerationSource, type index_ModerationOptions as ModerationOptions, type index_ModerationProvider as ModerationProvider, type index_ModerationResult as ModerationResult, index_SUMMARY_KEYWORD_LIMIT as SUMMARY_KEYWORD_LIMIT, type index_SummarizationOptions as SummarizationOptions, type index_SummarizationPromptOverrides as SummarizationPromptOverrides, type index_SummarizationPromptSections as SummarizationPromptSections, type index_SummaryAndTagsResult as SummaryAndTagsResult, type index_SummaryType as SummaryType, type index_ThumbnailModerationScore as ThumbnailModerationScore, type index_TranslationOptions as TranslationOptions, type index_TranslationPayload as TranslationPayload, type index_TranslationResult as TranslationResult, index_burnedInCaptionsSchema as burnedInCaptionsSchema, index_chapterSchema as chapterSchema, index_chaptersSchema as chaptersSchema, index_generateChapters as generateChapters, index_generateEmbeddings as generateEmbeddings, index_generateVideoEmbeddings as generateVideoEmbeddings, index_getModerationScores as getModerationScores, index_getSummaryAndTags as getSummaryAndTags, index_hasBurnedInCaptions as hasBurnedInCaptions, index_summarySchema as summarySchema, index_translateAudio as translateAudio, index_translateCaptions as translateCaptions, index_translationSchema as translationSchema };
766
+ export { type index_AskQuestionsOptions as AskQuestionsOptions, type index_AskQuestionsResult as AskQuestionsResult, type index_AskQuestionsType as AskQuestionsType, type index_AudioTranslationOptions as AudioTranslationOptions, type index_AudioTranslationResult as AudioTranslationResult, type index_BurnedInCaptionsAnalysis as BurnedInCaptionsAnalysis, type index_BurnedInCaptionsOptions as BurnedInCaptionsOptions, type index_BurnedInCaptionsPromptOverrides as BurnedInCaptionsPromptOverrides, type index_BurnedInCaptionsPromptSections as BurnedInCaptionsPromptSections, type index_BurnedInCaptionsResult as BurnedInCaptionsResult, type index_Chapter as Chapter, type index_ChapterSystemPromptSections as ChapterSystemPromptSections, type index_ChaptersOptions as ChaptersOptions, type index_ChaptersPromptOverrides as ChaptersPromptOverrides, type index_ChaptersPromptSections as ChaptersPromptSections, type index_ChaptersResult as ChaptersResult, type index_ChaptersType as ChaptersType, type index_EmbeddingsOptions as EmbeddingsOptions, type index_EmbeddingsResult as EmbeddingsResult, type index_HiveModerationOutput as HiveModerationOutput, type index_HiveModerationSource as HiveModerationSource, type index_ModerationOptions as ModerationOptions, type index_ModerationProvider as ModerationProvider, type index_ModerationResult as ModerationResult, type index_Question as Question, type index_QuestionAnswer as QuestionAnswer, type index_QuestionAnswerType as QuestionAnswerType, index_SUMMARY_KEYWORD_LIMIT as SUMMARY_KEYWORD_LIMIT, type index_SummarizationOptions as SummarizationOptions, type index_SummarizationPromptOverrides as SummarizationPromptOverrides, type index_SummarizationPromptSections as SummarizationPromptSections, type index_SummaryAndTagsResult as SummaryAndTagsResult, type index_SummaryType as SummaryType, type index_ThumbnailModerationScore as ThumbnailModerationScore, type index_TranslationOptions as TranslationOptions, type index_TranslationPayload as TranslationPayload, type index_TranslationResult as TranslationResult, index_askQuestions as askQuestions, index_burnedInCaptionsSchema as burnedInCaptionsSchema, index_chapterSchema as chapterSchema, index_chaptersSchema as chaptersSchema, index_generateChapters as generateChapters, index_generateEmbeddings as generateEmbeddings, index_generateVideoEmbeddings as generateVideoEmbeddings, index_getModerationScores as getModerationScores, index_getSummaryAndTags as getSummaryAndTags, index_hasBurnedInCaptions as hasBurnedInCaptions, index_questionAnswerSchema as questionAnswerSchema, index_summarySchema as summarySchema, index_translateAudio as translateAudio, index_translateCaptions as translateCaptions, index_translationSchema as translationSchema };
623
767
  }
624
768
 
625
- export { type SummarizationPromptSections as A, type BurnedInCaptionsResult as B, type Chapter as C, type SummarizationPromptOverrides as D, type EmbeddingsOptions as E, type SummarizationOptions as F, getSummaryAndTags as G, type HiveModerationSource as H, type AudioTranslationResult as I, type AudioTranslationOptions as J, translateAudio as K, type TranslationResult as L, type ModerationResult as M, type TranslationOptions as N, translationSchema as O, type TranslationPayload as P, translateCaptions as Q, SUMMARY_KEYWORD_LIMIT as S, type ThumbnailModerationScore as T, type BurnedInCaptionsPromptSections as a, type BurnedInCaptionsPromptOverrides as b, type BurnedInCaptionsOptions as c, burnedInCaptionsSchema as d, type BurnedInCaptionsAnalysis as e, chapterSchema as f, chaptersSchema as g, hasBurnedInCaptions as h, index as i, type ChaptersType as j, type ChaptersResult as k, type ChaptersPromptSections as l, type ChaptersPromptOverrides as m, type ChaptersOptions as n, type ChapterSystemPromptSections as o, generateChapters as p, type EmbeddingsResult as q, generateEmbeddings as r, generateVideoEmbeddings as s, type ModerationProvider as t, type HiveModerationOutput as u, type ModerationOptions as v, getModerationScores as w, summarySchema as x, type SummaryType as y, type SummaryAndTagsResult as z };
769
+ export { type AskQuestionsOptions as A, type BurnedInCaptionsResult as B, type Chapter as C, type HiveModerationOutput as D, type EmbeddingsOptions as E, type ModerationOptions as F, getModerationScores as G, type HiveModerationSource as H, summarySchema as I, type SummaryType as J, type SummaryAndTagsResult as K, type SummarizationPromptSections as L, type ModerationResult as M, type SummarizationPromptOverrides as N, type SummarizationOptions as O, getSummaryAndTags as P, type Question as Q, type AudioTranslationResult as R, SUMMARY_KEYWORD_LIMIT as S, type ThumbnailModerationScore as T, type AudioTranslationOptions as U, translateAudio as V, type TranslationResult as W, type TranslationOptions as X, translationSchema as Y, type TranslationPayload as Z, translateCaptions as _, type QuestionAnswer as a, type AskQuestionsResult as b, type QuestionAnswerType as c, type AskQuestionsType as d, askQuestions as e, type BurnedInCaptionsPromptSections as f, type BurnedInCaptionsPromptOverrides as g, type BurnedInCaptionsOptions as h, index as i, burnedInCaptionsSchema as j, type BurnedInCaptionsAnalysis as k, hasBurnedInCaptions as l, chapterSchema as m, chaptersSchema as n, type ChaptersType as o, type ChaptersResult as p, questionAnswerSchema as q, type ChaptersPromptSections as r, type ChaptersPromptOverrides as s, type ChaptersOptions as t, type ChapterSystemPromptSections as u, generateChapters as v, type EmbeddingsResult as w, generateEmbeddings as x, generateVideoEmbeddings as y, type ModerationProvider as z };
@@ -1,4 +1,4 @@
1
- import { b as WorkflowCredentialsInput, A as AssetTextTrack, c as MuxAsset, h as TextChunk, C as ChunkingStrategy } from './types-KcVfWtUl.js';
1
+ import { b as WorkflowCredentialsInput, A as AssetTextTrack, c as MuxAsset, h as TextChunk, C as ChunkingStrategy } from './types-CqjLMB84.js';
2
2
 
3
3
  declare const DEFAULT_STORYBOARD_WIDTH = 640;
4
4
  /**
package/dist/index.d.ts CHANGED
@@ -1,14 +1,14 @@
1
- import { W as WorkflowCredentials } from './types-KcVfWtUl.js';
2
- export { A as AssetTextTrack, i as ChunkEmbedding, C as ChunkingStrategy, E as Encrypted, a as EncryptedPayload, I as ImageSubmissionMode, M as MuxAIOptions, c as MuxAsset, f as PlaybackAsset, P as PlaybackPolicy, h as TextChunk, g as TokenChunkingConfig, k as TokenUsage, T as ToneType, V as VTTChunkingConfig, j as VideoEmbeddingsResult, b as WorkflowCredentialsInput, d as decryptFromWorkflow, e as encryptForWorkflow } from './types-KcVfWtUl.js';
3
- export { i as primitives } from './index-B5bQ-IQk.js';
4
- export { i as workflows } from './index-D4eU6RMH.js';
1
+ import { W as WorkflowCredentials } from './types-CqjLMB84.js';
2
+ export { A as AssetTextTrack, i as ChunkEmbedding, C as ChunkingStrategy, E as Encrypted, a as EncryptedPayload, I as ImageSubmissionMode, M as MuxAIOptions, c as MuxAsset, f as PlaybackAsset, P as PlaybackPolicy, h as TextChunk, g as TokenChunkingConfig, k as TokenUsage, T as ToneType, U as UsageMetadata, V as VTTChunkingConfig, j as VideoEmbeddingsResult, b as WorkflowCredentialsInput, d as decryptFromWorkflow, e as encryptForWorkflow } from './types-CqjLMB84.js';
3
+ export { i as primitives } from './index-DAlX4SNJ.js';
4
+ export { i as workflows } from './index-B6ro59tn.js';
5
5
  import '@mux/mux-node';
6
6
  import 'zod';
7
7
  import '@ai-sdk/anthropic';
8
8
  import '@ai-sdk/google';
9
9
  import '@ai-sdk/openai';
10
10
 
11
- var version = "0.5.1";
11
+ var version = "0.6.0";
12
12
 
13
13
  /**
14
14
  * A function that returns workflow credentials, either synchronously or asynchronously.