mulmocast 2.6.8 → 2.6.10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/agents/image_genai_agent.js +8 -6
- package/lib/agents/image_replicate_agent.js +30 -15
- package/lib/agents/movie_genai_agent.js +1 -1
- package/lib/agents/movie_replicate_agent.js +5 -0
- package/lib/agents/tts_google_agent.js +18 -2
- package/lib/types/provider2agent.d.ts +1 -0
- package/lib/types/provider2agent.js +52 -2
- package/package.json +5 -5
- package/scripts/test/README.md +1 -1
- package/scripts/test/test_flux2.json +29 -0
- package/scripts/test/{gpt.json → test_gpt_image.json} +8 -0
- package/scripts/test/test_replicate_images.json +79 -0
- package/scripts/test/test_replicate_new_movies.json +50 -0
|
@@ -60,14 +60,16 @@ export const imageGenAIAgent = async ({ namedInputs, params, config, }) => {
|
|
|
60
60
|
const model = params.model ?? provider2ImageAgent["google"].defaultModel;
|
|
61
61
|
const apiKey = config?.apiKey;
|
|
62
62
|
const ai = params.vertexai_project
|
|
63
|
-
?
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
63
|
+
? (() => {
|
|
64
|
+
const location = params.vertexai_location ?? "us-central1";
|
|
65
|
+
if (model === "gemini-3-pro-image-preview" && location !== "global") {
|
|
66
|
+
GraphAILogger.warn(`imageGenAIAgent: model "${model}" on Vertex AI is only available in location "global", but got "${location}". Set imageParams.vertexai_location to "global".`);
|
|
67
|
+
}
|
|
68
|
+
return new GoogleGenAI({ vertexai: true, project: params.vertexai_project, location });
|
|
69
|
+
})()
|
|
68
70
|
: (() => {
|
|
69
71
|
if (!apiKey) {
|
|
70
|
-
throw new Error("Google GenAI
|
|
72
|
+
throw new Error("Google GenAI authentication is required. Either set GEMINI_API_KEY (Gemini API) or specify imageParams.vertexai_project (Vertex AI). See docs/vertexai_en.md or docs/vertexai_ja.md.", {
|
|
71
73
|
cause: apiKeyMissingError("imageGenAIAgent", imageAction, "GEMINI_API_KEY"),
|
|
72
74
|
});
|
|
73
75
|
}
|
|
@@ -4,6 +4,22 @@ import Replicate from "replicate";
|
|
|
4
4
|
import { getAspectRatio } from "./movie_replicate_agent.js";
|
|
5
5
|
import { apiKeyMissingError, agentIncorrectAPIKeyError, agentGenerationError, agentInvalidResponseError, imageAction, imageFileTarget, hasCause, } from "../utils/error_cause.js";
|
|
6
6
|
import { provider2ImageAgent } from "../types/provider2agent.js";
|
|
7
|
+
// Replicate image models return one of: FileOutput (object with url() method),
|
|
8
|
+
// Array<FileOutput>, string URL, or { url: string }. Normalize to URL string/URL.
|
|
9
|
+
const extractImageUrl = (output) => {
|
|
10
|
+
if (typeof output === "string")
|
|
11
|
+
return output;
|
|
12
|
+
if (Array.isArray(output))
|
|
13
|
+
return output.length > 0 ? extractImageUrl(output[0]) : undefined;
|
|
14
|
+
if (output && typeof output === "object" && "url" in output) {
|
|
15
|
+
const url = output.url;
|
|
16
|
+
if (typeof url === "function")
|
|
17
|
+
return url();
|
|
18
|
+
if (typeof url === "string")
|
|
19
|
+
return url;
|
|
20
|
+
}
|
|
21
|
+
return undefined;
|
|
22
|
+
};
|
|
7
23
|
export const imageReplicateAgent = async ({ namedInputs, params, config, }) => {
|
|
8
24
|
const { prompt, referenceImages } = namedInputs;
|
|
9
25
|
const { canvasSize } = params;
|
|
@@ -29,22 +45,21 @@ export const imageReplicateAgent = async ({ namedInputs, params, config, }) => {
|
|
|
29
45
|
}
|
|
30
46
|
try {
|
|
31
47
|
const output = await replicate.run(model, { input });
|
|
32
|
-
|
|
33
|
-
if (
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
throw new Error(`Error downloading image: ${imageResponse.status} - ${imageResponse.statusText}`, {
|
|
38
|
-
cause: agentGenerationError("imageReplicateAgent", imageAction, imageFileTarget),
|
|
39
|
-
});
|
|
40
|
-
}
|
|
41
|
-
const arrayBuffer = await imageResponse.arrayBuffer();
|
|
42
|
-
const buffer = Buffer.from(arrayBuffer);
|
|
43
|
-
return { buffer };
|
|
48
|
+
const imageUrl = extractImageUrl(output);
|
|
49
|
+
if (!imageUrl) {
|
|
50
|
+
throw new Error("ERROR: generateImage returned undefined", {
|
|
51
|
+
cause: agentInvalidResponseError("imageReplicateAgent", imageAction, imageFileTarget),
|
|
52
|
+
});
|
|
44
53
|
}
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
54
|
+
const imageResponse = await fetch(imageUrl);
|
|
55
|
+
if (!imageResponse.ok) {
|
|
56
|
+
throw new Error(`Error downloading image: ${imageResponse.status} - ${imageResponse.statusText}`, {
|
|
57
|
+
cause: agentGenerationError("imageReplicateAgent", imageAction, imageFileTarget),
|
|
58
|
+
});
|
|
59
|
+
}
|
|
60
|
+
const arrayBuffer = await imageResponse.arrayBuffer();
|
|
61
|
+
const buffer = Buffer.from(arrayBuffer);
|
|
62
|
+
return { buffer };
|
|
48
63
|
}
|
|
49
64
|
catch (error) {
|
|
50
65
|
GraphAILogger.info("Replicate generation error:", error);
|
|
@@ -175,7 +175,7 @@ export const movieGenAIAgent = async ({ namedInputs, params, config, }) => {
|
|
|
175
175
|
})
|
|
176
176
|
: (() => {
|
|
177
177
|
if (!apiKey) {
|
|
178
|
-
throw new Error("Google GenAI
|
|
178
|
+
throw new Error("Google GenAI authentication is required. Either set GEMINI_API_KEY (Gemini API) or specify movieParams.vertexai_project (Vertex AI). See docs/vertexai_en.md or docs/vertexai_ja.md.", {
|
|
179
179
|
cause: apiKeyMissingError("movieGenAIAgent", imageAction, "GEMINI_API_KEY"),
|
|
180
180
|
});
|
|
181
181
|
}
|
|
@@ -136,6 +136,11 @@ export const movieReplicateAgent = async ({ namedInputs, params, config, }) => {
|
|
|
136
136
|
cause: agentGenerationError("movieReplicateAgent", imageAction, unsupportedModelTarget),
|
|
137
137
|
});
|
|
138
138
|
}
|
|
139
|
+
if (provider2MovieAgent.replicate.modelParams[model].start_image_required && !imagePath) {
|
|
140
|
+
throw new Error(`Model ${model} requires a start image (image-to-video only)`, {
|
|
141
|
+
cause: agentGenerationError("movieReplicateAgent", imageAction, unsupportedModelTarget),
|
|
142
|
+
});
|
|
143
|
+
}
|
|
139
144
|
const duration = getModelDuration("replicate", model, params.duration);
|
|
140
145
|
if (duration === undefined || !provider2MovieAgent.replicate.modelParams[model].durations.includes(duration)) {
|
|
141
146
|
throw new Error(`Duration ${duration} is not supported for model ${model}. Supported durations: ${provider2MovieAgent.replicate.modelParams[model].durations.join(", ")}`, {
|
|
@@ -2,6 +2,10 @@ import { GraphAILogger } from "graphai";
|
|
|
2
2
|
import * as textToSpeech from "@google-cloud/text-to-speech";
|
|
3
3
|
import { agentGenerationError, audioAction, audioFileTarget } from "../utils/error_cause.js";
|
|
4
4
|
const client = new textToSpeech.TextToSpeechClient();
|
|
5
|
+
// Hard cap so a hung Google TTS RPC can't pin a beat indefinitely.
|
|
6
|
+
// Most synthesizeSpeech calls return in seconds; 60s leaves headroom
|
|
7
|
+
// for long inputs and slow regions while still failing loud.
|
|
8
|
+
const SYNTHESIZE_TIMEOUT_MS = 60_000;
|
|
5
9
|
const getPrompt = (text, instructions) => {
|
|
6
10
|
if (instructions) {
|
|
7
11
|
return `### DIRECTOR'S NOTES\n${instructions}\n\n#### TRANSCRIPT\n${text}`;
|
|
@@ -37,7 +41,7 @@ export const ttsGoogleAgent = async ({ namedInputs, params }) => {
|
|
|
37
41
|
};
|
|
38
42
|
try {
|
|
39
43
|
// Call the Text-to-Speech API
|
|
40
|
-
const [response] = await client.synthesizeSpeech(request);
|
|
44
|
+
const [response] = await client.synthesizeSpeech(request, { timeout: SYNTHESIZE_TIMEOUT_MS });
|
|
41
45
|
return { buffer: response.audioContent };
|
|
42
46
|
}
|
|
43
47
|
catch (e) {
|
|
@@ -47,11 +51,23 @@ export const ttsGoogleAgent = async ({ namedInputs, params }) => {
|
|
|
47
51
|
};
|
|
48
52
|
}
|
|
49
53
|
GraphAILogger.info(e);
|
|
50
|
-
|
|
54
|
+
// gRPC errors from @google-cloud/text-to-speech are ServiceError
|
|
55
|
+
// (extends Error with a `details` string). Surface that human-readable
|
|
56
|
+
// text so callers don't see only "TTS Google Error".
|
|
57
|
+
throw new Error(`TTS Google Error: ${grpcErrorDetail(e)}`, {
|
|
51
58
|
cause: agentGenerationError("ttsGoogleAgent", audioAction, audioFileTarget),
|
|
52
59
|
});
|
|
53
60
|
}
|
|
54
61
|
};
|
|
62
|
+
const grpcErrorDetail = (e) => {
|
|
63
|
+
if (e instanceof Error) {
|
|
64
|
+
const details = e.details;
|
|
65
|
+
if (typeof details === "string" && details)
|
|
66
|
+
return details;
|
|
67
|
+
return e.message;
|
|
68
|
+
}
|
|
69
|
+
return String(e);
|
|
70
|
+
};
|
|
55
71
|
const ttsGoogleAgentInfo = {
|
|
56
72
|
name: "ttsGoogleAgent",
|
|
57
73
|
agent: ttsGoogleAgent,
|
|
@@ -44,7 +44,7 @@ export const provider2TTSAgent = {
|
|
|
44
44
|
models: ["mock-model"],
|
|
45
45
|
},
|
|
46
46
|
};
|
|
47
|
-
export const gptImages = ["gpt-image-1.5", "gpt-image-1", "gpt-image-1-mini"];
|
|
47
|
+
export const gptImages = ["gpt-image-2", "gpt-image-1.5", "gpt-image-1", "gpt-image-1-mini"];
|
|
48
48
|
export const provider2ImageAgent = {
|
|
49
49
|
openai: {
|
|
50
50
|
agentName: "imageOpenaiAgent",
|
|
@@ -69,7 +69,23 @@ export const provider2ImageAgent = {
|
|
|
69
69
|
replicate: {
|
|
70
70
|
agentName: "imageReplicateAgent",
|
|
71
71
|
defaultModel: "bytedance/seedream-4",
|
|
72
|
-
models: [
|
|
72
|
+
models: [
|
|
73
|
+
"bytedance/seedream-4",
|
|
74
|
+
"qwen/qwen-image",
|
|
75
|
+
"black-forest-labs/flux-2-pro",
|
|
76
|
+
"black-forest-labs/flux-2-dev",
|
|
77
|
+
"black-forest-labs/flux-1.1-pro",
|
|
78
|
+
"black-forest-labs/flux-1.1-pro-ultra",
|
|
79
|
+
"black-forest-labs/flux-pro",
|
|
80
|
+
"black-forest-labs/flux-dev",
|
|
81
|
+
"black-forest-labs/flux-schnell",
|
|
82
|
+
"ideogram-ai/ideogram-v3-turbo",
|
|
83
|
+
"ideogram-ai/ideogram-v3-balanced",
|
|
84
|
+
"ideogram-ai/ideogram-v3-quality",
|
|
85
|
+
"recraft-ai/recraft-v3",
|
|
86
|
+
"stability-ai/stable-diffusion-3.5-large",
|
|
87
|
+
"luma/photon",
|
|
88
|
+
],
|
|
73
89
|
keyName: "REPLICATE_API_TOKEN",
|
|
74
90
|
},
|
|
75
91
|
mock: {
|
|
@@ -112,6 +128,10 @@ export const provider2MovieAgent = {
|
|
|
112
128
|
"runwayml/gen-4.5",
|
|
113
129
|
"kwaivgi/kling-v3-omni-video",
|
|
114
130
|
"kwaivgi/kling-v3-video",
|
|
131
|
+
"alibaba/happyhorse-1.0",
|
|
132
|
+
"minimax/hailuo-2.3",
|
|
133
|
+
"minimax/hailuo-2.3-fast",
|
|
134
|
+
"pixverse/pixverse-v5",
|
|
115
135
|
],
|
|
116
136
|
modelParams: {
|
|
117
137
|
"bytedance/seedance-1-lite": {
|
|
@@ -273,6 +293,36 @@ export const provider2MovieAgent = {
|
|
|
273
293
|
audio: { mode: AUDIO_MODE_OPTIONAL, param: "generate_audio" },
|
|
274
294
|
price_per_sec: 0.3,
|
|
275
295
|
},
|
|
296
|
+
// TODO: price_per_sec for the models below is a coarse approximation.
|
|
297
|
+
// Actual Replicate pricing varies by resolution / duration / quality and
|
|
298
|
+
// cannot be expressed as a single per-second number. Verify each model at
|
|
299
|
+
// https://replicate.com/<owner>/<model> when this field starts being consumed.
|
|
300
|
+
"alibaba/happyhorse-1.0": {
|
|
301
|
+
durations: [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15],
|
|
302
|
+
start_image: "image",
|
|
303
|
+
audio: { mode: AUDIO_MODE_NEVER },
|
|
304
|
+
price_per_sec: 0.05,
|
|
305
|
+
},
|
|
306
|
+
"minimax/hailuo-2.3": {
|
|
307
|
+
durations: [6, 10],
|
|
308
|
+
start_image: "first_frame_image",
|
|
309
|
+
audio: { mode: AUDIO_MODE_NEVER },
|
|
310
|
+
price_per_sec: 0.1,
|
|
311
|
+
},
|
|
312
|
+
"minimax/hailuo-2.3-fast": {
|
|
313
|
+
durations: [6, 10],
|
|
314
|
+
start_image: "first_frame_image",
|
|
315
|
+
start_image_required: true,
|
|
316
|
+
audio: { mode: AUDIO_MODE_NEVER },
|
|
317
|
+
price_per_sec: 0.06,
|
|
318
|
+
},
|
|
319
|
+
"pixverse/pixverse-v5": {
|
|
320
|
+
durations: [5, 8],
|
|
321
|
+
start_image: "image",
|
|
322
|
+
last_image: "last_frame_image",
|
|
323
|
+
audio: { mode: AUDIO_MODE_NEVER },
|
|
324
|
+
price_per_sec: 0.12,
|
|
325
|
+
},
|
|
276
326
|
},
|
|
277
327
|
},
|
|
278
328
|
google: {
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "mulmocast",
|
|
3
|
-
"version": "2.6.
|
|
3
|
+
"version": "2.6.10",
|
|
4
4
|
"description": "",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "lib/index.node.js",
|
|
@@ -109,11 +109,11 @@
|
|
|
109
109
|
"dotenv": "^17.4.2",
|
|
110
110
|
"fluent-ffmpeg": "^2.1.3",
|
|
111
111
|
"graphai": "^2.0.16",
|
|
112
|
-
"jsdom": "^29.0
|
|
112
|
+
"jsdom": "^29.1.0",
|
|
113
113
|
"marked": "^18.0.2",
|
|
114
114
|
"mulmocast-vision": "^1.0.9",
|
|
115
|
-
"ora": "^9.
|
|
116
|
-
"puppeteer": "^24.
|
|
115
|
+
"ora": "^9.4.0",
|
|
116
|
+
"puppeteer": "^24.42.0",
|
|
117
117
|
"replicate": "^1.4.0",
|
|
118
118
|
"yaml": "^2.8.3",
|
|
119
119
|
"yargs": "^18.0.0",
|
|
@@ -136,7 +136,7 @@
|
|
|
136
136
|
"prettier": "^3.8.3",
|
|
137
137
|
"tsx": "^4.21.0",
|
|
138
138
|
"typescript": "6.0.3",
|
|
139
|
-
"typescript-eslint": "^8.
|
|
139
|
+
"typescript-eslint": "^8.59.1"
|
|
140
140
|
},
|
|
141
141
|
"engines": {
|
|
142
142
|
"node": ">=22.0.0"
|
package/scripts/test/README.md
CHANGED
|
@@ -106,7 +106,7 @@ Language setting tests
|
|
|
106
106
|
Provider-specific feature tests
|
|
107
107
|
|
|
108
108
|
- [**test_hello_google.json**](./test_hello_google.json) - Google TTS専用テスト / Google TTS specific test
|
|
109
|
-
- [**
|
|
109
|
+
- [**test_gpt_image.json**](./test_gpt_image.json) - GPT image model test
|
|
110
110
|
- [**mulmo_story.json**](./mulmo_story.json) - ストーリー形式テスト / Story format test
|
|
111
111
|
- [**nano_banana.json**](./nano_banana.json) - カスタムサンプル / Custom sample
|
|
112
112
|
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
{
|
|
2
|
+
"$mulmocast": {
|
|
3
|
+
"version": "1.1"
|
|
4
|
+
},
|
|
5
|
+
"canvasSize": {
|
|
6
|
+
"width": 1024,
|
|
7
|
+
"height": 1024
|
|
8
|
+
},
|
|
9
|
+
"imageParams": {
|
|
10
|
+
"provider": "replicate",
|
|
11
|
+
"model": "black-forest-labs/flux-2-pro"
|
|
12
|
+
},
|
|
13
|
+
"speechParams": {
|
|
14
|
+
"speakers": {
|
|
15
|
+
"Presenter": {
|
|
16
|
+
"voiceId": "shimmer",
|
|
17
|
+
"provider": "openai"
|
|
18
|
+
}
|
|
19
|
+
}
|
|
20
|
+
},
|
|
21
|
+
"lang": "en",
|
|
22
|
+
"beats": [
|
|
23
|
+
{
|
|
24
|
+
"speaker": "Presenter",
|
|
25
|
+
"text": "Testing flux-2-pro on Replicate.",
|
|
26
|
+
"imagePrompt": "A cinematic photo of a red panda holding a tiny espresso cup, soft window light, shallow depth of field"
|
|
27
|
+
}
|
|
28
|
+
]
|
|
29
|
+
}
|
|
@@ -0,0 +1,79 @@
|
|
|
1
|
+
{
|
|
2
|
+
"$mulmocast": { "version": "1.1" },
|
|
3
|
+
"canvasSize": { "width": 1024, "height": 1024 },
|
|
4
|
+
"imageParams": {
|
|
5
|
+
"provider": "replicate"
|
|
6
|
+
},
|
|
7
|
+
"speechParams": {
|
|
8
|
+
"speakers": {
|
|
9
|
+
"Presenter": {
|
|
10
|
+
"voiceId": "shimmer",
|
|
11
|
+
"provider": "openai"
|
|
12
|
+
}
|
|
13
|
+
}
|
|
14
|
+
},
|
|
15
|
+
"audioParams": { "bgmVolume": 0 },
|
|
16
|
+
"lang": "en",
|
|
17
|
+
"title": "Replicate Image Models Showcase",
|
|
18
|
+
"beats": [
|
|
19
|
+
{
|
|
20
|
+
"speaker": "Presenter",
|
|
21
|
+
"text": "PROMPT: A red panda holding a tiny espresso cup, soft window light, cinematic.",
|
|
22
|
+
"image": {
|
|
23
|
+
"type": "textSlide",
|
|
24
|
+
"slide": {
|
|
25
|
+
"title": "PROMPT: A red panda holding a tiny espresso cup, soft window light, cinematic."
|
|
26
|
+
}
|
|
27
|
+
}
|
|
28
|
+
},
|
|
29
|
+
{
|
|
30
|
+
"id": "flux-2-pro",
|
|
31
|
+
"speaker": "Presenter",
|
|
32
|
+
"text": "Flux 2 Pro from Black Forest Labs — high quality text-to-image.",
|
|
33
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
34
|
+
"imageParams": { "model": "black-forest-labs/flux-2-pro" }
|
|
35
|
+
},
|
|
36
|
+
{
|
|
37
|
+
"id": "flux-1.1-pro-ultra",
|
|
38
|
+
"speaker": "Presenter",
|
|
39
|
+
"text": "Flux 1.1 Pro Ultra — ultra-high resolution variant.",
|
|
40
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
41
|
+
"imageParams": { "model": "black-forest-labs/flux-1.1-pro-ultra" }
|
|
42
|
+
},
|
|
43
|
+
{
|
|
44
|
+
"id": "flux-schnell",
|
|
45
|
+
"speaker": "Presenter",
|
|
46
|
+
"text": "Flux Schnell — fastest open Flux model.",
|
|
47
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
48
|
+
"imageParams": { "model": "black-forest-labs/flux-schnell" }
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
"id": "ideogram-v3-turbo",
|
|
52
|
+
"speaker": "Presenter",
|
|
53
|
+
"text": "Ideogram v3 Turbo — strong at typography and posters.",
|
|
54
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
55
|
+
"imageParams": { "model": "ideogram-ai/ideogram-v3-turbo" }
|
|
56
|
+
},
|
|
57
|
+
{
|
|
58
|
+
"id": "recraft-v3",
|
|
59
|
+
"speaker": "Presenter",
|
|
60
|
+
"text": "Recraft v3 — strong at clean illustration and text in images.",
|
|
61
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
62
|
+
"imageParams": { "model": "recraft-ai/recraft-v3" }
|
|
63
|
+
},
|
|
64
|
+
{
|
|
65
|
+
"id": "sd-3.5-large",
|
|
66
|
+
"speaker": "Presenter",
|
|
67
|
+
"text": "Stable Diffusion 3.5 Large — open weights baseline.",
|
|
68
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
69
|
+
"imageParams": { "model": "stability-ai/stable-diffusion-3.5-large" }
|
|
70
|
+
},
|
|
71
|
+
{
|
|
72
|
+
"id": "luma-photon",
|
|
73
|
+
"speaker": "Presenter",
|
|
74
|
+
"text": "Luma Photon — strong photorealism and style transfer.",
|
|
75
|
+
"imagePrompt": "A red panda holding a tiny espresso cup, soft window light, cinematic",
|
|
76
|
+
"imageParams": { "model": "luma/photon" }
|
|
77
|
+
}
|
|
78
|
+
]
|
|
79
|
+
}
|
|
@@ -0,0 +1,50 @@
|
|
|
1
|
+
{
|
|
2
|
+
"$mulmocast": { "version": "1.1" },
|
|
3
|
+
"movieParams": {
|
|
4
|
+
"provider": "replicate"
|
|
5
|
+
},
|
|
6
|
+
"audioParams": { "bgmVolume": 0 },
|
|
7
|
+
"captionParams": { "lang": "en" },
|
|
8
|
+
"lang": "en",
|
|
9
|
+
"title": "Replicate New Movie Models Showcase",
|
|
10
|
+
"beats": [
|
|
11
|
+
{
|
|
12
|
+
"text": "PROMPT: a woman is walking through a busy Tokyo street at night, she is wearing dark sunglasses",
|
|
13
|
+
"image": {
|
|
14
|
+
"type": "textSlide",
|
|
15
|
+
"slide": {
|
|
16
|
+
"title": "PROMPT: a woman is walking through a busy Tokyo street at night, she is wearing dark sunglasses"
|
|
17
|
+
}
|
|
18
|
+
}
|
|
19
|
+
},
|
|
20
|
+
{
|
|
21
|
+
"id": "alibaba-happyhorse",
|
|
22
|
+
"text": "alibaba/happyhorse-1.0",
|
|
23
|
+
"duration": 5,
|
|
24
|
+
"moviePrompt": "a woman is walking through a busy Tokyo street at night, she is wearing dark sunglasses",
|
|
25
|
+
"movieParams": { "model": "alibaba/happyhorse-1.0" }
|
|
26
|
+
},
|
|
27
|
+
{
|
|
28
|
+
"id": "hailuo-23",
|
|
29
|
+
"text": "minimax/hailuo-2.3",
|
|
30
|
+
"duration": 6,
|
|
31
|
+
"moviePrompt": "a woman is walking through a busy Tokyo street at night, she is wearing dark sunglasses",
|
|
32
|
+
"movieParams": { "model": "minimax/hailuo-2.3" }
|
|
33
|
+
},
|
|
34
|
+
{
|
|
35
|
+
"id": "pixverse-v5",
|
|
36
|
+
"text": "pixverse/pixverse-v5",
|
|
37
|
+
"duration": 5,
|
|
38
|
+
"moviePrompt": "a woman is walking through a busy Tokyo street at night, she is wearing dark sunglasses",
|
|
39
|
+
"movieParams": { "model": "pixverse/pixverse-v5" }
|
|
40
|
+
},
|
|
41
|
+
{
|
|
42
|
+
"id": "hailuo-23-fast",
|
|
43
|
+
"text": "minimax/hailuo-2.3-fast (i2v: starts from a generated still image)",
|
|
44
|
+
"duration": 6,
|
|
45
|
+
"imagePrompt": "a woman wearing dark sunglasses standing on a busy Tokyo street at night, neon reflections, cinematic still",
|
|
46
|
+
"moviePrompt": "the woman starts walking forward, neon reflections shimmer on the wet pavement",
|
|
47
|
+
"movieParams": { "model": "minimax/hailuo-2.3-fast" }
|
|
48
|
+
}
|
|
49
|
+
]
|
|
50
|
+
}
|