@aigne/openai 1.74.0-beta.1 → 1.74.0-beta.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (35) hide show
  1. package/README.md +4 -4
  2. package/dist/index.cjs +2 -0
  3. package/dist/index.d.cts +2 -1
  4. package/dist/index.d.mts +2 -1
  5. package/dist/index.mjs +2 -1
  6. package/dist/openai-chat-model.cjs +18 -18
  7. package/dist/openai-chat-model.d.cts +23 -23
  8. package/dist/openai-chat-model.d.cts.map +1 -1
  9. package/dist/openai-chat-model.d.mts +23 -23
  10. package/dist/openai-chat-model.d.mts.map +1 -1
  11. package/dist/openai-chat-model.mjs +7 -7
  12. package/dist/openai-chat-model.mjs.map +1 -1
  13. package/dist/openai-embedding-model.cjs +68 -0
  14. package/dist/openai-embedding-model.d.cts +28 -0
  15. package/dist/openai-embedding-model.d.cts.map +1 -0
  16. package/dist/openai-embedding-model.d.mts +28 -0
  17. package/dist/openai-embedding-model.d.mts.map +1 -0
  18. package/dist/openai-embedding-model.mjs +68 -0
  19. package/dist/openai-embedding-model.mjs.map +1 -0
  20. package/dist/openai-image-model.cjs +7 -7
  21. package/dist/openai-image-model.d.cts +23 -23
  22. package/dist/openai-image-model.d.cts.map +1 -1
  23. package/dist/openai-image-model.d.mts +23 -23
  24. package/dist/openai-image-model.d.mts.map +1 -1
  25. package/dist/openai-image-model.mjs +3 -3
  26. package/dist/openai-image-model.mjs.map +1 -1
  27. package/dist/openai-video-model.cjs +9 -9
  28. package/dist/openai-video-model.d.cts +42 -42
  29. package/dist/openai-video-model.d.cts.map +1 -1
  30. package/dist/openai-video-model.d.mts +42 -42
  31. package/dist/openai-video-model.d.mts.map +1 -1
  32. package/dist/openai-video-model.mjs +3 -3
  33. package/dist/openai-video-model.mjs.map +1 -1
  34. package/package.json +3 -6
  35. package/LICENSE.md +0 -93
@@ -1,4 +1,4 @@
1
- import { AgentInvokeOptions, VideoModel, VideoModelInput, VideoModelOptions, VideoModelOutput } from "@aigne/core";
1
+ import { ModelInvokeOptions, VideoModel, VideoModelInput, VideoModelOptions, VideoModelOutput } from "@aigne/model-base";
2
2
  import OpenAI, { ClientOptions } from "openai";
3
3
 
4
4
  //#region src/openai-video-model.d.ts
@@ -7,28 +7,28 @@ import OpenAI, { ClientOptions } from "openai";
7
7
  */
8
8
  interface OpenAIVideoModelInput extends VideoModelInput {
9
9
  /**
10
- * Sora model to use for video generation
11
- *
12
- * - `sora-2`: Standard version, lower cost
13
- * - `sora-2-pro`: Pro version, higher quality
14
- *
15
- * @default "sora-2"
16
- */
10
+ * Sora model to use for video generation
11
+ *
12
+ * - `sora-2`: Standard version, lower cost
13
+ * - `sora-2-pro`: Pro version, higher quality
14
+ *
15
+ * @default "sora-2"
16
+ */
17
17
  model?: "sora-2" | "sora-2-pro";
18
18
  /**
19
- * Video resolution (width x height)
20
- *
21
- * - `720x1280`: Vertical video (9:16)
22
- * - `1280x720`: Horizontal video (16:9)
23
- * - `1024x1792`: Vertical video (9:16, higher resolution)
24
- * - `1792x1024`: Horizontal video (16:9, higher resolution)
25
- */
19
+ * Video resolution (width x height)
20
+ *
21
+ * - `720x1280`: Vertical video (9:16)
22
+ * - `1280x720`: Horizontal video (16:9)
23
+ * - `1024x1792`: Vertical video (9:16, higher resolution)
24
+ * - `1792x1024`: Horizontal video (16:9, higher resolution)
25
+ */
26
26
  size?: "720x1280" | "1280x720" | "1024x1792" | "1792x1024";
27
27
  /**
28
- * Video duration in seconds
29
- *
30
- * @default "4"
31
- */
28
+ * Video duration in seconds
29
+ *
30
+ * @default "4"
31
+ */
32
32
  seconds?: "4" | "8" | "12";
33
33
  }
34
34
  /**
@@ -40,44 +40,44 @@ interface OpenAIVideoModelOutput extends VideoModelOutput {}
40
40
  */
41
41
  interface OpenAIVideoModelOptions extends VideoModelOptions<OpenAIVideoModelInput, OpenAIVideoModelOutput> {
42
42
  /**
43
- * API key for OpenAI API
44
- *
45
- * If not provided, will look for OPENAI_API_KEY in environment variables
46
- */
43
+ * API key for OpenAI API
44
+ *
45
+ * If not provided, will look for OPENAI_API_KEY in environment variables
46
+ */
47
47
  apiKey?: string;
48
48
  /**
49
- * Base URL for OpenAI API
50
- *
51
- * Useful for proxies or alternate endpoints
52
- */
49
+ * Base URL for OpenAI API
50
+ *
51
+ * Useful for proxies or alternate endpoints
52
+ */
53
53
  baseURL?: string;
54
54
  /**
55
- * OpenAI model to use
56
- *
57
- * Defaults to 'sora-2'
58
- */
55
+ * OpenAI model to use
56
+ *
57
+ * Defaults to 'sora-2'
58
+ */
59
59
  model?: string;
60
60
  /**
61
- * Additional model options to control behavior
62
- */
61
+ * Additional model options to control behavior
62
+ */
63
63
  modelOptions?: Omit<Partial<OpenAIVideoModelInput>, "model">;
64
64
  /**
65
- * Client options for OpenAI API
66
- */
65
+ * Client options for OpenAI API
66
+ */
67
67
  clientOptions?: Partial<ClientOptions>;
68
68
  /**
69
- * Polling interval in milliseconds for checking video generation status
70
- *
71
- * Defaults to 2000ms (2 seconds)
72
- */
69
+ * Polling interval in milliseconds for checking video generation status
70
+ *
71
+ * Defaults to 2000ms (2 seconds)
72
+ */
73
73
  pollingInterval?: number;
74
74
  }
75
75
  declare class OpenAIVideoModel extends VideoModel<OpenAIVideoModelInput, OpenAIVideoModelOutput> {
76
76
  options?: OpenAIVideoModelOptions | undefined;
77
77
  constructor(options?: OpenAIVideoModelOptions | undefined);
78
78
  /**
79
- * @hidden
80
- */
79
+ * @hidden
80
+ */
81
81
  protected _client?: OpenAI;
82
82
  protected apiKeyEnvName: string;
83
83
  get client(): OpenAI;
@@ -88,7 +88,7 @@ declare class OpenAIVideoModel extends VideoModel<OpenAIVideoModelInput, OpenAIV
88
88
  };
89
89
  get modelOptions(): Omit<Partial<OpenAIVideoModelInput>, "model"> | undefined;
90
90
  downloadToFile(videoId: string): Promise<string>;
91
- process(input: OpenAIVideoModelInput, _options: AgentInvokeOptions): Promise<OpenAIVideoModelOutput>;
91
+ process(input: OpenAIVideoModelInput, _options: ModelInvokeOptions): Promise<OpenAIVideoModelOutput>;
92
92
  }
93
93
  //#endregion
94
94
  export { OpenAIVideoModel, OpenAIVideoModelInput, OpenAIVideoModelOptions, OpenAIVideoModelOutput };
@@ -1 +1 @@
1
- {"version":3,"file":"openai-video-model.d.mts","names":[],"sources":["../src/openai-video-model.ts"],"mappings":";;;;;AAqBA;AAgCA;UAhCiB,qBAAA,SAA8B,eAAA;EAAA;AAgC/C;AAKA;;;;;;EArC+C,KAAA;EAAA;AAgC/C;AAKA;;;;;;EArC+C,IAAA;EAAA;AAgC/C;AAKA;;;EArC+C,OAAA;AAAA;AAAA;AAgC/C;AAKA;AArC+C,UAgC9B,sBAAA,SAA+B,gBAAA;AAAA;AAKhD;;AALgD,UAK/B,uBAAA,SACP,iBAAA,CAAkB,qBAAA,EAAuB,sBAAA;EAAA;;;;;EAAA,MAAA;EAAA;;;;;EAAA,OAAA;EAAA;;;;;EAAA,KAAA;EAAA;;;EAAA,YAAA,GAyBlC,IAAA,CAAK,OAAA,CAAQ,qBAAA;EAAA;;;EAAA,aAAA,GAKZ,OAAA,CAAQ,aAAA;EAAA;;;;AAyB1B;EAzB0B,eAAA;AAAA;AAAA,cAyBb,gBAAA,SAAyB,UAAA,CAAW,qBAAA,EAAuB,sBAAA;EAAA,OAAA,GAChC,uBAAA;EAAA,YAAA,OAAA,GAAA,uBAAA;EAAA;;;EAAA,UAAA,OAAA,GAYlB,MAAA;EAAA,UAAA,aAAA;EAAA,IAAA,OAAA,GAIV,MAAA;EAAA,IAAA,WAAA;IAAA,GAAA;IAAA,MAAA;IAAA,KAAA;EAAA;EAAA,IAAA,aAAA,GAuBM,IAAA,CAAA,OAAA,CAAA,qBAAA;EAAA,eAAA,OAAA,WAIuB,OAAA;EAAA,QAAA,KAAA,EAS9B,qBAAA,EAAA,QAAA,EACG,kBAAA,GACT,OAAA,CAAQ,sBAAA;AAAA"}
1
+ {"version":3,"file":"openai-video-model.d.mts","names":[],"sources":["../src/openai-video-model.ts"],"mappings":";;;;;AAqBA;;UAAiB,qBAAA,SAA8B,eAAA;EAAe;;;;;;;AAgC9D;EAvBE,KAAA;;;;AA4BF;;;;;EAlBE,IAAA;EA4CoB;;;;;EArCpB,OAAA;AAAA;;;;UAMe,sBAAA,SAA+B,gBAAA;;;;UAK/B,uBAAA,SACP,iBAAA,CAAkB,qBAAA,EAAuB,sBAAA;EAyBrB;;;;;EAnB5B,MAAA;EA+Be;AAkBjB;;;;EA1CE,OAAA;EA2CsC;;;;;EApCtC,KAAA;EA2EgB;;;EAtEhB,YAAA,GAAe,IAAA,CAAK,OAAA,CAAQ,qBAAA;EAqFjB;;;EAhFX,aAAA,GAAgB,OAAA,CAAQ,aAAA;EAyBsB;;;;;EAlB9C,eAAA;AAAA;AAAA,cAkBW,gBAAA,SAAyB,UAAA,CAAW,qBAAA,EAAuB,sBAAA;EAC1C,OAAA,GAAU,uBAAA;cAAV,OAAA,GAAU,uBAAA;EAYlB;;;EAAA,UAAV,OAAA,GAAU,MAAA;EAAA,UAEV,aAAA;EAAA,IAEN,MAAA,CAAA,GAAM,MAAA;EAAA,IAeG,UAAA,CAAA;;;;;MAQT,YAAA,CAAA,GAAY,IAAA,CAAA,OAAA,CAAA,qBAAA;EAIV,cAAA,CAAe,OAAA,WAAkB,OAAA;EAQxB,OAAA,CACb,KAAA,EAAO,qBAAA,EACP,QAAA,EAAU,kBAAA,GACT,OAAA,CAAQ,sBAAA;AAAA"}
@@ -1,7 +1,7 @@
1
1
  import { CustomOpenAI } from "./openai.mjs";
2
- import { VideoModel, videoModelInputSchema } from "@aigne/core";
3
- import { logger } from "@aigne/core/utils/logger";
4
- import { checkArguments } from "@aigne/core/utils/type-utils";
2
+ import { VideoModel, videoModelInputSchema } from "@aigne/model-base";
3
+ import { logger } from "@aigne/model-base/utils/logger";
4
+ import { checkArguments } from "@aigne/model-base/utils/type-utils";
5
5
  import { z } from "zod";
6
6
 
7
7
  //#region src/openai-video-model.ts
@@ -1 +1 @@
1
- {"version":3,"file":"openai-video-model.mjs","names":[],"sources":["../src/openai-video-model.ts"],"sourcesContent":["import {\n type AgentInvokeOptions,\n VideoModel,\n type VideoModelInput,\n type VideoModelOptions,\n type VideoModelOutput,\n videoModelInputSchema,\n} from \"@aigne/core\";\nimport { logger } from \"@aigne/core/utils/logger\";\nimport { checkArguments } from \"@aigne/core/utils/type-utils\";\nimport type OpenAI from \"openai\";\nimport type { ClientOptions } from \"openai\";\nimport { type ZodType, z } from \"zod\";\nimport { CustomOpenAI } from \"./openai.js\";\n\nconst DEFAULT_MODEL = \"sora-2\";\nconst DEFAULT_SECONDS = 4;\n\n/**\n * Input options for OpenAI Video Model\n */\nexport interface OpenAIVideoModelInput extends VideoModelInput {\n /**\n * Sora model to use for video generation\n *\n * - `sora-2`: Standard version, lower cost\n * - `sora-2-pro`: Pro version, higher quality\n *\n * @default \"sora-2\"\n */\n model?: \"sora-2\" | \"sora-2-pro\";\n\n /**\n * Video resolution (width x height)\n *\n * - `720x1280`: Vertical video (9:16)\n * - `1280x720`: Horizontal video (16:9)\n * - `1024x1792`: Vertical video (9:16, higher resolution)\n * - `1792x1024`: Horizontal video (16:9, higher resolution)\n */\n size?: \"720x1280\" | \"1280x720\" | \"1024x1792\" | \"1792x1024\";\n\n /**\n * Video duration in seconds\n *\n * @default \"4\"\n */\n seconds?: \"4\" | \"8\" | \"12\";\n}\n\n/**\n * Output from OpenAI Video Model\n */\nexport interface OpenAIVideoModelOutput extends VideoModelOutput {}\n\n/**\n * Configuration options for OpenAI Video Model\n */\nexport interface OpenAIVideoModelOptions\n extends VideoModelOptions<OpenAIVideoModelInput, OpenAIVideoModelOutput> {\n /**\n * API key for OpenAI API\n *\n * If not provided, will look for OPENAI_API_KEY in environment variables\n */\n apiKey?: string;\n\n /**\n * Base URL for OpenAI API\n *\n * Useful for proxies or alternate endpoints\n */\n baseURL?: string;\n\n /**\n * OpenAI model to use\n *\n * Defaults to 'sora-2'\n */\n model?: string;\n\n /**\n * Additional model options to control behavior\n */\n modelOptions?: Omit<Partial<OpenAIVideoModelInput>, \"model\">;\n\n /**\n * Client options for OpenAI API\n */\n clientOptions?: Partial<ClientOptions>;\n\n /**\n * Polling interval in milliseconds for checking video generation status\n *\n * Defaults to 2000ms (2 seconds)\n */\n pollingInterval?: number;\n}\n\nconst openAIVideoModelInputSchema: ZodType<OpenAIVideoModelInput> = videoModelInputSchema.extend({\n model: z.enum([\"sora-2\", \"sora-2-pro\"]).optional(),\n seconds: z.enum([\"4\", \"8\", \"12\"]).optional(),\n size: z.enum([\"720x1280\", \"1280x720\", \"1024x1792\", \"1792x1024\"]).optional(),\n});\n\nconst openAIVideoModelOptionsSchema = z.object({\n apiKey: z.string().optional(),\n baseURL: z.string().optional(),\n model: z.string().optional(),\n modelOptions: z.object({}).optional(),\n clientOptions: z.object({}).optional(),\n pollingInterval: z.number().optional(),\n});\n\nexport class OpenAIVideoModel extends VideoModel<OpenAIVideoModelInput, OpenAIVideoModelOutput> {\n constructor(public override options?: OpenAIVideoModelOptions) {\n super({\n ...options,\n description: options?.description ?? \"Generate videos using OpenAI Sora models\",\n inputSchema: openAIVideoModelInputSchema,\n });\n if (options) checkArguments(this.name, openAIVideoModelOptionsSchema, options);\n }\n\n /**\n * @hidden\n */\n protected _client?: OpenAI;\n\n protected apiKeyEnvName = \"OPENAI_API_KEY\";\n\n get client() {\n const { apiKey, url } = this.credential;\n if (!apiKey)\n throw new Error(\n `${this.name} requires an API key. Please provide it via \\`options.apiKey\\`, or set the \\`${this.apiKeyEnvName}\\` environment variable`,\n );\n\n this._client ??= new CustomOpenAI({\n baseURL: url,\n apiKey,\n ...this.options?.clientOptions,\n });\n return this._client;\n }\n\n override get credential() {\n return {\n url: this.options?.baseURL || process.env.OPENAI_BASE_URL,\n apiKey: this.options?.apiKey || process.env[this.apiKeyEnvName],\n model: this.options?.model || DEFAULT_MODEL,\n };\n }\n\n get modelOptions() {\n return this.options?.modelOptions;\n }\n\n async downloadToFile(videoId: string): Promise<string> {\n logger.debug(\"Downloading video content...\");\n const content = await this.client.videos.downloadContent(videoId);\n const arrayBuffer = await content.arrayBuffer();\n const buffer = Buffer.from(arrayBuffer);\n return buffer.toString(\"base64\");\n }\n\n override async process(\n input: OpenAIVideoModelInput,\n _options: AgentInvokeOptions,\n ): Promise<OpenAIVideoModelOutput> {\n const model = input.model ?? input.modelOptions?.model ?? this.credential.model;\n\n const createParams: OpenAI.Videos.VideoCreateParams = {\n model: model as OpenAI.Videos.VideoModel,\n prompt: input.prompt,\n };\n\n if (input.seconds) createParams.seconds = input.seconds;\n if (input.size) createParams.size = input.size;\n if (input.image) {\n createParams.input_reference = await this.transformFileType(\"file\", input.image).then(\n (file) =>\n new File([Buffer.from(file.data, \"base64\")], file.filename || \"image.png\", {\n type: file.mimeType,\n }),\n );\n }\n\n let video = await this.client.videos.create(createParams);\n logger.debug(`Video generation started: ${video.id}`);\n\n const pollingInterval = this.options?.pollingInterval ?? 2000;\n while (video.status === \"in_progress\" || video.status === \"queued\") {\n await new Promise((resolve) => setTimeout(resolve, pollingInterval));\n video = await this.client.videos.retrieve(video.id);\n\n const progress = video.progress ?? 0;\n const statusText = video.status === \"queued\" ? \"Queued\" : \"Processing\";\n logger.debug(`${statusText}: ${progress.toFixed(1)}%`);\n }\n\n if (video.status === \"failed\") {\n throw new Error(`Video generation failed: ${video.error?.message || \"Unknown error\"}`);\n }\n\n if (video.status !== \"completed\") {\n throw new Error(`Unexpected video status: ${video.status}`);\n }\n\n return {\n videos: [\n {\n type: \"file\",\n data: await this.downloadToFile(video.id),\n mimeType: \"video/mp4\",\n filename: `${video.id}.mp4`,\n },\n ],\n usage: {\n inputTokens: 0,\n outputTokens: 0,\n },\n model,\n seconds: input.seconds ? parseInt(input.seconds, 10) : DEFAULT_SECONDS,\n };\n }\n}\n"],"mappings":";;;;;;;AAeA,MAAM,gBAAgB;AACtB,MAAM,kBAAkB;AAmFxB,MAAM,8BAA8D,sBAAsB,OAAO;CAC/F,OAAO,EAAE,KAAK,CAAC,UAAU,aAAa,CAAC,CAAC,UAAU;CAClD,SAAS,EAAE,KAAK;EAAC;EAAK;EAAK;EAAK,CAAC,CAAC,UAAU;CAC5C,MAAM,EAAE,KAAK;EAAC;EAAY;EAAY;EAAa;EAAY,CAAC,CAAC,UAAU;CAC5E,CAAC;AAEF,MAAM,gCAAgC,EAAE,OAAO;CAC7C,QAAQ,EAAE,QAAQ,CAAC,UAAU;CAC7B,SAAS,EAAE,QAAQ,CAAC,UAAU;CAC9B,OAAO,EAAE,QAAQ,CAAC,UAAU;CAC5B,cAAc,EAAE,OAAO,EAAE,CAAC,CAAC,UAAU;CACrC,eAAe,EAAE,OAAO,EAAE,CAAC,CAAC,UAAU;CACtC,iBAAiB,EAAE,QAAQ,CAAC,UAAU;CACvC,CAAC;AAEF,IAAa,mBAAb,cAAsC,WAA0D;CAC9F,YAAY,AAAgB,SAAmC;AAC7D,QAAM;GACJ,GAAG;GACH,aAAa,SAAS,eAAe;GACrC,aAAa;GACd,CAAC;EALwB;AAM1B,MAAI,QAAS,gBAAe,KAAK,MAAM,+BAA+B,QAAQ;;;;;CAMhF,AAAU;CAEV,AAAU,gBAAgB;CAE1B,IAAI,SAAS;EACX,MAAM,EAAE,QAAQ,QAAQ,KAAK;AAC7B,MAAI,CAAC,OACH,OAAM,IAAI,MACR,GAAG,KAAK,KAAK,+EAA+E,KAAK,cAAc,yBAChH;AAEH,OAAK,YAAY,IAAI,aAAa;GAChC,SAAS;GACT;GACA,GAAG,KAAK,SAAS;GAClB,CAAC;AACF,SAAO,KAAK;;CAGd,IAAa,aAAa;AACxB,SAAO;GACL,KAAK,KAAK,SAAS,WAAW,QAAQ,IAAI;GAC1C,QAAQ,KAAK,SAAS,UAAU,QAAQ,IAAI,KAAK;GACjD,OAAO,KAAK,SAAS,SAAS;GAC/B;;CAGH,IAAI,eAAe;AACjB,SAAO,KAAK,SAAS;;CAGvB,MAAM,eAAe,SAAkC;AACrD,SAAO,MAAM,+BAA+B;EAE5C,MAAM,cAAc,OADJ,MAAM,KAAK,OAAO,OAAO,gBAAgB,QAAQ,EAC/B,aAAa;AAE/C,SADe,OAAO,KAAK,YAAY,CACzB,SAAS,SAAS;;CAGlC,MAAe,QACb,OACA,UACiC;EACjC,MAAM,QAAQ,MAAM,SAAS,MAAM,cAAc,SAAS,KAAK,WAAW;EAE1E,MAAM,eAAgD;GAC7C;GACP,QAAQ,MAAM;GACf;AAED,MAAI,MAAM,QAAS,cAAa,UAAU,MAAM;AAChD,MAAI,MAAM,KAAM,cAAa,OAAO,MAAM;AAC1C,MAAI,MAAM,MACR,cAAa,kBAAkB,MAAM,KAAK,kBAAkB,QAAQ,MAAM,MAAM,CAAC,MAC9E,SACC,IAAI,KAAK,CAAC,OAAO,KAAK,KAAK,MAAM,SAAS,CAAC,EAAE,KAAK,YAAY,aAAa,EACzE,MAAM,KAAK,UACZ,CAAC,CACL;EAGH,IAAI,QAAQ,MAAM,KAAK,OAAO,OAAO,OAAO,aAAa;AACzD,SAAO,MAAM,6BAA6B,MAAM,KAAK;EAErD,MAAM,kBAAkB,KAAK,SAAS,mBAAmB;AACzD,SAAO,MAAM,WAAW,iBAAiB,MAAM,WAAW,UAAU;AAClE,SAAM,IAAI,SAAS,YAAY,WAAW,SAAS,gBAAgB,CAAC;AACpE,WAAQ,MAAM,KAAK,OAAO,OAAO,SAAS,MAAM,GAAG;GAEnD,MAAM,WAAW,MAAM,YAAY;GACnC,MAAM,aAAa,MAAM,WAAW,WAAW,WAAW;AAC1D,UAAO,MAAM,GAAG,WAAW,IAAI,SAAS,QAAQ,EAAE,CAAC,GAAG;;AAGxD,MAAI,MAAM,WAAW,SACnB,OAAM,IAAI,MAAM,4BAA4B,MAAM,OAAO,WAAW,kBAAkB;AAGxF,MAAI,MAAM,WAAW,YACnB,OAAM,IAAI,MAAM,4BAA4B,MAAM,SAAS;AAG7D,SAAO;GACL,QAAQ,CACN;IACE,MAAM;IACN,MAAM,MAAM,KAAK,eAAe,MAAM,GAAG;IACzC,UAAU;IACV,UAAU,GAAG,MAAM,GAAG;IACvB,CACF;GACD,OAAO;IACL,aAAa;IACb,cAAc;IACf;GACD;GACA,SAAS,MAAM,UAAU,SAAS,MAAM,SAAS,GAAG,GAAG;GACxD"}
1
+ {"version":3,"file":"openai-video-model.mjs","names":[],"sources":["../src/openai-video-model.ts"],"sourcesContent":["import {\n type ModelInvokeOptions,\n VideoModel,\n type VideoModelInput,\n type VideoModelOptions,\n type VideoModelOutput,\n videoModelInputSchema,\n} from \"@aigne/model-base\";\nimport { logger } from \"@aigne/model-base/utils/logger\";\nimport { checkArguments } from \"@aigne/model-base/utils/type-utils\";\nimport type OpenAI from \"openai\";\nimport type { ClientOptions } from \"openai\";\nimport { type ZodType, z } from \"zod\";\nimport { CustomOpenAI } from \"./openai.js\";\n\nconst DEFAULT_MODEL = \"sora-2\";\nconst DEFAULT_SECONDS = 4;\n\n/**\n * Input options for OpenAI Video Model\n */\nexport interface OpenAIVideoModelInput extends VideoModelInput {\n /**\n * Sora model to use for video generation\n *\n * - `sora-2`: Standard version, lower cost\n * - `sora-2-pro`: Pro version, higher quality\n *\n * @default \"sora-2\"\n */\n model?: \"sora-2\" | \"sora-2-pro\";\n\n /**\n * Video resolution (width x height)\n *\n * - `720x1280`: Vertical video (9:16)\n * - `1280x720`: Horizontal video (16:9)\n * - `1024x1792`: Vertical video (9:16, higher resolution)\n * - `1792x1024`: Horizontal video (16:9, higher resolution)\n */\n size?: \"720x1280\" | \"1280x720\" | \"1024x1792\" | \"1792x1024\";\n\n /**\n * Video duration in seconds\n *\n * @default \"4\"\n */\n seconds?: \"4\" | \"8\" | \"12\";\n}\n\n/**\n * Output from OpenAI Video Model\n */\nexport interface OpenAIVideoModelOutput extends VideoModelOutput {}\n\n/**\n * Configuration options for OpenAI Video Model\n */\nexport interface OpenAIVideoModelOptions\n extends VideoModelOptions<OpenAIVideoModelInput, OpenAIVideoModelOutput> {\n /**\n * API key for OpenAI API\n *\n * If not provided, will look for OPENAI_API_KEY in environment variables\n */\n apiKey?: string;\n\n /**\n * Base URL for OpenAI API\n *\n * Useful for proxies or alternate endpoints\n */\n baseURL?: string;\n\n /**\n * OpenAI model to use\n *\n * Defaults to 'sora-2'\n */\n model?: string;\n\n /**\n * Additional model options to control behavior\n */\n modelOptions?: Omit<Partial<OpenAIVideoModelInput>, \"model\">;\n\n /**\n * Client options for OpenAI API\n */\n clientOptions?: Partial<ClientOptions>;\n\n /**\n * Polling interval in milliseconds for checking video generation status\n *\n * Defaults to 2000ms (2 seconds)\n */\n pollingInterval?: number;\n}\n\nconst openAIVideoModelInputSchema: ZodType<OpenAIVideoModelInput> = videoModelInputSchema.extend({\n model: z.enum([\"sora-2\", \"sora-2-pro\"]).optional(),\n seconds: z.enum([\"4\", \"8\", \"12\"]).optional(),\n size: z.enum([\"720x1280\", \"1280x720\", \"1024x1792\", \"1792x1024\"]).optional(),\n});\n\nconst openAIVideoModelOptionsSchema = z.object({\n apiKey: z.string().optional(),\n baseURL: z.string().optional(),\n model: z.string().optional(),\n modelOptions: z.object({}).optional(),\n clientOptions: z.object({}).optional(),\n pollingInterval: z.number().optional(),\n});\n\nexport class OpenAIVideoModel extends VideoModel<OpenAIVideoModelInput, OpenAIVideoModelOutput> {\n constructor(public override options?: OpenAIVideoModelOptions) {\n super({\n ...options,\n description: options?.description ?? \"Generate videos using OpenAI Sora models\",\n inputSchema: openAIVideoModelInputSchema,\n });\n if (options) checkArguments(this.name, openAIVideoModelOptionsSchema, options);\n }\n\n /**\n * @hidden\n */\n protected _client?: OpenAI;\n\n protected apiKeyEnvName = \"OPENAI_API_KEY\";\n\n get client() {\n const { apiKey, url } = this.credential;\n if (!apiKey)\n throw new Error(\n `${this.name} requires an API key. Please provide it via \\`options.apiKey\\`, or set the \\`${this.apiKeyEnvName}\\` environment variable`,\n );\n\n this._client ??= new CustomOpenAI({\n baseURL: url,\n apiKey,\n ...this.options?.clientOptions,\n });\n return this._client;\n }\n\n override get credential() {\n return {\n url: this.options?.baseURL || process.env.OPENAI_BASE_URL,\n apiKey: this.options?.apiKey || process.env[this.apiKeyEnvName],\n model: this.options?.model || DEFAULT_MODEL,\n };\n }\n\n get modelOptions() {\n return this.options?.modelOptions;\n }\n\n async downloadToFile(videoId: string): Promise<string> {\n logger.debug(\"Downloading video content...\");\n const content = await this.client.videos.downloadContent(videoId);\n const arrayBuffer = await content.arrayBuffer();\n const buffer = Buffer.from(arrayBuffer);\n return buffer.toString(\"base64\");\n }\n\n override async process(\n input: OpenAIVideoModelInput,\n _options: ModelInvokeOptions,\n ): Promise<OpenAIVideoModelOutput> {\n const model = input.model ?? input.modelOptions?.model ?? this.credential.model;\n\n const createParams: OpenAI.Videos.VideoCreateParams = {\n model: model as OpenAI.Videos.VideoModel,\n prompt: input.prompt,\n };\n\n if (input.seconds) createParams.seconds = input.seconds;\n if (input.size) createParams.size = input.size;\n if (input.image) {\n createParams.input_reference = await this.transformFileType(\"file\", input.image).then(\n (file) =>\n new File([Buffer.from(file.data, \"base64\")], file.filename || \"image.png\", {\n type: file.mimeType,\n }),\n );\n }\n\n let video = await this.client.videos.create(createParams);\n logger.debug(`Video generation started: ${video.id}`);\n\n const pollingInterval = this.options?.pollingInterval ?? 2000;\n while (video.status === \"in_progress\" || video.status === \"queued\") {\n await new Promise((resolve) => setTimeout(resolve, pollingInterval));\n video = await this.client.videos.retrieve(video.id);\n\n const progress = video.progress ?? 0;\n const statusText = video.status === \"queued\" ? \"Queued\" : \"Processing\";\n logger.debug(`${statusText}: ${progress.toFixed(1)}%`);\n }\n\n if (video.status === \"failed\") {\n throw new Error(`Video generation failed: ${video.error?.message || \"Unknown error\"}`);\n }\n\n if (video.status !== \"completed\") {\n throw new Error(`Unexpected video status: ${video.status}`);\n }\n\n return {\n videos: [\n {\n type: \"file\",\n data: await this.downloadToFile(video.id),\n mimeType: \"video/mp4\",\n filename: `${video.id}.mp4`,\n },\n ],\n usage: {\n inputTokens: 0,\n outputTokens: 0,\n },\n model,\n seconds: input.seconds ? parseInt(input.seconds, 10) : DEFAULT_SECONDS,\n };\n }\n}\n"],"mappings":";;;;;;;AAeA,MAAM,gBAAgB;AACtB,MAAM,kBAAkB;AAmFxB,MAAM,8BAA8D,sBAAsB,OAAO;CAC/F,OAAO,EAAE,KAAK,CAAC,UAAU,aAAa,CAAC,CAAC,UAAU;CAClD,SAAS,EAAE,KAAK;EAAC;EAAK;EAAK;EAAK,CAAC,CAAC,UAAU;CAC5C,MAAM,EAAE,KAAK;EAAC;EAAY;EAAY;EAAa;EAAY,CAAC,CAAC,UAAU;CAC5E,CAAC;AAEF,MAAM,gCAAgC,EAAE,OAAO;CAC7C,QAAQ,EAAE,QAAQ,CAAC,UAAU;CAC7B,SAAS,EAAE,QAAQ,CAAC,UAAU;CAC9B,OAAO,EAAE,QAAQ,CAAC,UAAU;CAC5B,cAAc,EAAE,OAAO,EAAE,CAAC,CAAC,UAAU;CACrC,eAAe,EAAE,OAAO,EAAE,CAAC,CAAC,UAAU;CACtC,iBAAiB,EAAE,QAAQ,CAAC,UAAU;CACvC,CAAC;AAEF,IAAa,mBAAb,cAAsC,WAA0D;CAC9F,YAAY,AAAgB,SAAmC;AAC7D,QAAM;GACJ,GAAG;GACH,aAAa,SAAS,eAAe;GACrC,aAAa;GACd,CAAC;EALwB;AAM1B,MAAI,QAAS,gBAAe,KAAK,MAAM,+BAA+B,QAAQ;;;;;CAMhF,AAAU;CAEV,AAAU,gBAAgB;CAE1B,IAAI,SAAS;EACX,MAAM,EAAE,QAAQ,QAAQ,KAAK;AAC7B,MAAI,CAAC,OACH,OAAM,IAAI,MACR,GAAG,KAAK,KAAK,+EAA+E,KAAK,cAAc,yBAChH;AAEH,OAAK,YAAY,IAAI,aAAa;GAChC,SAAS;GACT;GACA,GAAG,KAAK,SAAS;GAClB,CAAC;AACF,SAAO,KAAK;;CAGd,IAAa,aAAa;AACxB,SAAO;GACL,KAAK,KAAK,SAAS,WAAW,QAAQ,IAAI;GAC1C,QAAQ,KAAK,SAAS,UAAU,QAAQ,IAAI,KAAK;GACjD,OAAO,KAAK,SAAS,SAAS;GAC/B;;CAGH,IAAI,eAAe;AACjB,SAAO,KAAK,SAAS;;CAGvB,MAAM,eAAe,SAAkC;AACrD,SAAO,MAAM,+BAA+B;EAE5C,MAAM,cAAc,OADJ,MAAM,KAAK,OAAO,OAAO,gBAAgB,QAAQ,EAC/B,aAAa;AAE/C,SADe,OAAO,KAAK,YAAY,CACzB,SAAS,SAAS;;CAGlC,MAAe,QACb,OACA,UACiC;EACjC,MAAM,QAAQ,MAAM,SAAS,MAAM,cAAc,SAAS,KAAK,WAAW;EAE1E,MAAM,eAAgD;GAC7C;GACP,QAAQ,MAAM;GACf;AAED,MAAI,MAAM,QAAS,cAAa,UAAU,MAAM;AAChD,MAAI,MAAM,KAAM,cAAa,OAAO,MAAM;AAC1C,MAAI,MAAM,MACR,cAAa,kBAAkB,MAAM,KAAK,kBAAkB,QAAQ,MAAM,MAAM,CAAC,MAC9E,SACC,IAAI,KAAK,CAAC,OAAO,KAAK,KAAK,MAAM,SAAS,CAAC,EAAE,KAAK,YAAY,aAAa,EACzE,MAAM,KAAK,UACZ,CAAC,CACL;EAGH,IAAI,QAAQ,MAAM,KAAK,OAAO,OAAO,OAAO,aAAa;AACzD,SAAO,MAAM,6BAA6B,MAAM,KAAK;EAErD,MAAM,kBAAkB,KAAK,SAAS,mBAAmB;AACzD,SAAO,MAAM,WAAW,iBAAiB,MAAM,WAAW,UAAU;AAClE,SAAM,IAAI,SAAS,YAAY,WAAW,SAAS,gBAAgB,CAAC;AACpE,WAAQ,MAAM,KAAK,OAAO,OAAO,SAAS,MAAM,GAAG;GAEnD,MAAM,WAAW,MAAM,YAAY;GACnC,MAAM,aAAa,MAAM,WAAW,WAAW,WAAW;AAC1D,UAAO,MAAM,GAAG,WAAW,IAAI,SAAS,QAAQ,EAAE,CAAC,GAAG;;AAGxD,MAAI,MAAM,WAAW,SACnB,OAAM,IAAI,MAAM,4BAA4B,MAAM,OAAO,WAAW,kBAAkB;AAGxF,MAAI,MAAM,WAAW,YACnB,OAAM,IAAI,MAAM,4BAA4B,MAAM,SAAS;AAG7D,SAAO;GACL,QAAQ,CACN;IACE,MAAM;IACN,MAAM,MAAM,KAAK,eAAe,MAAM,GAAG;IACzC,UAAU;IACV,UAAU,GAAG,MAAM,GAAG;IACvB,CACF;GACD,OAAO;IACL,aAAa;IACb,cAAc;IACf;GACD;GACA,SAAS,MAAM,UAAU,SAAS,MAAM,SAAS,GAAG,GAAG;GACxD"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@aigne/openai",
3
- "version": "1.74.0-beta.1",
3
+ "version": "1.74.0-beta.10",
4
4
  "description": "AIGNE OpenAI SDK for integrating with OpenAI's GPT models and API services",
5
5
  "license": "Elastic-2.0",
6
6
  "publishConfig": {
@@ -36,8 +36,7 @@
36
36
  "@aigne/uuid": "^13.0.1",
37
37
  "openai": "^6.14.0",
38
38
  "zod": "^3.25.67",
39
- "@aigne/utils": "^1.74.0-beta.1",
40
- "@aigne/core": "^1.74.0-beta.1"
39
+ "@aigne/model-base": "^1.74.0-beta.10"
41
40
  },
42
41
  "devDependencies": {
43
42
  "@types/bun": "^1.3.6",
@@ -45,9 +44,7 @@
45
44
  "rimraf": "^6.1.2",
46
45
  "tsdown": "0.20.0-beta.3",
47
46
  "typescript": "5.9.2",
48
- "@aigne/scripts": "0.0.0",
49
- "@aigne/typescript-config": "0.0.0",
50
- "@aigne/utils": "1.74.0-beta.1"
47
+ "@aigne/typescript-config": "0.0.0"
51
48
  },
52
49
  "scripts": {
53
50
  "build": "tsdown",
package/LICENSE.md DELETED
@@ -1,93 +0,0 @@
1
- Elastic License 2.0
2
-
3
- URL: https://www.elastic.co/licensing/elastic-license
4
-
5
- ## Acceptance
6
-
7
- By using the software, you agree to all of the terms and conditions below.
8
-
9
- ## Copyright License
10
-
11
- The licensor grants you a non-exclusive, royalty-free, worldwide,
12
- non-sublicensable, non-transferable license to use, copy, distribute, make
13
- available, and prepare derivative works of the software, in each case subject to
14
- the limitations and conditions below.
15
-
16
- ## Limitations
17
-
18
- You may not provide the software to third parties as a hosted or managed
19
- service, where the service provides users with access to any substantial set of
20
- the features or functionality of the software.
21
-
22
- You may not move, change, disable, or circumvent the license key functionality
23
- in the software, and you may not remove or obscure any functionality in the
24
- software that is protected by the license key.
25
-
26
- You may not alter, remove, or obscure any licensing, copyright, or other notices
27
- of the licensor in the software. Any use of the licensor’s trademarks is subject
28
- to applicable law.
29
-
30
- ## Patents
31
-
32
- The licensor grants you a license, under any patent claims the licensor can
33
- license, or becomes able to license, to make, have made, use, sell, offer for
34
- sale, import and have imported the software, in each case subject to the
35
- limitations and conditions in this license. This license does not cover any
36
- patent claims that you cause to be infringed by modifications or additions to
37
- the software. If you or your company make any written claim that the software
38
- infringes or contributes to infringement of any patent, your patent license for
39
- the software granted under these terms ends immediately. If your company makes
40
- such a claim, your patent license ends immediately for work on behalf of your
41
- company.
42
-
43
- ## Notices
44
-
45
- You must ensure that anyone who gets a copy of any part of the software from you
46
- also gets a copy of these terms.
47
-
48
- If you modify the software, you must include in any modified copies of the
49
- software prominent notices stating that you have modified the software.
50
-
51
- ## No Other Rights
52
-
53
- These terms do not imply any licenses other than those expressly granted in
54
- these terms.
55
-
56
- ## Termination
57
-
58
- If you use the software in violation of these terms, such use is not licensed,
59
- and your licenses will automatically terminate. If the licensor provides you
60
- with a notice of your violation, and you cease all violation of this license no
61
- later than 30 days after you receive that notice, your licenses will be
62
- reinstated retroactively. However, if you violate these terms after such
63
- reinstatement, any additional violation of these terms will cause your licenses
64
- to terminate automatically and permanently.
65
-
66
- ## No Liability
67
-
68
- *As far as the law allows, the software comes as is, without any warranty or
69
- condition, and the licensor will not be liable to you for any damages arising
70
- out of these terms or the use or nature of the software, under any kind of
71
- legal claim.*
72
-
73
- ## Definitions
74
-
75
- The **licensor** is the entity offering these terms, and the **software** is the
76
- software the licensor makes available under these terms, including any portion
77
- of it.
78
-
79
- **you** refers to the individual or entity agreeing to these terms.
80
-
81
- **your company** is any legal entity, sole proprietorship, or other kind of
82
- organization that you work for, plus all organizations that have control over,
83
- are under the control of, or are under common control with that
84
- organization. **control** means ownership of substantially all the assets of an
85
- entity, or the power to direct its management and policies by vote, contract, or
86
- otherwise. Control can be direct or indirect.
87
-
88
- **your licenses** are all the licenses granted to you for the software under
89
- these terms.
90
-
91
- **use** means anything you do with the software requiring one of your licenses.
92
-
93
- **trademark** means trademarks, service marks, and similar rights.